Re:flink sql作业如何支持配置流?

2023-11-20 文章 Xuyang
Hi, 
是否可以将这个”配置维表“换成流表,利用flink cdc,改动这个配置表的时候,监听字段cdc变化,同时下游上流join呢?




--

Best!
Xuyang





在 2023-11-20 19:24:47,"casel.chen"  写道:
>我有一个flink 
>sql作业过滤条件customer_id是需要根据用户配置来定的,类似于Apollo配置中心,是否可以通过定义一张配置维表来实现呢?设置TTL定期去获取最新配置。
>
>
>create table customer_conf_tbl (
>  customer_id STRING
>) with (
>  'connector' = 'apollo',
>  '其他属性' 
>);
>select * from biz_table where customer_id in (select string_split(customer_id, 
>',') from customer_conf_tbl)
>
>
>如果要做成配置实时更新作用于sql作业的话又该如何实现呢?


flink sql作业如何支持配置流?

2023-11-20 文章 casel.chen
我有一个flink 
sql作业过滤条件customer_id是需要根据用户配置来定的,类似于Apollo配置中心,是否可以通过定义一张配置维表来实现呢?设置TTL定期去获取最新配置。


create table customer_conf_tbl (
  customer_id STRING
) with (
  'connector' = 'apollo',
  '其他属性' 
);
select * from biz_table where customer_id in (select string_split(customer_id, 
',') from customer_conf_tbl)


如果要做成配置实时更新作用于sql作业的话又该如何实现呢?

Re: [DISCUSS] Change the default restart-strategy to exponential-delay

2023-11-19 文章 Rui Fan
Hi David and Mason,

Thanks for your feedback!

To David:

> Given that the new default feels more complex than the current behavior,
if we decide to do this I think it will be important to include the
rationale you've shared in the documentation.

Sounds make sense to me, I will add the related doc if we
update the default strategy.

To Mason:

> I suppose we could do some benchmarking on what works well for the
resource providers that Flink relies on e.g. Kubernetes. Based on
conferences and blogs,
> it seems most people are relying on Kubernetes to deploy Flink and the
restart strategy has a large dependency on how well Kubernetes can scale to
requests to redeploy the job.

Sorry, I didn't understand what type of benchmarking
we should do, could you elaborate on it? Thanks a lot.

Best,
Rui

On Sat, Nov 18, 2023 at 3:32 AM Mason Chen  wrote:

> Hi Rui,
>
> I suppose we could do some benchmarking on what works well for the
> resource providers that Flink relies on e.g. Kubernetes. Based on
> conferences and blogs, it seems most people are relying on Kubernetes to
> deploy Flink and the restart strategy has a large dependency on how well
> Kubernetes can scale to requests to redeploy the job.
>
> Best,
> Mason
>
> On Fri, Nov 17, 2023 at 10:07 AM David Anderson 
> wrote:
>
>> Rui,
>>
>> I don't have any direct experience with this topic, but given the
>> motivation you shared, the proposal makes sense to me. Given that the new
>> default feels more complex than the current behavior, if we decide to do
>> this I think it will be important to include the rationale you've shared in
>> the documentation.
>>
>> David
>>
>> On Wed, Nov 15, 2023 at 10:17 PM Rui Fan <1996fan...@gmail.com> wrote:
>>
>>> Hi dear flink users and devs:
>>>
>>> FLIP-364[1] intends to make some improvements to restart-strategy
>>> and discuss updating some of the default values of exponential-delay,
>>> and whether exponential-delay can be used as the default
>>> restart-strategy.
>>> After discussing at dev mail list[2], we hope to collect more feedback
>>> from Flink users.
>>>
>>> # Why does the default restart-strategy need to be updated?
>>>
>>> If checkpointing is enabled, the default value is fixed-delay with
>>> Integer.MAX_VALUE restart attempts and '1 s' delay[3]. It means
>>> the job will restart infinitely with high frequency when a job
>>> continues to fail.
>>>
>>> When the Kafka cluster fails, a large number of flink jobs will be
>>> restarted frequently. After the kafka cluster is recovered, a large
>>> number of high-frequency restarts of flink jobs may cause the
>>> kafka cluster to avalanche again.
>>>
>>> Considering the exponential-delay as the default strategy with
>>> a couple of reasons:
>>>
>>> - The exponential-delay can reduce the restart frequency when
>>>   a job continues to fail.
>>> - It can restart a job quickly when a job fails occasionally.
>>> - The restart-strategy.exponential-delay.jitter-factor can avoid r
>>>   estarting multiple jobs at the same time. It’s useful to prevent
>>>   avalanches.
>>>
>>> # What are the current default values[4] of exponential-delay?
>>>
>>> restart-strategy.exponential-delay.initial-backoff : 1s
>>> restart-strategy.exponential-delay.backoff-multiplier : 2.0
>>> restart-strategy.exponential-delay.jitter-factor : 0.1
>>> restart-strategy.exponential-delay.max-backoff : 5 min
>>> restart-strategy.exponential-delay.reset-backoff-threshold : 1h
>>>
>>> backoff-multiplier=2 means that the delay time of each restart
>>> will be doubled. The delay times are:
>>> 1s, 2s, 4s, 8s, 16s, 32s, 64s, 128s, 256s, 300s, 300s, etc.
>>>
>>> The delay time is increased rapidly, it will affect the recover
>>> time for flink jobs.
>>>
>>> # Option improvements
>>>
>>> We think the backoff-multiplier between 1 and 2 is more sensible,
>>> such as:
>>>
>>> restart-strategy.exponential-delay.backoff-multiplier : 1.2
>>> restart-strategy.exponential-delay.max-backoff : 1 min
>>>
>>> After updating, the delay times are:
>>>
>>> 1s, 1.2s, 1.44s, 1.728s, 2.073s, 2.488s, 2.985s, 3.583s, 4.299s,
>>> 5.159s, 6.191s, 7.430s, 8.916s, 10.699s, 12.839s, 15.407s, 18.488s,
>>> 22.186s, 26.623s, 31.948s, 38.337s, etc
>>>
>>> They achieve the following goals:
>>> - When restarts are infrequent in a short period of time, flink can
>>>   quickly restart the job. (For example: the retry delay time when
>>>   restarting 5 times is 2.073s)
>>> - When restarting frequently in a short period of time, flink can
>>>   slightly reduce the restart frequency to prevent avalanches.
>>>   (For example: the retry delay time when retrying 10 times is 5.1 s,
>>>   and the retry delay time when retrying 20 times is 38s, which is not
>>> very
>>> large.)
>>>
>>> As @Mingliang Liu   mentioned at dev mail list: the
>>> one-size-fits-all
>>> default values do not exist. So our goal is that the default values
>>> can be suitable for most jobs.
>>>
>>> Looking forward to your thoughts and feedback, thanks~
>>>
>>> [1] 

Re:Re:flink的sql gateway支持自定义的UDF吗?

2023-11-19 文章 RS
Hi,
这种ADD JAR的方式测试了也可以用,谢谢了老哥


Thanks





在 2023-11-01 17:34:48,"Xuyang"  写道:
>Hi, 
>你指的是sql gateway上 ADD JAR这种方式来上传自定义UDF的jar包[1]么?
>
>
>
>
>[1] 
>https://nightlies.apache.org/flink/flink-docs-release-1.17/docs/dev/table/sql/jar/
>
>--
>
>Best!
>Xuyang
>
>
>
>
>
>在 2023-11-01 14:21:04,"RS"  写道:
>>Hi
>>flink的sql gateway支持自定义的UDF吗?包括java和python的,有示例可以参考下吗?


Re:Re:flink的sql gateway支持自定义的UDF吗?

2023-11-19 文章 RS
Hi,
是的,自定义的UDF比较多,或者实现方式不同,所以加载的时候,想单独加载下,
sql-client有个参数就可以支持,-j 
sql gateway为什么不提供了?


Thanks





在 2023-11-01 17:34:48,"Xuyang"  写道:
>Hi, 
>你指的是sql gateway上 ADD JAR这种方式来上传自定义UDF的jar包[1]么?
>
>
>
>
>[1] 
>https://nightlies.apache.org/flink/flink-docs-release-1.17/docs/dev/table/sql/jar/
>
>--
>
>Best!
>Xuyang
>
>
>
>
>
>在 2023-11-01 14:21:04,"RS"  写道:
>>Hi
>>flink的sql gateway支持自定义的UDF吗?包括java和python的,有示例可以参考下吗?


Re: [DISCUSS] Change the default restart-strategy to exponential-delay

2023-11-17 文章 David Anderson
Rui,

I don't have any direct experience with this topic, but given the
motivation you shared, the proposal makes sense to me. Given that the new
default feels more complex than the current behavior, if we decide to do
this I think it will be important to include the rationale you've shared in
the documentation.

David

On Wed, Nov 15, 2023 at 10:17 PM Rui Fan <1996fan...@gmail.com> wrote:

> Hi dear flink users and devs:
>
> FLIP-364[1] intends to make some improvements to restart-strategy
> and discuss updating some of the default values of exponential-delay,
> and whether exponential-delay can be used as the default restart-strategy.
> After discussing at dev mail list[2], we hope to collect more feedback
> from Flink users.
>
> # Why does the default restart-strategy need to be updated?
>
> If checkpointing is enabled, the default value is fixed-delay with
> Integer.MAX_VALUE restart attempts and '1 s' delay[3]. It means
> the job will restart infinitely with high frequency when a job
> continues to fail.
>
> When the Kafka cluster fails, a large number of flink jobs will be
> restarted frequently. After the kafka cluster is recovered, a large
> number of high-frequency restarts of flink jobs may cause the
> kafka cluster to avalanche again.
>
> Considering the exponential-delay as the default strategy with
> a couple of reasons:
>
> - The exponential-delay can reduce the restart frequency when
>   a job continues to fail.
> - It can restart a job quickly when a job fails occasionally.
> - The restart-strategy.exponential-delay.jitter-factor can avoid r
>   estarting multiple jobs at the same time. It’s useful to prevent
>   avalanches.
>
> # What are the current default values[4] of exponential-delay?
>
> restart-strategy.exponential-delay.initial-backoff : 1s
> restart-strategy.exponential-delay.backoff-multiplier : 2.0
> restart-strategy.exponential-delay.jitter-factor : 0.1
> restart-strategy.exponential-delay.max-backoff : 5 min
> restart-strategy.exponential-delay.reset-backoff-threshold : 1h
>
> backoff-multiplier=2 means that the delay time of each restart
> will be doubled. The delay times are:
> 1s, 2s, 4s, 8s, 16s, 32s, 64s, 128s, 256s, 300s, 300s, etc.
>
> The delay time is increased rapidly, it will affect the recover
> time for flink jobs.
>
> # Option improvements
>
> We think the backoff-multiplier between 1 and 2 is more sensible,
> such as:
>
> restart-strategy.exponential-delay.backoff-multiplier : 1.2
> restart-strategy.exponential-delay.max-backoff : 1 min
>
> After updating, the delay times are:
>
> 1s, 1.2s, 1.44s, 1.728s, 2.073s, 2.488s, 2.985s, 3.583s, 4.299s,
> 5.159s, 6.191s, 7.430s, 8.916s, 10.699s, 12.839s, 15.407s, 18.488s,
> 22.186s, 26.623s, 31.948s, 38.337s, etc
>
> They achieve the following goals:
> - When restarts are infrequent in a short period of time, flink can
>   quickly restart the job. (For example: the retry delay time when
>   restarting 5 times is 2.073s)
> - When restarting frequently in a short period of time, flink can
>   slightly reduce the restart frequency to prevent avalanches.
>   (For example: the retry delay time when retrying 10 times is 5.1 s,
>   and the retry delay time when retrying 20 times is 38s, which is not very
> large.)
>
> As @Mingliang Liu   mentioned at dev mail list: the
> one-size-fits-all
> default values do not exist. So our goal is that the default values
> can be suitable for most jobs.
>
> Looking forward to your thoughts and feedback, thanks~
>
> [1] https://cwiki.apache.org/confluence/x/uJqzDw
> [2] https://lists.apache.org/thread/5cgrft73kgkzkgjozf9zfk0w2oj7rjym
> [3]
>
> https://nightlies.apache.org/flink/flink-docs-release-1.18/docs/deployment/config/#restart-strategy-type
> [4]
>
> https://nightlies.apache.org/flink/flink-docs-master/docs/ops/state/task_failure_recovery/#exponential-delay-restart-strategy
>
> Best,
> Rui
>


[DISCUSS] Change the default restart-strategy to exponential-delay

2023-11-15 文章 Rui Fan
Hi dear flink users and devs:

FLIP-364[1] intends to make some improvements to restart-strategy
and discuss updating some of the default values of exponential-delay,
and whether exponential-delay can be used as the default restart-strategy.
After discussing at dev mail list[2], we hope to collect more feedback
from Flink users.

# Why does the default restart-strategy need to be updated?

If checkpointing is enabled, the default value is fixed-delay with
Integer.MAX_VALUE restart attempts and '1 s' delay[3]. It means
the job will restart infinitely with high frequency when a job
continues to fail.

When the Kafka cluster fails, a large number of flink jobs will be
restarted frequently. After the kafka cluster is recovered, a large
number of high-frequency restarts of flink jobs may cause the
kafka cluster to avalanche again.

Considering the exponential-delay as the default strategy with
a couple of reasons:

- The exponential-delay can reduce the restart frequency when
  a job continues to fail.
- It can restart a job quickly when a job fails occasionally.
- The restart-strategy.exponential-delay.jitter-factor can avoid r
  estarting multiple jobs at the same time. It’s useful to prevent
  avalanches.

# What are the current default values[4] of exponential-delay?

restart-strategy.exponential-delay.initial-backoff : 1s
restart-strategy.exponential-delay.backoff-multiplier : 2.0
restart-strategy.exponential-delay.jitter-factor : 0.1
restart-strategy.exponential-delay.max-backoff : 5 min
restart-strategy.exponential-delay.reset-backoff-threshold : 1h

backoff-multiplier=2 means that the delay time of each restart
will be doubled. The delay times are:
1s, 2s, 4s, 8s, 16s, 32s, 64s, 128s, 256s, 300s, 300s, etc.

The delay time is increased rapidly, it will affect the recover
time for flink jobs.

# Option improvements

We think the backoff-multiplier between 1 and 2 is more sensible,
such as:

restart-strategy.exponential-delay.backoff-multiplier : 1.2
restart-strategy.exponential-delay.max-backoff : 1 min

After updating, the delay times are:

1s, 1.2s, 1.44s, 1.728s, 2.073s, 2.488s, 2.985s, 3.583s, 4.299s,
5.159s, 6.191s, 7.430s, 8.916s, 10.699s, 12.839s, 15.407s, 18.488s,
22.186s, 26.623s, 31.948s, 38.337s, etc

They achieve the following goals:
- When restarts are infrequent in a short period of time, flink can
  quickly restart the job. (For example: the retry delay time when
  restarting 5 times is 2.073s)
- When restarting frequently in a short period of time, flink can
  slightly reduce the restart frequency to prevent avalanches.
  (For example: the retry delay time when retrying 10 times is 5.1 s,
  and the retry delay time when retrying 20 times is 38s, which is not very
large.)

As @Mingliang Liu   mentioned at dev mail list: the
one-size-fits-all
default values do not exist. So our goal is that the default values
can be suitable for most jobs.

Looking forward to your thoughts and feedback, thanks~

[1] https://cwiki.apache.org/confluence/x/uJqzDw
[2] https://lists.apache.org/thread/5cgrft73kgkzkgjozf9zfk0w2oj7rjym
[3]
https://nightlies.apache.org/flink/flink-docs-release-1.18/docs/deployment/config/#restart-strategy-type
[4]
https://nightlies.apache.org/flink/flink-docs-master/docs/ops/state/task_failure_recovery/#exponential-delay-restart-strategy

Best,
Rui


[SUMMARY] Flink 1.19 Release Sync 11/14/2023

2023-11-15 文章 Lincoln Lee
Hi devs and users,

Yesterday was the first release sync of Flink 1.19, I’d like to share the
summary:

- Sync meeting
We switched back to google meet because there's some account limitation for
zoom on some region and the google meet is available when creator is not
online.
The meeting will happen every 2 weeks and switch to weekly after the
feature freeze.

- Feature freezing date
Jan 26, 2024

- Features & issues tracking
The community has collected many features on the 1.19 wiki page[1] and it
is encouraged to continuously updating the page for contributors, also
there exists large amounts of jira issues[2].
Please be aware that, for all `@Public` APIs that are intended to be
changed / removed in release 2.0, the deprecation work should be completed
in 1.19.
Another important thing is that since a lot of the work in 1.19 is also
related to the 2.0 release, tagging related issues with '2.0-related' tag
will make it easier for the 2.0 release managers to track progress.

- Daily work divisions
In general, every release manager will be working on all daily issues. For
some major tasks, in order to make sure there will at least always be
someone to take care of them, they have been assigned to specific release
managers[1]. If you need support in each of these areas, please don't
hesitate to contact us.

- Blockers
  - FLINK-31449 Remove DeclarativeSlotManager related logic @Xintong will
track it
  - FLINK-33531 Nightly Python fails @Dian Fu will look at this
  - FLINK-18356 flink-table-planner Exit code 137 on ci pipeline @Matthias
pr reviewing

- Retrospective of 1.18 release
Thanks for the efforts from previous release managers and also several
valuable thoughts and suggestions:
  - The release process now has a jira template, which will make the work
easier for the new release managers, and the overall steps will still
documented on the wiki page and continuously updated in the next releases.
We'll also be looking at automation to continue to streamline releases.
  - 1.18 experienced relatively long release testing, We found that finding
volunteers to join the testing after rc is ready can be a long wait. So in
1.19 we will try to find volunteers earlier(we added a new column:
volunteers for testing on the wiki page[1]), and before release testing,
let the feature developers describe the detailed testing steps, so that
subsequent testing can go faster.
  - The documentation build and flink-docker CI have been migrated to
GHA(Github actions), there's still a lot of work to be done to migrate the
CI pipeline from azure to GHA[3], and welcome to join in for our goal of
improving the experience of our contributors!

The next release sync will be on November 28th, 2023.

Google Meet: https://meet.google.com/vcx-arzs-trv

[1] https://cwiki.apache.org/confluence/display/FLINK/1.19+Release
[2] https://issues.apache.org/jira/secure/RapidBoard.jspa?rapidView=592
[3] https://issues.apache.org/jira/browse/FLINK-27075

Best regards,
Yun, Jing, Martijn and Lincoln


Re:Flink sql 1.17.1 字段类型 DECIMAL(10, 0) 无法执行sql

2023-11-14 文章 Xuyang
Hi, 你的图挂了,可以贴一下图床链接或者直接贴一下代码。




--

Best!
Xuyang




在 2023-11-15 09:39:22,"刘聪聪"  写道:

Flink 1.17.1 遇到  DECIMAL(10, 0)类型字段,直接无法运行,我用强转都不行,还是报数组越界,去除 DECIMAL(10, 
0)类型字段,sql运行都正常。













Re: 关于Apache Flink源码贡献流程

2023-11-14 文章 shi peng7
hi,
你可以发给dev那个邮件列表咨询,这个邮件列表是用户交流用的开发者没那个多

Best regards,
k...@jiayeli.cn

tanjialiang  于 4月24日 下午4:06写道:

Hello,everyone.
我想向apache 
flink贡献源码,由于修复这个issue需要新增一些API,按照流程需要发起邮件讨论,但这个topic只得到一名开发者关注,这样的情况下我应该如何进行后面的流程?期待有熟悉flink源码贡献的开发者可以提供帮助


issue: https://issues.apache.org/jira/browse/FLINK-31686
discuss邮件标题: EncodingFormat and DecondingFormat provide copy API


Best regrads
tanjialiang.


Flink sql 1.17.1 字段类型 DECIMAL(10, 0) 无法执行sql

2023-11-14 文章 刘聪聪
Flink 1.17.1 遇到  DECIMAL(10, 0)类型字段,直接无法运行,我用强转都不行,还是报数组越界,去除 DECIMAL(10, 
0)类型字段,sql运行都正常。













Re: Canal-json格式下乱码导致结果不符合预期

2023-11-13 文章 Feng Jin
hi

这个看起来不像是乱码造成的。

你可以尝试加上去重,还原出正确的CDC 再看下结果。

具体步骤如下:
1. 给 source 设置主键
2. table config 中设置 table.exec.source.cdc-events-duplicate 参数为 true
或者 set 'table.exec.source.cdc-events-duplicate'='true'

Best,
Feng



On Mon, Nov 13, 2023 at 4:09 PM yawning  wrote:

> mysql里面字段:
>
> `encrypted_xx` blob
>
> Canal-json
> "encrypted_xx":"\u0003üUãcA\u0018\u001A}àh\u0013\u001F æÉ"
>
>
>
> 乱码中会有}]这种特殊符号
>
>
> 普通查询:
> select * from tbl
>
> 结果符合预期:
>
> -U[273426307, xxx, u°àÈ;óX«V, üUãcA}àh æÉ, 1699359473865, 2]
> +U[273426307, xxx,
> u°àÈ;óX«V, üUãcA}àh æÉ, 1699359473865, 2]
> -U[399385197, yyy, nì¡… ^³Ø )½ÍU¸, ÉîA6³gŽŽMõýNêfŠ], 1697648682595, 0]
> +U[399385197, yyy, nì¡… ^³Ø )½ÍU¸, ÉîA6³gŽŽMõýNêfŠ], 1699359694026, 0]
> -U[399385197, yyy, nì¡… ^³Ø )½ÍU¸, ÉîA6³gŽŽMõýNêfŠ], 1699359694026, 0]
> +U[399385197, yyy, nì¡… ^³Ø )½ÍU¸, ÉîA6³gŽŽMõýNêfŠ], 1699359694242, 0]
>
> 聚合查询:
> select count(*) from tbl
>
>
> 结果错误
> +I[1]-D[1]
> +I[1]
> -D[1]
> +I[1]


Re: Re: Re: 关于 flink connect jar release 使用问题

2023-11-09 文章 zhhui yan
感谢

Xuyang  于2023年11月10日周五 15:19写道:

> 另外,我确认了一下,你贴的kafka connector官方文档下载链接和maven依赖确实都是有问题的,社区已经有issue了[1]。
> 具体connector版本和flink版本对应关系就参考链接[2]吧,实际上kafka connector针对flink
> 1.18的jar包已经有了,对应的版本是maven仓库[3]里的‘3.0.1-1.18’。
>
>
> [1] https://issues.apache.org/jira/browse/FLINK-33512
> [2] https://flink.apache.org/downloads/
> [3]
> https://repo.maven.apache.org/maven2/org/apache/flink/flink-sql-connector-kafka/
>
>
> --
> Best!
> Xuyang
>
>
>
> 在 2023-11-10 14:43:27,"zhhui yan"  写道:
> >谢谢了,这些文档我都看到了,看起来 要使用1.18 还需要再等等了
> >
> >Xuyang  于2023年11月10日周五 14:32写道:
> >
> >> Hi,
> >>   可以关注下这个讨论[1],1.18的connector还没有release出来。
> >>   在flink 1.17的时候,flink connector基本上都从主仓库移出去了,参考kafka connector [1]。
> >>   connector的下载和兼容的flink版本可以看下这个界面[3]。
> >>
> >>
> >>
> >>
> >> [1] https://lists.apache.org/thread/r31f988m57rtjy4s75030pzwrlqybpq2
> >>
> >> [2] https://issues.apache.org/jira/browse/FLINK-30859
> >> [3] https://flink.apache.org/downloads/
> >>
> >> --
> >>
> >> Best!
> >> Xuyang
> >>
> >>
> >>
> >>
> >> 在 2023-11-10 11:52:18,"zhhui yan"  写道:
> >>
> >> 所有的指向1.18 的 几乎都是,另外是 对应的 connector 以后是不是不要依赖 flink的具体版本了?,不然这个升级难道老大了
> >> 文档:
> >> https://nightlies.apache.org/flink/flink-docs-release-1.18/docs/connectors/datastream/kafka/
> >>
> >>
> >>
> >>
> >>
> >>
> >> Xuyang  于2023年11月10日周五 11:13写道:
> >>
> >> Hi, 可以贴一下release文档的地址吗?应该不需要自己编译才对。
> >>
> >>
> >>
> >>
> >> --
> >>
> >> Best!
> >> Xuyang
> >>
> >>
> >>
> >>
> >>
> >> 在 2023-11-09 15:46:06,"zhhui yan"  写道:
> >> >我们这边想测试一下 升级flink 到1.18 但是我们的release 文档里面给的 connect 包都是404 ,这个是用1.17
> >> >的包,还是需要自己编译?
> >>
> >>
> >>
> >>
> >>
> >> --
> >>
> >> All with you!
> >>
> >>From:  zhhuiyan
> >>
> >>  E-Mail:yzh...@gmail.com
> >>
> >>  QQ:451722401
> >>
> >>  Phone:13146724775
> >>
> >> 公司:九瑞网络科技有限公司
> >
> >
> >
> >--
> >All with you!
> >
> > From:  zhhuiyan
> >
> > E-Mail:yzh...@gmail.com
> >
> >   QQ:451722401
> >
> > Phone:13146724775
> >
> >  公司:九瑞网络科技有限公司
>
>

-- 
best with you!
zhhuiyan


Re:Re: Re: 关于 flink connect jar release 使用问题

2023-11-09 文章 Xuyang
另外,我确认了一下,你贴的kafka connector官方文档下载链接和maven依赖确实都是有问题的,社区已经有issue了[1]。
具体connector版本和flink版本对应关系就参考链接[2]吧,实际上kafka connector针对flink 
1.18的jar包已经有了,对应的版本是maven仓库[3]里的‘3.0.1-1.18’。





[1] https://issues.apache.org/jira/browse/FLINK-33512

[2] https://flink.apache.org/downloads/
[3] 
https://repo.maven.apache.org/maven2/org/apache/flink/flink-sql-connector-kafka/




--

Best!
Xuyang





在 2023-11-10 14:43:27,"zhhui yan"  写道:
>谢谢了,这些文档我都看到了,看起来 要使用1.18 还需要再等等了
>
>Xuyang  于2023年11月10日周五 14:32写道:
>
>> Hi,
>>   可以关注下这个讨论[1],1.18的connector还没有release出来。
>>   在flink 1.17的时候,flink connector基本上都从主仓库移出去了,参考kafka connector [1]。
>>   connector的下载和兼容的flink版本可以看下这个界面[3]。
>>
>>
>>
>>
>> [1] https://lists.apache.org/thread/r31f988m57rtjy4s75030pzwrlqybpq2
>>
>> [2] https://issues.apache.org/jira/browse/FLINK-30859
>> [3] https://flink.apache.org/downloads/
>>
>> --
>>
>> Best!
>> Xuyang
>>
>>
>>
>>
>> 在 2023-11-10 11:52:18,"zhhui yan"  写道:
>>
>> 所有的指向1.18 的 几乎都是,另外是 对应的 connector 以后是不是不要依赖 flink的具体版本了?,不然这个升级难道老大了
>> 文档:
>> https://nightlies.apache.org/flink/flink-docs-release-1.18/docs/connectors/datastream/kafka/
>>
>>
>>
>>
>>
>>
>> Xuyang  于2023年11月10日周五 11:13写道:
>>
>> Hi, 可以贴一下release文档的地址吗?应该不需要自己编译才对。
>>
>>
>>
>>
>> --
>>
>> Best!
>> Xuyang
>>
>>
>>
>>
>>
>> 在 2023-11-09 15:46:06,"zhhui yan"  写道:
>> >我们这边想测试一下 升级flink 到1.18 但是我们的release 文档里面给的 connect 包都是404 ,这个是用1.17
>> >的包,还是需要自己编译?
>>
>>
>>
>>
>>
>> --
>>
>> All with you!
>>
>>From:  zhhuiyan
>>
>>  E-Mail:yzh...@gmail.com
>>
>>  QQ:451722401
>>
>>  Phone:13146724775
>>
>> 公司:九瑞网络科技有限公司
>
>
>
>-- 
>All with you!
>
> From:  zhhuiyan
>
> E-Mail:yzh...@gmail.com
>
>   QQ:451722401
>
> Phone:13146724775
>
>  公司:九瑞网络科技有限公司


Re: Re: 关于 flink connect jar release 使用问题

2023-11-09 文章 zhhui yan
谢谢了,这些文档我都看到了,看起来 要使用1.18 还需要再等等了

Xuyang  于2023年11月10日周五 14:32写道:

> Hi,
>   可以关注下这个讨论[1],1.18的connector还没有release出来。
>   在flink 1.17的时候,flink connector基本上都从主仓库移出去了,参考kafka connector [1]。
>   connector的下载和兼容的flink版本可以看下这个界面[3]。
>
>
>
>
> [1] https://lists.apache.org/thread/r31f988m57rtjy4s75030pzwrlqybpq2
>
> [2] https://issues.apache.org/jira/browse/FLINK-30859
> [3] https://flink.apache.org/downloads/
>
> --
>
> Best!
> Xuyang
>
>
>
>
> 在 2023-11-10 11:52:18,"zhhui yan"  写道:
>
> 所有的指向1.18 的 几乎都是,另外是 对应的 connector 以后是不是不要依赖 flink的具体版本了?,不然这个升级难道老大了
> 文档:
> https://nightlies.apache.org/flink/flink-docs-release-1.18/docs/connectors/datastream/kafka/
>
>
>
>
>
>
> Xuyang  于2023年11月10日周五 11:13写道:
>
> Hi, 可以贴一下release文档的地址吗?应该不需要自己编译才对。
>
>
>
>
> --
>
> Best!
> Xuyang
>
>
>
>
>
> 在 2023-11-09 15:46:06,"zhhui yan"  写道:
> >我们这边想测试一下 升级flink 到1.18 但是我们的release 文档里面给的 connect 包都是404 ,这个是用1.17
> >的包,还是需要自己编译?
>
>
>
>
>
> --
>
> All with you!
>
>From:  zhhuiyan
>
>  E-Mail:yzh...@gmail.com
>
>  QQ:451722401
>
>  Phone:13146724775
>
> 公司:九瑞网络科技有限公司



-- 
All with you!

 From:  zhhuiyan

 E-Mail:yzh...@gmail.com

   QQ:451722401

 Phone:13146724775

  公司:九瑞网络科技有限公司


Re:Re: 关于 flink connect jar release 使用问题

2023-11-09 文章 Xuyang
Hi, 
  可以关注下这个讨论[1],1.18的connector还没有release出来。
  在flink 1.17的时候,flink connector基本上都从主仓库移出去了,参考kafka connector [1]。
  connector的下载和兼容的flink版本可以看下这个界面[3]。




[1] https://lists.apache.org/thread/r31f988m57rtjy4s75030pzwrlqybpq2

[2] https://issues.apache.org/jira/browse/FLINK-30859
[3] https://flink.apache.org/downloads/

--

Best!
Xuyang




在 2023-11-10 11:52:18,"zhhui yan"  写道:

所有的指向1.18 的 几乎都是,另外是 对应的 connector 以后是不是不要依赖 flink的具体版本了?,不然这个升级难道老大了
文档: 
https://nightlies.apache.org/flink/flink-docs-release-1.18/docs/connectors/datastream/kafka/
 






Xuyang  于2023年11月10日周五 11:13写道:

Hi, 可以贴一下release文档的地址吗?应该不需要自己编译才对。




--

Best!
Xuyang





在 2023-11-09 15:46:06,"zhhui yan"  写道:
>我们这边想测试一下 升级flink 到1.18 但是我们的release 文档里面给的 connect 包都是404 ,这个是用1.17
>的包,还是需要自己编译?





--

All with you!
 
From:  zhhuiyan
   
E-Mail:yzh...@gmail.com
   
QQ:451722401
   
Phone:13146724775
  
公司:九瑞网络科技有限公司

Re: 关于 flink connect jar release 使用问题

2023-11-09 文章 zhhui yan
所有的指向1.18 的 几乎都是,另外是 对应的 connector 以后是不是不要依赖 flink的具体版本了?,不然这个升级难道老大了
文档:
https://nightlies.apache.org/flink/flink-docs-release-1.18/docs/connectors/datastream/kafka/

[image: image.png][image: image.png]


Xuyang  于2023年11月10日周五 11:13写道:

> Hi, 可以贴一下release文档的地址吗?应该不需要自己编译才对。
>
>
>
>
> --
>
> Best!
> Xuyang
>
>
>
>
>
> 在 2023-11-09 15:46:06,"zhhui yan"  写道:
> >我们这边想测试一下 升级flink 到1.18 但是我们的release 文档里面给的 connect 包都是404 ,这个是用1.17
> >的包,还是需要自己编译?
>


-- 
All with you!

 From:  zhhuiyan

 E-Mail:yzh...@gmail.com

   QQ:451722401

 Phone:13146724775

  公司:九瑞网络科技有限公司


Re:关于 flink connect jar release 使用问题

2023-11-09 文章 Xuyang
Hi, 可以贴一下release文档的地址吗?应该不需要自己编译才对。




--

Best!
Xuyang





在 2023-11-09 15:46:06,"zhhui yan"  写道:
>我们这边想测试一下 升级flink 到1.18 但是我们的release 文档里面给的 connect 包都是404 ,这个是用1.17
>的包,还是需要自己编译?


关于 flink connect jar release 使用问题

2023-11-08 文章 zhhui yan
我们这边想测试一下 升级flink 到1.18 但是我们的release 文档里面给的 connect 包都是404 ,这个是用1.17
的包,还是需要自己编译?


Re:FLINK-33365 - Missing filter condition in execution plan containing lookup join with mysql jdbc connector

2023-11-07 文章 Xuyang
Hi, 
看了下发现这个jira下面已经有人在尝试复现但是没有成功。
如果可以的话,可以在jira下面留言回复一起多提供一些可以复现的case,帮助assigner复现这个问题,从而更快的定位+修复。




--

Best!
Xuyang





在 2023-11-07 15:59:53,"casel.chen"  写道:
>这个critical issue有人fix吗?我们线上使用flink 1.17.1版本有使用jdbc维表查询on带and过滤条件,发现and过滤条件不起作用
>
>
>例如
>select xxx from a left join b on a.id = b.id and b.type = 'xxx'
>发现b.type='xxx'这个过滤条件不起作用


FLINK-33365 - Missing filter condition in execution plan containing lookup join with mysql jdbc connector

2023-11-07 文章 casel.chen
这个critical issue有人fix吗?我们线上使用flink 1.17.1版本有使用jdbc维表查询on带and过滤条件,发现and过滤条件不起作用


例如
select xxx from a left join b on a.id = b.id and b.type = 'xxx'
发现b.type='xxx'这个过滤条件不起作用

Re: 退订

2023-11-06 文章 Yunfeng Zhou
Hi,

请发送任意内容的邮件到 user-zh-unsubscr...@flink.apache.org 来取消订阅邮件。

Best
Yunfeng Zhou

On Mon, Nov 6, 2023 at 5:30 PM maozhaolin  wrote:
>
> 退订


退订

2023-11-06 文章 maozhaolin
退订

Flink-1.15版本

2023-11-04 文章 Ray
各位专家:当前遇到如下问题1、场景:在使用Yarn场景下提交flink任务2、版本:Flink1.15.03、日志:查看yarn上的日志发下,版本上的问题2023-11-04
 15:04:42,313 ERROR org.apache.flink.util.FatalExitExceptionHandler 
 [] - FATAL: Thread 'flink-akka.actor.internal-dispatcher-3' produced an 
uncaught exception. Stopping the process...java.lang.NoClassDefFoundError: 
akka/actor/dungeon/FaultHandling$$anonfun$handleNonFatalOrInterruptedException$1
at 
akka.actor.dungeon.FaultHandling.handleNonFatalOrInterruptedException(FaultHandling.scala:334)
 ~[flink-rpc-akka_abb25197-9eee-499b-8c1f-f52d0afc4374.jar:1.15.0]
at 
akka.actor.dungeon.FaultHandling.handleNonFatalOrInterruptedException$(FaultHandling.scala:334)
 ~[flink-rpc-akka_abb25197-9eee-499b-8c1f-f52d0afc4374.jar:1.15.0]
at 
akka.actor.ActorCell.handleNonFatalOrInterruptedException(ActorCell.scala:411) 
~[flink-rpc-akka_abb25197-9eee-499b-8c1f-f52d0afc4374.jar:1.15.0]
at akka.actor.ActorCell.invoke(ActorCell.scala:551) 
~[flink-rpc-akka_abb25197-9eee-499b-8c1f-f52d0afc4374.jar:1.15.0]
at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:270) 
~[flink-rpc-akka_abb25197-9eee-499b-8c1f-f52d0afc4374.jar:1.15.0]
at akka.dispatch.Mailbox.run(Mailbox.scala:231) 
~[flink-rpc-akka_abb25197-9eee-499b-8c1f-f52d0afc4374.jar:1.15.0]
at akka.dispatch.Mailbox.exec(Mailbox.scala:243) 
[flink-rpc-akka_abb25197-9eee-499b-8c1f-f52d0afc4374.jar:1.15.0]
at java.util.concurrent.ForkJoinTask.doExec(ForkJoinTask.java:289) 
[?:1.8.0_181]
at 
java.util.concurrent.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1056) 
[?:1.8.0_181]
at java.util.concurrent.ForkJoinPool.runWorker(ForkJoinPool.java:1692) 
[?:1.8.0_181]
at 
java.util.concurrent.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:157) 
[?:1.8.0_181]
Caused by: java.lang.ClassNotFoundException: 
akka.actor.dungeon.FaultHandling$$anonfun$handleNonFatalOrInterruptedException$1
at java.net.URLClassLoader.findClass(URLClassLoader.java:381) 
~[?:1.8.0_181]
at java.lang.ClassLoader.loadClass(ClassLoader.java:424) ~[?:1.8.0_181]
at 
org.apache.flink.core.classloading.ComponentClassLoader.loadClassFromComponentOnly(ComponentClassLoader.java:149)
 ~[flink-dist-1.15.0.jar:1.15.0]
at 
org.apache.flink.core.classloading.ComponentClassLoader.loadClass(ComponentClassLoader.java:112)
 ~[flink-dist-1.15.0.jar:1.15.0]
at java.lang.ClassLoader.loadClass(ClassLoader.java:357) ~[?:1.8.0_181]
... 11 more
2023-11-04 15:04:42,324 ERROR 
org.apache.flink.runtime.util.ClusterUncaughtExceptionHandler [] - WARNING: 
Thread 'flink-shutdown-hook-1' produced an uncaught exception. If you want to 
fail on uncaught exceptions, then configure cluster.uncaught-exception-handling 
accordingly
java.lang.NoClassDefFoundError: 
scala/collection/convert/Wrappers$MutableSetWrapper
at 
scala.collection.convert.AsScalaConverters.asScalaSet(AsScalaConverters.scala:126)
 ~[flink-rpc-akka_abb25197-9eee-499b-8c1f-f52d0afc4374.jar:1.15.0]
at 
scala.collection.convert.AsScalaConverters.asScalaSet$(AsScalaConverters.scala:124)
 ~[flink-rpc-akka_abb25197-9eee-499b-8c1f-f52d0afc4374.jar:1.15.0]
at 
akka.util.ccompat.package$JavaConverters$.asScalaSet(package.scala:86) 
~[flink-rpc-akka_abb25197-9eee-499b-8c1f-f52d0afc4374.jar:1.15.0]
at 
scala.collection.convert.DecorateAsScala.$anonfun$asScalaSetConverter$1(DecorateAsScala.scala:59)
 ~[flink-rpc-akka_abb25197-9eee-499b-8c1f-f52d0afc4374.jar:1.15.0]
at 
scala.collection.convert.Decorators$AsScala.asScala(Decorators.scala:25) 
~[flink-rpc-akka_abb25197-9eee-499b-8c1f-f52d0afc4374.jar:1.15.0]
at 
akka.actor.CoordinatedShutdown$tasks$.totalDuration(CoordinatedShutdown.scala:481)
 ~[flink-rpc-akka_abb25197-9eee-499b-8c1f-f52d0afc4374.jar:1.15.0]
at 
akka.actor.CoordinatedShutdown.totalTimeout(CoordinatedShutdown.scala:784) 
~[flink-rpc-akka_abb25197-9eee-499b-8c1f-f52d0afc4374.jar:1.15.0]
at 
akka.actor.CoordinatedShutdown$.$anonfun$initJvmHook$1(CoordinatedShutdown.scala:271)
 ~[flink-rpc-akka_abb25197-9eee-499b-8c1f-f52d0afc4374.jar:1.15.0]
at 
akka.actor.CoordinatedShutdown$$anon$3.run(CoordinatedShutdown.scala:814) 
~[flink-rpc-akka_abb25197-9eee-499b-8c1f-f52d0afc4374.jar:1.15.0]
Caused by: java.lang.ClassNotFoundException: 
scala.collection.convert.Wrappers$MutableSetWrapper
at java.net.URLClassLoader.findClass(URLClassLoader.java:381) 
~[?:1.8.0_181]
at java.lang.ClassLoader.loadClass(ClassLoader.java:424) ~[?:1.8.0_181]
at 
org.apache.flink.core.classloading.ComponentClassLoader.loadClassFromComponentOnly(ComponentClassLoader.java:149)
 ~[flink-dist-1.15.0.jar:1.15.0]
at 
org.apache.flink.core.classloading.ComponentClassLoader.loadClass(ComponentClassLoader.java:112)
 ~[flink-dist-1.15.0.jar:1.15.0]
at java.lang.ClassLoader.loadClass(ClassLoader.java:357) ~[?:1.8.0_181]
... 9 more

4、查看Issue上,[FLINK-25998] Flink 

Re:疑似BUG: 在滑动窗口中使用reduce()聚合时数据被多次处理

2023-11-03 文章 Xuyang
Hi,
   验证了下,问题疑似出现在reduce函数中,复用了下wordCount1这个对象。我试了下new一个新的WordCount作为输出应该就可以了。
猜测这可能和基于Heap的state backend有关,多个窗口的heap state可能直接使用的是一个对象的地址。


```
.reduce(
(wordCount1, wordCount2) -> {
WordCount newWC =
new WordCount(
wordCount1.word, wordCount1.count + wordCount2.count);
System.err.println(newWC);
return newWC;
})
```

--

Best!
Xuyang





在 2023-11-03 10:53:37,"tao zhang"  写道:
>reduce()方法的状态在窗口间未被隔离,多个窗口聚合时使用的是同一对象.一个数据进入时,被重复累加
>是reduce的特性吗? 还是reduce中的窗口间隔离出现问题? 希望得到回复
>
>测试输入如下:
>1001,/home,1000
>1002,/home,2000
>
>输出如下:
>input> test.Event(user=1001, page=/home, ts=1000)
>input> test.Event(user=1002, page=/home, ts=2000)
>test.WordCount(word=/home, count=2)
>test.WordCount(word=/home, count=3)
>
>代码如下:
>
>import lombok.AllArgsConstructor;
>import lombok.Data;
>import lombok.NoArgsConstructor;
>import org.apache.flink.api.common.eventtime.WatermarkStrategy;
>import org.apache.flink.streaming.api.datastream.SingleOutputStreamOperator;
>import org.apache.flink.streaming.api.environment.StreamExecutionEnvironment;
>import 
>org.apache.flink.streaming.api.windowing.assigners.SlidingEventTimeWindows;
>import org.apache.flink.streaming.api.windowing.time.Time;
>import java.io.Serializable;
>import java.time.Duration;
>
>public class test {
>public static void main(String[] args) {
>//准备环境
>StreamExecutionEnvironment env = 
> StreamExecutionEnvironment.getExecutionEnvironment();
>env.setParallelism(1);
>
>//从端口读数据
>SingleOutputStreamOperator ds1 = 
> env.socketTextStream("hadoop102", 5).map(
>value->{
>String[] strings = value.split(",");
>return new 
> Event(strings[0].trim(),strings[1].trim(),Long.valueOf(strings[2].trim()) );
>}
>
>).assignTimestampsAndWatermarks(
>//增加水位线策略
>
> WatermarkStrategy.forBoundedOutOfOrderness(Duration.ZERO).withTimestampAssigner((Event,
>  l) -> Event.getTs())
>);
>//检查输入流
>ds1.print("input");
>
>
>ds1.map(event -> new WordCount(event.getPage(), 1)
>).keyBy(WordCount::getWord
>//按键分组
>).window(
>//TumblingEventTimeWindows.of(Time.seconds(10))
>SlidingEventTimeWindows.of(Time.seconds(10), Time.seconds(5))
>//size为10步长为5的滑动窗口
>).reduce(
>//先增量聚合.将多个数据处理为一个中间结果
>
>(wordCount1, wordCount2) -> {
>
>Integer count = wordCount1.getCount();
>
>wordCount1.setCount(count + 1);
>
>System.out.println(wordCount1);
>
>return wordCount1;
>}
>
>
>);
>
>try {
>env.execute();
>} catch (Exception e) {
>throw new RuntimeException(e);
>}
>}
>
>@Data
>@AllArgsConstructor
>@NoArgsConstructor
>public static class Event {
>private String user;
>private String page;
>private Long ts;
>
>}
>
>@Data
>@AllArgsConstructor
>@NoArgsConstructor
>
>public static class WordCount implements Serializable {
>private String word;
>private Integer count;
>
>}
>
>
>
>}
>


疑似BUG: 在滑动窗口中使用reduce()聚合时数据被多次处理

2023-11-02 文章 tao zhang
reduce()方法的状态在窗口间未被隔离,多个窗口聚合时使用的是同一对象.一个数据进入时,被重复累加
是reduce的特性吗? 还是reduce中的窗口间隔离出现问题? 希望得到回复

测试输入如下:
1001,/home,1000
1002,/home,2000

输出如下:
input> test.Event(user=1001, page=/home, ts=1000)
input> test.Event(user=1002, page=/home, ts=2000)
test.WordCount(word=/home, count=2)
test.WordCount(word=/home, count=3)

代码如下:

import lombok.AllArgsConstructor;
import lombok.Data;
import lombok.NoArgsConstructor;
import org.apache.flink.api.common.eventtime.WatermarkStrategy;
import org.apache.flink.streaming.api.datastream.SingleOutputStreamOperator;
import org.apache.flink.streaming.api.environment.StreamExecutionEnvironment;
import 
org.apache.flink.streaming.api.windowing.assigners.SlidingEventTimeWindows;
import org.apache.flink.streaming.api.windowing.time.Time;
import java.io.Serializable;
import java.time.Duration;

public class test {
public static void main(String[] args) {
//准备环境
StreamExecutionEnvironment env = 
StreamExecutionEnvironment.getExecutionEnvironment();
env.setParallelism(1);

//从端口读数据
SingleOutputStreamOperator ds1 = 
env.socketTextStream("hadoop102", 5).map(
value->{
String[] strings = value.split(",");
return new 
Event(strings[0].trim(),strings[1].trim(),Long.valueOf(strings[2].trim()) );
}

).assignTimestampsAndWatermarks(
//增加水位线策略

WatermarkStrategy.forBoundedOutOfOrderness(Duration.ZERO).withTimestampAssigner((Event,
 l) -> Event.getTs())
);
//检查输入流
ds1.print("input");


ds1.map(event -> new WordCount(event.getPage(), 1)
).keyBy(WordCount::getWord
//按键分组
).window(
//TumblingEventTimeWindows.of(Time.seconds(10))
SlidingEventTimeWindows.of(Time.seconds(10), Time.seconds(5))
//size为10步长为5的滑动窗口
).reduce(
//先增量聚合.将多个数据处理为一个中间结果

(wordCount1, wordCount2) -> {

Integer count = wordCount1.getCount();

wordCount1.setCount(count + 1);

System.out.println(wordCount1);

return wordCount1;
}


);

try {
env.execute();
} catch (Exception e) {
throw new RuntimeException(e);
}
}

@Data
@AllArgsConstructor
@NoArgsConstructor
public static class Event {
private String user;
private String page;
private Long ts;

}

@Data
@AllArgsConstructor
@NoArgsConstructor

public static class WordCount implements Serializable {
private String word;
private Integer count;

}



}



Re:flink的sql gateway支持自定义的UDF吗?

2023-11-01 文章 Xuyang
Hi, 
你指的是sql gateway上 ADD JAR这种方式来上传自定义UDF的jar包[1]么?




[1] 
https://nightlies.apache.org/flink/flink-docs-release-1.17/docs/dev/table/sql/jar/

--

Best!
Xuyang





在 2023-11-01 14:21:04,"RS"  写道:
>Hi
>flink的sql gateway支持自定义的UDF吗?包括java和python的,有示例可以参考下吗?


flink的sql gateway支持自定义的UDF吗?

2023-11-01 文章 RS
Hi
flink的sql gateway支持自定义的UDF吗?包括java和python的,有示例可以参考下吗?

Re:flink 1.18.0的sql gateway支持提交作业到k8s上运行吗?

2023-11-01 文章 RS
Hi,
提交到本地是flink配置文件里面配置的jobmanager的地址,所以肯定也是提交到K8S的吧
yarn的不太清楚。





在 2023-10-30 14:36:23,"casel.chen"  写道:
>想问一下目前flink 1.18.0的sql gateway只支持提交作业到本地运行吗?能否支持提交作业到yarn或k8s上运行呢?


[ANNOUNCE] Apache Flink Kubernetes Operator 1.6.1 released

2023-10-30 文章 Rui Fan
The Apache Flink community is very happy to announce the release of Apache
Flink Kubernetes Operator 1.6.1.

Please check out the release blog post for an overview of the release:
https://flink.apache.org/2023/10/27/apache-flink-kubernetes-operator-1.6.1-release-announcement/

The release is available for download at:
https://flink.apache.org/downloads.html

Maven artifacts for Flink Kubernetes Operator can be found at:
https://search.maven.org/artifact/org.apache.flink/flink-kubernetes-operator

Official Docker image for Flink Kubernetes Operator applications can be
found at:
https://hub.docker.com/r/apache/flink-kubernetes-operator

The full release notes are available in Jira:
https://issues.apache.org/jira/secure/ReleaseNote.jspa?projectId=12315522=12353784

We would like to thank all contributors of the Apache Flink community who
made this release possible!

Regards,
Rui Fan


Re: Re: flink 1.18.0 使用 hive beeline + jdbc driver连接sql gateway失败

2023-10-30 文章 Benchao Li
hiveserver2 endpoint 就是让 flink gateway 直接变成 hive server2,对外来讲它就是 hive
server2 了,它可以直接跟已有的跟 hive server2 的工具配合一起使用。

但是现在你其实用的是 flink jdbc driver,这个并不是跟 hive server2 交互,它就是跟 flink gateway
交互,所以你用hive server2的模式启动,它就不认识了。

casel.chen  于2023年10月30日周一 14:36写道:
>
> 果然不指定endpoint为hiveserver2类型后使用hive beeline工具连接上了。感谢!
> 不过我仍然有个疑问,看官网文档上有写提供 hiveserver2 endpoint 
> 是为了兼容hive方言,按理也应该可以使用beeline连接上,因为原本beeline支持连接hiveserver2
> 以下是原文:
> HiveServer2 Endpoint is compatible with HiveServer2 wire protocol and allows 
> users to interact (e.g. submit Hive SQL) with Flink SQL Gateway with existing 
> Hive clients, such as Hive JDBC, Beeline, DBeaver, Apache Superset and so on.
> 这里有提到Beeline工具,难道不是 beeline> !connect jdbc:flink://localhost:8083 这样的连接方式了?
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
> 在 2023-10-30 11:27:15,"Benchao Li"  写道:
> >Hi casel,
> >
> >Flink JDBC 链接到 gateway 目前使用的是 flink 的 gateway 接口,所以你在启动 gateway
> >的时候不用指定 endpoint 为 hiveserver2 类型,用 Flink 默认的 gateway endpoint 类型即可。
> >
> >casel.chen  于2023年10月29日周日 17:24写道:
> >>
> >> 1. 启动flink集群
> >> bin/start-cluster.sh
> >>
> >>
> >> 2. 启动sql gateway
> >> bin/sql-gateway.sh start -Dsql-gateway.endpoint.type=hiveserver2
> >>
> >>
> >> 3. 将flink-sql-jdbc-driver-bundle-1.18.0.jar放到apache-hive-3.1.2-bin/lib目录下
> >>
> >>
> >> 4. 到apache-hive-3.1.2-bin目录下启动beeline连接sql gateway,提示输入用户名和密码时直接按的回车
> >> $ bin/beeline
> >> SLF4J: Class path contains multiple SLF4J bindings.
> >> SLF4J: Found binding in 
> >> [jar:file:/Users/admin/dev/hadoop-3.3.4/share/hadoop/common/lib/slf4j-reload4j-1.7.36.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> >> SLF4J: Found binding in 
> >> [jar:file:/Users/admin/dev/apache-hive-3.1.2-bin/lib/log4j-slf4j-impl-2.10.0.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> >> SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an 
> >> explanation.
> >> SLF4J: Actual binding is of type [org.slf4j.impl.Reload4jLoggerFactory]
> >> Beeline version 3.1.2 by Apache Hive
> >> beeline> !connect jdbc:flink://localhost:8083
> >> Connecting to jdbc:flink://localhost:8083
> >> Enter username for jdbc:flink://localhost:8083:
> >> Enter password for jdbc:flink://localhost:8083:
> >> Failed to create the executor.
> >> 0: jdbc:flink://localhost:8083 (closed)> CREATE TABLE T(
> >> . . . . . . . . . . . . . . . . . . . .>   a INT,
> >> . . . . . . . . . . . . . . . . . . . .>   b VARCHAR(10)
> >> . . . . . . . . . . . . . . . . . . . .> ) WITH (
> >> . . . . . . . . . . . . . . . . . . . .>   'connector' = 'filesystem',
> >> . . . . . . . . . . . . . . . . . . . .>   'path' = 'file:///tmp/T.csv',
> >> . . . . . . . . . . . . . . . . . . . .>   'format' = 'csv'
> >> . . . . . . . . . . . . . . . . . . . .> );
> >> Failed to create the executor.
> >> Connection is already closed.
> >>
> >
> >
> >--
> >
> >Best,
> >Benchao Li



-- 

Best,
Benchao Li


Re:Re:Re:回复: flink sql不支持show create catalog 吗?

2023-10-30 文章 casel.chen
谢谢解答,我查了一下目前有两种CatalogStore实现,一个是基于内存的,另一个是基于文件系统的。
请问要如何配置基于文件系统的CatalogStore?这个文件可以在对象存储上吗?flink sql client要如何使用这个CatalogStore? 
谢谢!

















在 2023-10-30 10:28:34,"Xuyang"  写道:
>Hi, CatalogStore 的引入我理解是为了Catalog能被更好地管理、注册和元数据存储,具体motivation可以参考Flip295[1].
>我的理解是倒不是说“引入CatalogStore后才可以提供show create 
>catalog语法支持”,而是之前没有直接存储catalog配置的地方和能力,在CatalogStore之后,天然支持了对catalog配置的存储,因此这个feat就可以直接快速的支持了。
>
>
>
>
>[1] 
>https://cwiki.apache.org/confluence/display/FLINK/FLIP-295%3A+Support+lazy+initialization+of+catalogs+and+persistence+of+catalog+configurations
>
>
>
>
>--
>
>Best!
>Xuyang
>
>
>
>
>
>在 2023-10-29 20:34:52,"casel.chen"  写道:
>>请问flink 1.18引入的CatalogStore是为了解决什么问题呢?为什么引入CatalogStore后才可以提供show create 
>>catalog语法支持?
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>在 2023-10-20 17:03:46,"李宇彬"  写道:
>>>Hi Feng
>>>
>>>
>>>我之前建过一个jira(https://issues.apache.org/jira/browse/FLINK-24939),引入CatalogStore后,实现这个特性的时机应该已经成熟了,我们这边业务场景里用到了很多catalog,管理起来很麻烦,有这个特性会好很多。
>>>| |
>>> 回复的原邮件 
>>>| 发件人 | Feng Jin |
>>>| 发送日期 | 2023年10月20日 13:18 |
>>>| 收件人 |  |
>>>| 主题 | Re: flink sql不支持show create catalog 吗? |
>>>hi casel
>>>
>>>
>>>从 1.18 开始,引入了 CatalogStore,持久化了 Catalog 的配置,确实可以支持 show create catalog 了。
>>>
>>>
>>>Best,
>>>Feng
>>>
>>>On Fri, Oct 20, 2023 at 11:55 AM casel.chen  wrote:
>>>
>>>之前在flink sql中创建过一个catalog,现在想查看当初创建catalog的语句复制并修改一下另存为一个新的catalog,发现flink
>>>sql不支持show create catalog 。
>>>而据我所知doris是支持show create catalog语句的。flink sql能否支持一下呢?


flink 1.18.0的sql gateway支持提交作业到k8s上运行吗?

2023-10-30 文章 casel.chen
想问一下目前flink 1.18.0的sql gateway只支持提交作业到本地运行吗?能否支持提交作业到yarn或k8s上运行呢?

Re:Re: flink 1.18.0 使用 hive beeline + jdbc driver连接sql gateway失败

2023-10-30 文章 casel.chen
果然不指定endpoint为hiveserver2类型后使用hive beeline工具连接上了。感谢!
不过我仍然有个疑问,看官网文档上有写提供 hiveserver2 endpoint 
是为了兼容hive方言,按理也应该可以使用beeline连接上,因为原本beeline支持连接hiveserver2
以下是原文:
HiveServer2 Endpoint is compatible with HiveServer2 wire protocol and allows 
users to interact (e.g. submit Hive SQL) with Flink SQL Gateway with existing 
Hive clients, such as Hive JDBC, Beeline, DBeaver, Apache Superset and so on.
这里有提到Beeline工具,难道不是 beeline> !connect jdbc:flink://localhost:8083 这样的连接方式了?



















在 2023-10-30 11:27:15,"Benchao Li"  写道:
>Hi casel,
>
>Flink JDBC 链接到 gateway 目前使用的是 flink 的 gateway 接口,所以你在启动 gateway
>的时候不用指定 endpoint 为 hiveserver2 类型,用 Flink 默认的 gateway endpoint 类型即可。
>
>casel.chen  于2023年10月29日周日 17:24写道:
>>
>> 1. 启动flink集群
>> bin/start-cluster.sh
>>
>>
>> 2. 启动sql gateway
>> bin/sql-gateway.sh start -Dsql-gateway.endpoint.type=hiveserver2
>>
>>
>> 3. 将flink-sql-jdbc-driver-bundle-1.18.0.jar放到apache-hive-3.1.2-bin/lib目录下
>>
>>
>> 4. 到apache-hive-3.1.2-bin目录下启动beeline连接sql gateway,提示输入用户名和密码时直接按的回车
>> $ bin/beeline
>> SLF4J: Class path contains multiple SLF4J bindings.
>> SLF4J: Found binding in 
>> [jar:file:/Users/admin/dev/hadoop-3.3.4/share/hadoop/common/lib/slf4j-reload4j-1.7.36.jar!/org/slf4j/impl/StaticLoggerBinder.class]
>> SLF4J: Found binding in 
>> [jar:file:/Users/admin/dev/apache-hive-3.1.2-bin/lib/log4j-slf4j-impl-2.10.0.jar!/org/slf4j/impl/StaticLoggerBinder.class]
>> SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an 
>> explanation.
>> SLF4J: Actual binding is of type [org.slf4j.impl.Reload4jLoggerFactory]
>> Beeline version 3.1.2 by Apache Hive
>> beeline> !connect jdbc:flink://localhost:8083
>> Connecting to jdbc:flink://localhost:8083
>> Enter username for jdbc:flink://localhost:8083:
>> Enter password for jdbc:flink://localhost:8083:
>> Failed to create the executor.
>> 0: jdbc:flink://localhost:8083 (closed)> CREATE TABLE T(
>> . . . . . . . . . . . . . . . . . . . .>   a INT,
>> . . . . . . . . . . . . . . . . . . . .>   b VARCHAR(10)
>> . . . . . . . . . . . . . . . . . . . .> ) WITH (
>> . . . . . . . . . . . . . . . . . . . .>   'connector' = 'filesystem',
>> . . . . . . . . . . . . . . . . . . . .>   'path' = 'file:///tmp/T.csv',
>> . . . . . . . . . . . . . . . . . . . .>   'format' = 'csv'
>> . . . . . . . . . . . . . . . . . . . .> );
>> Failed to create the executor.
>> Connection is already closed.
>>
>
>
>-- 
>
>Best,
>Benchao Li


Re: flink sql如何处理脏数据问题?

2023-10-29 文章 ying lin
还有一种做法就是使用datastream,datastream支持sideoutput,但 flink
sql不支持,不过有一种迂回的做法就是flinksql -> datastream -> flink
sql,可以查一下官网资料,flinksql和datastream可以互相转换。

Xuyang  于2023年10月30日周一 10:17写道:

> Flink SQL目前对于脏数据没有类似side output的机制来输出,这个需求用自定义connector应该可以实现。
>
>
>
>
>
>
>
> --
>
> Best!
> Xuyang
>
>
>
>
>
> 在 2023-10-29 10:23:38,"casel.chen"  写道:
> >场景:使用flink
> sql将数据写入下游OLAP系统,如doris,遇到一些异常情况,比如字段值超长或者分区字段值为当前doris表不存在的分区(需要先人为创建)等等,当前写入这些脏数据会使得作业写入报错,进而导致作业失败。我们是希望能够将这些“脏”数据单独发到一个kafka
> topic或者写入一个文件便于事后审查。这个目前有办法做到吗?
>


Re: flink 1.18.0 使用 hive beeline + jdbc driver连接sql gateway失败

2023-10-29 文章 Benchao Li
Hi casel,

Flink JDBC 链接到 gateway 目前使用的是 flink 的 gateway 接口,所以你在启动 gateway
的时候不用指定 endpoint 为 hiveserver2 类型,用 Flink 默认的 gateway endpoint 类型即可。

casel.chen  于2023年10月29日周日 17:24写道:
>
> 1. 启动flink集群
> bin/start-cluster.sh
>
>
> 2. 启动sql gateway
> bin/sql-gateway.sh start -Dsql-gateway.endpoint.type=hiveserver2
>
>
> 3. 将flink-sql-jdbc-driver-bundle-1.18.0.jar放到apache-hive-3.1.2-bin/lib目录下
>
>
> 4. 到apache-hive-3.1.2-bin目录下启动beeline连接sql gateway,提示输入用户名和密码时直接按的回车
> $ bin/beeline
> SLF4J: Class path contains multiple SLF4J bindings.
> SLF4J: Found binding in 
> [jar:file:/Users/admin/dev/hadoop-3.3.4/share/hadoop/common/lib/slf4j-reload4j-1.7.36.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: Found binding in 
> [jar:file:/Users/admin/dev/apache-hive-3.1.2-bin/lib/log4j-slf4j-impl-2.10.0.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an 
> explanation.
> SLF4J: Actual binding is of type [org.slf4j.impl.Reload4jLoggerFactory]
> Beeline version 3.1.2 by Apache Hive
> beeline> !connect jdbc:flink://localhost:8083
> Connecting to jdbc:flink://localhost:8083
> Enter username for jdbc:flink://localhost:8083:
> Enter password for jdbc:flink://localhost:8083:
> Failed to create the executor.
> 0: jdbc:flink://localhost:8083 (closed)> CREATE TABLE T(
> . . . . . . . . . . . . . . . . . . . .>   a INT,
> . . . . . . . . . . . . . . . . . . . .>   b VARCHAR(10)
> . . . . . . . . . . . . . . . . . . . .> ) WITH (
> . . . . . . . . . . . . . . . . . . . .>   'connector' = 'filesystem',
> . . . . . . . . . . . . . . . . . . . .>   'path' = 'file:///tmp/T.csv',
> . . . . . . . . . . . . . . . . . . . .>   'format' = 'csv'
> . . . . . . . . . . . . . . . . . . . .> );
> Failed to create the executor.
> Connection is already closed.
>


-- 

Best,
Benchao Li


Re:Re:回复: flink sql不支持show create catalog 吗?

2023-10-29 文章 Xuyang
Hi, CatalogStore 的引入我理解是为了Catalog能被更好地管理、注册和元数据存储,具体motivation可以参考Flip295[1].
我的理解是倒不是说“引入CatalogStore后才可以提供show create 
catalog语法支持”,而是之前没有直接存储catalog配置的地方和能力,在CatalogStore之后,天然支持了对catalog配置的存储,因此这个feat就可以直接快速的支持了。




[1] 
https://cwiki.apache.org/confluence/display/FLINK/FLIP-295%3A+Support+lazy+initialization+of+catalogs+and+persistence+of+catalog+configurations




--

Best!
Xuyang





在 2023-10-29 20:34:52,"casel.chen"  写道:
>请问flink 1.18引入的CatalogStore是为了解决什么问题呢?为什么引入CatalogStore后才可以提供show create 
>catalog语法支持?
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>在 2023-10-20 17:03:46,"李宇彬"  写道:
>>Hi Feng
>>
>>
>>我之前建过一个jira(https://issues.apache.org/jira/browse/FLINK-24939),引入CatalogStore后,实现这个特性的时机应该已经成熟了,我们这边业务场景里用到了很多catalog,管理起来很麻烦,有这个特性会好很多。
>>| |
>> 回复的原邮件 
>>| 发件人 | Feng Jin |
>>| 发送日期 | 2023年10月20日 13:18 |
>>| 收件人 |  |
>>| 主题 | Re: flink sql不支持show create catalog 吗? |
>>hi casel
>>
>>
>>从 1.18 开始,引入了 CatalogStore,持久化了 Catalog 的配置,确实可以支持 show create catalog 了。
>>
>>
>>Best,
>>Feng
>>
>>On Fri, Oct 20, 2023 at 11:55 AM casel.chen  wrote:
>>
>>之前在flink sql中创建过一个catalog,现在想查看当初创建catalog的语句复制并修改一下另存为一个新的catalog,发现flink
>>sql不支持show create catalog 。
>>而据我所知doris是支持show create catalog语句的。flink sql能否支持一下呢?


Re:flink sql如何处理脏数据问题?

2023-10-29 文章 Xuyang
Flink SQL目前对于脏数据没有类似side output的机制来输出,这个需求用自定义connector应该可以实现。







--

Best!
Xuyang





在 2023-10-29 10:23:38,"casel.chen"  写道:
>场景:使用flink 
>sql将数据写入下游OLAP系统,如doris,遇到一些异常情况,比如字段值超长或者分区字段值为当前doris表不存在的分区(需要先人为创建)等等,当前写入这些脏数据会使得作业写入报错,进而导致作业失败。我们是希望能够将这些“脏”数据单独发到一个kafka
> topic或者写入一个文件便于事后审查。这个目前有办法做到吗?


Re:回复: flink sql不支持show create catalog 吗?

2023-10-29 文章 casel.chen
请问flink 1.18引入的CatalogStore是为了解决什么问题呢?为什么引入CatalogStore后才可以提供show create 
catalog语法支持?

















在 2023-10-20 17:03:46,"李宇彬"  写道:
>Hi Feng
>
>
>我之前建过一个jira(https://issues.apache.org/jira/browse/FLINK-24939),引入CatalogStore后,实现这个特性的时机应该已经成熟了,我们这边业务场景里用到了很多catalog,管理起来很麻烦,有这个特性会好很多。
>| |
> 回复的原邮件 
>| 发件人 | Feng Jin |
>| 发送日期 | 2023年10月20日 13:18 |
>| 收件人 |  |
>| 主题 | Re: flink sql不支持show create catalog 吗? |
>hi casel
>
>
>从 1.18 开始,引入了 CatalogStore,持久化了 Catalog 的配置,确实可以支持 show create catalog 了。
>
>
>Best,
>Feng
>
>On Fri, Oct 20, 2023 at 11:55 AM casel.chen  wrote:
>
>之前在flink sql中创建过一个catalog,现在想查看当初创建catalog的语句复制并修改一下另存为一个新的catalog,发现flink
>sql不支持show create catalog 。
>而据我所知doris是支持show create catalog语句的。flink sql能否支持一下呢?


flink 1.18.0 使用 hive beeline + jdbc driver连接sql gateway失败

2023-10-29 文章 casel.chen
1. 启动flink集群
bin/start-cluster.sh


2. 启动sql gateway
bin/sql-gateway.sh start -Dsql-gateway.endpoint.type=hiveserver2


3. 将flink-sql-jdbc-driver-bundle-1.18.0.jar放到apache-hive-3.1.2-bin/lib目录下


4. 到apache-hive-3.1.2-bin目录下启动beeline连接sql gateway,提示输入用户名和密码时直接按的回车
$ bin/beeline 
SLF4J: Class path contains multiple SLF4J bindings.
SLF4J: Found binding in 
[jar:file:/Users/admin/dev/hadoop-3.3.4/share/hadoop/common/lib/slf4j-reload4j-1.7.36.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: Found binding in 
[jar:file:/Users/admin/dev/apache-hive-3.1.2-bin/lib/log4j-slf4j-impl-2.10.0.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation.
SLF4J: Actual binding is of type [org.slf4j.impl.Reload4jLoggerFactory]
Beeline version 3.1.2 by Apache Hive
beeline> !connect jdbc:flink://localhost:8083
Connecting to jdbc:flink://localhost:8083
Enter username for jdbc:flink://localhost:8083: 
Enter password for jdbc:flink://localhost:8083: 
Failed to create the executor.
0: jdbc:flink://localhost:8083 (closed)> CREATE TABLE T(
. . . . . . . . . . . . . . . . . . . .>   a INT,
. . . . . . . . . . . . . . . . . . . .>   b VARCHAR(10)
. . . . . . . . . . . . . . . . . . . .> ) WITH (
. . . . . . . . . . . . . . . . . . . .>   'connector' = 'filesystem',
. . . . . . . . . . . . . . . . . . . .>   'path' = 'file:///tmp/T.csv',
. . . . . . . . . . . . . . . . . . . .>   'format' = 'csv'
. . . . . . . . . . . . . . . . . . . .> );
Failed to create the executor.
Connection is already closed.



flink sql如何处理脏数据问题?

2023-10-28 文章 casel.chen
场景:使用flink 
sql将数据写入下游OLAP系统,如doris,遇到一些异常情况,比如字段值超长或者分区字段值为当前doris表不存在的分区(需要先人为创建)等等,当前写入这些脏数据会使得作业写入报错,进而导致作业失败。我们是希望能够将这些“脏”数据单独发到一个kafka
 topic或者写入一个文件便于事后审查。这个目前有办法做到吗?

FW: Unable to achieve Flink kafka connector exactly once delivery semantics.

2023-10-27 文章 Gopal Chennupati (gchennup)
Hi,
Can someone please help me to resolve the below issue while running flink job.
Or provide me any doc/example which describe the exactly-once delivery 
guarantee semantics.

Thanks,
Gopal.

From: Gopal Chennupati (gchennup) 
Date: Friday, 27 October 2023 at 11:00 AM
To: commun...@flink.apache.org , 
u...@flink.apache.org 
Subject: Unable to achieve Flink kafka connector exactly once delivery 
semantics.
Hi Team,


I am trying to configure my kafka sink connector with “exactly-once” delivery 
guarantee, however it’s failing when I run the flink job with this 
configuration, here is the full exception stack trace from the job logs.


[Source: SG-SGT-TransformerJob -> Map -> Sink: Writer -> Sink: Committer 
(5/10)#12] WARN org.apache.kafka.common.utils.AppInfoParser - Error registering 
AppInfo mbean

javax.management.InstanceAlreadyExistsException: 
kafka.producer:type=app-info,id=producer-sgt-4-1

  at 
java.management/com.sun.jmx.mbeanserver.Repository.addMBean(Repository.java:436)

  at 
java.management/com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.registerWithRepository(DefaultMBeanServerInterceptor.java:1855)

  at 
java.management/com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.registerDynamicMBean(DefaultMBeanServerInterceptor.java:955)

  at 
java.management/com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.registerObject(DefaultMBeanServerInterceptor.java:890)

  at 
java.management/com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.registerMBean(DefaultMBeanServerInterceptor.java:320)

  at 
java.management/com.sun.jmx.mbeanserver.JmxMBeanServer.registerMBean(JmxMBeanServer.java:522)

  at 
org.apache.kafka.common.utils.AppInfoParser.registerAppInfo(AppInfoParser.java:64)

  at 
org.apache.kafka.clients.producer.KafkaProducer.(KafkaProducer.java:433)

  at 
org.apache.kafka.clients.producer.KafkaProducer.(KafkaProducer.java:289)

  at 
org.apache.kafka.clients.producer.KafkaProducer.(KafkaProducer.java:316)

  at 
org.apache.kafka.clients.producer.KafkaProducer.(KafkaProducer.java:301)

  at 
org.apache.flink.connector.kafka.sink.FlinkKafkaInternalProducer.(FlinkKafkaInternalProducer.java:55)

  at 
org.apache.flink.connector.kafka.sink.KafkaWriter.getOrCreateTransactionalProducer(KafkaWriter.java:332)

  at 
org.apache.flink.connector.kafka.sink.TransactionAborter.abortTransactionOfSubtask(TransactionAborter.java:104)

  at 
org.apache.flink.connector.kafka.sink.TransactionAborter.abortTransactionsWithPrefix(TransactionAborter.java:82)

  at 
org.apache.flink.connector.kafka.sink.TransactionAborter.abortLingeringTransactions(TransactionAborter.java:66)

  at 
org.apache.flink.connector.kafka.sink.KafkaWriter.abortLingeringTransactions(KafkaWriter.java:295)

  at 
org.apache.flink.connector.kafka.sink.KafkaWriter.(KafkaWriter.java:176)

  at 
org.apache.flink.connector.kafka.sink.KafkaSink.createWriter(KafkaSink.java:111)

  at 
org.apache.flink.connector.kafka.sink.KafkaSink.createWriter(KafkaSink.java:57)

  at 
org.apache.flink.streaming.runtime.operators.sink.StatefulSinkWriterStateHandler.createWriter(StatefulSinkWriterStateHandler.java:117)

  at 
org.apache.flink.streaming.runtime.operators.sink.SinkWriterOperator.initializeState(SinkWriterOperator.java:146)

  at 
org.apache.flink.streaming.api.operators.StreamOperatorStateHandler.initializeOperatorState(StreamOperatorStateHandler.java:122)

  at 
org.apache.flink.streaming.api.operators.AbstractStreamOperator.initializeState(AbstractStreamOperator.java:274)

  at 
org.apache.flink.streaming.runtime.tasks.RegularOperatorChain.initializeStateAndOpenOperators(RegularOperatorChain.java:106)

  at 
org.apache.flink.streaming.runtime.tasks.StreamTask.restoreGates(StreamTask.java:734)

  at 
org.apache.flink.streaming.runtime.tasks.StreamTaskActionExecutor$1.call(StreamTaskActionExecutor.java:55)

  at 
org.apache.flink.streaming.runtime.tasks.StreamTask.restoreInternal(StreamTask.java:709)

  at 
org.apache.flink.streaming.runtime.tasks.StreamTask.restore(StreamTask.java:675)

  at 
org.apache.flink.runtime.taskmanager.Task.runWithSystemExitMonitoring(Task.java:952)

  at 
org.apache.flink.runtime.taskmanager.Task.restoreAndInvoke(Task.java:921)

  at org.apache.flink.runtime.taskmanager.Task.doRun(Task.java:745)

  at org.apache.flink.runtime.taskmanager.Task.run(Task.java:562)

  at java.base/java.lang.Thread.run(Thread.java:834)


And here is the producer configuration,
KafkaSink sink = KafkaSink
.builder()

.setBootstrapServers(producerConfig.getProperty("bootstrap.servers"))
.setKafkaProducerConfig(producerConfig)
.setRecordSerializer(new 
GenericMessageSerialization<>(generic_key.class,
generic_value.class, 
producerConfig.getProperty("topic"),

Re: [ANNOUNCE] Apache Flink 1.18.0 released

2023-10-26 文章 Jark Wu
Congratulations and thanks release managers and everyone who has
contributed!

Best,
Jark

On Fri, 27 Oct 2023 at 12:25, Hang Ruan  wrote:

> Congratulations!
>
> Best,
> Hang
>
> Samrat Deb  于2023年10月27日周五 11:50写道:
>
> > Congratulations on the great release
> >
> > Bests,
> > Samrat
> >
> > On Fri, 27 Oct 2023 at 7:59 AM, Yangze Guo  wrote:
> >
> > > Great work! Congratulations to everyone involved!
> > >
> > > Best,
> > > Yangze Guo
> > >
> > > On Fri, Oct 27, 2023 at 10:23 AM Qingsheng Ren 
> wrote:
> > > >
> > > > Congratulations and big THANK YOU to everyone helping with this
> > release!
> > > >
> > > > Best,
> > > > Qingsheng
> > > >
> > > > On Fri, Oct 27, 2023 at 10:18 AM Benchao Li 
> > > wrote:
> > > >>
> > > >> Great work, thanks everyone involved!
> > > >>
> > > >> Rui Fan <1996fan...@gmail.com> 于2023年10月27日周五 10:16写道:
> > > >> >
> > > >> > Thanks for the great work!
> > > >> >
> > > >> > Best,
> > > >> > Rui
> > > >> >
> > > >> > On Fri, Oct 27, 2023 at 10:03 AM Paul Lam 
> > > wrote:
> > > >> >
> > > >> > > Finally! Thanks to all!
> > > >> > >
> > > >> > > Best,
> > > >> > > Paul Lam
> > > >> > >
> > > >> > > > 2023年10月27日 03:58,Alexander Fedulov <
> > alexander.fedu...@gmail.com>
> > > 写道:
> > > >> > > >
> > > >> > > > Great work, thanks everyone!
> > > >> > > >
> > > >> > > > Best,
> > > >> > > > Alexander
> > > >> > > >
> > > >> > > > On Thu, 26 Oct 2023 at 21:15, Martijn Visser <
> > > martijnvis...@apache.org>
> > > >> > > > wrote:
> > > >> > > >
> > > >> > > >> Thank you all who have contributed!
> > > >> > > >>
> > > >> > > >> Op do 26 okt 2023 om 18:41 schreef Feng Jin <
> > > jinfeng1...@gmail.com>
> > > >> > > >>
> > > >> > > >>> Thanks for the great work! Congratulations
> > > >> > > >>>
> > > >> > > >>>
> > > >> > > >>> Best,
> > > >> > > >>> Feng Jin
> > > >> > > >>>
> > > >> > > >>> On Fri, Oct 27, 2023 at 12:36 AM Leonard Xu <
> > xbjt...@gmail.com>
> > > wrote:
> > > >> > > >>>
> > > >> > >  Congratulations, Well done!
> > > >> > > 
> > > >> > >  Best,
> > > >> > >  Leonard
> > > >> > > 
> > > >> > >  On Fri, Oct 27, 2023 at 12:23 AM Lincoln Lee <
> > > lincoln.8...@gmail.com>
> > > >> > >  wrote:
> > > >> > > 
> > > >> > > > Thanks for the great work! Congrats all!
> > > >> > > >
> > > >> > > > Best,
> > > >> > > > Lincoln Lee
> > > >> > > >
> > > >> > > >
> > > >> > > > Jing Ge  于2023年10月27日周五
> > 00:16写道:
> > > >> > > >
> > > >> > > >> The Apache Flink community is very happy to announce the
> > > release of
> > > >> > > > Apache
> > > >> > > >> Flink 1.18.0, which is the first release for the Apache
> > > Flink 1.18
> > > >> > > > series.
> > > >> > > >>
> > > >> > > >> Apache Flink® is an open-source unified stream and batch
> > data
> > > >> > >  processing
> > > >> > > >> framework for distributed, high-performing,
> > > always-available, and
> > > >> > > > accurate
> > > >> > > >> data applications.
> > > >> > > >>
> > > >> > > >> The release is available for download at:
> > > >> > > >> https://flink.apache.org/downloads.html
> > > >> > > >>
> > > >> > > >> Please check out the release blog post for an overview of
> > the
> > > >> > > > improvements
> > > >> > > >> for this release:
> > > >> > > >>
> > > >> > > >>
> > > >> > > >
> > > >> > > 
> > > >> > > >>>
> > > >> > > >>
> > > >> > >
> > >
> >
> https://flink.apache.org/2023/10/24/announcing-the-release-of-apache-flink-1.18/
> > > >> > > >>
> > > >> > > >> The full release notes are available in Jira:
> > > >> > > >>
> > > >> > > >>
> > > >> > > >
> > > >> > > 
> > > >> > > >>>
> > > >> > > >>
> > > >> > >
> > >
> >
> https://issues.apache.org/jira/secure/ReleaseNote.jspa?projectId=12315522=12352885
> > > >> > > >>
> > > >> > > >> We would like to thank all contributors of the Apache
> Flink
> > > >> > > >> community
> > > >> > >  who
> > > >> > > >> made this release possible!
> > > >> > > >>
> > > >> > > >> Best regards,
> > > >> > > >> Konstantin, Qingsheng, Sergey, and Jing
> > > >> > > >>
> > > >> > > >
> > > >> > > 
> > > >> > > >>>
> > > >> > > >>
> > > >> > >
> > > >> > >
> > > >>
> > > >>
> > > >>
> > > >> --
> > > >>
> > > >> Best,
> > > >> Benchao Li
> > >
> >
>


Re: [ANNOUNCE] Apache Flink 1.18.0 released

2023-10-26 文章 Hang Ruan
Congratulations!

Best,
Hang

Samrat Deb  于2023年10月27日周五 11:50写道:

> Congratulations on the great release
>
> Bests,
> Samrat
>
> On Fri, 27 Oct 2023 at 7:59 AM, Yangze Guo  wrote:
>
> > Great work! Congratulations to everyone involved!
> >
> > Best,
> > Yangze Guo
> >
> > On Fri, Oct 27, 2023 at 10:23 AM Qingsheng Ren  wrote:
> > >
> > > Congratulations and big THANK YOU to everyone helping with this
> release!
> > >
> > > Best,
> > > Qingsheng
> > >
> > > On Fri, Oct 27, 2023 at 10:18 AM Benchao Li 
> > wrote:
> > >>
> > >> Great work, thanks everyone involved!
> > >>
> > >> Rui Fan <1996fan...@gmail.com> 于2023年10月27日周五 10:16写道:
> > >> >
> > >> > Thanks for the great work!
> > >> >
> > >> > Best,
> > >> > Rui
> > >> >
> > >> > On Fri, Oct 27, 2023 at 10:03 AM Paul Lam 
> > wrote:
> > >> >
> > >> > > Finally! Thanks to all!
> > >> > >
> > >> > > Best,
> > >> > > Paul Lam
> > >> > >
> > >> > > > 2023年10月27日 03:58,Alexander Fedulov <
> alexander.fedu...@gmail.com>
> > 写道:
> > >> > > >
> > >> > > > Great work, thanks everyone!
> > >> > > >
> > >> > > > Best,
> > >> > > > Alexander
> > >> > > >
> > >> > > > On Thu, 26 Oct 2023 at 21:15, Martijn Visser <
> > martijnvis...@apache.org>
> > >> > > > wrote:
> > >> > > >
> > >> > > >> Thank you all who have contributed!
> > >> > > >>
> > >> > > >> Op do 26 okt 2023 om 18:41 schreef Feng Jin <
> > jinfeng1...@gmail.com>
> > >> > > >>
> > >> > > >>> Thanks for the great work! Congratulations
> > >> > > >>>
> > >> > > >>>
> > >> > > >>> Best,
> > >> > > >>> Feng Jin
> > >> > > >>>
> > >> > > >>> On Fri, Oct 27, 2023 at 12:36 AM Leonard Xu <
> xbjt...@gmail.com>
> > wrote:
> > >> > > >>>
> > >> > >  Congratulations, Well done!
> > >> > > 
> > >> > >  Best,
> > >> > >  Leonard
> > >> > > 
> > >> > >  On Fri, Oct 27, 2023 at 12:23 AM Lincoln Lee <
> > lincoln.8...@gmail.com>
> > >> > >  wrote:
> > >> > > 
> > >> > > > Thanks for the great work! Congrats all!
> > >> > > >
> > >> > > > Best,
> > >> > > > Lincoln Lee
> > >> > > >
> > >> > > >
> > >> > > > Jing Ge  于2023年10月27日周五
> 00:16写道:
> > >> > > >
> > >> > > >> The Apache Flink community is very happy to announce the
> > release of
> > >> > > > Apache
> > >> > > >> Flink 1.18.0, which is the first release for the Apache
> > Flink 1.18
> > >> > > > series.
> > >> > > >>
> > >> > > >> Apache Flink® is an open-source unified stream and batch
> data
> > >> > >  processing
> > >> > > >> framework for distributed, high-performing,
> > always-available, and
> > >> > > > accurate
> > >> > > >> data applications.
> > >> > > >>
> > >> > > >> The release is available for download at:
> > >> > > >> https://flink.apache.org/downloads.html
> > >> > > >>
> > >> > > >> Please check out the release blog post for an overview of
> the
> > >> > > > improvements
> > >> > > >> for this release:
> > >> > > >>
> > >> > > >>
> > >> > > >
> > >> > > 
> > >> > > >>>
> > >> > > >>
> > >> > >
> >
> https://flink.apache.org/2023/10/24/announcing-the-release-of-apache-flink-1.18/
> > >> > > >>
> > >> > > >> The full release notes are available in Jira:
> > >> > > >>
> > >> > > >>
> > >> > > >
> > >> > > 
> > >> > > >>>
> > >> > > >>
> > >> > >
> >
> https://issues.apache.org/jira/secure/ReleaseNote.jspa?projectId=12315522=12352885
> > >> > > >>
> > >> > > >> We would like to thank all contributors of the Apache Flink
> > >> > > >> community
> > >> > >  who
> > >> > > >> made this release possible!
> > >> > > >>
> > >> > > >> Best regards,
> > >> > > >> Konstantin, Qingsheng, Sergey, and Jing
> > >> > > >>
> > >> > > >
> > >> > > 
> > >> > > >>>
> > >> > > >>
> > >> > >
> > >> > >
> > >>
> > >>
> > >>
> > >> --
> > >>
> > >> Best,
> > >> Benchao Li
> >
>


Re: 退订

2023-10-26 文章 Junrui Lee
Hi,

请发送任意内容的邮件到 user-zh-unsubscr...@flink.apache.org 来取消订阅邮件。

Best,
Junrui

13430298988 <13430298...@163.com> 于2023年10月27日周五 11:00写道:

> 退订


Re: 退订

2023-10-26 文章 Junrui Lee
Hi,

请发送任意内容的邮件到 user-zh-unsubscr...@flink.apache.org 来取消订阅邮件。

Best,
Junrui

chenyu_opensource  于2023年10月27日周五 10:20写道:

> 退订


退订

2023-10-26 文章 13430298988
退订

Re: [ANNOUNCE] Apache Flink 1.18.0 released

2023-10-26 文章 Yangze Guo
Great work! Congratulations to everyone involved!

Best,
Yangze Guo

On Fri, Oct 27, 2023 at 10:23 AM Qingsheng Ren  wrote:
>
> Congratulations and big THANK YOU to everyone helping with this release!
>
> Best,
> Qingsheng
>
> On Fri, Oct 27, 2023 at 10:18 AM Benchao Li  wrote:
>>
>> Great work, thanks everyone involved!
>>
>> Rui Fan <1996fan...@gmail.com> 于2023年10月27日周五 10:16写道:
>> >
>> > Thanks for the great work!
>> >
>> > Best,
>> > Rui
>> >
>> > On Fri, Oct 27, 2023 at 10:03 AM Paul Lam  wrote:
>> >
>> > > Finally! Thanks to all!
>> > >
>> > > Best,
>> > > Paul Lam
>> > >
>> > > > 2023年10月27日 03:58,Alexander Fedulov  写道:
>> > > >
>> > > > Great work, thanks everyone!
>> > > >
>> > > > Best,
>> > > > Alexander
>> > > >
>> > > > On Thu, 26 Oct 2023 at 21:15, Martijn Visser 
>> > > > wrote:
>> > > >
>> > > >> Thank you all who have contributed!
>> > > >>
>> > > >> Op do 26 okt 2023 om 18:41 schreef Feng Jin 
>> > > >>
>> > > >>> Thanks for the great work! Congratulations
>> > > >>>
>> > > >>>
>> > > >>> Best,
>> > > >>> Feng Jin
>> > > >>>
>> > > >>> On Fri, Oct 27, 2023 at 12:36 AM Leonard Xu  
>> > > >>> wrote:
>> > > >>>
>> > >  Congratulations, Well done!
>> > > 
>> > >  Best,
>> > >  Leonard
>> > > 
>> > >  On Fri, Oct 27, 2023 at 12:23 AM Lincoln Lee 
>> > >  
>> > >  wrote:
>> > > 
>> > > > Thanks for the great work! Congrats all!
>> > > >
>> > > > Best,
>> > > > Lincoln Lee
>> > > >
>> > > >
>> > > > Jing Ge  于2023年10月27日周五 00:16写道:
>> > > >
>> > > >> The Apache Flink community is very happy to announce the release 
>> > > >> of
>> > > > Apache
>> > > >> Flink 1.18.0, which is the first release for the Apache Flink 1.18
>> > > > series.
>> > > >>
>> > > >> Apache Flink® is an open-source unified stream and batch data
>> > >  processing
>> > > >> framework for distributed, high-performing, always-available, and
>> > > > accurate
>> > > >> data applications.
>> > > >>
>> > > >> The release is available for download at:
>> > > >> https://flink.apache.org/downloads.html
>> > > >>
>> > > >> Please check out the release blog post for an overview of the
>> > > > improvements
>> > > >> for this release:
>> > > >>
>> > > >>
>> > > >
>> > > 
>> > > >>>
>> > > >>
>> > > https://flink.apache.org/2023/10/24/announcing-the-release-of-apache-flink-1.18/
>> > > >>
>> > > >> The full release notes are available in Jira:
>> > > >>
>> > > >>
>> > > >
>> > > 
>> > > >>>
>> > > >>
>> > > https://issues.apache.org/jira/secure/ReleaseNote.jspa?projectId=12315522=12352885
>> > > >>
>> > > >> We would like to thank all contributors of the Apache Flink
>> > > >> community
>> > >  who
>> > > >> made this release possible!
>> > > >>
>> > > >> Best regards,
>> > > >> Konstantin, Qingsheng, Sergey, and Jing
>> > > >>
>> > > >
>> > > 
>> > > >>>
>> > > >>
>> > >
>> > >
>>
>>
>>
>> --
>>
>> Best,
>> Benchao Li


Re: [ANNOUNCE] Apache Flink 1.18.0 released

2023-10-26 文章 Qingsheng Ren
Congratulations and big THANK YOU to everyone helping with this release!

Best,
Qingsheng

On Fri, Oct 27, 2023 at 10:18 AM Benchao Li  wrote:

> Great work, thanks everyone involved!
>
> Rui Fan <1996fan...@gmail.com> 于2023年10月27日周五 10:16写道:
> >
> > Thanks for the great work!
> >
> > Best,
> > Rui
> >
> > On Fri, Oct 27, 2023 at 10:03 AM Paul Lam  wrote:
> >
> > > Finally! Thanks to all!
> > >
> > > Best,
> > > Paul Lam
> > >
> > > > 2023年10月27日 03:58,Alexander Fedulov 
> 写道:
> > > >
> > > > Great work, thanks everyone!
> > > >
> > > > Best,
> > > > Alexander
> > > >
> > > > On Thu, 26 Oct 2023 at 21:15, Martijn Visser <
> martijnvis...@apache.org>
> > > > wrote:
> > > >
> > > >> Thank you all who have contributed!
> > > >>
> > > >> Op do 26 okt 2023 om 18:41 schreef Feng Jin 
> > > >>
> > > >>> Thanks for the great work! Congratulations
> > > >>>
> > > >>>
> > > >>> Best,
> > > >>> Feng Jin
> > > >>>
> > > >>> On Fri, Oct 27, 2023 at 12:36 AM Leonard Xu 
> wrote:
> > > >>>
> > >  Congratulations, Well done!
> > > 
> > >  Best,
> > >  Leonard
> > > 
> > >  On Fri, Oct 27, 2023 at 12:23 AM Lincoln Lee <
> lincoln.8...@gmail.com>
> > >  wrote:
> > > 
> > > > Thanks for the great work! Congrats all!
> > > >
> > > > Best,
> > > > Lincoln Lee
> > > >
> > > >
> > > > Jing Ge  于2023年10月27日周五 00:16写道:
> > > >
> > > >> The Apache Flink community is very happy to announce the
> release of
> > > > Apache
> > > >> Flink 1.18.0, which is the first release for the Apache Flink
> 1.18
> > > > series.
> > > >>
> > > >> Apache Flink® is an open-source unified stream and batch data
> > >  processing
> > > >> framework for distributed, high-performing, always-available,
> and
> > > > accurate
> > > >> data applications.
> > > >>
> > > >> The release is available for download at:
> > > >> https://flink.apache.org/downloads.html
> > > >>
> > > >> Please check out the release blog post for an overview of the
> > > > improvements
> > > >> for this release:
> > > >>
> > > >>
> > > >
> > > 
> > > >>>
> > > >>
> > >
> https://flink.apache.org/2023/10/24/announcing-the-release-of-apache-flink-1.18/
> > > >>
> > > >> The full release notes are available in Jira:
> > > >>
> > > >>
> > > >
> > > 
> > > >>>
> > > >>
> > >
> https://issues.apache.org/jira/secure/ReleaseNote.jspa?projectId=12315522=12352885
> > > >>
> > > >> We would like to thank all contributors of the Apache Flink
> > > >> community
> > >  who
> > > >> made this release possible!
> > > >>
> > > >> Best regards,
> > > >> Konstantin, Qingsheng, Sergey, and Jing
> > > >>
> > > >
> > > 
> > > >>>
> > > >>
> > >
> > >
>
>
>
> --
>
> Best,
> Benchao Li
>


退订

2023-10-26 文章 chenyu_opensource
退订

Re: [ANNOUNCE] Apache Flink 1.18.0 released

2023-10-26 文章 Benchao Li
Great work, thanks everyone involved!

Rui Fan <1996fan...@gmail.com> 于2023年10月27日周五 10:16写道:
>
> Thanks for the great work!
>
> Best,
> Rui
>
> On Fri, Oct 27, 2023 at 10:03 AM Paul Lam  wrote:
>
> > Finally! Thanks to all!
> >
> > Best,
> > Paul Lam
> >
> > > 2023年10月27日 03:58,Alexander Fedulov  写道:
> > >
> > > Great work, thanks everyone!
> > >
> > > Best,
> > > Alexander
> > >
> > > On Thu, 26 Oct 2023 at 21:15, Martijn Visser 
> > > wrote:
> > >
> > >> Thank you all who have contributed!
> > >>
> > >> Op do 26 okt 2023 om 18:41 schreef Feng Jin 
> > >>
> > >>> Thanks for the great work! Congratulations
> > >>>
> > >>>
> > >>> Best,
> > >>> Feng Jin
> > >>>
> > >>> On Fri, Oct 27, 2023 at 12:36 AM Leonard Xu  wrote:
> > >>>
> >  Congratulations, Well done!
> > 
> >  Best,
> >  Leonard
> > 
> >  On Fri, Oct 27, 2023 at 12:23 AM Lincoln Lee 
> >  wrote:
> > 
> > > Thanks for the great work! Congrats all!
> > >
> > > Best,
> > > Lincoln Lee
> > >
> > >
> > > Jing Ge  于2023年10月27日周五 00:16写道:
> > >
> > >> The Apache Flink community is very happy to announce the release of
> > > Apache
> > >> Flink 1.18.0, which is the first release for the Apache Flink 1.18
> > > series.
> > >>
> > >> Apache Flink® is an open-source unified stream and batch data
> >  processing
> > >> framework for distributed, high-performing, always-available, and
> > > accurate
> > >> data applications.
> > >>
> > >> The release is available for download at:
> > >> https://flink.apache.org/downloads.html
> > >>
> > >> Please check out the release blog post for an overview of the
> > > improvements
> > >> for this release:
> > >>
> > >>
> > >
> > 
> > >>>
> > >>
> > https://flink.apache.org/2023/10/24/announcing-the-release-of-apache-flink-1.18/
> > >>
> > >> The full release notes are available in Jira:
> > >>
> > >>
> > >
> > 
> > >>>
> > >>
> > https://issues.apache.org/jira/secure/ReleaseNote.jspa?projectId=12315522=12352885
> > >>
> > >> We would like to thank all contributors of the Apache Flink
> > >> community
> >  who
> > >> made this release possible!
> > >>
> > >> Best regards,
> > >> Konstantin, Qingsheng, Sergey, and Jing
> > >>
> > >
> > 
> > >>>
> > >>
> >
> >



-- 

Best,
Benchao Li


Re: [ANNOUNCE] Apache Flink 1.18.0 released

2023-10-26 文章 Rui Fan
Thanks for the great work!

Best,
Rui

On Fri, Oct 27, 2023 at 10:03 AM Paul Lam  wrote:

> Finally! Thanks to all!
>
> Best,
> Paul Lam
>
> > 2023年10月27日 03:58,Alexander Fedulov  写道:
> >
> > Great work, thanks everyone!
> >
> > Best,
> > Alexander
> >
> > On Thu, 26 Oct 2023 at 21:15, Martijn Visser 
> > wrote:
> >
> >> Thank you all who have contributed!
> >>
> >> Op do 26 okt 2023 om 18:41 schreef Feng Jin 
> >>
> >>> Thanks for the great work! Congratulations
> >>>
> >>>
> >>> Best,
> >>> Feng Jin
> >>>
> >>> On Fri, Oct 27, 2023 at 12:36 AM Leonard Xu  wrote:
> >>>
>  Congratulations, Well done!
> 
>  Best,
>  Leonard
> 
>  On Fri, Oct 27, 2023 at 12:23 AM Lincoln Lee 
>  wrote:
> 
> > Thanks for the great work! Congrats all!
> >
> > Best,
> > Lincoln Lee
> >
> >
> > Jing Ge  于2023年10月27日周五 00:16写道:
> >
> >> The Apache Flink community is very happy to announce the release of
> > Apache
> >> Flink 1.18.0, which is the first release for the Apache Flink 1.18
> > series.
> >>
> >> Apache Flink® is an open-source unified stream and batch data
>  processing
> >> framework for distributed, high-performing, always-available, and
> > accurate
> >> data applications.
> >>
> >> The release is available for download at:
> >> https://flink.apache.org/downloads.html
> >>
> >> Please check out the release blog post for an overview of the
> > improvements
> >> for this release:
> >>
> >>
> >
> 
> >>>
> >>
> https://flink.apache.org/2023/10/24/announcing-the-release-of-apache-flink-1.18/
> >>
> >> The full release notes are available in Jira:
> >>
> >>
> >
> 
> >>>
> >>
> https://issues.apache.org/jira/secure/ReleaseNote.jspa?projectId=12315522=12352885
> >>
> >> We would like to thank all contributors of the Apache Flink
> >> community
>  who
> >> made this release possible!
> >>
> >> Best regards,
> >> Konstantin, Qingsheng, Sergey, and Jing
> >>
> >
> 
> >>>
> >>
>
>


Re: [ANNOUNCE] Apache Flink 1.18.0 released

2023-10-26 文章 Paul Lam
Finally! Thanks to all!

Best,
Paul Lam

> 2023年10月27日 03:58,Alexander Fedulov  写道:
> 
> Great work, thanks everyone!
> 
> Best,
> Alexander
> 
> On Thu, 26 Oct 2023 at 21:15, Martijn Visser 
> wrote:
> 
>> Thank you all who have contributed!
>> 
>> Op do 26 okt 2023 om 18:41 schreef Feng Jin 
>> 
>>> Thanks for the great work! Congratulations
>>> 
>>> 
>>> Best,
>>> Feng Jin
>>> 
>>> On Fri, Oct 27, 2023 at 12:36 AM Leonard Xu  wrote:
>>> 
 Congratulations, Well done!
 
 Best,
 Leonard
 
 On Fri, Oct 27, 2023 at 12:23 AM Lincoln Lee 
 wrote:
 
> Thanks for the great work! Congrats all!
> 
> Best,
> Lincoln Lee
> 
> 
> Jing Ge  于2023年10月27日周五 00:16写道:
> 
>> The Apache Flink community is very happy to announce the release of
> Apache
>> Flink 1.18.0, which is the first release for the Apache Flink 1.18
> series.
>> 
>> Apache Flink® is an open-source unified stream and batch data
 processing
>> framework for distributed, high-performing, always-available, and
> accurate
>> data applications.
>> 
>> The release is available for download at:
>> https://flink.apache.org/downloads.html
>> 
>> Please check out the release blog post for an overview of the
> improvements
>> for this release:
>> 
>> 
> 
 
>>> 
>> https://flink.apache.org/2023/10/24/announcing-the-release-of-apache-flink-1.18/
>> 
>> The full release notes are available in Jira:
>> 
>> 
> 
 
>>> 
>> https://issues.apache.org/jira/secure/ReleaseNote.jspa?projectId=12315522=12352885
>> 
>> We would like to thank all contributors of the Apache Flink
>> community
 who
>> made this release possible!
>> 
>> Best regards,
>> Konstantin, Qingsheng, Sergey, and Jing
>> 
> 
 
>>> 
>> 



Re: [ANNOUNCE] Apache Flink 1.18.0 released

2023-10-26 文章 liu ron
Great work, thanks everyone!

Best,
Ron

Alexander Fedulov  于2023年10月27日周五 04:00写道:

> Great work, thanks everyone!
>
> Best,
> Alexander
>
> On Thu, 26 Oct 2023 at 21:15, Martijn Visser 
> wrote:
>
> > Thank you all who have contributed!
> >
> > Op do 26 okt 2023 om 18:41 schreef Feng Jin 
> >
> > > Thanks for the great work! Congratulations
> > >
> > >
> > > Best,
> > > Feng Jin
> > >
> > > On Fri, Oct 27, 2023 at 12:36 AM Leonard Xu  wrote:
> > >
> > > > Congratulations, Well done!
> > > >
> > > > Best,
> > > > Leonard
> > > >
> > > > On Fri, Oct 27, 2023 at 12:23 AM Lincoln Lee  >
> > > > wrote:
> > > >
> > > > > Thanks for the great work! Congrats all!
> > > > >
> > > > > Best,
> > > > > Lincoln Lee
> > > > >
> > > > >
> > > > > Jing Ge  于2023年10月27日周五 00:16写道:
> > > > >
> > > > > > The Apache Flink community is very happy to announce the release
> of
> > > > > Apache
> > > > > > Flink 1.18.0, which is the first release for the Apache Flink
> 1.18
> > > > > series.
> > > > > >
> > > > > > Apache Flink® is an open-source unified stream and batch data
> > > > processing
> > > > > > framework for distributed, high-performing, always-available, and
> > > > > accurate
> > > > > > data applications.
> > > > > >
> > > > > > The release is available for download at:
> > > > > > https://flink.apache.org/downloads.html
> > > > > >
> > > > > > Please check out the release blog post for an overview of the
> > > > > improvements
> > > > > > for this release:
> > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> https://flink.apache.org/2023/10/24/announcing-the-release-of-apache-flink-1.18/
> > > > > >
> > > > > > The full release notes are available in Jira:
> > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> https://issues.apache.org/jira/secure/ReleaseNote.jspa?projectId=12315522=12352885
> > > > > >
> > > > > > We would like to thank all contributors of the Apache Flink
> > community
> > > > who
> > > > > > made this release possible!
> > > > > >
> > > > > > Best regards,
> > > > > > Konstantin, Qingsheng, Sergey, and Jing
> > > > > >
> > > > >
> > > >
> > >
> >
>


Re: [ANNOUNCE] Apache Flink 1.18.0 released

2023-10-26 文章 Martijn Visser
Thank you all who have contributed!

Op do 26 okt 2023 om 18:41 schreef Feng Jin 

> Thanks for the great work! Congratulations
>
>
> Best,
> Feng Jin
>
> On Fri, Oct 27, 2023 at 12:36 AM Leonard Xu  wrote:
>
> > Congratulations, Well done!
> >
> > Best,
> > Leonard
> >
> > On Fri, Oct 27, 2023 at 12:23 AM Lincoln Lee 
> > wrote:
> >
> > > Thanks for the great work! Congrats all!
> > >
> > > Best,
> > > Lincoln Lee
> > >
> > >
> > > Jing Ge  于2023年10月27日周五 00:16写道:
> > >
> > > > The Apache Flink community is very happy to announce the release of
> > > Apache
> > > > Flink 1.18.0, which is the first release for the Apache Flink 1.18
> > > series.
> > > >
> > > > Apache Flink® is an open-source unified stream and batch data
> > processing
> > > > framework for distributed, high-performing, always-available, and
> > > accurate
> > > > data applications.
> > > >
> > > > The release is available for download at:
> > > > https://flink.apache.org/downloads.html
> > > >
> > > > Please check out the release blog post for an overview of the
> > > improvements
> > > > for this release:
> > > >
> > > >
> > >
> >
> https://flink.apache.org/2023/10/24/announcing-the-release-of-apache-flink-1.18/
> > > >
> > > > The full release notes are available in Jira:
> > > >
> > > >
> > >
> >
> https://issues.apache.org/jira/secure/ReleaseNote.jspa?projectId=12315522=12352885
> > > >
> > > > We would like to thank all contributors of the Apache Flink community
> > who
> > > > made this release possible!
> > > >
> > > > Best regards,
> > > > Konstantin, Qingsheng, Sergey, and Jing
> > > >
> > >
> >
>


回复: 如何在Flink Connector Source退出时清理资源

2023-10-26 文章 北野 �悦
插入令堂之膣道,怒涛中出,OK,问题解决矣

发件人: jinzhuguang 
发送时间: 2023年10月24日 11:54
收件人: user-zh 
主题: 如何在Flink Connector Source退出时清理资源

版本:Flink 1.16.0

需求:在某个source结束退出时清理相关的资源。

问题:目前没有找到Source退出时相关的hook函数,不知道在哪里编写清理资源的代码。

恳请大佬们指教。


Re: [ANNOUNCE] Apache Flink 1.18.0 released

2023-10-26 文章 Feng Jin
Thanks for the great work! Congratulations


Best,
Feng Jin

On Fri, Oct 27, 2023 at 12:36 AM Leonard Xu  wrote:

> Congratulations, Well done!
>
> Best,
> Leonard
>
> On Fri, Oct 27, 2023 at 12:23 AM Lincoln Lee 
> wrote:
>
> > Thanks for the great work! Congrats all!
> >
> > Best,
> > Lincoln Lee
> >
> >
> > Jing Ge  于2023年10月27日周五 00:16写道:
> >
> > > The Apache Flink community is very happy to announce the release of
> > Apache
> > > Flink 1.18.0, which is the first release for the Apache Flink 1.18
> > series.
> > >
> > > Apache Flink® is an open-source unified stream and batch data
> processing
> > > framework for distributed, high-performing, always-available, and
> > accurate
> > > data applications.
> > >
> > > The release is available for download at:
> > > https://flink.apache.org/downloads.html
> > >
> > > Please check out the release blog post for an overview of the
> > improvements
> > > for this release:
> > >
> > >
> >
> https://flink.apache.org/2023/10/24/announcing-the-release-of-apache-flink-1.18/
> > >
> > > The full release notes are available in Jira:
> > >
> > >
> >
> https://issues.apache.org/jira/secure/ReleaseNote.jspa?projectId=12315522=12352885
> > >
> > > We would like to thank all contributors of the Apache Flink community
> who
> > > made this release possible!
> > >
> > > Best regards,
> > > Konstantin, Qingsheng, Sergey, and Jing
> > >
> >
>


Re: [ANNOUNCE] Apache Flink 1.18.0 released

2023-10-26 文章 Leonard Xu
Congratulations, Well done!

Best,
Leonard

On Fri, Oct 27, 2023 at 12:23 AM Lincoln Lee  wrote:

> Thanks for the great work! Congrats all!
>
> Best,
> Lincoln Lee
>
>
> Jing Ge  于2023年10月27日周五 00:16写道:
>
> > The Apache Flink community is very happy to announce the release of
> Apache
> > Flink 1.18.0, which is the first release for the Apache Flink 1.18
> series.
> >
> > Apache Flink® is an open-source unified stream and batch data processing
> > framework for distributed, high-performing, always-available, and
> accurate
> > data applications.
> >
> > The release is available for download at:
> > https://flink.apache.org/downloads.html
> >
> > Please check out the release blog post for an overview of the
> improvements
> > for this release:
> >
> >
> https://flink.apache.org/2023/10/24/announcing-the-release-of-apache-flink-1.18/
> >
> > The full release notes are available in Jira:
> >
> >
> https://issues.apache.org/jira/secure/ReleaseNote.jspa?projectId=12315522=12352885
> >
> > We would like to thank all contributors of the Apache Flink community who
> > made this release possible!
> >
> > Best regards,
> > Konstantin, Qingsheng, Sergey, and Jing
> >
>


Re: [ANNOUNCE] Apache Flink 1.18.0 released

2023-10-26 文章 Lincoln Lee
Thanks for the great work! Congrats all!

Best,
Lincoln Lee


Jing Ge  于2023年10月27日周五 00:16写道:

> The Apache Flink community is very happy to announce the release of Apache
> Flink 1.18.0, which is the first release for the Apache Flink 1.18 series.
>
> Apache Flink® is an open-source unified stream and batch data processing
> framework for distributed, high-performing, always-available, and accurate
> data applications.
>
> The release is available for download at:
> https://flink.apache.org/downloads.html
>
> Please check out the release blog post for an overview of the improvements
> for this release:
>
> https://flink.apache.org/2023/10/24/announcing-the-release-of-apache-flink-1.18/
>
> The full release notes are available in Jira:
>
> https://issues.apache.org/jira/secure/ReleaseNote.jspa?projectId=12315522=12352885
>
> We would like to thank all contributors of the Apache Flink community who
> made this release possible!
>
> Best regards,
> Konstantin, Qingsheng, Sergey, and Jing
>


[ANNOUNCE] Apache Flink 1.18.0 released

2023-10-26 文章 Jing Ge
The Apache Flink community is very happy to announce the release of Apache
Flink 1.18.0, which is the first release for the Apache Flink 1.18 series.

Apache Flink® is an open-source unified stream and batch data processing
framework for distributed, high-performing, always-available, and accurate
data applications.

The release is available for download at:
https://flink.apache.org/downloads.html

Please check out the release blog post for an overview of the improvements
for this release:
https://flink.apache.org/2023/10/24/announcing-the-release-of-apache-flink-1.18/

The full release notes are available in Jira:
https://issues.apache.org/jira/secure/ReleaseNote.jspa?projectId=12315522=12352885

We would like to thank all contributors of the Apache Flink community who
made this release possible!

Best regards,
Konstantin, Qingsheng, Sergey, and Jing


Re: Flink1.17.1 yarn token 过期问题

2023-10-26 文章 Paul Lam
Hello,

这个问题解决了吗?我遇到相同的问题,还没定为到原因。

Best,
Paul Lam

> 2023年7月20日 12:04,王刚  写道:
> 
> 异常栈信息
> ```
> 
> 2023-07-20 11:43:01,627 ERROR 
> org.apache.flink.runtime.taskexecutor.TaskManagerRunner  [] - Terminating 
> TaskManagerRunner with exit code 1.
> org.apache.flink.util.FlinkException: Failed to start the TaskManagerRunner.
>at 
> org.apache.flink.runtime.taskexecutor.TaskManagerRunner.runTaskManager(TaskManagerRunner.java:488)
>  ~[flink-dist-1.17-SNAPSHOT.jar:1.17-SNAPSHOT]
>at 
> org.apache.flink.runtime.taskexecutor.TaskManagerRunner.lambda$runTaskManagerProcessSecurely$5(TaskManagerRunner.java:530)
>  ~[flink-dist-1.17-SNAPSHOT.jar:1.17-SNAPSHOT]
>at java.security.AccessController.doPrivileged(Native Method) 
> ~[?:1.8.0_92]
>at javax.security.auth.Subject.doAs(Subject.java:422) ~[?:1.8.0_92]
>at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1730)
>  ~[hadoop-common-3.2.1.jar:?]
>at 
> org.apache.flink.runtime.security.contexts.HadoopSecurityContext.runSecured(HadoopSecurityContext.java:41)
>  ~[flink-dist-1.17-SNAPSHOT.jar:1.17-SNAPSHOT]
>at 
> org.apache.flink.runtime.taskexecutor.TaskManagerRunner.runTaskManagerProcessSecurely(TaskManagerRunner.java:530)
>  [flink-dist-1.17-SNAPSHOT.jar:1.17-SNAPSHOT]
>at 
> org.apache.flink.yarn.YarnTaskExecutorRunner.runTaskManagerSecurely(YarnTaskExecutorRunner.java:94)
>  [flink-dist-1.17-SNAPSHOT.jar:1.17-SNAPSHOT]
>at 
> org.apache.flink.yarn.YarnTaskExecutorRunner.main(YarnTaskExecutorRunner.java:68)
>  [flink-dist-1.17-SNAPSHOT.jar:1.17-SNAPSHOT]
> Caused by: org.apache.hadoop.ipc.RemoteException: token (token for flink: 
> HDFS_DELEGATION_TOKEN 
> owner=flink/lf-client-flink-28-243-196.hadoop.local@HADOOP.LOCAL, renewer=, 
> realUser=, issueDate=1689734389821, maxDate=1690339189821, 
> sequenceNumber=266208479, masterKeyId=1131) can't be found in cache
>at org.apache.hadoop.ipc.Client.getRpcResponse(Client.java:1557) 
> ~[hadoop-common-3.2.1.jar:?]
>at org.apache.hadoop.ipc.Client.call(Client.java:1494) 
> ~[hadoop-common-3.2.1.jar:?]
>at org.apache.hadoop.ipc.Client.call(Client.java:1391) 
> ~[hadoop-common-3.2.1.jar:?]
>at 
> org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:233)
>  ~[hadoop-common-3.2.1.jar:?]
>at 
> org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:118)
>  ~[hadoop-common-3.2.1.jar:?]
>at com.sun.proxy.$Proxy26.mkdirs(Unknown Source) ~[?:?]
>at 
> org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.mkdirs(ClientNamenodeProtocolTranslatorPB.java:660)
>  ~[hadoop-hdfs-client-3.2.1.jar:?]
>at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) 
> ~[?:1.8.0_92]
>at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) 
> ~[?:1.8.0_92]
>at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>  ~[?:1.8.0_92]
>at java.lang.reflect.Method.invoke(Method.java:498) ~[?:1.8.0_92]
>at 
> org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:422)
>  ~[hadoop-common-3.2.1.jar:?]
>at 
> org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeMethod(RetryInvocationHandler.java:165)
>  ~[hadoop-common-3.2.1.jar:?]
>at 
> org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invoke(RetryInvocationHandler.java:157)
>  ~[hadoop-common-3.2.1.jar:?]
>at 
> org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeOnce(RetryInvocationHandler.java:95)
>  ~[hadoop-common-3.2.1.jar:?]
>at 
> org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:359)
>  ~[hadoop-common-3.2.1.jar:?]
>at com.sun.proxy.$Proxy27.mkdirs(Unknown Source) ~[?:?]
>at 
> org.apache.hadoop.hdfs.DFSClient.primitiveMkdir(DFSClient.java:2425) 
> ~[hadoop-hdfs-client-3.2.1.jar:?]
>at org.apache.hadoop.hdfs.DFSClient.mkdirs(DFSClient.java:2401) 
> ~[hadoop-hdfs-client-3.2.1.jar:?]
>at 
> org.apache.hadoop.hdfs.DistributedFileSystem$27.doCall(DistributedFileSystem.java:1318)
>  ~[hadoop-hdfs-client-3.2.1.jar:?]
>at 
> org.apache.hadoop.hdfs.DistributedFileSystem$27.doCall(DistributedFileSystem.java:1315)
>  ~[hadoop-hdfs-client-3.2.1.jar:?]
>at 
> org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)
>  ~[hadoop-common-3.2.1.jar:?]
>at 
> org.apache.hadoop.hdfs.DistributedFileSystem.mkdirsInternal(DistributedFileSystem.java:1332)
>  ~[hadoop-hdfs-client-3.2.1.jar:?]
>at 
> org.apache.hadoop.hdfs.DistributedFileSystem.mkdirs(DistributedFileSystem.java:1307)
>  ~[hadoop-hdfs-client-3.2.1.jar:?]
>at org.apache.hadoop.fs.FileSystem.mkdirs(FileSystem.java:2275) 
> ~[hadoop-common-3.2.1.jar:?]
>at 
> 

FLINK-24035, how can this issue be repeated?

2023-10-24 文章 rui chen
   We encountered similar problems in production, and we want to integrate
   FLINK-24035 to solve them, but we don't know how to repeat the problem.


Re:如何在Flink Connector Source退出时清理资源

2023-10-23 文章 Xuyang
Hi, 
看一下你的DynamicTableSource实现的类,如果你用的是InputFormat的旧source(用的是类似InputFormatProvider.of),可以使用InputFormat里的close方法;
如果用的是flip-27的source(用的是类似SourceProvider.of),SplitReader里也有一个close方法










--

Best!
Xuyang





在 2023-10-24 11:54:36,"jinzhuguang"  写道:
>版本:Flink 1.16.0
>
>需求:在某个source结束退出时清理相关的资源。
>
>问题:目前没有找到Source退出时相关的hook函数,不知道在哪里编写清理资源的代码。
>
>恳请大佬们指教。


如何在Flink Connector Source退出时清理资源

2023-10-23 文章 jinzhuguang
版本:Flink 1.16.0

需求:在某个source结束退出时清理相关的资源。

问题:目前没有找到Source退出时相关的hook函数,不知道在哪里编写清理资源的代码。

恳请大佬们指教。

Re: Problems with the state.backend.fs.memory-threshold parameter

2023-10-23 文章 rui chen
Hi,Zakelly
Thank you for your answer.

Best,
rui


Zakelly Lan  于2023年10月13日周五 19:12写道:

> Hi rui,
>
> The 'state.backend.fs.memory-threshold' configures the threshold below
> which state is stored as part of the metadata, rather than in separate
> files. So as a result the JM will use its memory to merge small
> checkpoint files and write them into one file. Currently the
> FLIP-306[1][2] is proposed to merge small checkpoint files without
> consuming JM memory. This feature is currently being worked on and is
> targeted for the next minor release (1.19).
>
>
> Best,
> Zakelly
>
> [1]
> https://cwiki.apache.org/confluence/display/FLINK/FLIP-306%3A+Unified+File+Merging+Mechanism+for+Checkpoints
> [2] https://issues.apache.org/jira/browse/FLINK-32070
>
> On Fri, Oct 13, 2023 at 6:28 PM rui chen  wrote:
> >
> > We found that for some tasks, the JM memory continued to increase. I set
> > the parameter of state.backend.fs.memory-threshold to 0, and the JM
> memory
> > would no longer increase, but many small files might be written in this
> > way. Does the community have any optimization plan for this area?
>


回复: flink sql不支持show create catalog 吗?

2023-10-20 文章 李宇彬
Hi Feng


我之前建过一个jira(https://issues.apache.org/jira/browse/FLINK-24939),引入CatalogStore后,实现这个特性的时机应该已经成熟了,我们这边业务场景里用到了很多catalog,管理起来很麻烦,有这个特性会好很多。
| |
 回复的原邮件 
| 发件人 | Feng Jin |
| 发送日期 | 2023年10月20日 13:18 |
| 收件人 |  |
| 主题 | Re: flink sql不支持show create catalog 吗? |
hi casel


从 1.18 开始,引入了 CatalogStore,持久化了 Catalog 的配置,确实可以支持 show create catalog 了。


Best,
Feng

On Fri, Oct 20, 2023 at 11:55 AM casel.chen  wrote:

之前在flink sql中创建过一个catalog,现在想查看当初创建catalog的语句复制并修改一下另存为一个新的catalog,发现flink
sql不支持show create catalog 。
而据我所知doris是支持show create catalog语句的。flink sql能否支持一下呢?


jobmanager 开启ha模式重复启用的问题

2023-10-20 文章 s
flink onYarn集群,运行模式为yarn application
开启zookeeper  high 
availability后,JobManager由于异常退出后,会重新启动一个jm在其他节点,这个没有问题,但是之前的那个tm的信息在zk中并不会删除,导致出现了两个job在,并且后续启动的那个job长期属于悬挂状态,不会正常工作
flink 版本为1.15.3,日志中的错误如下:


页面上的现象如下:


zk中的状态看上tm的信息不会删除,仍然是被locks的状态






请教下各位,是否有遇到类似的情况,zk中的过久信息是否可以配置删除?



Re: kafka_appender收集flink任务日志连接数过多问题

2023-10-19 文章 Feng Jin
可以考虑在每台 yarn 机器部署日志服务(可收集本机的日志到 kafka)
yarn container -> 单机的日志服务 -> kafka.



On Mon, Oct 16, 2023 at 3:58 PM 阿华田  wrote:

>
> Flink集群使用kafka_appender收集flink产生的日志,但是现在实时运行的任务超过了三千个,运行的yarn-container有20万+。导致存储日志的kafka集群连接数过多,kafka集群压力有点大,请教各位大佬flink日志采集这块还有什么好的方式吗
>
>
> | |
> 阿华田
> |
> |
> a15733178...@163.com
> |
> 签名由网易邮箱大师定制
>
>


Re: flink sql不支持show create catalog 吗?

2023-10-19 文章 Feng Jin
hi casel


从 1.18 开始,引入了 CatalogStore,持久化了 Catalog 的配置,确实可以支持 show create catalog 了。


Best,
Feng

On Fri, Oct 20, 2023 at 11:55 AM casel.chen  wrote:

> 之前在flink sql中创建过一个catalog,现在想查看当初创建catalog的语句复制并修改一下另存为一个新的catalog,发现flink
> sql不支持show create catalog 。
> 而据我所知doris是支持show create catalog语句的。flink sql能否支持一下呢?


flink sql不支持show create catalog 吗?

2023-10-19 文章 casel.chen
之前在flink sql中创建过一个catalog,现在想查看当初创建catalog的语句复制并修改一下另存为一个新的catalog,发现flink 
sql不支持show create catalog 。
而据我所知doris是支持show create catalog语句的。flink sql能否支持一下呢?

Re: Flink SQL的状态清理

2023-10-17 文章 Jane Chan
Hi, 你好

如果使用的是 standalone session cluster, 想要在 JM/TM 日志中看到参数打印出来, 需要在集群启动前在
flink-conf.yaml 配置 table.exec.state.ttl: '${TTL}', 再启动集群;
集群启动后再修改的话, 日志不会打印出来, 可以通过 SET; 命令查看当前 SQL CLI 中配置的参数.
另外, 需要先执行 SET 'table.exec.state.ttl' = '${TTL}' , 然后提交作业, 可以确认下操作顺序是否有误.

祝好!
Jane

On Mon, Oct 9, 2023 at 6:01 PM 小昌同学  wrote:

> 你好,老师,我也是这样设置的,我这边是flink sql client,但是我去flink web ui界面并没有看到这个配置生效。
>
>
> | |
> 小昌同学
> |
> |
> ccc0606fight...@163.com
> |
>  回复的原邮件 
> | 发件人 | Jane Chan |
> | 发送日期 | 2023年9月25日 11:24 |
> | 收件人 |  |
> | 主题 | Re: Flink SQL的状态清理 |
> Hi,
>
> 可以通过设置 table.exec.state.ttl 来控制状态算子的 state TTL. 更多信息请参阅 [1]
>
> [1]
>
> https://nightlies.apache.org/flink/flink-docs-master/zh/docs/dev/table/concepts/overview/#%e7%8a%b6%e6%80%81%e7%ae%a1%e7%90%86
>
> Best,
> Jane
>
> On Thu, Sep 21, 2023 at 5:17 PM faronzz  wrote:
>
> 试试这个 t_env.get_config().set("table.exec.state.ttl", "86400 s")
>
>
>
>
> | |
> faronzz
> |
> |
> faro...@163.com
> |
>
>
>  回复的原邮件 
> | 发件人 | 小昌同学 |
> | 发送日期 | 2023年09月21日 17:06 |
> | 收件人 | user-zh |
> | 主题 | Flink SQL的状态清理 |
>
>
> 各位老师好,请教一下大家关于flink sql的状态清理问题,我百度的话只找到相关的minbath设置,sql是没有配置state的ttl设置嘛
> | |
> 小昌同学
> |
> |
> ccc0606fight...@163.com
> |
>


Re: 请问1.18什么时候可以发布呢,想体验1.17jdk

2023-10-16 文章 Jing Ge
快了,已经开始voting了 :-))

On Sun, Oct 15, 2023 at 5:55 AM kcz <573693...@qq.com.invalid> wrote:

>


kafka_appender收集flink任务日志连接数过多问题

2023-10-16 文章 阿华田
Flink集群使用kafka_appender收集flink产生的日志,但是现在实时运行的任务超过了三千个,运行的yarn-container有20万+。导致存储日志的kafka集群连接数过多,kafka集群压力有点大,请教各位大佬flink日志采集这块还有什么好的方式吗


| |
阿华田
|
|
a15733178...@163.com
|
签名由网易邮箱大师定制



Flink启动失败问题

2023-10-16 文章 qwemssd
Flink版本


到bin目录执行:  ./start-cluster.sh




报错信息:








看是端口占用其实没有




尝试修改端口,还是一样报错端口问题,  任意修改一个端口多是报同上面一样的错误。




端口也是开放的






希望大佬们能帮忙看看是什么问题呢???


| |
qwemssd
|
|
qwem...@163.com
|

请问1.18什么时候可以发布呢,想体验1.17jdk

2023-10-14 文章 kcz


Re: Problems with the state.backend.fs.memory-threshold parameter

2023-10-13 文章 Zakelly Lan
Hi rui,

The 'state.backend.fs.memory-threshold' configures the threshold below
which state is stored as part of the metadata, rather than in separate
files. So as a result the JM will use its memory to merge small
checkpoint files and write them into one file. Currently the
FLIP-306[1][2] is proposed to merge small checkpoint files without
consuming JM memory. This feature is currently being worked on and is
targeted for the next minor release (1.19).


Best,
Zakelly

[1] 
https://cwiki.apache.org/confluence/display/FLINK/FLIP-306%3A+Unified+File+Merging+Mechanism+for+Checkpoints
[2] https://issues.apache.org/jira/browse/FLINK-32070

On Fri, Oct 13, 2023 at 6:28 PM rui chen  wrote:
>
> We found that for some tasks, the JM memory continued to increase. I set
> the parameter of state.backend.fs.memory-threshold to 0, and the JM memory
> would no longer increase, but many small files might be written in this
> way. Does the community have any optimization plan for this area?


Real-time task blocking problem

2023-10-13 文章 rui chen
After the task restart of our 1.13 version, kakfa consumption zero problem
occurred. Have you ever encountered it?


Problems with the state.backend.fs.memory-threshold parameter

2023-10-13 文章 rui chen
We found that for some tasks, the JM memory continued to increase. I set
the parameter of state.backend.fs.memory-threshold to 0, and the JM memory
would no longer increase, but many small files might be written in this
way. Does the community have any optimization plan for this area?


Re: Cannot find metata file metadats in directory

2023-10-13 文章 rui chen
After the task is restarted for several times, it is found that the
supported cp is deleted. I view the audit log of HDFS and find that the
deletion request comes from JM

Hangxiang Yu  于2023年9月30日周六 17:10写道:

> Hi,
> How did you point out the checkpoint path you restored from ?
>
> Seems that you are trying to restore from a not completed or failed
> checkpoint.
>
> On Thu, Sep 28, 2023 at 6:09 PM rui chen  wrote:
>
> > When we use 1.13.2,we have the following error:
> > FileNotFoundException: Cannot find metata file metadats in directory
> > 'hdfs://xx/f408dbe327f9e5053e76d7b5323d6e81/chk-173'.
> >
>
>
> --
> Best,
> Hangxiang.
>


Re: 关于Flink SQL无法对带metadata列的表在执行insert into时,指定具体的列名的问题。

2023-10-12 文章 jinzhuguang
感谢大佬!!!

> 2023年10月13日 10:44,tanjialiang  写道:
> 
> Hi, 
> 这个问题已经在1.17.0修复,详细可以看https://issues.apache.org/jira/browse/FLINK-30922
> 
> 
> best wishes,
> tanjialiang.
> 
> 
>  回复的原邮件 
> | 发件人 | jinzhuguang |
> | 发送日期 | 2023年10月13日 10:39 |
> | 收件人 | user-zh |
> | 主题 | 关于Flink SQL无法对带metadata列的表在执行insert into时,指定具体的列名的问题。 |
> 首先,我的Flink版本为1.16.0
> 为了方便理解,我以Kafka作为案例来描述:
> 我有以下两个表:
> CREATE TABLE orders(
> user_id BIGINT,
> name STRING,
> timestamp TIMESTAMP(3) METADATA VIRTUAL
> )WITH(
> 'connector'='kafka',
> 'topic'='orders',
> 'properties.group.id' = 'test_join_tempral',
> 'scan.startup.mode'='earliest-offset',
> 'properties.bootstrap.servers'='localhost:9092',
> 'format'='json',
> 'json.ignore-parse-errors' = 'true'
> );
> CREATE TABLE kafka_sink(
> user_id BIGINT,
> name STRING,
> timestamp TIMESTAMP(3) METADATA FROM 'timestamp'
> )WITH(
> 'connector'='kafka',
> 'topic'='kafka_sink',
> 'properties.group.id' = 'test_join_tempral',
> 'scan.startup.mode'='earliest-offset',
> 'properties.bootstrap.servers'='localhost:9092',
> 'format'='json',
> 'json.ignore-parse-errors' = 'true'
> );
> 
> 正常情况:
> Flink SQL> insert into kafka_sink select user_id,name,`timestamp` from orders;
> [INFO] Submitting SQL update statement to the cluster...
> [INFO] SQL update statement has been successfully submitted to the cluster:
> Job ID: e419ae9d2cad4c3c2a2c1150c1a86653
> 
> 
> 异常情况:
> Flink SQL> insert into kafka_sink(user_id,name,`timestamp`) select 
> user_id,name,`timestamp` from orders;
> [ERROR] Could not execute SQL statement. Reason:
> org.apache.calcite.sql.validate.SqlValidatorException: Unknown target column 
> 'timestamp'
> 很奇怪,为什么指定列名就不行了呢?而且还是识别不到”ts”列,kafka_sink schema如下:
> Flink SQL> describe kafka_sink;
> +---+--+--+-+---+---+
> |  name | type | null | key |extras | 
> watermark |
> +---+--+--+-+---+---+
> |   user_id |   BIGINT | TRUE | |   | 
>   |
> |  name |   STRING | TRUE | |   | 
>   |
> | timestamp | TIMESTAMP(3) | TRUE | | METADATA FROM 'timestamp' | 
>   |
> +---+--+--+-+---+---+
> 
> 
> 
> 恳请解答!



回复:关于Flink SQL无法对带metadata列的表在执行insert into时,指定具体的列名的问题。

2023-10-12 文章 tanjialiang
Hi, 
这个问题已经在1.17.0修复,详细可以看https://issues.apache.org/jira/browse/FLINK-30922


best wishes,
tanjialiang.


 回复的原邮件 
| 发件人 | jinzhuguang |
| 发送日期 | 2023年10月13日 10:39 |
| 收件人 | user-zh |
| 主题 | 关于Flink SQL无法对带metadata列的表在执行insert into时,指定具体的列名的问题。 |
首先,我的Flink版本为1.16.0
为了方便理解,我以Kafka作为案例来描述:
我有以下两个表:
CREATE TABLE orders(
user_id BIGINT,
name STRING,
timestamp TIMESTAMP(3) METADATA VIRTUAL
)WITH(
'connector'='kafka',
'topic'='orders',
'properties.group.id' = 'test_join_tempral',
'scan.startup.mode'='earliest-offset',
'properties.bootstrap.servers'='localhost:9092',
'format'='json',
'json.ignore-parse-errors' = 'true'
);
CREATE TABLE kafka_sink(
user_id BIGINT,
name STRING,
timestamp TIMESTAMP(3) METADATA FROM 'timestamp'
)WITH(
'connector'='kafka',
'topic'='kafka_sink',
'properties.group.id' = 'test_join_tempral',
'scan.startup.mode'='earliest-offset',
'properties.bootstrap.servers'='localhost:9092',
'format'='json',
'json.ignore-parse-errors' = 'true'
);

正常情况:
Flink SQL> insert into kafka_sink select user_id,name,`timestamp` from orders;
[INFO] Submitting SQL update statement to the cluster...
[INFO] SQL update statement has been successfully submitted to the cluster:
Job ID: e419ae9d2cad4c3c2a2c1150c1a86653


异常情况:
Flink SQL> insert into kafka_sink(user_id,name,`timestamp`) select 
user_id,name,`timestamp` from orders;
[ERROR] Could not execute SQL statement. Reason:
org.apache.calcite.sql.validate.SqlValidatorException: Unknown target column 
'timestamp'
很奇怪,为什么指定列名就不行了呢?而且还是识别不到”ts”列,kafka_sink schema如下:
Flink SQL> describe kafka_sink;
+---+--+--+-+---+---+
|  name | type | null | key |extras | watermark 
|
+---+--+--+-+---+---+
|   user_id |   BIGINT | TRUE | |   |   
|
|  name |   STRING | TRUE | |   |   
|
| timestamp | TIMESTAMP(3) | TRUE | | METADATA FROM 'timestamp' |   
|
+---+--+--+-+---+---+



恳请解答!

关于Flink SQL无法对带metadata列的表在执行insert into时,指定具体的列名的问题。

2023-10-12 文章 jinzhuguang
首先,我的Flink版本为1.16.0
为了方便理解,我以Kafka作为案例来描述:
我有以下两个表:
CREATE TABLE orders(
user_id BIGINT,
name STRING,
timestamp TIMESTAMP(3) METADATA VIRTUAL
)WITH(
'connector'='kafka',
'topic'='orders',
'properties.group.id' = 'test_join_tempral',
'scan.startup.mode'='earliest-offset',
'properties.bootstrap.servers'='localhost:9092',
'format'='json',
'json.ignore-parse-errors' = 'true'
);
CREATE TABLE kafka_sink(
user_id BIGINT,
name STRING,
timestamp TIMESTAMP(3) METADATA FROM 'timestamp'
)WITH(
'connector'='kafka',
'topic'='kafka_sink',
'properties.group.id' = 'test_join_tempral',
'scan.startup.mode'='earliest-offset',
'properties.bootstrap.servers'='localhost:9092',
'format'='json',
'json.ignore-parse-errors' = 'true'
);

正常情况: 
Flink SQL> insert into kafka_sink select user_id,name,`timestamp` from orders;
[INFO] Submitting SQL update statement to the cluster...
[INFO] SQL update statement has been successfully submitted to the cluster:
Job ID: e419ae9d2cad4c3c2a2c1150c1a86653


异常情况:
Flink SQL> insert into kafka_sink(user_id,name,`timestamp`) select 
user_id,name,`timestamp` from orders;
[ERROR] Could not execute SQL statement. Reason:
org.apache.calcite.sql.validate.SqlValidatorException: Unknown target column 
'timestamp'
很奇怪,为什么指定列名就不行了呢?而且还是识别不到”ts”列,kafka_sink schema如下:
Flink SQL> describe kafka_sink;
+---+--+--+-+---+---+
|  name | type | null | key |extras | watermark 
|
+---+--+--+-+---+---+
|   user_id |   BIGINT | TRUE | |   |   
|
|  name |   STRING | TRUE | |   |   
|
| timestamp | TIMESTAMP(3) | TRUE | | METADATA FROM 'timestamp' |   
|
+---+--+--+-+---+---+



恳请解答!

回复: Flink SQL的状态清理

2023-10-09 文章 小昌同学
你好,老师,我也是这样设置的,我这边是flink sql client,但是我去flink web ui界面并没有看到这个配置生效。


| |
小昌同学
|
|
ccc0606fight...@163.com
|
 回复的原邮件 
| 发件人 | Jane Chan |
| 发送日期 | 2023年9月25日 11:24 |
| 收件人 |  |
| 主题 | Re: Flink SQL的状态清理 |
Hi,

可以通过设置 table.exec.state.ttl 来控制状态算子的 state TTL. 更多信息请参阅 [1]

[1]
https://nightlies.apache.org/flink/flink-docs-master/zh/docs/dev/table/concepts/overview/#%e7%8a%b6%e6%80%81%e7%ae%a1%e7%90%86

Best,
Jane

On Thu, Sep 21, 2023 at 5:17 PM faronzz  wrote:

试试这个 t_env.get_config().set("table.exec.state.ttl", "86400 s")




| |
faronzz
|
|
faro...@163.com
|


 回复的原邮件 
| 发件人 | 小昌同学 |
| 发送日期 | 2023年09月21日 17:06 |
| 收件人 | user-zh |
| 主题 | Flink SQL的状态清理 |


各位老师好,请教一下大家关于flink sql的状态清理问题,我百度的话只找到相关的minbath设置,sql是没有配置state的ttl设置嘛
| |
小昌同学
|
|
ccc0606fight...@163.com
|


toolongframeexception

2023-10-08 文章 Jiacheng Jiang
大家好:

   我搭建了一个单机standealone 的flink用来测试mysql 
CDC,我发现每天晚上4点和4.30,我的taskmanager和jobmanager日志都有akka的TooLongFrameException,是BlobServerConnection报出来的,那个时候可能有定时任务在操作大量操作mysql数据库,我想问:

  1.  TooLongFrameException可能是mysql操作比较大造成的吗?如果是,有解决方法吗?
  2.  BlobServer是干嘛的?我一直认为blobserver是用来放job jar之类的,应该和cdc job的数据无关吧
  3.  哪些情况可能造成TooLongFrameException?

感谢大家


回复:flink两阶段提交

2023-10-08 文章 海风
多谢啦



 回复的原邮件 
| 发件人 | Feng Jin |
| 日期 | 2023年10月08日 13:17 |
| 收件人 | user-zh@flink.apache.org |
| 抄送至 | |
| 主题 | Re: flink两阶段提交 |
hi,

可以参考这篇博客,描述的非常清晰:
https://flink.apache.org/2018/02/28/an-overview-of-end-to-end-exactly-once-processing-in-apache-flink-with-apache-kafka-too/


Best,
Feng

On Sun, Sep 24, 2023 at 9:54 PM 海风 <18751805...@163.com> wrote:

> 请教一下,flink的两阶段提交对于sink算子,预提交是在做检查点的哪个阶段触发的?预提交时具体是做了什么工作?
>
>
>


Re: flink两阶段提交

2023-10-07 文章 Feng Jin
hi,

可以参考这篇博客,描述的非常清晰:
https://flink.apache.org/2018/02/28/an-overview-of-end-to-end-exactly-once-processing-in-apache-flink-with-apache-kafka-too/


Best,
Feng

On Sun, Sep 24, 2023 at 9:54 PM 海风 <18751805...@163.com> wrote:

> 请教一下,flink的两阶段提交对于sink算子,预提交是在做检查点的哪个阶段触发的?预提交时具体是做了什么工作?
>
>
>


Re: Flink CDC消费Apache Paimon表

2023-10-07 文章 Feng Jin
hi casel

Flink 实时消费 paimon,默认情况就是全量 + 增量的方式。

具体可以参考:  https://paimon.apache.org/docs/master/maintenance/configurations/
中的 scan.mode 参数


best,
Feng

On Fri, Sep 29, 2023 at 5:50 PM casel.chen  wrote:

> 目前想要通过Flink全量+增量消费Apache Paimon表需要分别起离线和增量消费两个作业,比较麻烦,而且无法无缝衔接,能否通过类似Flink
> CDC消费mysql表的方式消费Apache Paimon表?


Re: 退订

2023-10-06 文章 Yunfeng Zhou
Hi,

请发送任意内容的邮件到 user-zh-unsubscr...@flink.apache.org 来取消订阅邮件。

Best,
Yunfeng

On Wed, Oct 4, 2023 at 10:07 AM 1  wrote:
>
>


退订

2023-10-03 文章 1



Re: Cannot find metata file metadats in directory

2023-09-30 文章 Hangxiang Yu
Hi,
How did you point out the checkpoint path you restored from ?

Seems that you are trying to restore from a not completed or failed
checkpoint.

On Thu, Sep 28, 2023 at 6:09 PM rui chen  wrote:

> When we use 1.13.2,we have the following error:
> FileNotFoundException: Cannot find metata file metadats in directory
> 'hdfs://xx/f408dbe327f9e5053e76d7b5323d6e81/chk-173'.
>


-- 
Best,
Hangxiang.


Flink CDC消费Apache Paimon表

2023-09-29 文章 casel.chen
目前想要通过Flink全量+增量消费Apache Paimon表需要分别起离线和增量消费两个作业,比较麻烦,而且无法无缝衔接,能否通过类似Flink 
CDC消费mysql表的方式消费Apache Paimon表?

Cannot find metata file metadats in directory

2023-09-28 文章 rui chen
When we use 1.13.2,we have the following error:
FileNotFoundException: Cannot find metata file metadats in directory
'hdfs://xx/f408dbe327f9e5053e76d7b5323d6e81/chk-173'.


回复: 退订

2023-09-26 文章 Chen Zhanghao
Hi,

请发送任意内容的邮件到 user-zh-unsubscr...@flink.apache.org 来取消订阅邮件。

Best,
Zhanghao Chen

发件人: guangyu05 
发送时间: 2023年9月26日 18:45
收件人: user-zh@flink.apache.org 
主题: 退订

退订


退订

2023-09-26 文章 guangyu05
退订

Re: 退订

2023-09-26 文章 Xuannan Su
Hi,

请发送任意内容的邮件到 user-zh-unsubscr...@flink.apache.org 
 地址来取消订阅来自
user-zh@flink.apache.org  
mailto:u...@flink.apache.org>> 邮件组的邮件,你可以参考[1][2]
管理你的邮件订阅。
Please send email to user-zh-unsubscr...@flink.apache.org 
 if you want to
unsubscribe the mail from user-zh@flink.apache.org 
 mailto:u...@flink.apache.org>>,
and you can refer [1][2] for more details.

Best,
Xuannan

[1] https://flink.apache.org/zh/community/#%e9%82%ae%e4%bb%b6%e5%88%97%e8%a1%a8
[2] https://flink.apache.org/community.html#mailing-lists

> On Sep 26, 2023, at 16:44, 王唯  wrote:
> 
> 退订



Re: 退订

2023-09-26 文章 Xuannan Su
Hi,

请发送任意内容的邮件到 user-zh-unsubscr...@flink.apache.org 
 地址来取消订阅来自
user-zh@flink.apache.org  
mailto:u...@flink.apache.org>> 邮件组的邮件,你可以参考[1][2]
管理你的邮件订阅。
Please send email to user-zh-unsubscr...@flink.apache.org 
 if you want to
unsubscribe the mail from user-zh@flink.apache.org 
 mailto:u...@flink.apache.org>>,
and you can refer [1][2] for more details.

Best,
Xuannan

[1] https://flink.apache.org/zh/community/#%e9%82%ae%e4%bb%b6%e5%88%97%e8%a1%a8
[2] https://flink.apache.org/community.html#mailing-lists

> On Sep 26, 2023, at 14:43, 陈建华  wrote:
> 
> 退订



Re: 退订

2023-09-26 文章 Xuannan Su
Hi,

请发送任意内容的邮件到 user-zh-unsubscr...@flink.apache.org 
 地址来取消订阅来自
user-zh@flink.apache.org  
mailto:u...@flink.apache.org>> 邮件组的邮件,你可以参考[1][2]
管理你的邮件订阅。
Please send email to user-zh-unsubscr...@flink.apache.org 
 if you want to
unsubscribe the mail from user-zh@flink.apache.org 
 mailto:u...@flink.apache.org>>,
and you can refer [1][2] for more details.

Best,
Xuannan

[1] https://flink.apache.org/zh/community/#%e9%82%ae%e4%bb%b6%e5%88%97%e8%a1%a8
[2] https://flink.apache.org/community.html#mailing-lists

> On Sep 26, 2023, at 14:37, 悠舒 刘  wrote:
> 
> 退订



退订

2023-09-26 文章 王唯
退订

退订

2023-09-26 文章 陈建华
退订

退订

2023-09-26 文章 悠舒 刘
退订


<    1   2   3   4   5   6   7   8   9   10   >