Re: Metrics & Monitoring in Flink SQL

2022-07-12 Thread yuxia
With Flink SQL, You can define your own source/sink metrics [1], but you can't define the metrics for the intermediate operators. [1] https://nightlies.apache.org/flink/flink-docs-master/docs/dev/table/sourcessinks/ Best regards, Yuxia 发件人: "Salva Alcántara" 收件人: "User" 发送时间: 星期三, 20

Re: Parsing a JSON array string as a Flink SQL Array data type

2022-07-12 Thread yuxia
I'm afraid of there's no build-in function on the hand. But you can write a UDF[1] to convert the JSON array string to Flink's array. [1] https://nightlies.apache.org/flink/flink-docs-master/docs/dev/table/functions/udfs/ Best regards, Yuxia - 原始邮件 - 发件人: "Abhishek Rai" 收件人: "User" , "

Parsing a JSON array string as a Flink SQL Array data type

2022-07-12 Thread Abhishek Rai
Hello! I'm trying to use the new JSON functions in Flink 1.15 to parse JSON input data. In particular, using JSON_QUERY, I'm able to extract out JSON array elements from a larger JSON record. However, this function returns the JSON array as a string. I'd like to run this array through the SQL U

Metrics & Monitoring in Flink SQL

2022-07-12 Thread Salva Alcántara
I have a question regarding Flink SQL, which I'm lately getting into. So far, my experience is with the DataStream API mostly. In that context, it's easy for me to generate metrics for my operators. However, I'm just wondering which level of control there is regarding monitoring & metrics when work

Re: Does Table API connector, csv, has some option to ignore some columns

2022-07-12 Thread yuxia
For Json format, you only need to define the parital columns to be selected in Flink DDL. But for csv format, it's not supported. In csv file, if there's no header, how can you mapping the incomplete columns defined in Flink DDL to the origin fields in the csv file? Thus, you need to write the

[ANNOUNCE] Apache Flink ML 2.1.0 released

2022-07-12 Thread Zhipeng Zhang
The Apache Flink community is excited to announce the release of Flink ML 2.1.0! This release focuses on improving Flink ML's infrastructure, such as Python SDK, memory management, and benchmark framework, to facilitate the development of performant, memory-safe, and easy-to-use algorithm librari

Re: Re: [ANNOUNCE] Apache Flink 1.15.1 released

2022-07-12 Thread Gabor Somogyi
In order to provide a hotfix please set "taskmanager.resource-id" to something which doesn't contain special any character. G On Tue, Jul 12, 2022 at 11:59 AM Gabor Somogyi wrote: > Flink tried to create the following dir: tm_localhost:50329-fc0146 > Colon is allowed on linux but not on window

Re: Re: [ANNOUNCE] Apache Flink 1.15.1 released

2022-07-12 Thread Gabor Somogyi
Flink tried to create the following dir: tm_localhost:50329-fc0146 Colon is allowed on linux but not on windows and that's the reason of the exception. BR, G On Tue, Jul 12, 2022 at 11:30 AM wrote: > ... > 2022-07-12 11:25:08,448 INFO > akka.remote.Remoting

Re: Re: [ANNOUNCE] Apache Flink 1.15.1 released

2022-07-12 Thread podunk
... 2022-07-12 11:25:08,448 INFO  akka.remote.Remoting [] - Remoting started; listening on addresses :[akka.tcp://flink@localhost:50329] 2022-07-12 11:25:08,658 INFO  org.apache.flink.runtime.rpc.akka.AkkaRpcServiceUtils    [] - Actor system started at a

Re: Re: Does Table API connector, csv, has some option to ignore some columns

2022-07-12 Thread podunk
This is really surprising. When you import data from a file, you really rarely need to import everything from that file. Most often it is several columns. So the program that reads the file should be able to do this - this is the ABC of working with data.   Often the suggestion is "you can w

Re: Data is lost in the ListState

2022-07-12 Thread Arthur Li
Thanks David, it’s clear now. In my case, the x2 will be recovered by the last checkpoint if it was processed and wait few seconds to be checkpointed Or my source can be rewind and it will replay the x2. > 2022年7月12日 09:10,David Anderson 写道: > > This is, in fact, the expected behavior. Let

DataSet API, chaining database access

2022-07-12 Thread bastien dine
Hello, I am struggling with DataSet API, I need to chain 2 db access, so I can easily do a source1 -> map2 The fact is that map should and can not be used for dbaccess, when the request is taking too long, we have timeout in akka connection between TM & JM I know we can not chain sources : source1