Custom metricReporterFactory could not be found in k8s deployment.

2024-03-22 Thread longfeng Xu
hi all, bg: create a custom metric reporter via kafka, it works in ideaj local environment. but failed when packaged and deployed in k8s env (ververica by alibaba) flink 1.12 config: metrics.reporter.kafka.factory.class: org.apache.flink.metrics.kafka.KafkaReporterFactory

Fwd: Global connection open and close

2024-03-22 Thread Alex Cruise
[previous didn't cc list, sorry for dupes] The classic connection pool pattern, where expensive connections are created relatively few times and used by lots of transient short-lived tasks, each of which borrows a connection from the pool and returns it when done, would still be usable here, but

Re: Global connection open and close

2024-03-22 Thread Alex Cruise
The classic connection pool pattern, where expensive connections are created relatively few times and used by lots of transient short-lived tasks, each of which borrows a connection from the pool and returns it when done, would still be usable here, but as Péter points out, you can't rely on a

Re: Understanding RocksDBStateBackend in Flink on Yarn on AWS EMR

2024-03-22 Thread Sachin Mittal
So, when we create an EMR cluster the NN service runs on the primary node of the cluster. Now at the time of creating the cluster, how can we specify the name of this NN in format hdfs://*namenode-host*:8020/. Is there a standard name by which we can identify the NN server ? Thanks Sachin On

Re: Flink unable to read from kafka source due to starting offset strategy

2024-03-22 Thread Sachin Mittal
Hi, After some debugging I see these in the logs: 2024-03-22 14:25:47,555 INFO org.apache.kafka.clients.NetworkClient [] - [Consumer clientId=qubit-data-consumer-0, groupId=spflink] Disconnecting from node 11 due to request timeout. 2024-03-22 14:25:47,647 INFO

Re:Re: Re: Read data from elasticsearch using Java flink

2024-03-22 Thread Xuyang
Hi, Nida. About using connector with Java by Flink DataStream api, you can mainly reference these doc[1][2]. However, Elaticsearch connector only supports sink currently. What you need it to build a custom ES connector with that patch[3] yourself. The following steps may help you. 1.

Re: Flink unable to read from kafka source due to starting offset strategy

2024-03-22 Thread Sachin Mittal
Hi, I was experimenting with different starting offset strategies for my Flink job, especially in cases where jobs are canceled and scheduled again and I would like to start with the last committed offset and if the same is not available then start from the latest. So I decided to use this:

Re: Understanding RocksDBStateBackend in Flink on Yarn on AWS EMR

2024-03-22 Thread Asimansu Bera
Hello Sachin, Typically, Cloud VMs are ephemeral, meaning that if the EMR cluster goes down or VMs are required to be shut down for security updates or due to faults, new VMs will be added to the cluster. As a result, any data stored in the local file system, such as file://tmp, would be lost. To