Re: Difference between Session Mode and Session Job(Flink Opearator)

2022-07-07 Thread bat man
k job. > > I hope that helps :) > Gyula > > On Thu, Jul 7, 2022 at 7:42 AM bat man wrote: > >> Hi, >> >> I want to understand the difference between session mode and the new >> deployment mode - Flink Session Job which I believe is newly introduced as >

Difference between Session Mode and Session Job(Flink Opearator)

2022-07-06 Thread bat man
Hi, I want to understand the difference between session mode and the new deployment mode - Flink Session Job which I believe is newly introduced as part of the Flink Operator(1.15) release. What's the benefit of using this mode as opposed to session mode as both run sessions to which flink jobs

Re: context.timestamp null in keyedprocess function

2022-06-15 Thread bat man
Has anyone experienced this or has any clue? On Tue, Jun 14, 2022 at 6:21 PM bat man wrote: > Hi, > > We are using flink 12.1 on AWS EMR. The job reads the event stream and > enrich stream from another topic. > We extend AssignerWithPeriodicWatermarks to assign waterma

context.timestamp null in keyedprocess function

2022-06-14 Thread bat man
Hi, We are using flink 12.1 on AWS EMR. The job reads the event stream and enrich stream from another topic. We extend AssignerWithPeriodicWatermarks to assign watermarks and extract timestamp from the event and handle idle source partitions. AutoWatermarkInterval set to 5000L. The timestamp

Re: Flink on Native Kubernetes S3 checkpointing error

2021-11-22 Thread bat man
On Mon, Nov 22, 2021 at 7:26 PM Matthias Pohl wrote: > Hi bat man, > this feature seems to be tied to a certain AWS SDK version [1] which you > already considered. But I checked the version used in Flink 1.13.1 for the > s3 filesystem. It seems like the version that's used (1.11.

Flink on Native Kubernetes S3 checkpointing error

2021-11-21 Thread bat man
Hi, I am using flink 1.13.1 to use checkpointing(RocksDB) on s3 with native kubernetes. Passing in this parameter to job - *-Dfs.s3a.aws.credentials.provider=com.amazonaws.auth.WebIdentityTokenCredentialsProvider* I am getting this error in job-manager logs - *Caused by:

Re: Flink S3 Presto Checkpointing Permission Forbidden

2021-11-21 Thread bat man
Hi Dennis, Were you able to use checkpointing on s3 with native kubernetes. I am using flink 1.13.1 and did tried your solution of passing the webidentitytokencredentialsprovider. *-Dfs.s3a.aws.credentials.provider=com.amazonaws.auth.WebIdentityTokenCredentialsProvider* I am getting this error

Re: Flink Native Kubernetes - Configuration kubernetes.flink.log.dir not working

2021-09-18 Thread bat man
i Ma 于2021年9月14日周二 下午4:48写道: > >> Hi >> >> Maybe you could try the `kubectl describe pod -n ${namespace} >> ${podname}` to see what happened atm. >> >> Best, >> Guowei >> >> >> On Tue, Sep 14, 2021 at 2:58 PM bat man wrote: >> >>

Re: Flink Native Kubernetes - Configuration kubernetes.flink.log.dir not working

2021-09-14 Thread bat man
; > On Mon, Sep 13, 2021 at 10:59 PM bat man wrote: > >> Hi, >> >> I am running a POC to evaluate Flink on Native Kubernetes. I tried >> changing the default log location by using the configuration - >> kubernetes.flink.log.dir >> However, the j

Flink Native Kubernetes - Configuration kubernetes.flink.log.dir not working

2021-09-13 Thread bat man
Hi, I am running a POC to evaluate Flink on Native Kubernetes. I tried changing the default log location by using the configuration - kubernetes.flink.log.dir However, the job in application mode fails after bringing up the task manager. This is the command I use - ./bin/flink run-application

KafkaFetcher [] - Committing offsets to Kafka failed.

2021-08-26 Thread bat man
Hi, I am using flink 12.1 to consume data from kafka in a streaming job. Using the flink-connector-kafka_2.12:1.12.1. Kafka broker version is 2.2.1 In logs I see warnings like this - 2021-08-26 13:36:49,903 WARN org.apache.flink.streaming.connectors.kafka.internals.KafkaFetcher [] - Committing

Re: High DirectByteBuffer Usage

2021-07-15 Thread bat man
e/KAFKA-4090 > ------ > *From:* bat man > *Sent:* Wednesday, July 14, 2021 10:55:54 AM > *To:* Timo Walther > *Cc:* user > *Subject:* Re: High DirectByteBuffer Usage > > Hi Timo, > > I am looking at these options. > However, I had a couple of questions - &

Re: High DirectByteBuffer Usage

2021-07-14 Thread bat man
ble/#container-memory-exceeded > > It is likely that the high number of output streams could cause your > issues. > > Regards, > Timo > > > > > On 14.07.21 08:46, bat man wrote: > > Hi, > > I have a job which reads different streams from 5 kafka topics. It > >

Re: Flink 1.13 Native Kubernetes - Custom Pod Templates

2021-07-13 Thread bat man
; https://ci.apache.org/projects/flink/flink-docs-master/docs/deployment/config/#kubernetes > [2]. > https://ci.apache.org/projects/flink/flink-docs-master/docs/deployment/resource-providers/native_kubernetes/#fields-overwritten-by-flink > [3]. > https://ci.apache.org/projects/flink/flink

Re: Flink 1.13 Native Kubernetes - Custom Pod Templates

2021-07-13 Thread bat man
flink-docs-master/docs/deployment/resource-providers/native_kubernetes/#fields-overwritten-by-flink > [3]. > https://ci.apache.org/projects/flink/flink-docs-master/docs/deployment/metric_reporters/#prometheus > > Best, > Yang > > > bat man 于2021年7月11日周日 下午5:27写道: > >&

Flink 1.13 Native Kubernetes - Custom Pod Templates

2021-07-11 Thread bat man
Hi, I am running a POC to run flink 1.13 on Native Kubernetes. Per documentation [1] this supports custom pod templates. Are there any references for sample pod templates which can be used as a baseline to build upon. Could not find any references, documentation[1] has one sample for

Re: Memory Usage - Total Memory Usage on UI and Metric

2021-07-03 Thread bat man
erties": { > "yarn.nodemanager.pmem-check-enabled": "false", > "yarn.nodemanager.vmem-check-enabled": "false" > } > } > ] > > > …might help. > > — Ken > > > On Jul 2, 2021, at 8:36 AM, bat man wrote: > &g

Memory Usage - Total Memory Usage on UI and Metric

2021-07-02 Thread bat man
Hi, I am running a streaming job (Flink 1.9) on EMR on yarn. Flink web UI or metrics reported from prometheus shows total memory usage within specified task manager memory - 3GB. Metrics shows below numbers(in MB) - Heap - 577 Non Heap - 241 DirectMemoryUsed - 852 Non-heap does rise gradually,

Re: Read kafka offsets from checkpoint - state processor

2021-05-07 Thread bat man
Anyone who has tried this or can help on this. Thanks. On Thu, May 6, 2021 at 10:34 AM bat man wrote: > Hi Users, > > Is there a way that Flink 1.9 the checkpointed data can be read using the > state processor api. > Docs [1] says - When reading operator state, users specify the

Read kafka offsets from checkpoint - state processor

2021-05-05 Thread bat man
Hi Users, Is there a way that Flink 1.9 the checkpointed data can be read using the state processor api. Docs [1] says - When reading operator state, users specify the operator uid, the state name, and the type information. What is the type for the kafka operator, which needs to be specified

Re: Get consumed Kafka offsets from Flink kafka source

2021-04-14 Thread bat man
L and use SQL to query/analyse the state. > > Best > Piotrek > > [1] > https://ci.apache.org/projects/flink/flink-docs-stable/dev/libs/state_processor_api.html > > śr., 14 kwi 2021 o 11:25 bat man napisał(a): > >> Hi All, >> >> Is there any way I can in

Get consumed Kafka offsets from Flink kafka source

2021-04-14 Thread bat man
Hi All, Is there any way I can inspect/query the checkpointed data. Scenario is like this - We have a high volume of data coming in the data stream pipeline for which kafka is source, in case if fails bcoz of bad data I want to analyse the data which caused the issue. It could be that some data

Re: flink cdc postgres connector 1.1 - heartbeat.interval.ms - WAL consumption

2021-04-08 Thread bat man
ther the fix can be backported. > > Cheers, > Till > > On Thu, Apr 8, 2021 at 10:42 AM bat man wrote: > >> Anyone who has faced similar issues with cdc with Postgres. >> >> I see the restart_lsn and confirmed_flush_lsn constant since the snapshot >> repl

Re: flink cdc postgres connector 1.1 - heartbeat.interval.ms - WAL consumption

2021-04-08 Thread bat man
Thanks. On Wed, Apr 7, 2021 at 12:51 PM bat man wrote: > Hi there, > > I am using flink 1.11 and cdc connector 1.1 to stream changes from a > postgres table. I see the WAL consumption is increasing gradually even > though the writes to tables are very less. > > I am usin

flink cdc postgres connector 1.1 - heartbeat.interval.ms - WAL consumption

2021-04-07 Thread bat man
Hi there, I am using flink 1.11 and cdc connector 1.1 to stream changes from a postgres table. I see the WAL consumption is increasing gradually even though the writes to tables are very less. I am using AWS RDS, from [1] I understand that setting the parameter heartbeat.interval.ms solves this

io.netty.channel.epoll.EpollEventLoopGroup cannot be cast to io.netty.channel.MultithreadEventLoopGroup

2021-03-10 Thread bat man
Hi, I am using this library -* jasync-postgresql* [1] for async calls to postgres in asyncio operator. I am using running Flink 1.9 on EMR. I am facing this error - Caused by: java.lang.ClassCastException: io.netty.channel.epoll.EpollEventLoopGroup cannot be cast to

Re: java options to generate heap dump in EMR not working

2021-03-08 Thread bat man
acter for the configuration ? > > For the case of configuring `env.java.opts`, it seems the JobManager also > could not be launched with > this configuration. > > Best, > Yun > > ----------Original Mail -- > *Sender:*bat man > *Send Date:*Sat Mar 6

Re: java.lang.OutOfMemoryError: GC overhead limit exceeded

2021-03-08 Thread bat man
Mar 2021 at 7:48 AM, Xintong Song wrote: > Hi Hemant, > I don't see any problem in your settings. Any exceptions suggesting why TM > containers are not coming up? > > Thank you~ > > Xintong Song > > > > On Sat, Mar 6, 2021 at 3:53 PM bat man wrote: > >>

java options to generate heap dump in EMR not working

2021-03-06 Thread bat man
Hi, I am trying to generate a heap dump to debug a GC overhead OOM. For that I added the below java options in flink-conf.yaml, however after adding this the yarn is not able to launch the containers. The job logs show it goes on requesting for containers from yarn and it gets them, again

Re: java.lang.OutOfMemoryError: GC overhead limit exceeded

2021-03-05 Thread bat man
t; memory usages leveraging profiling tools. > > Thank you~ > > Xintong Song > > > [1] > https://docs.oracle.com/javase/8/docs/technotes/guides/troubleshoot/memleaks002.html > > [2] > https://ci.apache.org/projects/flink/flink-docs-release-1.12/deployment/memory/m

java.lang.OutOfMemoryError: GC overhead limit exceeded

2021-03-05 Thread bat man
Hi, Getting the below OOM but the job failed 4-5 times and recovered from there. j *ava.lang.Exception: java.lang.OutOfMemoryError: GC overhead limit exceededat

Re: Watermark doesn't progress after job restore from savepoint

2021-03-04 Thread bat man
gt; its state, and re-emits it upon recovery. > > Best, > Piotrek > > czw., 4 mar 2021 o 15:43 bat man napisał(a): > >> Hi All, >> >> I have a job where my source is kafka. Stream1 is partition the data on >> dynamic key, join the data with static rules(source ka

Watermark doesn't progress after job restore from savepoint

2021-03-04 Thread bat man
Hi All, I have a job where my source is kafka. Stream1 is partition the data on dynamic key, join the data with static rules(source kafka).I use KeyedCoProcessFunction to join the Steam1 with Stream2(source kafka). All works fine in a normal run. For changing the watermark generation interval I

Re: BroadcastState dropped when data deleted in Kafka

2021-03-03 Thread bat man
uot;).uid("rt").setParallelism(5); On Fri, Feb 26, 2021 at 3:38 PM Arvid Heise wrote: > Hi, > > I have no idea what's going on. There is no mechanism in DataStream to > react to deleted records. > > Can you reproduce it locally and debug through it? > > > > O

Re: BroadcastState dropped when data deleted in Kafka

2021-02-24 Thread bat man
cess function [1]. If you have done that, I'm a bit lost. > > [1] > https://ci.apache.org/projects/flink/flink-docs-stable/dev/stream/state/broadcast_state.html > > On Wed, Feb 24, 2021 at 7:30 AM bat man wrote: > >> Hi, >> >> This is my code below - >>

Re: BroadcastState dropped when data deleted in Kafka

2021-02-23 Thread bat man
ka shouldn't affect Flink state in general. > Probably, some operator in your pipeline is re-reading the topic > and overwrites the state, dropping what was deleted by Kafka. > Could you share the code? > > Regards, > Roman > > > On Tue, Feb 23, 2021 at 7:12 AM bat man wr

BroadcastState dropped when data deleted in Kafka

2021-02-22 Thread bat man
Hi, I have 2 streams one event data and the other rules. I broadcast the rules stream and then key the data stream on event type. The connected stream is processed thereafter. We faced an issue where the rules data in the topic got deleted because of Kafka retention policy. Post this the existing

Re: Tag flink metrics to job name

2021-02-19 Thread bat man
we don't differentiate between the two when it > comes to metrics. > > On 2/19/2021 3:59 AM, bat man wrote: > > I meant the Flink jobname. I’m using the below reporter - > > metrics.reporter.prom.class: > org.apache.flink.metrics.prometheus.PrometheusReporter > > Is there any w

Re: Tag flink metrics to job name

2021-02-18 Thread bat man
hen you mean "job_name", are you referring to the Prometheus concept of > jobs, of the one of Flink? > > Which of Flink prometheus reporters are you using? > > On 2/17/2021 7:37 PM, bat man wrote: > > Hello there, > > > > I am using prometheus to push

Tag flink metrics to job name

2021-02-17 Thread bat man
Hello there, I am using prometheus to push metrics to prometheus and then use grafana for visualization. There are metrics like - flink_taskmanager_Status_JVM_CPU_Load, flink_taskmanager_Status_JVM_CPU_Load, flink_taskmanager_Status_JVM_CPU_Time etc which do not gives job_name. It is tied to an

Re: GC overhead limit exceeded when using Prometheus exporter

2021-02-16 Thread bat man
ant, > > Have you tried running a new Flink version? Can you create a heap dump > when the process fails? This could help us digging into whether there is > some memory leak. > > Cheers, > Till > > On Tue, Feb 16, 2021 at 5:21 PM bat man wrote: &g

GC overhead limit exceeded when using Prometheus exporter

2021-02-16 Thread bat man
Hi there, I am facing *java.lang.OutOfMemoryError: GC overhead limit exceeded *when using prometheus exporter with* Flink 1.9 *on *AWS EMR *emr-5.28.1. I have other jobs which run fine. tihs specific job fails with the below error stack. Exception in thread "pool-3-thread-2"

Re: Flink app logs to Elastic Search

2021-01-14 Thread bat man
I was able to make it work with a fresh Elastic installation. Now taskmanager and jobmanager logs are available in elastic. Thanks for the pointers. -Hemant. On Wed, Jan 13, 2021 at 6:21 PM Aljoscha Krettek wrote: > On 2021/01/11 01:29, bat man wrote: > >Yes, no entries to the elast

Re: Main class logs in Yarn Mode

2021-01-12 Thread bat man
be located in your console output and the "log/" > directory of your Flink distribution. > > Best, > Yangze Guo > > On Tue, Jan 12, 2021 at 4:50 PM bat man wrote: > > > > Hi, > > > > I am running a sample job as below - > > > >

Main class logs in Yarn Mode

2021-01-12 Thread bat man
Hi, I am running a sample job as below - public class WordCountExample { static Logger logger = LoggerFactory.getLogger(WordCountExample.class); public static void main(String[] args) throws Exception { final ExecutionEnvironment env = ExecutionEnvironment. getExecutionEnvironment();

Flink app logs to Elastic Search

2021-01-07 Thread bat man
Hi Team, I have a requirement to push the flink app logs to Elastic Search for log management. Can anyone guide if you are already doing this. I have tried this - https://cristian.io/post/flink-log4j/ I’m not getting any error for a sample job I tried. I am using EMR to run Flink 1.9 and Elastic

Re: Pushing metrics to Influx from Flink 1.9 on AWS EMR(5.28)

2020-08-14 Thread bat man
influxdb.username: > flink-metricsmetrics.reporter.influxdb.password: > qwertymetrics.reporter.influxdb.retentionPolicy: one_hour > > > [1] > https://ci.apache.org/projects/flink/flink-docs-release-1.9/monitoring/metrics.html#influxdb-orgapacheflinkmetricsinfluxdbinfluxdbreporter > > On Thu, Aug

Re: Pushing metrics to Influx from Flink 1.9 on AWS EMR(5.28)

2020-08-13 Thread bat man
Anyone who has made metrics integration to external systems for flink running on AWS EMR, can you share if its a configuration issue or EMR specific issue. Thanks, Hemant On Wed, Aug 12, 2020 at 9:55 PM bat man wrote: > An update in the yarn logs I could see the below - > > Classpat

Re: Pushing metrics to Influx from Flink 1.9 on AWS EMR(5.28)

2020-08-12 Thread bat man
reported but then org.apache.flink.runtime.metrics.ReporterSetup complains of not finding it. Can anyone guide what I am missing here. Thanks, Hemant On Wed, Aug 12, 2020 at 9:15 PM bat man wrote: > Hello Experts, > > I am running Flink - 1.9.0 on AWS EMR(emr-5.28.1). I want to push

Pushing metrics to Influx from Flink 1.9 on AWS EMR(5.28)

2020-08-12 Thread bat man
Hello Experts, I am running Flink - 1.9.0 on AWS EMR(emr-5.28.1). I want to push metrics to Influxdb. I followed the documentation[1]. I added the configuration to /usr/lib/flink/conf/flink-conf.yaml and copied the jar to /usr/lib/flink//lib folder on master node. However, I also understand that

Re: Handle idle kafka source in Flink 1.9

2020-08-05 Thread bat man
ttps://github.com/ververica/flink-training-troubleshooting/blob/master/src/main/java/com/ververica/flinktraining/solutions/troubleshoot/TroubledStreamingJobSolution2.java#L129-L187 > > On Thu, Jul 23, 2020 at 8:48 PM bat man wrote: > >> Thanks Niels for a great talk. You have covered

Re: Handle idle kafka source in Flink 1.9

2020-07-23 Thread bat man
it to the next stream(next kafka topic) after enrichment. Have you got any solution for this? -Hemant On Thu, Jul 23, 2020 at 12:05 PM Niels Basjes wrote: > Have a look at this presentation I gave a few weeks ago. > https://youtu.be/bQmz7JOmE_4 > > Niels Basjes > > On Wed, 22 Jul 2

Re: Handle idle kafka source in Flink 1.9

2020-07-22 Thread bat man
Hi Team, Can someone share their experiences handling this. Thanks. On Tue, Jul 21, 2020 at 11:30 AM bat man wrote: > Hello, > > I have a pipeline which consumes data from a Kafka source. Since, the > partitions are partitioned by device_id in case a group of devices is down >

Handle idle kafka source in Flink 1.9

2020-07-21 Thread bat man
Hello, I have a pipeline which consumes data from a Kafka source. Since, the partitions are partitioned by device_id in case a group of devices is down some partitions will not get normal flow of data. I understand from documentation here[1] in flink 1.11 one can declare the source idle -