Re: [Flink-1.14.3] Restart of pod due to duplicatejob submission

2022-02-25 Thread Parag Somani
it would be equivalent to that of k8s based resiliency. On Fri, Feb 25, 2022 at 11:55 AM Yang Wang wrote: > This might be related with FLINK-21928 and seems already fixed in 1.14.0. > But it will have some limitations and users need to manually clean up the > HA entries. > > > Best

[Flink-1.14.3] Restart of pod due to duplicatejob submission

2022-02-23 Thread Parag Somani
Hello, Recently due to log4j vulnerabilities, we have upgraded to Apache Flink 1.14.3. What we observed we are getting following exception, and because of it pod gets in crashloopback. We have seen this issues esp. during the time of upgrade or deployment time when existing pod is already

Duplicate job submission error

2022-01-27 Thread Parag Somani
Hello All, While deploying on our one of environment, we encountered crashloopback of job manager pod. Env: K8s Flink: 1.14.2 Could you suggest, how can we troubleshoot this and possible handling of this? exception snipper as follows: 2022-01-27 06:58:07.326 ERROR 44 --- [lt-dispatcher-4]

Re: CVE-2021-44228 - Log4j2 vulnerability

2021-12-15 Thread Parag Somani
Thank you Chesnay for expediting this fix...! Can you suggest, when can I get binaries for 1.14.2 flink version? On Thu, Dec 16, 2021 at 5:52 AM Chesnay Schepler wrote: > We will push docker images for all new releases, yes. > > On 16/12/2021 01:16, Michael Guterl wrote: > > Will you all be

Re: s3 access denied with flink-s3-fs-presto

2021-10-26 Thread Parag Somani
Hello, I have successfully been able to store data on S3 bucket. Earlier, I used to have a similar issue. What you need to confirm: 1. S3 bucket is created with RW access(irrespective if it is minio or AWS S3) 2. "flink/opt/flink-s3-fs-presto-1.14.0.jar" jar is copied to plugin directory of

Re: Troubleshooting checkpoint timeout

2021-10-20 Thread Parag Somani
I had similar problem, where i have concurrent two checkpoints were configured. Also, i used to save it in S3(using minio) on k8s 1.18 env. Flink service were getting restarted and timeout was happening. It got resolved: 1. As minio ran out of disk space, caused failure of checkpoints(this was

Re: Issues while upgrading from 1.12.1 to 1.14.0

2021-10-06 Thread Parag Somani
tup you are using, but did you check out > [1]? Maybe the parameter > [--fromSavepoint /path/to/savepoint [--allowNonRestoredState]] > is what you are looking for? > > Best regards, > Nico > > [1] > https://ci.apache.org/projects/flink/flink-docs-master/docs/deployment/resource-pro

Issues while upgrading from 1.12.1 to 1.14.0

2021-10-05 Thread Parag Somani
Hello, We are currently using Apache flink 1.12.0 deployed on k8s cluster of 1.18 with zk for HA. Due to certain vulnerabilities in container related with few jar(like netty-*, meso), we are forced to upgrade. While upgrading flink to 1.14.0, faced NPE,