Flink InfluxDb connector not present in Maven

2021-04-17 Thread Vinay Patil
Hi Team, Flink influx db connector `flink-connector-influxdb_2.1` is not present in Maven , can you please upload the same https://repo.maven.apache.org/maven2/org/apache/bahir/ Regards, Vinay Patil

Re: Flink Deployment on Kubernetes session Cluster

2020-07-29 Thread Vinay Patil
and Regards, Vinay Patil On Wed, Jul 29, 2020 at 11:08 AM Yang Wang wrote: > Hi Vinay Patil, > > You are right. Flink does not provide any isolation between different jobs > in the same Flink session cluster. > You could use Flink job cluster or application cluster(from 1.11) to get >

Flink Deployment on Kubernetes session Cluster

2020-07-28 Thread Vinay Patil
Regards, Vinay Patil

Re: Timer metric in Flink

2020-06-11 Thread Vinay Patil
Ohh Okay, basically implement the Gauge and add timer functionality to it for now. Is there a plan or JIRA ticket to add Timer metric in future release, I think it is good to have Regards, Vinay Patil On Wed, Jun 10, 2020 at 5:55 PM Chesnay Schepler wrote: > You cannot add custom met

Timer metric in Flink

2020-06-10 Thread Vinay Patil
will also have to create Timer interface and add it to the metric group. Is this possible? I want to have a timer to check Hbase lookup time. Regards, Vinay Patil

Re: Handling stale data enrichment

2020-04-24 Thread Vinay Patil
dle late record). I think the downstream consumer of enriched data will have to dedup the duplicate records or else we will end up having stale enrichment. Regards, Vinay Patil On Fri, Apr 24, 2020 at 12:14 PM Konstantin Knauf wrote: > Hi Vinay, > > I assume your subscription updat

Handling stale data enrichment

2020-04-23 Thread Vinay Patil
if it is stale data or not based on timestamp (watermark) as it can happen that a particular enriched record is not updated for 6 hrs. Regards, Vinay Patil

Streaming File Sink - Parquet File Writer

2019-10-29 Thread Vinay Patil
RowFormatBuilder. P.S. Curious to know Why was the RollingPolicy not exposed in case of BulkFormat ? Regards, Vinay Patil

Re: Using STSAssumeRoleSessionCredentialsProvider for cross account access

2019-10-29 Thread Vinay Patil
Thanks Fabian, @Gordon - Can you please help here. Regards, Vinay Patil On Fri, Oct 25, 2019 at 9:11 PM Fabian Hueske wrote: > Hi Vinay, > > Maybe Gordon (in CC) has an idea about this issue. > > Best, Fabian > > Am Do., 24. Okt. 2019 um 14:50 Uhr schrieb Vinay

DynamoStreams Consumer millisBehindLatest metric

2019-10-28 Thread Vinay Patil
supports DynamoStreams Regards, Vinay Patil

Re: Using STSAssumeRoleSessionCredentialsProvider for cross account access

2019-10-24 Thread Vinay Patil
Hi, Can someone pls help here , facing issues in Prod . I see the following ticket in unresolved state. https://issues.apache.org/jira/browse/FLINK-8417 Regards, Vinay Patil On Thu, Oct 24, 2019 at 11:01 AM Vinay Patil wrote: > Hi, > > I am trying to access dynamo streams from a

Using STSAssumeRoleSessionCredentialsProvider for cross account access

2019-10-23 Thread Vinay Patil
assume that the credentials are not required to be passed : https://github.com/apache/flink/blob/abbd6b02d743486f3c0c1336139dd6b3edd20840/flink-connectors/flink-connector-kinesis/src/main/java/org/apache/flink/streaming/connectors/kinesis/util/AWSUtil.java#L164 Regards, Vinay Patil

Re: Consuming data from dynamoDB streams to flink

2019-08-08 Thread Vinay Patil
Hello, For anyone looking for setting up alerts for flink application ,here is good blog by Flink itself : https://www.ververica.com/blog/monitoring-apache-flink-applications-101 So, for dynamoDb streams we can set the alert on millisBehindLatest Regards, Vinay Patil On Wed, Aug 7, 2019 at 2

Re: Consuming data from dynamoDB streams to flink

2019-08-07 Thread Vinay Patil
the consumer is lagging behind. Regards, Vinay Patil On Fri, Jul 19, 2019 at 10:40 PM Andrey Zagrebin wrote: > Hi Vinay, > > 1. I would assume it works similar to kinesis connector (correct me if > wrong, people who actually developed it) > 2. If you have activated jus

Re: StackOverflow Error

2019-07-21 Thread Vinay Patil
Hi Ravi, The uber jar was correct, adding ClosureCleanerLevel to TOP_LEVEL resolved this issue. Thanks a lot. Is there any disadvantage of explicitly setting this ? Regards, Vinay Patil On Sat, Jul 20, 2019 at 10:23 PM Ravi Bhushan Ratnakar < ravibhushanratna...@gmail.com> wrote: >

Re: StackOverflow Error

2019-07-20 Thread Vinay Patil
Hi Vinay, > > Please make sure that all your custom code is serializable. You can run > this using new mode. > > Thanks, > Ravi > > On Sat 20 Jul, 2019, 08:13 Vinay Patil, wrote: > >> Hi, >> >> I am trying to run a pipel

StackOverflow Error

2019-07-20 Thread Vinay Patil
ms-dynamo-streams", new JsonSerializationSchema())) .name("Kafka Sink"); try { env.execute(); } catch (Exception e) { System.out.println("Caught exception for pipeline" + e.getMessage()); e.printStackTrace(); } } Regards, Vinay Patil

Re: Consuming data from dynamoDB streams to flink

2019-07-19 Thread Vinay Patil
is the recommended parallelism to be set for the source , should it be one to one mapping , for example if there are 3 shards , then parallelism should be 3 ? Regards, Vinay Patil On Wed, Aug 1, 2018 at 3:42 PM Ying Xu [via Apache Flink Mailing List archive.] wrote: > Thank you so much Fab

Re: EXT :Re: StreamingFileSink cannot get AWS S3 credentials

2019-01-17 Thread Vinay Patil
Hi Stephan., Yes, we tried setting fs.s3a.aws.credentials.provider but we are getting class not found exception for InstanceProfileCredentialsProvider because of shading issue. Regards, Vinay Patil On Thu, Jan 17, 2019 at 3:02 PM Stephan Ewen wrote: > Regarding configurations: Accord

Re: StreamingFileSink cannot get AWS S3 credentials

2019-01-16 Thread Vinay Patil
Hi Till, Can you please let us know the configurations that we need to set for Profile based credential provider in flink-conf.yaml Exporting AWS_PROFILE property on EMR did not work. Regards, Vinay Patil On Wed, Jan 16, 2019 at 3:05 PM Till Rohrmann wrote: > The old BucketingSink was us

Re: StreamingFileSink cannot get AWS S3 credentials

2019-01-16 Thread Vinay Patil
. We tried adding fs.s3a.impl to core-site.xml when the default configurations were not working. Regards, Vinay Patil On Wed, Jan 16, 2019 at 2:55 PM Till Rohrmann wrote: > Hi Vinay, > > Flink's file systems are self contained and won't respect the > core-site.xml if I'm not mistaken.

Re: StreamingFileSink cannot get AWS S3 credentials

2019-01-15 Thread Vinay Patil
Hi, Can someone please help on this issue. We have even tried to set fs.s3a.impl in core-site.xml, still its not working. Regards, Vinay Patil On Fri, Jan 11, 2019 at 5:03 PM Taher Koitawala [via Apache Flink User Mailing List archive.] wrote: > Hi All, > We have implemented S

Re: AvroInputFormat Serialisation Issue

2018-11-09 Thread Vinay Patil
Hi, Changing the classloader config to parent-first solved the issue. Regards, Vinay Patil On Wed, Nov 7, 2018 at 7:25 AM Vinay Patil wrote: > Hi, > > Can someone please help here. > > On Nov 6, 2018 10:46 PM, "Vinay Patil [via Apache Flink User Mailing List > arc

Re: AvroInputFormat Serialisation Issue

2018-11-06 Thread Vinay Patil
rializer.java:93) at com.esotericsoftware.kryo.serializers.CollectionSerializer.read(CollectionSerializer.java:22) at com.esotericsoftware.kryo.Kryo.readObject(Kryo.java:679) Please let me know if there is a fix for this issue as I have not faced this problem for DataStreams. Regards, Vinay Patil -- Sent fro

Re: Unable to start session cluster using Docker

2018-10-06 Thread Vinay Patil
Thank you Till, I am able to start the session-cluster now. Regards, Vinay Patil On Fri, Oct 5, 2018 at 8:15 PM Till Rohrmann wrote: > Hi Vinay, > > are you referring to flink-contrib/docker-flink/docker-compose.yml? We > recently fixed the command line parsing with Flink 1.5.4 an

Unable to start session cluster using Docker

2018-10-04 Thread Vinay Patil
lhost in /etc/hosts file. Can you please let me know what is the issue here. Regards, Vinay Patil

Re: Checkpointing not happening in Standalone HA mode

2018-07-27 Thread Vinay Patil
Hi Vino, Yes, Job runs successfully, however, no checkpoints are successful. I will update the source Regards, Vinay Patil On Fri, Jul 27, 2018 at 2:00 PM vino yang wrote: > Hi Vinay, > > Oh! You use a collection source? That's the problem. Please use a general > source like Kaf

Re: Checkpointing not happening in Standalone HA mode

2018-07-27 Thread Vinay Patil
Source is not being executed at the moment. Aborting checkpoint. In the pipeline I have a stream initialized using "fromCollection". I think I will have to get rid of this. What do you suggest Regards, Vinay Patil On Thu, Jul 26, 2018 at 12:04 PM vino yang wrote: > Hi Vinay: > >

Re: Checkpointing not happening in Standalone HA mode

2018-07-25 Thread Vinay Patil
Hi Chesnay, No error in the logs. That is why I am not able to understand why checkpoints are getting triggered. Regards, Vinay Patil On Wed, Jul 25, 2018 at 4:36 PM Chesnay Schepler wrote: > Please check the job- and taskmanager logs for anything suspicious. > > On 25.07.2018 12:

Re: Checkpointing not happening in Standalone HA mode

2018-07-25 Thread Vinay Patil
No error in the logs. That is why I am not able to understand why checkpoints are not getting triggered. Regards, Vinay Patil On Wed, Jul 25, 2018 at 4:44 PM Vinay Patil wrote: > Hi Chesnay, > > No error in the logs. That is why I am not able to understand why > checkpoints

Checkpointing not happening in Standalone HA mode

2018-07-25 Thread Vinay Patil
do not see any checkpoints triggered on Flink UI. Am I missing any configurations to be set for the RemoteExecutionEnvironment for checkpointing to work. Regards, Vinay Patil

Re: Query regarding rest.port property

2018-07-25 Thread Vinay Patil
, but still not able to hit the rest api's, Is there anything else I can do here ? Yes, you are right about separating the API's into two parts. Regards, Vinay Patil On Sat, Jul 21, 2018 at 1:46 AM Chesnay Schepler wrote: > Something that I was thinking about a while ago was to separate the REST &g

Query regarding rest.port property

2018-07-20 Thread Vinay Patil
Web UI running or am I missing any configuration ? Regards, Vinay Patil

Re: Strictly use TLSv1.2

2018-06-22 Thread Vinay Patil
Hi Fabian, Created a JIRA ticket : https://issues.apache.org/jira/browse/FLINK-9643 Regards, Vinay Patil On Fri, Jun 22, 2018 at 1:25 PM Fabian Hueske wrote: > Hi Vinay, > > This looks like a bug. > Would you mind creating a Jira ticket [1] for this issue? > > Thank you v

Re: Strictly use TLSv1.2

2018-06-21 Thread Vinay Patil
ink@taskmanager1:port/user/taskmanager) Now, when I hit the above command for the data port, it does not allow TLSv1.1 and only allows TLSv1.2 Can you please let me know how can I enforce all the flink ports to use TLSv1.2. Regards, Vinay Patil

Re: Strictly use TLSv1.2

2018-06-21 Thread Vinay Patil
Hi, Can someone please help me with this issue. Regards, Vinay Patil -- Sent from: http://apache-flink-user-mailing-list-archive.2336050.n4.nabble.com/

Re: Plain text SSL passwords in Log file

2018-03-29 Thread Vinay Patil
I have created FLINK-9111 <https://issues.apache.org/jira/browse/FLINK-9111> as this is not handled in the latest code of GlobalConfiguration. Regards, Vinay Patil On Thu, Mar 29, 2018 at 8:33 AM, Vinay Patil <vinay18.pa...@gmail.com> wrote: > Hi, > > If this is

Re: Plain text SSL passwords in Log file

2018-03-29 Thread Vinay Patil
Hi, If this is not part of Flink 1.5 or not handled in latest 1.4.2 release, I can open a JIRA. Should be a small change. What do you think ? Regards, Vinay Patil On Wed, Mar 28, 2018 at 4:11 PM, Vinay Patil <vinay18.pa...@gmail.com> wrote: > Hi Greg, > > I am not conce

Re: Plain text SSL passwords in Log file

2018-03-28 Thread Vinay Patil
to not print the ssl passwords . Regards, Vinay Patil On Wed, Mar 28, 2018 at 3:53 PM, Greg Hogan <c...@greghogan.com> wrote: > With the current method you always have the risk, no matter which keywords > you filter on ("secret", "password", etc.), that the key name

Plain text SSL passwords in Log file

2018-03-28 Thread Vinay Patil
release ? (I am using Flink 1.3.2) Regards, Vinay Patil

Unable to see more than 5 jobs on Flink Dashboard

2018-03-28 Thread Vinay Patil
Hi, I am not able to see more than 5 jobs on Flink Dashboard. I have set web.history to 50 in flink-conf.yaml file. Is there any other configuration I have to set to see more jobs on Flink Dashboard Regards, Vinay Patil

Re: Flink SSL Setup on a standalone cluster

2018-03-23 Thread Vinay Patil
Hi, The passwords are shown in plain text in logs , is this fixed in newer versions of flink (I am using 1.3.2) Also, please let me know the answer to my previous queries in this mail chain Regards, Vinay Patil On Mon, Mar 19, 2018 at 7:35 PM, Vinay Patil <vinay18.pa...@gmail.com> wrote:

Re: Flink SSL Setup on a standalone cluster

2018-03-19 Thread Vinay Patil
to my previous mail Regards, Vinay Patil On Fri, Mar 16, 2018 at 10:15 AM, Vinay Patil <vinay18.pa...@gmail.com> wrote: > Hi Chesnay, > > After setting the configurations for Remote Execution Environment the job > gets submitted ,I had to set ssl-verify-hostname to false.

Re: Flink SSL Setup on a standalone cluster

2018-03-15 Thread Vinay Patil
get a Lost to Job Manager Exception. This only happens when SSL is enabled. Regards, Vinay Patil On Thu, Mar 15, 2018 at 10:28 AM, Vinay Patil <vinay18.pa...@gmail.com> wrote: > Just an update, I am submitting the job from the master node, not using > the normal flink run comma

Re: Flink SSL Setup on a standalone cluster

2018-03-15 Thread Vinay Patil
Just an update, I am submitting the job from the master node, not using the normal flink run command to submit the job , but using Remote Execution Environment in code to do this. And in that I am passing the hostname which is same as provided in flink-conf.yaml Regards, Vinay Patil On Thu

Flink SSL Setup on a standalone cluster

2018-03-14 Thread Vinay Patil
to import the certificate to the java default trustore, so I have provided the trustore and keystore as jvm args to the job. Is there any other configuration I should do so that the job is submitted Regards, Vinay Patil

Regarding Task Slots allocation

2018-02-16 Thread Vinay Patil
? Regards, Vinay Patil

Concurrent modification Exception when submitting multiple jobs

2018-02-15 Thread Vinay Patil
1.3.2, and I am making sure that the job name is different for each job. Can you please let me know if I am doing something wrong. Regards, Vinay Patil -- Sent from: http://apache-flink-user-mailing-list-archive.2336050.n4.nabble.com/

Re: Get the JobID when Flink job fails

2018-02-06 Thread Vinay Patil
Hi, I see we can generate our own JobID, but how do I use it to submit the job to the cluster. I am using remoteExecutionEnvironment to submit the job to the cluster. Also, can you please answer the query of earlier mail. Regards, Vinay Patil On Thu, Feb 1, 2018 at 1:50 PM, Vinay Patil

Get the JobID when Flink job fails

2018-02-01 Thread Vinay Patil
Hi, When the Flink job executes successfully I get the jobID, however when the Flink job fails the jobID is not returned. How do I get the jobId in this case ? Do I need to call /joboverview REST api to get the job ID by looking for the Job Name ? Regards, Vinay Patil

Re: Send ACK when all records of file are processed

2018-01-25 Thread Vinay Patil
that we can have a next ack operator which will generate the response. Also, how do I get/access the Watermark value in the ack operator ? It will be a simple map operator, right ? Regards, Vinay Patil On Thu, Jan 25, 2018 at 4:50 AM, Piotr Nowojski <pi...@data-artisans.com> wrote:

Send ACK when all records of file are processed

2018-01-24 Thread Vinay Patil
slot. So, when EOF dummy records is read I can generate a response/ack. Is there a better way I can deal with this ? Regards, Vinay Patil

Re: S3 Write Execption

2017-12-05 Thread vinay patil
) Regards, Vinay Patil -- Sent from: http://apache-flink-user-mailing-list-archive.2336050.n4.nabble.com/

Re: S3 Write Execption

2017-12-05 Thread vinay patil
) Regards, Vinay Patil -- Sent from: http://apache-flink-user-mailing-list-archive.2336050.n4.nabble.com/

Re: Serialization issues with DataStreamUtils

2017-08-31 Thread vinay patil
Hi, After adding the following two lines the serialization trace does not show the Schema related classes: env.getConfig().registerTypeWithKryoSerializer(GenericData.Array.class, Serializers.SpecificInstanceCollectionSerializerForArrayList.class);

Re: Flink Elastic Sink AWS ES

2017-08-27 Thread vinay patil
ESSink) You can read this blog post: https://aws.amazon.com/blogs/big-data/build-a-real-time-stream-processing-pipeline-with-apache-flink-on-aws/ Regards, Vinay Patil On Sun, Aug 27, 2017 at 7:02 PM, ant burton [via Apache Flink User Mailing List archive.] <ml+s2336050n15173...@n4.nabble.com>

Re: Serialization issues with DataStreamUtils

2017-08-26 Thread vinay patil
Hi Robert, The test case code is as follows: GenericRecord testData = new GenericData.Record(avroSchema); SingleOutputStreamOperator testStream = env.fromElements(testData) .map(new DummyOperator(...));

Serialization issues with DataStreamUtils

2017-08-25 Thread vinay patil
) at org.apache.flink.contrib.streaming.SocketStreamIterator.hasNext(SocketStreamIterator.java:114) I tried to to register the above classes but it did not work. Also this error comes randomly for some tests while some test pass. What could be the issue ? Regards, Vinay Patil -- View this message

Re: Flink Data Streaming to S3

2017-08-14 Thread vinay patil
Hi, Yes, I am able to write to S3 using DataStream API. I have answered you the approach on SO Regards, Vinay Patil On Mon, Aug 14, 2017 at 4:21 AM, ant burton [via Apache Flink User Mailing List archive.] <ml+s2336050n14871...@n4.nabble.com> wrote: > Hello, > > Has anybody bee

Re: No file system found with scheme s3

2017-08-12 Thread vinay patil
Hi, The config should be *fs.s3a.impl* instead of *fs.s3.impl* Also when you are providing the S3 write path in config file or directly in code start with *s3a://* Regards, Vinay Patil On Sat, Aug 12, 2017 at 6:07 AM, ant burton [via Apache Flink User Mailing List archive.] <ml+s2336050n1

Re: [EXTERNAL] Re: Help required - "BucketingSink" usage to write HDFS Files

2017-08-07 Thread vinay patil
/connectors/fs/bucketing/BucketingSink.html Regards, Vinay Patil On Mon, Aug 7, 2017 at 9:15 AM, Raja.Aravapalli [via Apache Flink User Mailing List archive.] <ml+s2336050n14716...@n4.nabble.com> wrote: > Hi Vinay, > > > > Thanks for the response. > > > > I h

Re: Help required - "BucketingSink" usage to write HDFS Files

2017-08-06 Thread vinay patil
Hi Raja, Have you enabled checkpointing? The files will be rolled to complete state when the batch size is reached (in your case 2 MB) or when the bucket is inactive for a certain amount of time. Regards, Vinay Patil On Mon, Aug 7, 2017 at 7:53 AM, Raja.Aravapalli [via Apache Flink User

Re: Memory Leak - Flink / RocksDB ?

2017-07-25 Thread vinay patil
Tuning Guide:* https://github.com/facebook/rocksdb/wiki/RocksDB-Tuning-Guide Hope it helps. Regards, Vinay Patil On Tue, Jul 25, 2017 at 6:51 PM, Shashwat Rastogi [via Apache Flink User Mailing List archive.] <ml+s2336050n14439...@n4.nabble.com> wrote: > Hi, > > We have several

Re: Why would a kafka source checkpoint take so long?

2017-07-13 Thread Vinay Patil
Hi Stephan, Sure will do that next time when I observe it. Regards, Vinay Patil On Thu, Jul 13, 2017 at 8:09 PM, Stephan Ewen <se...@apache.org> wrote: > Is there any way you can pull a thread dump from the TMs at the point when > that happens? > > On Wed, Jul 12, 2017 at 8:

Re: Why would a kafka source checkpoint take so long?

2017-07-12 Thread vinay patil
Hi Gyula, I have observed similar issue with FlinkConsumer09 and 010 and posted it to the mailing list as well . This issue is not consistent, however whenever it happens it leads to checkpoints getting failed or taking a long time to complete. Regards, Vinay Patil On Wed, Jul 12, 2017 at 7:00

Re: Is there some metric info about RocksdbBackend?

2017-06-30 Thread vinay patil
try to set createStatistics() as well. By the way I was able to get rid of memory consumption issue. Did you try using FLASH_SSD_OPTION ? Regards, Vinay Patil On Fri, Jun 30, 2017 at 2:49 PM, gerryzhou [via Apache Flink User Mailing List archive.] <ml+s2336050n14081...@n4.nabble.com&

Re: Checkpointing with RocksDB as statebackend

2017-06-29 Thread Vinay Patil
s flushing is taking place at regular intervals ) Regards, Vinay Patil On Thu, Jun 29, 2017 at 9:13 PM, Vinay Patil <vinay18.pa...@gmail.com> wrote: > The state size is not that huge. On the Flink UI when it showed the data > sent as 4GB , the physical memory usage was close to 90G

Re: Checkpointing with RocksDB as statebackend

2017-06-29 Thread Vinay Patil
The state size is not that huge. On the Flink UI when it showed the data sent as 4GB , the physical memory usage was close to 90GB .. I will re-run by setting the Flushing options of RocksDB because I am facing this issue on 1.2.0 as well. Regards, Vinay Patil On Thu, Jun 29, 2017 at 9:03 PM

Re: Checkpointing with RocksDB as statebackend

2017-06-29 Thread Vinay Patil
Hi Aljoscha, Yes I have tried with 1.2.1 and 1.3.0 , facing the same issue. The issue is not of Heap memory , it is of the Off-Heap memory that is getting used ( please refer to the earlier snapshot I have attached in which the graph keeps on growing ). Regards, Vinay Patil On Thu, Jun 29

Re: Checkpointing with RocksDB as statebackend

2017-06-29 Thread Vinay Patil
you please help in resolving this issue Regards, Vinay Patil On Thu, Jun 29, 2017 at 6:01 PM, gerryzhou [via Apache Flink User Mailing List archive.] <ml+s2336050n1406...@n4.nabble.com> wrote: > Hi, Vinay, > I observed a similar problem in flink 1.3.0 with rocksdb. I wonder

Re: Checkpointing with RocksDB as statebackend

2017-06-29 Thread Vinay Patil
? Regards, Vinay Patil On Thu, Jun 29, 2017 at 7:30 AM, SHI Xiaogang <shixiaoga...@gmail.com> wrote: > Hi Vinay, > > We observed a similar problem before. We found that RocksDB keeps a lot of > index and filter blocks in memory. With the growth in state size (in our > cases,

Re: Checkpointing with RocksDB as statebackend

2017-06-28 Thread Vinay Patil
I had attached is of Off-heap memory, I have only assigned 12GB heap memory per TM Regards, Vinay Patil On Wed, Jun 28, 2017 at 8:43 PM, Aljoscha Krettek <aljos...@apache.org> wrote: > Hi, > > Just a quick question, because I’m not sure whether this came up in the > discussion

Re: Checkpointing with RocksDB as statebackend

2017-06-27 Thread vinay patil
RocksDB configurations <http://apache-flink-user-mailing-list-archive.2336050.n4.nabble.com/file/n14013/TM_Memory_Usage.png> Regards, Vinay Patil -- View this message in context: http://apache-flink-user-mailing-list-archive.2336050.n4.nabble.com/Re-Checkpointing-with-RocksDB-as-stateb

Re: Checkpointing with RocksDB as statebackend

2017-06-26 Thread Vinay Patil
all memory. This was not happening with previous version ; maximum 30GB was getting utilized. Because of this issue the job manager was killed and the job failed. Is there any other configurations I have to do ? P.S I am currently using FRocksDB Regards, Vinay Patil On Fri, May 5, 2017 at 1:01

Re: In-transit Data Encryption in EMR

2017-06-09 Thread vinay patil
Hi Guys, Can anyone please provide me solution to my queries. On Jun 8, 2017 11:30 PM, "Vinay Patil" <vinay18.pa...@gmail.com> wrote: > Hi Guys, > > I am able to setup SSL correctly, however the following command does not > work correctly and results in the error I

Re: In-transit Data Encryption in EMR

2017-06-08 Thread vinay patil
? Currently I am just relying on the logs. 2. Wild Card is not working with the keytool command, can you please let me know what is the issue with the following command: keytool -genkeypair -alias ca -keystore: -ext SAN=dns:node1.* Regards, Vinay Patil On Mon, Jun 5, 2017 at 8:43 PM, vinay patil [via

Re: In-transit Data Encryption in EMR

2017-06-05 Thread vinay patil
org.apache.flink.configuration.GlobalConfiguration- Loading configuration property: security.ssl.truststore-password, password*/ Regards, Vinay Patil -- View this message in context: http://apache-flink-user-mailing-list-archive.2336050.n4.nabble.com/In-transit-Data-Encryption-in-EMR-tp13455p13490

Re: In-transit Data Encryption in EMR

2017-06-05 Thread vinay patil
/application_1496660166576_0001/flink-dist_2.10-1.2.0.jar, expected: file:/// I see a JIRA ticket regarding the same but did not find any solution to this. Regards, Vinay Patil -- View this message in context: http://apache-flink-user-mailing-list-archive.2336050.n4.nabble.com/In-transit-Data-Encryption

Re: In-transit Data Encryption in EMR

2017-06-05 Thread vinay patil
Thank you Till. Gordon can you please help. Regards, Vinay Patil On Fri, Jun 2, 2017 at 9:10 PM, Till Rohrmann [via Apache Flink User Mailing List archive.] <ml+s2336050n13459...@n4.nabble.com> wrote: > Hi Vinay, > > I've pulled my colleague Gordon into the conversation who can

In-transit Data Encryption in EMR

2017-06-02 Thread vinay patil
, Vinay Patil -- View this message in context: http://apache-flink-user-mailing-list-archive.2336050.n4.nabble.com/In-transit-Data-Encryption-in-EMR-tp13455.html Sent from the Apache Flink User Mailing List archive. mailing list archive at Nabble.com.

Re: Queries regarding Historical Reprocessing

2017-05-03 Thread Vinay Patil
Hi Guys, Can someone please help me in understanding this ? Regards, Vinay Patil On Thu, Apr 27, 2017 at 12:36 PM, Vinay Patil <vinay18.pa...@gmail.com> wrote: > Hi Guys, > > For historical reprocessing , I am reading the avro data from S3 and > passing these records to

Queries regarding Historical Reprocessing

2017-04-27 Thread Vinay Patil
after processing , is this because Flink treats the S3 source as finite data ? What will happen if the data is continuously written to S3 from one pipeline and from the second pipeline I am doing historical re-processing ? Regards, Vinay Patil

Queries regarding Historical Reprocessing

2017-04-26 Thread vinay patil
after processing , is this because Flink treats the S3 source as finite data ? What will happen if the data is continuously written to S3 from one pipeline and from the second pipeline I am doing historical re-processing ? Regards, Vinay Patil -- View this message in context: http://apache-flin

Re: Checkpointing with RocksDB as statebackend

2017-03-27 Thread vinay patil
the RocksDB fix in 1.2.1 so that I can test it out. Regards, Vinay Patil On Sat, Mar 18, 2017 at 12:25 AM, Stephan Ewen [via Apache Flink User Mailing List archive.] <ml-node+s2336050n1227...@n4.nabble.com> wrote: > @vinay Let's see how fast we get this fix in - I hope yes. It may depe

Re: Checkpointing with RocksDB as statebackend

2017-03-17 Thread vinay patil
Hi Stephan, Is the performance related change of RocksDB going to be part of Flink 1.2.1 ? Regards, Vinay Patil On Thu, Mar 16, 2017 at 6:13 PM, Stephan Ewen [via Apache Flink User Mailing List archive.] <ml-node+s2336050n12243...@n4.nabble.com> wrote: > The only immediate w

Re: Checkpointing with RocksDB as statebackend

2017-03-16 Thread vinay patil
for Streaming application (running on YARN - EMR ) ? Regards, Vinay Patil On Thu, Mar 16, 2017 at 6:36 PM, rmetzger0 [via Apache Flink User Mailing List archive.] <ml-node+s2336050n12244...@n4.nabble.com> wrote: > Yes, you can change the GC using the env.java.opts parameter. > We are not se

Re: Checkpointing with RocksDB as statebackend

2017-03-16 Thread vinay patil
Hi Stephan, What can be the workaround for this ? Also need one confirmation : Is G1 GC used by default when running the pipeline on YARN. (I see a thread of 2015 where G1 is used by default for JAVA8) Regards, Vinay Patil On Wed, Mar 15, 2017 at 10:32 PM, Stephan Ewen [via Apache Flink User

Re: Checkpointing with RocksDB as statebackend

2017-03-15 Thread vinay patil
, not sure how this will affect in production as we are going to get above 200 million data. As a workaround can I take the savepoint while the pipeline is running ? Let's say if I take savepoint after every 30minutes, will it work ? Regards, Vinay Patil On Tue, Mar 14, 2017 at 10:02 PM, Stephan Ewen

Re: Frequent Full GC's in case of FSStateBackend

2017-03-08 Thread vinay patil
don't see Full GC happening, also I am using Flink >> 1.2.0 version and I have set the statebackend in flink-conf.yaml file to >> rocksdb, so by default does this do asynchronous checkpointing or I have to >> specify it at the job level ? >> >> Regards, >> Vinay Patil &g

Re: Integrate Flink with S3 on EMR cluster

2017-03-08 Thread vinay patil
Hi , @Shannon - I am not facing any issue while writing to S3, was getting NoClassDef errors when reading the file from S3. ''Hadoop File System" - I mean I am using FileSystem class of Hadoop to read the file from S3. @Stephan - I tried with 1.1.4 , was getting the same issue. The easiest way

Re: Integrate Flink with S3 on EMR cluster

2017-03-07 Thread vinay patil
Hi Guys, Has anyone got this error before ? If yes, have you found any other solution apart from copying the jar files to flink lib folder Regards, Vinay Patil On Mon, Mar 6, 2017 at 8:21 PM, vinay patil [via Apache Flink User Mailing List archive.] <ml-node+s2336050n12053...@n4.nabble.

Re: Integrate Flink with S3 on EMR cluster

2017-03-06 Thread vinay patil
Hi Guys, I am getting the same exception: EMRFileSystem not Found I am trying to read encrypted S3 file using Hadoop File System class. (using Flink 1.2.0) When I copy all the libs from /usr/share/aws/emrfs/lib and /usr/lib/hadoop to Flink lib folder , it works. However I see that all these

Re: Checkpointing with RocksDB as statebackend

2017-02-27 Thread vinay patil
to 10minutes, I have observed that nothing gets written to sink (tried with S3 as well as HDFS), atleast I was expecting pending files here. This issue gets worst when checkpointing is disabled as nothing is written. Regards, Vinay Patil On Mon, Feb 27, 2017 at 10:55 PM, Stephan Ewen [via

Re: Checkpointing with RocksDB as statebackend

2017-02-25 Thread vinay patil
, now within 5minutes of run the state size grows to 30GB , after checkpointing the 30GB state that is maintained in rocksDB has to be copied to HDFS, right ? is this causing the pipeline to stall ? Regards, Vinay Patil On Sat, Feb 25, 2017 at 12:22 AM, Vinay Patil <vinay18.pa...@gmail.com>

Re: Checkpointing with RocksDB as statebackend

2017-02-24 Thread vinay patil
uploads of even data amounts as > low as kilobytes due to its throttling policies. > > That would be a super important fix to add! > > Best, > Stephan > > > On Fri, Feb 24, 2017 at 2:58 PM, vinay patil <[hidden email] > <http:///user/SendEmail.jtp?type=node=11885=0>>

Re: Checkpointing with RocksDB as statebackend

2017-02-24 Thread vinay patil
Hi, I have attached a snapshot for reference: As you can see all the 3 checkpointins failed , for checkpoint ID 2 and 3 it is stuck at the Kafka source after 50% (The data sent till now by Kafka source 1 is 65GB and sent by source 2 is 15GB ) Within 10minutes 15M records were processed, and for

Re: Checkpointing with RocksDB as statebackend

2017-02-24 Thread vinay patil
Hi Stephan, Thank you for the brief explanation. Yes I have already enabled Object Reuse mode because of which I see significant improvement. I am currently running on r3.4xlarge having 122GB memory, as you suggested I had increased the checkpoint interval to 10minutes and minimum pause between

Re: Checkpointing with RocksDB as statebackend

2017-02-23 Thread vinay patil
with 122GB RAM, in case of FSStatebackend does it use TM heap memory or physical memory to store the state ? Regards, Vinay Patil On Thu, Feb 23, 2017 at 7:50 PM, vinay patil [via Apache Flink User Mailing List archive.] <ml-node+s2336050n11831...@n4.nabble.com> wrote: > Hi Stephan, > > An

Re: Checkpointing with RocksDB as statebackend

2017-02-23 Thread vinay patil
Hi Stephan, Anyways the Async exception is gone. I have increased my instance type to r3.2xlarge having 60GB of memory. BUt what I have observed here is that for two task managers the memory usage is close to 30GB but for other two it goes up to 55GB, the load is equally distributed among all

Re: Flink checkpointing gets stuck

2017-02-21 Thread vinay patil
to TM as rocksDB also uses physical memory Regards, Vinay Patil On Tue, Feb 21, 2017 at 8:03 PM, Shai Kaplan [via Apache Flink User Mailing List archive.] <ml-node+s2336050n11780...@n4.nabble.com> wrote: > Hi Vinay. > > > > I couldn't understand from the thread, what conf

  1   2   >