Hi Stephan,

Is there a ticket number/link to track this, My job has all the conditions
you mentioned.

Thanks,
Vishnu

On Tue, Mar 14, 2017 at 7:13 AM, Stephan Ewen <se...@apache.org> wrote:

> Hi Vinay!
>
> We just discovered a bug in RocksDB. The bug affects windows without
> reduce() or fold(), windows with evictors, and ListState.
>
> A certain access pattern in RocksDB starts being so slow after a certain
> size-per-key that it basically brings down the streaming program and the
> snapshots.
>
> We are reaching out to the RocksDB folks and looking for workarounds in
> Flink.
>
> Greetings,
> Stephan
>
>
> On Wed, Mar 1, 2017 at 12:10 PM, Stephan Ewen <se...@apache.org> wrote:
>
>> @vinay  Can you try to not set the buffer timeout at all? I am actually
>> not sure what would be the effect of setting it to a negative value, that
>> can be a cause of problems...
>>
>>
>> On Mon, Feb 27, 2017 at 7:44 PM, Seth Wiesman <swies...@mediamath.com>
>> wrote:
>>
>>> Vinay,
>>>
>>>
>>>
>>> The bucketing sink performs rename operations during the checkpoint and
>>> if it tries to rename a file that is not yet consistent that would cause a
>>> FileNotFound exception which would fail the checkpoint.
>>>
>>>
>>>
>>> Stephan,
>>>
>>>
>>>
>>> Currently my aws fork contains some very specific assumptions about the
>>> pipeline that will in general only hold for my pipeline. This is because
>>> there were still some open questions that  I had about how to solve
>>> consistency issues in the general case. I will comment on the Jira issue
>>> with more specific.
>>>
>>>
>>>
>>> Seth Wiesman
>>>
>>>
>>>
>>> *From: *vinay patil <vinay18.pa...@gmail.com>
>>> *Reply-To: *"user@flink.apache.org" <user@flink.apache.org>
>>> *Date: *Monday, February 27, 2017 at 1:05 PM
>>> *To: *"user@flink.apache.org" <user@flink.apache.org>
>>>
>>> *Subject: *Re: Checkpointing with RocksDB as statebackend
>>>
>>>
>>>
>>> Hi Seth,
>>>
>>> Thank you for your suggestion.
>>>
>>> But if the issue is only related to S3, then why does this happen when I
>>> replace the S3 sink  to HDFS as well (for checkpointing I am using HDFS
>>> only )
>>>
>>> Stephan,
>>>
>>> Another issue I see is when I set env.setBufferTimeout(-1) , and keep
>>> the checkpoint interval to 10minutes, I have observed that nothing gets
>>> written to sink (tried with S3 as well as HDFS), atleast I was expecting
>>> pending files here.
>>>
>>> This issue gets worst when checkpointing is disabled  as nothing is
>>> written.
>>>
>>>
>>>
>>>
>>> Regards,
>>>
>>> Vinay Patil
>>>
>>>
>>>
>>> On Mon, Feb 27, 2017 at 10:55 PM, Stephan Ewen [via Apache Flink User
>>> Mailing List archive.] <[hidden email]> wrote:
>>>
>>> Hi Seth!
>>>
>>>
>>>
>>> Wow, that is an awesome approach.
>>>
>>>
>>>
>>> We have actually seen these issues as well and we are looking to
>>> eventually implement our own S3 file system (and circumvent Hadoop's S3
>>> connector that Flink currently relies on): https://issues.apache.org
>>> /jira/browse/FLINK-5706
>>>
>>>
>>>
>>> Do you think your patch would be a good starting point for that and
>>> would you be willing to share it?
>>>
>>>
>>>
>>> The Amazon AWS SDK for Java is Apache 2 licensed, so that is possible to
>>> fork officially, if necessary...
>>>
>>>
>>>
>>> Greetings,
>>>
>>> Stephan
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>> On Mon, Feb 27, 2017 at 5:15 PM, Seth Wiesman <[hidden email]
>>> <http://user/SendEmail.jtp?type=node&node=11943&i=0>> wrote:
>>>
>>> Just wanted to throw in my 2cts.
>>>
>>>
>>>
>>> I’ve been running pipelines with similar state size using rocksdb which
>>> externalize to S3 and bucket to S3. I was getting stalls like this and
>>> ended up tracing the problem to S3 and the bucketing sink. The solution was
>>> two fold:
>>>
>>>
>>>
>>> 1)       I forked hadoop-aws and have it treat flink as a source of
>>> truth. Emr uses a dynamodb table to determine if S3 is inconsistent.
>>> Instead I say that if flink believes that a file exists on S3 and we don’t
>>> see it then I am going to trust that flink is in a consistent state and S3
>>> is not. In this case, various operations will perform a back off and retry
>>> up to a certain number of times.
>>>
>>>
>>>
>>> 2)       The bucketing sink performs multiple renames over the lifetime
>>> of a file, occurring when a checkpoint starts and then again on
>>> notification after it completes. Due to S3’s consistency guarantees the
>>> second rename of file can never be assured to work and will eventually fail
>>> either during or after a checkpoint. Because there is no upper bound on the
>>> time it will take for a file on S3 to become consistent, retries cannot
>>> solve this specific problem as it could take upwards of many minutes to
>>> rename which would stall the entire pipeline. The only viable solution I
>>> could find was to write a custom sink which understands S3. Each writer
>>> will write file locally and then copy it to S3 on checkpoint. By only
>>> interacting with S3 once per file it can circumvent consistency issues all
>>> together.
>>>
>>>
>>>
>>> Hope this helps,
>>>
>>>
>>>
>>> Seth Wiesman
>>>
>>>
>>>
>>> *From: *vinay patil <[hidden email]
>>> <http://user/SendEmail.jtp?type=node&node=11943&i=1>>
>>> *Reply-To: *"[hidden email]
>>> <http://user/SendEmail.jtp?type=node&node=11943&i=2>" <[hidden email]
>>> <http://user/SendEmail.jtp?type=node&node=11943&i=3>>
>>> *Date: *Saturday, February 25, 2017 at 10:50 AM
>>> *To: *"[hidden email]
>>> <http://user/SendEmail.jtp?type=node&node=11943&i=4>" <[hidden email]
>>> <http://user/SendEmail.jtp?type=node&node=11943&i=5>>
>>> *Subject: *Re: Checkpointing with RocksDB as statebackend
>>>
>>>
>>>
>>> HI Stephan,
>>>
>>> Just to avoid the confusion here, I am using S3 sink for writing the
>>> data, and using HDFS for storing checkpoints.
>>>
>>> There are 2 core nodes (HDFS) and two task nodes on EMR
>>>
>>>
>>> I replaced s3 sink with HDFS for writing data in my last test.
>>>
>>> Let's say the checkpoint interval is 5 minutes, now within 5minutes of
>>> run the state size grows to 30GB ,  after checkpointing the 30GB state that
>>> is maintained in rocksDB has to be copied to HDFS, right ?  is this causing
>>> the pipeline to stall ?
>>>
>>>
>>> Regards,
>>>
>>> Vinay Patil
>>>
>>>
>>>
>>> On Sat, Feb 25, 2017 at 12:22 AM, Vinay Patil <[hidden email]> wrote:
>>>
>>> Hi Stephan,
>>>
>>> To verify if S3 is making teh pipeline stall, I have replaced the S3
>>> sink with HDFS and kept minimum pause between checkpoints to 5minutes,
>>> still I see the same issue with checkpoints getting failed.
>>>
>>> If I keep the  pause time to 20 seconds, all checkpoints are completed ,
>>> however there is a hit in overall throughput.
>>>
>>>
>>>
>>>
>>> Regards,
>>>
>>> Vinay Patil
>>>
>>>
>>>
>>> On Fri, Feb 24, 2017 at 10:09 PM, Stephan Ewen [via Apache Flink User
>>> Mailing List archive.] <[hidden email]> wrote:
>>>
>>> Flink's state backends currently do a good number of "make sure this
>>> exists" operations on the file systems. Through Hadoop's S3 filesystem,
>>> that translates to S3 bucket list operations, where there is a limit in how
>>> many operation may happen per time interval. After that, S3 blocks.
>>>
>>>
>>>
>>> It seems that operations that are totally cheap on HDFS are hellishly
>>> expensive (and limited) on S3. It may be that you are affected by that.
>>>
>>>
>>>
>>> We are gradually trying to improve the behavior there and be more S3
>>> aware.
>>>
>>>
>>>
>>> Both 1.3-SNAPSHOT and 1.2-SNAPSHOT already contain improvements there.
>>>
>>>
>>>
>>> Best,
>>>
>>> Stephan
>>>
>>>
>>>
>>>
>>>
>>> On Fri, Feb 24, 2017 at 4:42 PM, vinay patil <[hidden email]
>>> <http://user/SendEmail.jtp?type=node&node=11891&i=0>> wrote:
>>>
>>> Hi Stephan,
>>>
>>> So do you mean that S3 is causing the stall , as I have mentioned in my
>>> previous mail, I could not see any progress for 16minutes as checkpoints
>>> were getting failed continuously.
>>>
>>>
>>>
>>> On Feb 24, 2017 8:30 PM, "Stephan Ewen [via Apache Flink User Mailing
>>> List archive.]" <[hidden email]
>>> <http://user/SendEmail.jtp?type=node&node=11887&i=0>> wrote:
>>>
>>> Hi Vinay!
>>>
>>>
>>>
>>> True, the operator state (like Kafka) is currently not asynchronously
>>> checkpointed.
>>>
>>>
>>>
>>> While it is rather small state, we have seen before that on S3 it can
>>> cause trouble, because S3 frequently stalls uploads of even data amounts as
>>> low as kilobytes due to its throttling policies.
>>>
>>>
>>>
>>> That would be a super important fix to add!
>>>
>>>
>>>
>>> Best,
>>>
>>> Stephan
>>>
>>>
>>>
>>>
>>>
>>> On Fri, Feb 24, 2017 at 2:58 PM, vinay patil <[hidden email]
>>> <http://user/SendEmail.jtp?type=node&node=11885&i=0>> wrote:
>>>
>>> Hi,
>>>
>>> I have attached a snapshot for reference:
>>> As you can see all the 3 checkpointins failed , for checkpoint ID 2 and
>>> 3 it
>>> is stuck at the Kafka source after 50%
>>> (The data sent till now by Kafka source 1 is 65GB and sent by source 2 is
>>> 15GB )
>>>
>>> Within 10minutes 15M records were processed, and for the next 16minutes
>>> the
>>> pipeline is stuck , I don't see any progress beyond 15M because of
>>> checkpoints getting failed consistently.
>>>
>>> <http://apache-flink-user-mailing-list-archive.2336050.n4.na
>>> bble.com/file/n11882/Checkpointing_Failed.png>
>>>
>>>
>>>
>>> --
>>> View this message in context: http://apache-flink-user-maili
>>> ng-list-archive.2336050.n4.nabble.com/Re-Checkpointing-with-
>>> RocksDB-as-statebackend-tp11752p11882.html
>>>
>>> Sent from the Apache Flink User Mailing List archive. mailing list
>>> archive at Nabble.com.
>>>
>>>
>>>
>>>
>>> ------------------------------
>>>
>>> *If you reply to this email, your message will be added to the
>>> discussion below:*
>>>
>>> http://apache-flink-user-mailing-list-archive.2336050.n4.nab
>>> ble.com/Re-Checkpointing-with-RocksDB-as-statebackend-tp11752p11885.html
>>>
>>> To start a new topic under Apache Flink User Mailing List archive.,
>>> email [hidden email]
>>> <http://user/SendEmail.jtp?type=node&node=11887&i=1>
>>> To unsubscribe from Apache Flink User Mailing List archive., click here.
>>> NAML
>>> <http://apache-flink-user-mailing-list-archive.2336050.n4.nabble.com/template/NamlServlet.jtp?macro=macro_viewer&id=instant_html%21nabble%3Aemail.naml&base=nabble.naml.namespaces.BasicNamespace-nabble.view.web.template.NabbleNamespace-nabble.view.web.template.NodeNamespace&breadcrumbs=notify_subscribers%21nabble%3Aemail.naml-instant_emails%21nabble%3Aemail.naml-send_instant_email%21nabble%3Aemail.naml>
>>>
>>>
>>> ------------------------------
>>>
>>> View this message in context: Re: Checkpointing with RocksDB as
>>> statebackend
>>> <http://apache-flink-user-mailing-list-archive.2336050.n4.nabble.com/Re-Checkpointing-with-RocksDB-as-statebackend-tp11752p11887.html>
>>>
>>> Sent from the Apache Flink User Mailing List archive. mailing list
>>> archive
>>> <http://apache-flink-user-mailing-list-archive.2336050.n4.nabble.com/>
>>> at Nabble.com.
>>>
>>>
>>>
>>>
>>> ------------------------------
>>>
>>> *If you reply to this email, your message will be added to the
>>> discussion below:*
>>>
>>> http://apache-flink-user-mailing-list-archive.2336050.n4.nab
>>> ble.com/Re-Checkpointing-with-RocksDB-as-statebackend-tp11752p11891.html
>>>
>>> To start a new topic under Apache Flink User Mailing List archive.,
>>> email [hidden email]
>>> To unsubscribe from Apache Flink User Mailing List archive., click here.
>>> NAML
>>> <http://apache-flink-user-mailing-list-archive.2336050.n4.nabble.com/template/NamlServlet.jtp?macro=macro_viewer&id=instant_html%21nabble%3Aemail.naml&base=nabble.naml.namespaces.BasicNamespace-nabble.view.web.template.NabbleNamespace-nabble.view.web.template.NodeNamespace&breadcrumbs=notify_subscribers%21nabble%3Aemail.naml-instant_emails%21nabble%3Aemail.naml-send_instant_email%21nabble%3Aemail.naml>
>>>
>>>
>>>
>>>
>>>
>>>
>>> ------------------------------
>>>
>>> View this message in context: Re: Checkpointing with RocksDB as
>>> statebackend
>>> <http://apache-flink-user-mailing-list-archive.2336050.n4.nabble.com/Re-Checkpointing-with-RocksDB-as-statebackend-tp11752p11913.html>
>>> Sent from the Apache Flink User Mailing List archive. mailing list
>>> archive
>>> <http://apache-flink-user-mailing-list-archive.2336050.n4.nabble.com/>
>>> at Nabble.com.
>>>
>>>
>>>
>>>
>>> ------------------------------
>>>
>>> *If you reply to this email, your message will be added to the
>>> discussion below:*
>>>
>>> http://apache-flink-user-mailing-list-archive.2336050.n4.nab
>>> ble.com/Re-Checkpointing-with-RocksDB-as-statebackend-tp11752p11943.html
>>>
>>> To start a new topic under Apache Flink User Mailing List archive.,
>>> email [hidden email]
>>> To unsubscribe from Apache Flink User Mailing List archive., click here.
>>> NAML
>>> <http://apache-flink-user-mailing-list-archive.2336050.n4.nabble.com/template/NamlServlet.jtp?macro=macro_viewer&id=instant_html%21nabble%3Aemail.naml&base=nabble.naml.namespaces.BasicNamespace-nabble.view.web.template.NabbleNamespace-nabble.view.web.template.NodeNamespace&breadcrumbs=notify_subscribers%21nabble%3Aemail.naml-instant_emails%21nabble%3Aemail.naml-send_instant_email%21nabble%3Aemail.naml>
>>>
>>>
>>>
>>>
>>> ------------------------------
>>>
>>> View this message in context: Re: Checkpointing with RocksDB as
>>> statebackend
>>> <http://apache-flink-user-mailing-list-archive.2336050.n4.nabble.com/Re-Checkpointing-with-RocksDB-as-statebackend-tp11752p11949.html>
>>> Sent from the Apache Flink User Mailing List archive. mailing list
>>> archive
>>> <http://apache-flink-user-mailing-list-archive.2336050.n4.nabble.com/>
>>> at Nabble.com.
>>>
>>>
>>
>

Reply via email to