pnowojski commented on pull request #2:
URL: https://github.com/apache/flink-benchmarks/pull/2#issuecomment-678248194


   > I think my current way is exactly for this purpose, just set N =5 and the 
default concurrent cp should be 1 if I remembered correctly. I am not quite 
sure whether it is really necessary to set even smaller interval since 100ms is 
already small enough in most practical cases. If we give much smaller values, 
the backpressure might not be triggered before N checkpoints finished.
   
   But the problem with the current setup is that if checkpointing time is 
small, you will be measuring just the time to complete last checkpoint, instead 
of 5. And the problem is that even if the current interval makes sense, the 
checkpointing time can change in the future, as we improve things, which would 
render this benchmark obsolete.
   
   That's why I suggested to just trigger checkpoint as quickly as possible, to 
avoid this problem.
   
   >  If we give much smaller values, the backpressure might not be triggered 
before N checkpoints finished.
   
   Do you think this can happen? Sources will produce huge number of records 
immediately (millions of records per second). Even if checkpoint is triggered 
just a couple of `ms` after the source started, the task should be already 
fully back pressured.
   
   If this worries you, maybe there is a way of delaying the first checkpoint?


----------------------------------------------------------------
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Reply via email to