Hi Yu,

Thanks for the answers, it makes sense to me :)

Piotrek

> On 31 Oct 2019, at 11:25, Yu Li <car...@gmail.com> wrote:
> 
> Hi Piotr,
> 
> Thanks for the comments!
> 
> bq. How are you planning to execute the end-to-end benchmarks and integrate
> them with our build process?
> Great question! We plan to execute the end-to-end benchmark in a small
> cluster (like 3 vm nodes) to better reflect network cost, triggering it
> through our Jenkins service for micro benchmark and show the result on
> code-speed center. Will add these into FLIP document if no objections.
> 
> bq. Are you planning to monitor the throughput and latency at the same time?
> Good question. And you're right, we will stress the cluster to
> back-pressure and watch the throughput, latency doesn't mean much in the
> first test suites. Let me refine the document.
> 
> Thanks.
> 
> Best Regards,
> Yu
> 
> 
> On Wed, 30 Oct 2019 at 19:07, Piotr Nowojski <pi...@ververica.com> wrote:
> 
>> Hi Yu,
>> 
>> Thanks for bringing this up.
>> 
>> +1 for the idea and the proposal from my side.
>> 
>> I think that the proposed Test Job List might be a bit
>> redundant/excessive, but:
>> - we can always adjust this later, once we have the infrastructure in place
>> - as long as we have the computing resources and ability to quickly
>> interpret the results/catch regressions, it doesn’t hurt to have more
>> benchmarks/tests then strictly necessary.
>> 
>> Which brings me to a question. How are you planning to execute the
>> end-to-end benchmarks and integrate them with our build process?
>> 
>> Another smaller question:
>> 
>>> In this initial stage we will only monitor and display job throughput
>> and latency.
>> 
>> Are you planning to monitor the throughput and latency at the same time?
>> It might be a bit problematic, as when measuring the throughput you want to
>> saturate the system and hit some bottleneck, which will cause a
>> back-pressure (measuring latency at the same time when system is back
>> pressured doesn’t make much sense).
>> 
>> Piotrek
>> 
>>> On 30 Oct 2019, at 11:54, Yu Li <car...@gmail.com> wrote:
>>> 
>>> Hi everyone,
>>> 
>>> We would like to propose FLIP-83 that adds an end-to-end performance
>>> testing framework for Flink. We discovered some potential problems
>> through
>>> such an internal end-to-end performance testing framework before the
>>> release of 1.9.0 [1], so we'd like to contribute it to Flink community
>> as a
>>> supplement to the existing daily run micro performance benchmark [2] and
>>> nightly run end-to-end stability test [3].
>>> 
>>> The FLIP document could be found here:
>>> 
>> https://cwiki.apache.org/confluence/display/FLINK/FLIP-83%3A+Flink+End-to-end+Performance+Testing+Framework
>>> 
>>> Please kindly review the FLIP document and let us know if you have any
>>> comments/suggestions, thanks!
>>> 
>>> [1] https://s.apache.org/m8kcq
>>> [2] https://github.com/dataArtisans/flink-benchmarks
>>> [3] https://github.com/apache/flink/tree/master/flink-end-to-end-tests
>>> 
>>> Best Regards,
>>> Yu
>> 
>> 

Reply via email to