Re: Dynamic Flink SQL

2020-03-27 Thread Krzysztof Zarzycki
I want to do a bit different hacky PoC: * I will write a sink, that caches the results in "JVM global" memory. Then I will write a source, that reads this cache. * I will launch one job, that reads from Kafka source, shuffles the data to the desired partitioning and then sinks to that cache. * Then

Re: Dynamic Flink SQL

2020-03-25 Thread Arvid Heise
I saw that requirement but I'm not sure if you really need to modify the query at runtime. Unless you need reprocessing for newly added rules, I'd probably just cancel with savepoint and restart the application with the new rules. Of course, it depends on the rules themselves and how much state th

Re: Dynamic Flink SQL

2020-03-25 Thread Krzysztof Zarzycki
Hello Arvid, Thanks for joining to the thread! First, did you take into consideration that I would like to dynamically add queries on the same source? That means first define one query, later the day add another one , then another one, and so on. A Week later kill one of those, start yet another on

Re: Dynamic Flink SQL

2020-03-24 Thread Arvid Heise
Hi Krzysztof, from my past experience as data engineer, I can safely say that users often underestimate the optimization potential and techniques of the used systems. I implemented a similar thing in the past, where I parsed up to 500 rules reading from up to 10 data sources. The basic idea was to

Dynamic Flink SQL

2020-03-23 Thread Krzysztof Zarzycki
Dear Flink community! In our company we have implemented a system that realize the dynamic business rules pattern. We spoke about it during Flink Forward 2019 https://www.youtube.com/watch?v=CyrQ5B0exqU. The system is a great success and we would like to improve it. Let me shortly mention what the