Hello,

I found this question in the Nabble archive 
(http://apache-flink-user-mailing-list-archive.2336050.n4.nabble.com/Maintaining-watermarks-per-key-instead-of-per-operator-instance-tp7288.html
 
<http://apache-flink-user-mailing-list-archive.2336050.n4.nabble.com/Maintaining-watermarks-per-key-instead-of-per-operator-instance-tp7288.html>)
 but was unable/dont know how to reply.

Here is my question regarding the mentioned thread:

> Hello, 
> 
> I have similar requirements (see StackOverflor 
> http://stackoverflow.com/questions/40465335/apache-flink-multiple-window-aggregations-and-late-data
>  
> <http://stackoverflow.com/questions/40465335/apache-flink-multiple-window-aggregations-and-late-data>).
>  I am pretty new to flink, could you elaborate on a possible solution? We can 
> guarantee good ordering by sensor_id, thus watermarking by key would be the 
> only reasonable way for us 
> (sensorData.keyBy('id').timeWindow(1.minute).sum('value')), could I do my own 
> watermarking aftersensorData.keyBy('id').overwriteWatermarking()... per key? 
> Or maybe using custom state plus a custom trigger? What happens if a sensor 
> dies or is being removed completely, how can this be detected as watermarks 
> would be ignored for window garbage collection. Or could we dynamically 
> schedule a job of each sensor? Which would result in 1000 Jobs.


Thanks,
Stephan


Reply via email to