Jeff Hu created FLINK-23132:
-------------------------------

             Summary: flink upgrade issue(1.11.3->1.13.0)
                 Key: FLINK-23132
                 URL: https://issues.apache.org/jira/browse/FLINK-23132
             Project: Flink
          Issue Type: Bug
            Reporter: Jeff Hu


 
In order to improve the performance of data process, we store events to a map 
and do not process them untill event count reaches 100. in the meantime, start 
a timer in open method, so data is processed every 60 seconds

this works when flink version is *1.11.3*,

after upgrading flink version to *1.13.0*

I found sometimes events were consumed from Kafka continuously, but were not 
processed in RichFlatMapFunction, it means data was missing. after restarting 
service, it works well, but several hours later the same thing happened again.

any known issue for this flink version? any suggestions are appreciated.
 {{public class MyJob \{
    public static void main(String[] args) throws Exception {
        ...
        DataStream<String> rawEventSource = env.addSource(flinkKafkaConsumer);
        ...
    }}} {{public class MyMapFunction extends RichFlatMapFunction<String, 
String> implements Serializable \{

    @Override
    public void open(Configuration parameters) {
        ...
        long periodTimeout = 60;
        pool.scheduleAtFixedRate(() -> {
            // processing data
        }, periodTimeout, periodTimeout, TimeUnit.SECONDS);
    }
    
    @Override
    public void flatMap(String message, Collector<String> out) \{
        // store event to map 
        // count event, 
        // when count = 100, start data processing
    }
}}}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

Reply via email to