[ https://issues.apache.org/jira/browse/FLINK-3188?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Robert Metzger resolved FLINK-3188. ----------------------------------- Resolution: Fixed Fix Version/s: 1.0.0 Resolved for 1.0.0 in http://git-wip-us.apache.org/repos/asf/flink/commit/dcf86c27 > Deletes in Kafka source should be passed on to KeyedDeserializationSchema > ------------------------------------------------------------------------- > > Key: FLINK-3188 > URL: https://issues.apache.org/jira/browse/FLINK-3188 > Project: Flink > Issue Type: Bug > Components: Kafka Connector > Affects Versions: 1.0.0 > Reporter: Sebastian Klemke > Assignee: Robert Metzger > Fix For: 1.0.0 > > Attachments: kafka-deletions.patch, testcase.patch > > > When keys are deleted in the kafka queue, they show up as keys with null > payload. Currently in Flink 1.0-SNAPSHOT, these deletions are silently > skipped, without increasing current offset. > This leads to two problems: > 1. When a fetch window contains only deletions, LegacyFetcher gets stuck > 2. For KeyedDeserializationSchemas, it would make sense to pass deletions to > the deserializer, so that it can decide to wrap deleted keys as a deletion > command. This is also more consistent with the semantics of keys in Kafka > queues: When compaction is activated, only the latest message with the same > key needs to be kept by Kafka. > We propose the attached patch as a workaround for both issues. -- This message was sent by Atlassian JIRA (v6.3.4#6332)