Done. https://issues.apache.org/jira/browse/NIFI-2298 <https://issues.apache.org/jira/browse/NIFI-2298>
Thanks -Sumo > On Jul 17, 2016, at 1:37 PM, Sumanth Chinthagunta <xmlk...@gmail.com> wrote: > > Env: Linux , Kafka 9.0.x, NiFi: 0.7 > > We like to use new ConsumeKafka processor but found critical limitations > compared to old getKafka processor. > New ConsumeKafka is not writing critical Kafka attributes i.e., kafka.key, > kafka.offset, kafka.partition etc into flowFile attributes. > > > Old getKafka processor: > > Standard FlowFile Attributes > Key: 'entryDate' > Value: 'Sun Jul 17 15:17:00 CDT 2016' > Key: 'lineageStartDate' > Value: 'Sun Jul 17 15:17:00 CDT 2016' > Key: 'fileSize' > Value: '183' > FlowFile Attribute Map Content > Key: 'filename' > Value: '19709945781167274' > Key: 'kafka.key' > Value: '{"database":"test","table":"sc_job","pk.systemid":1}' > Key: 'kafka.offset' > Value: '1184010261' > Key: 'kafka.partition' > Value: '0' > Key: 'kafka.topic' > Value: ‘data' > Key: 'path' > Value: './' > Key: 'uuid' > Value: '244059bb-9ad9-4d74-b1fb-312eee72124a' > > ————————— > new ConsumeKafka processor : > > Standard FlowFile Attributes > Key: 'entryDate' > Value: 'Sun Jul 17 15:18:41 CDT 2016' > Key: 'lineageStartDate' > Value: 'Sun Jul 17 15:18:41 CDT 2016' > Key: 'fileSize' > Value: '183' > FlowFile Attribute Map Content > Key: 'filename' > Value: '19710046870478139' > Key: 'path' > Value: './' > Key: 'uuid' > Value: '349fbeb3-e342-4533-be4c-424793fa5c59’ > > Thanks > Sumo