Tim Tattersall created KAFKA-10735:
--------------------------------------

             Summary: Kafka producer producing corrupted avro values when 
confluent cluster is recreated and producer application is not restarted
                 Key: KAFKA-10735
                 URL: https://issues.apache.org/jira/browse/KAFKA-10735
             Project: Kafka
          Issue Type: Bug
            Reporter: Tim Tattersall


*Our Environment (AWS):*

1 x EC2 instance running 4 docker containers (using docker-compose)
 * cp-kafka 5.5.1
 * cp-zookeeper 5.5.1
 * cp-schema-registry 5.5.1
 * cp-enterprise-control-center 5.5.1

1 x ECS service running a single java application with spring-kafka producer

Topics are using String key and Avro value

 

*Problem:*
 * Avro values published after confluent cluster is recreated are corrupted. 
Expecting Avro json structure, received string value with corrupted Avro details
 ** Expected: {"metadata":{"nabEventVersion":"1.0","type":"Kafka IBMMQ sink 
connector","schemaUrl": ...*ongoing*
 ** Actual: 1.08Kafka IBMMQ source 
connector^kafka-conector-ibm-mq-source-entitlements-check\Kafka IBMMQ source 
connector - sourced....*ongoing*

 

*How to Reproduce*
 # Using an existing confluent cluster
 # Start a kafka producer java application (ours running with spring-kafka)
 # Destroy the existing confluent cluster (using docker-compose down)
 # Recreate the confluent cluster (using docker-compose up)
 # Add the topic back onto the new cluster
 # Trigger a message to be produced by the running Kafka producer

 

*Current Workaround*
 * Killing running tasks on ECS service and allowing AWS to start new ones



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

Reply via email to