Hey folks, I am getting the below error when reading data from a kafka topic . I have used confluent serializers to serialize this data but when trying to consume using confuent deserializer running into the below error.Any idea on what the issue could be here ?.Also how do I skip this record and read the next one ?
10:17:32.924 [kafka-coordinator-heartbeat-thread | cgroupId9] DEBUG org.apache.kafka.clients.consumer.internals.AbstractCoordinator - [Consumer clientId=consumer-1, groupId=cgroupId9] Sending Heartbeat request to coordinator xxx.com:9092 (id: 2147483645 rack: null) 10:17:33.144 [kafka-coordinator-heartbeat-thread | cgroupId9] DEBUG org.apache.kafka.clients.consumer.internals.AbstractCoordinator - [Consumer clientId=consumer-1, groupId=cgroupId9] Received successful Heartbeat response 10:17:34.872 [pool-1-thread-1] DEBUG io.confluent.kafka.schemaregistry.client.rest.RestService - Sending GET with input null to http://xxx.yy.com:8081/schemas/ids/321 10:17:35.983 [kafka-coordinator-heartbeat-thread | cgroupId9] DEBUG org.apache.kafka.clients.consumer.internals.AbstractCoordinator - [Consumer clientId=consumer-1, groupId=cgroupId9] Sending Heartbeat request to coordinator xxx.com:9092 (id: 2147483645 rack: null) 10:17:36.203 [kafka-coordinator-heartbeat-thread | cgroupId9] DEBUG org.apache.kafka.clients.consumer.internals.AbstractCoordinator - [Consumer clientId=consumer-1, groupId=cgroupId9] Received successful Heartbeat response 10:17:39.039 [kafka-coordinator-heartbeat-thread | cgroupId9] DEBUG org.apache.kafka.clients.consumer.internals.AbstractCoordinator - [Consumer clientId=consumer-1, groupId=cgroupId9] Sending Heartbeat request to coordinator xxx.com:9092 (id: 2147483645 rack: null) 10:17:39.257 [kafka-coordinator-heartbeat-thread | cgroupId9] DEBUG org.apache.kafka.clients.consumer.internals.AbstractCoordinator - [Consumer clientId=consumer-1, groupId=cgroupId9] Received successful Heartbeat response org.apache.kafka.common.errors.SerializationException: Error deserializing key/value for partition Logs-0 at offset 25106200. If needed, please seek past the record to continue consumption. Caused by: org.apache.kafka.common.errors.SerializationException: Error deserializing Avro message for id 321 Caused by: org.apache.kafka.common.errors.SerializationException: Could not find class com.test.model.avro.Log specified in writer's schema whilst finding reader's schema for a SpecificRecord. //Consumer Configs: Properties props = new Properties(); props.put("bootstrap.servers", "xxx:9092,yyy:9092"); props.put("group.id", groupId); props.put("enable.auto.commit", "false"); props.put("session.timeout.ms", "30000"); props.put("key.deserializer", "org.apache.kafka.common.serialization.StringDeserializer"); props.put("value.deserializer", "io.confluent.kafka.serializers.KafkaAvroDeserializer"); props.put("schema.registry.url", "http://xxx:8081"); props.put(KafkaAvroDeserializerConfig.SPECIFIC_AVRO_READER_CONFIG, true); props.put(ConsumerConfig.AUTO_OFFSET_RESET_CONFIG, "earliest"); props.put("max.poll.records", "100");
