My application did not have the  com.test.model.avro.Log  class during run
time . The log messages indicated this,it was just my oversight.

On Wed, Jan 23, 2019 at 10:28 AM chinchu chinchu <[email protected]>
wrote:

> Hey folks,
> I am getting the below error when reading data  from a kafka topic . I
> have used  confluent serializers to  serialize this data  but when trying
> to consume  using  confuent deserializer  running into the below error.Any
> idea on what the issue could be here ?.Also how do I skip this record and
> read the next one ?
>
>
> 10:17:32.924 [kafka-coordinator-heartbeat-thread | cgroupId9] DEBUG
> org.apache.kafka.clients.consumer.internals.AbstractCoordinator - [Consumer
> clientId=consumer-1, groupId=cgroupId9] Sending Heartbeat request to
> coordinator xxx.com:9092 (id: 2147483645 rack: null)
> 10:17:33.144 [kafka-coordinator-heartbeat-thread | cgroupId9] DEBUG
> org.apache.kafka.clients.consumer.internals.AbstractCoordinator - [Consumer
> clientId=consumer-1, groupId=cgroupId9] Received successful Heartbeat
> response
> 10:17:34.872 [pool-1-thread-1] DEBUG
> io.confluent.kafka.schemaregistry.client.rest.RestService - Sending GET
> with input null to http://xxx.yy.com:8081/schemas/ids/321
> 10:17:35.983 [kafka-coordinator-heartbeat-thread | cgroupId9] DEBUG
> org.apache.kafka.clients.consumer.internals.AbstractCoordinator - [Consumer
> clientId=consumer-1, groupId=cgroupId9] Sending Heartbeat request to
> coordinator xxx.com:9092 (id: 2147483645 rack: null)
> 10:17:36.203 [kafka-coordinator-heartbeat-thread | cgroupId9] DEBUG
> org.apache.kafka.clients.consumer.internals.AbstractCoordinator - [Consumer
> clientId=consumer-1, groupId=cgroupId9] Received successful Heartbeat
> response
> 10:17:39.039 [kafka-coordinator-heartbeat-thread | cgroupId9] DEBUG
> org.apache.kafka.clients.consumer.internals.AbstractCoordinator - [Consumer
> clientId=consumer-1, groupId=cgroupId9] Sending Heartbeat request to
> coordinator xxx.com:9092 (id: 2147483645 rack: null)
> 10:17:39.257 [kafka-coordinator-heartbeat-thread | cgroupId9] DEBUG
> org.apache.kafka.clients.consumer.internals.AbstractCoordinator - [Consumer
> clientId=consumer-1, groupId=cgroupId9] Received successful Heartbeat
> response
>
> org.apache.kafka.common.errors.SerializationException: Error deserializing
> key/value for partition Logs-0 at offset 25106200.
> If needed, please seek past the record to continue consumption.
> Caused by: org.apache.kafka.common.errors.SerializationException: Error
> deserializing Avro message for id 321
> Caused by: org.apache.kafka.common.errors.SerializationException: Could
> not find class com.test.model.avro.Log specified in writer's schema whilst
> finding reader's schema for a SpecificRecord.
>
>
> //Consumer Configs:
>             Properties props = new Properties();
>     props.put("bootstrap.servers", "xxx:9092,yyy:9092");
>     props.put("group.id", groupId);
>     props.put("enable.auto.commit", "false");
>     props.put("session.timeout.ms", "30000");
>     props.put("key.deserializer",
> "org.apache.kafka.common.serialization.StringDeserializer");
>     props.put("value.deserializer",
> "io.confluent.kafka.serializers.KafkaAvroDeserializer");
>     props.put("schema.registry.url", "http://xxx:8081";);
>     props.put(KafkaAvroDeserializerConfig.SPECIFIC_AVRO_READER_CONFIG,
> true);
>     props.put(ConsumerConfig.AUTO_OFFSET_RESET_CONFIG, "earliest");
>     props.put("max.poll.records", "100");
>

Reply via email to