Mohammad Tanvir Tasnim created KAFKA-6887:
---------------------------------------------

             Summary: Error in I/O with host (java.io.EOFException) raised in 
producer
                 Key: KAFKA-6887
                 URL: https://issues.apache.org/jira/browse/KAFKA-6887
             Project: Kafka
          Issue Type: Bug
          Components: clients
    Affects Versions: 0.8.2.1, 0.9.0.0
            Reporter: Mohammad Tanvir Tasnim


In a situation with a Kafka broker in 0.9 and producers still in 0.8.2.x, 
producers seems to raise the following after a variable amount of time since 
start :

{noformat}
2016-01-29 14:33:13,066 WARN [] o.a.k.c.n.Selector: Error in I/O with 
172.22.2.170
java.io.EOFException: null
        at 
org.apache.kafka.common.network.NetworkReceive.readFrom(NetworkReceive.java:62) 
~[org.apache.kafka.kafka-clients-0.8.2.0.jar:na]
        at org.apache.kafka.common.network.Selector.poll(Selector.java:248) 
~[org.apache.kafka.kafka-clients-0.8.2.0.jar:na]
        at org.apache.kafka.clients.NetworkClient.poll(NetworkClient.java:192) 
[org.apache.kafka.kafka-clients-0.8.2.0.jar:na]
        at 
org.apache.kafka.clients.producer.internals.Sender.run(Sender.java:191) 
[org.apache.kafka.kafka-clients-0.8.2.0.jar:na]
        at 
org.apache.kafka.clients.producer.internals.Sender.run(Sender.java:122) 
[org.apache.kafka.kafka-clients-0.8.2.0.jar:na]
        at java.lang.Thread.run(Thread.java:745) [na:1.8.0_66-internal]
{noformat}

This can be reproduced successfully by doing the following :

 * Start a 0.8.2 producer connected to the 0.9 broker
 * Wait 15 minutes, exactly
 * See the error in the producer logs.

Oddly, this also shows up in an active producer but after 10 minutes of 
activity.

Kafka's server.properties :

{noformat}
broker.id=1
listeners=PLAINTEXT://:9092
port=9092
num.network.threads=2
num.io.threads=2
socket.send.buffer.bytes=1048576
socket.receive.buffer.bytes=1048576
socket.request.max.bytes=104857600
log.dirs=/mnt/data/kafka
num.partitions=4
auto.create.topics.enable=false
delete.topic.enable=true
num.recovery.threads.per.data.dir=1
log.retention.hours=48
log.retention.bytes=524288000
log.segment.bytes=52428800
log.retention.check.interval.ms=60000
log.roll.hours=24
log.cleanup.policy=delete
log.cleaner.enable=true
zookeeper.connect=127.0.0.1:2181
zookeeper.connection.timeout.ms=1000000
{noformat}

Producer's configuration :

{noformat}
        compression.type = none
        metric.reporters = []
        metadata.max.age.ms = 300000
        metadata.fetch.timeout.ms = 60000
        acks = all
        batch.size = 16384
        reconnect.backoff.ms = 10
        bootstrap.servers = [127.0.0.1:9092]
        receive.buffer.bytes = 32768
        retry.backoff.ms = 500
        buffer.memory = 33554432
        timeout.ms = 30000
        key.serializer = class 
org.apache.kafka.common.serialization.StringSerializer
        retries = 3
        max.request.size = 5000000
        block.on.buffer.full = true
        value.serializer = class 
org.apache.kafka.common.serialization.StringSerializer
        metrics.sample.window.ms = 30000
        send.buffer.bytes = 131072
        max.in.flight.requests.per.connection = 5
        metrics.num.samples = 2
        linger.ms = 0
        client.id = 
{noformat}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

Reply via email to