We are using the Kafka sink in Flume 1.6 to forward messages to our cluster consisting of 3 hosts. All three of these hosts are configured in the Flume config. If one of the hosts dies (i.e. we kill the Kafka daemon), the Flume sink works reasonably well in that it starts publishing data to one of the other hosts. However, if we only interrupt the connection to a particular host, for example by closing the firewall, the Kafka sink cannot cope. The problem here seems to be that the host is still part of the Kafka cluster, as the communication there works fine. It cannot be reached from the host running Flume however. The Kafka sink should recognize that and switch to one of the other hosts. Instead it keeps on trying to connect to the unreachable host until the channel overflows and we start losing messages.
Am I misunderstanding some concepts here? Is there something I can configure to enable a more desirable behavior? Thanks tobias -- Tobias Heintz Teamlead Core Telefon: +49 30 47375370 | Fax: +49 30 484984411 E-Mail: tobias.hei...@plista.com Links: www.plista.com plista GmbH | Torstraße 33-35 | 10119 Berlin | Deutschland Amtsgericht Berlin-Charlottenburg | HRB 114726 B Geschäftsführer: Andreas Richter | Christian Laase | Dr. Dominik Matyka | Jana Kusick