Thanks for the quick reply John! I should also add that the root issue is
not so much the logging, rather that these log messages seem to be
correlated with periods where producers are not able to publish data to
kafka. The issue of not being able to publish data does not seem to
resolve until restarting or updating the pipeline.
Here's my publisher config map:
.withProducerConfigUpdates(
Map.ofEntries(
Map.entry(
ProducerConfig.PARTITIONER_CLASS_CONFIG,
DefaultPartitioner.class),
Map.entry(
ProducerConfig.COMPRESSION_TYPE_CONFIG,
CompressionType.GZIP.name),
Map.entry(
CommonClientConfigs.SECURITY_PROTOCOL_CONFIG,
SecurityProtocol.SASL_SSL.name),
Map.entry(
SaslConfigs.SASL_MECHANISM,
PlainSaslServer.PLAIN_MECHANISM),
Map.entry(
SaslConfigs.SASL_JAAS_CONFIG,
"org.apache.kafka.common.security.plain.PlainLoginModule required
username=\"<api_key>\" password=\"<api_secret>\";")))
Thanks,
Evan
On Tue, Sep 13, 2022 at 10:30 AM John Casey <[email protected]> wrote:
> Hi Evan,
>
> I haven't seen this before. Can you share your Kafka write configuration,
> and any other stack traces that could be relevant?
>
> John
>
> On Tue, Sep 13, 2022 at 10:23 AM Evan Galpin <[email protected]> wrote:
>
>> Hi all,
>>
>> I've recently started using the KafkaIO connector as a sink, and am new
>> to Kafka in general. My kafka clusters are hosted by Confluent Cloud. I'm
>> using Beam SDK 2.41.0. At least daily, the producers in my Beam pipeline
>> are getting stuck in a loop frantically logging this message:
>>
>> Node n disconnected.
>>
>> Resetting the last seen epoch of partition <topic-partition> to x since
>> the associated topicId changed from null to <topic_id>
>>
>> Updating the running pipeline "resolves" the issue I believe as a result
>> of recreating the Kafka producer clients, but it seems that as-is the
>> KafkaIO producer clients are not resilient to node disconnects. Might I be
>> missing a configuration option, or are there any known issues like this?
>>
>> Thanks,
>> Evan
>>
>