[
https://issues.apache.org/jira/browse/KAFKA-10363?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Alexey Kornev updated KAFKA-10363:
----------------------------------
Description:
We've just successfully set up a Kafka cluster consists of 3 brokers and faced
with the following issue: when we change order of zookeeper servers in
zookeeper.connect property in server.properties files and restart Kafka broker
then this Kafka broker tries to connect to a new Kafka cluster. As a result,
Kafka broker throws an error and shutdown.
For example, config server.properties on first broker:
{code:java}
broker.id=-1
...
zookeeper.connect=node_1:2181/kafka,node_2:2181/kafka,node_3:2181/kafka
{code}
We changed it to
{code:java}
broker.id=-1
...
zookeeper.connect=node_2:2181/kafka,node_3:2181/kafka,node_1:2181/kafka {code}
and restart Kafka broker.
Logs:
{code:java}
[2020-08-05 09:07:55,658] INFO [ExpirationReaper-0-Heartbeat]: Starting
(kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper)[2020-08-05
09:07:55,658] INFO [ExpirationReaper-0-Heartbeat]: Starting
(kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper)[2020-08-05
09:07:55,658] INFO [ExpirationReaper-0-topic]: Starting
(kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper)[2020-08-05
09:07:57,070] INFO Registered kafka:type=kafka.Log4jController MBean
(kafka.utils.Log4jControllerRegistration$)[2020-08-05 09:07:57,656] INFO
Registered signal handlers for TERM, INT, HUP
(org.apache.kafka.common.utils.LoggingSignalHandler)[2020-08-05 09:07:57,657]
INFO starting (kafka.server.KafkaServer)[2020-08-05 09:07:57,658] INFO
Connecting to zookeeper on
node_2:2181/kafka,node_3:2181/kafka,node_1:2181/kafka
(kafka.server.KafkaServer)[2020-08-05 09:07:57,685] INFO [ZooKeeperClient Kafka
server] Initializing a new session to node_2:2181.
(kafka.zookeeper.ZooKeeperClient)[2020-08-05 09:07:57,690] INFO Client
environment:zookeeper.version=3.4.14-4c25d480e66aadd371de8bd2fd8da255ac140bcf,
built on 03/06/2019 16:18 GMT (org.apache.zookeeper.ZooKeeper)[2020-08-05
09:07:57,693] INFO Client environment:host.name=localhost
(org.apache.zookeeper.ZooKeeper)[2020-08-05 09:07:57,693] INFO Client
environment:java.version=11.0.8 (org.apache.zookeeper.ZooKeeper)[2020-08-05
09:07:57,696] INFO Client environment:java.vendor=Ubuntu
(org.apache.zookeeper.ZooKeeper)[2020-08-05 09:07:57,696] INFO Client
environment:java.home=/usr/lib/jvm/java-11-openjdk-amd64
(org.apache.zookeeper.ZooKeeper)[2020-08-05 09:07:57,696] INFO Client
environment:java.class.path=/opt/kafka/current/bin/../libs/activation-1.1.1.jar:/opt/kafka/current/bin/../libs/aopalliance-repackaged-2.5.0.jar:/opt/kafka/current/bin/../libs/argparse4j-0.7.0.jar:/opt/kafka/current/bin/../libs/audience-annotations-0.5.0.jar:/opt/kafka/current/bin/../libs/commons-lang3-3.8.1.jar:/opt/kafka/current/bin/../libs/connect-api-2.3.1.jar:/opt/kafka/current/bin/../libs/connect-basic-auth-extension-2.3.1.jar:/opt/kafka/current/bin/../libs/connect-file-2.3.1.jar:/opt/kafka/current/bin/../libs/connect-json-2.3.1.jar:/opt/kafka/current/bin/../libs/connect-runtime-2.3.1.jar:/opt/kafka/current/bin/../libs/connect-transforms-2.3.1.jar:/opt/kafka/current/bin/../libs/guava-20.0.jar:/opt/kafka/current/bin/../libs/hk2-api-2.5.0.jar:/opt/kafka/current/bin/../libs/hk2-locator-2.5.0.jar:/opt/kafka/current/bin/../libs/hk2-utils-2.5.0.jar:/opt/kafka/current/bin/../libs/jackson-annotations-2.10.0.jar:/opt/kafka/current/bin/../libs/jackson-core-2.10.0.jar:/opt/kafka/current/bin/../libs/jackson-databind-2.10.0.jar:/opt/kafka/current/bin/../libs/jackson-dataformat-csv-2.10.0.jar:/opt/kafka/current/bin/../libs/jackson-datatype-jdk8-2.10.0.jar:/opt/kafka/current/bin/../libs/jackson-jaxrs-base-2.10.0.jar:/opt/kafka/current/bin/../libs/jackson-jaxrs-json-provider-2.10.0.jar:/opt/kafka/current/bin/../libs/jackson-module-jaxb-annotations-2.10.0.jar:/opt/kafka/current/bin/../libs/jackson-module-paranamer-2.10.0.jar:/opt/kafka/current/bin/../libs/jackson-module-scala_2.12-2.10.0.jar:/opt/kafka/current/bin/../libs/jakarta.activation-api-1.2.1.jar:/opt/kafka/current/bin/../libs/jakarta.annotation-api-1.3.4.jar:/opt/kafka/current/bin/../libs/jakarta.inject-2.5.0.jar:/opt/kafka/current/bin/../libs/jakarta.ws.rs-api-2.1.5.jar:/opt/kafka/current/bin/../libs/jakarta.xml.bind-api-2.3.2.jar:/opt/kafka/current/bin/../libs/javassist-3.22.0-CR2.jar:/opt/kafka/current/bin/../libs/javax.servlet-api-3.1.0.jar:/opt/kafka/current/bin/../libs/javax.ws.rs-api-2.1.1.jar:/opt/kafka/current/bin/../libs/jaxb-api-2.3.0.jar:/opt/kafka/current/bin/../libs/jersey-client-2.28.jar:/opt/kafka/current/bin/../libs/jersey-common-2.28.jar:/opt/kafka/current/bin/../libs/jersey-container-servlet-2.28.jar:/opt/kafka/current/bin/../libs/jersey-container-servlet-core-2.28.jar:/opt/kafka/current/bin/../libs/jersey-hk2-2.28.jar:/opt/kafka/current/bin/../libs/jersey-media-jaxb-2.28.jar:/opt/kafka/current/bin/../libs/jersey-server-2.28.jar:/opt/kafka/current/bin/../libs/jetty-client-9.4.18.v20190429.jar:/opt/kafka/current/bin/../libs/jetty-continuation-9.4.18.v20190429.jar:/opt/kafka/current/bin/../libs/jetty-http-9.4.18.v20190429.jar:/opt/kafka/current/bin/../libs/jetty-io-9.4.18.v20190429.jar:/opt/kafka/current/bin/../libs/jetty-security-9.4.18.v20190429.jar:/opt/kafka/current/bin/../libs/jetty-server-9.4.18.v20190429.jar:/opt/kafka/current/bin/../libs/jetty-servlet-9.4.18.v20190429.jar:/opt/kafka/current/bin/../libs/jetty-servlets-9.4.18.v20190429.jar:/opt/kafka/current/bin/../libs/jetty-util-9.4.18.v20190429.jar:/opt/kafka/current/bin/../libs/jopt-simple-5.0.4.jar:/opt/kafka/current/bin/../libs/jsr305-3.0.2.jar:/opt/kafka/current/bin/../libs/kafka-clients-2.3.1.jar:/opt/kafka/current/bin/../libs/kafka-log4j-appender-2.3.1.jar:/opt/kafka/current/bin/../libs/kafka-streams-2.3.1.jar:/opt/kafka/current/bin/../libs/kafka-streams-examples-2.3.1.jar:/opt/kafka/current/bin/../libs/kafka-streams-scala_2.12-2.3.1.jar:/opt/kafka/current/bin/../libs/kafka-streams-test-utils-2.3.1.jar:/opt/kafka/current/bin/../libs/kafka-tools-2.3.1.jar:/opt/kafka/current/bin/../libs/kafka_2.12-2.3.1-sources.jar:/opt/kafka/current/bin/../libs/kafka_2.12-2.3.1.jar:/opt/kafka/current/bin/../libs/log4j-1.2.17.jar:/opt/kafka/current/bin/../libs/lz4-java-1.6.0.jar:/opt/kafka/current/bin/../libs/maven-artifact-3.6.1.jar:/opt/kafka/current/bin/../libs/metrics-core-2.2.0.jar:/opt/kafka/current/bin/../libs/osgi-resource-locator-1.0.1.jar:/opt/kafka/current/bin/../libs/paranamer-2.8.jar:/opt/kafka/current/bin/../libs/plexus-utils-3.2.0.jar:/opt/kafka/current/bin/../libs/reflections-0.9.11.jar:/opt/kafka/current/bin/../libs/rocksdbjni-5.18.3.jar:/opt/kafka/current/bin/../libs/scala-library-2.12.10.jar:/opt/kafka/current/bin/../libs/scala-library-2.12.8.jar:/opt/kafka/current/bin/../libs/scala-logging_2.12-3.9.0.jar:/opt/kafka/current/bin/../libs/scala-reflect-2.12.8.jar:/opt/kafka/current/bin/../libs/slf4j-api-1.7.26.jar:/opt/kafka/current/bin/../libs/slf4j-log4j12-1.7.26.jar:/opt/kafka/current/bin/../libs/snappy-java-1.1.7.3.jar:/opt/kafka/current/bin/../libs/spotbugs-annotations-3.1.9.jar:/opt/kafka/current/bin/../libs/validation-api-2.0.1.Final.jar:/opt/kafka/current/bin/../libs/zkclient-0.11.jar:/opt/kafka/current/bin/../libs/zookeeper-3.4.14.jar:/opt/kafka/current/bin/../libs/zstd-jni-1.4.0-1.jar
(org.apache.zookeeper.ZooKeeper)[2020-08-05 09:07:57,697] INFO Client
environment:java.library.path=/usr/java/packages/lib:/usr/lib/x86_64-linux-gnu/jni:/lib/x86_64-linux-gnu:/usr/lib/x86_64-linux-gnu:/usr/lib/jni:/lib:/usr/lib
(org.apache.zookeeper.ZooKeeper)[2020-08-05 09:07:57,697] INFO Client
environment:java.io.tmpdir=/tmp (org.apache.zookeeper.ZooKeeper)[2020-08-05
09:07:57,697] INFO Client environment:java.compiler=<NA>
(org.apache.zookeeper.ZooKeeper)[2020-08-05 09:07:57,697] INFO Client
environment:os.name=Linux (org.apache.zookeeper.ZooKeeper)[2020-08-05
09:07:57,697] INFO Client environment:os.arch=amd64
(org.apache.zookeeper.ZooKeeper)[2020-08-05 09:07:57,698] INFO Client
environment:os.version=4.15.0-66-generic
(org.apache.zookeeper.ZooKeeper)[2020-08-05 09:07:57,698] INFO Client
environment:user.name=kafka (org.apache.zookeeper.ZooKeeper)[2020-08-05
09:07:57,698] INFO Client environment:user.home=/opt/kafka
(org.apache.zookeeper.ZooKeeper)[2020-08-05 09:07:57,698] INFO Client
environment:user.dir=/ (org.apache.zookeeper.ZooKeeper)[2020-08-05
09:07:57,699] INFO Initiating client connection, connectString=node_2:2181
sessionTimeout=6000
watcher=kafka.zookeeper.ZooKeeperClient$ZooKeeperClientWatcher$@74bada02
(org.apache.zookeeper.ZooKeeper)[2020-08-05 09:07:57,718] INFO [ZooKeeperClient
Kafka server] Waiting until connected.
(kafka.zookeeper.ZooKeeperClient)[2020-08-05 09:07:57,818] INFO Client
successfully logged in. (org.apache.zookeeper.Login)[2020-08-05 09:07:57,821]
INFO Client will use DIGEST-MD5 as SASL mechanism.
(org.apache.zookeeper.client.ZooKeeperSaslClient)[2020-08-05 09:07:57,826] INFO
Opening socket connection to server node_2/node_2:2181. Will attempt to
SASL-authenticate using Login Context section 'Client'
(org.apache.zookeeper.ClientCnxn)[2020-08-05 09:07:57,832] INFO Socket
connection established to node_2/node_2:2181, initiating session
(org.apache.zookeeper.ClientCnxn)[2020-08-05 09:07:57,841] INFO Session
establishment complete on server node_2/node_2:2181, sessionid =
0x373bdbbd3b00002, negotiated timeout = 6000
(org.apache.zookeeper.ClientCnxn)[2020-08-05 09:07:57,847] INFO
[ZooKeeperClient Kafka server] Connected.
(kafka.zookeeper.ZooKeeperClient)[2020-08-05 09:07:57,925] INFO Created
zookeeper path /kafka,node_3:2181/kafka,node_1:2181/kafka
(kafka.server.KafkaServer)[2020-08-05 09:07:57,926] INFO [ZooKeeperClient Kafka
server] Closing. (kafka.zookeeper.ZooKeeperClient)[2020-08-05 09:07:57,933]
INFO Session: 0x373bdbbd3b00002 closed
(org.apache.zookeeper.ZooKeeper)[2020-08-05 09:07:57,934] INFO EventThread shut
down for session: 0x373bdbbd3b00002
(org.apache.zookeeper.ClientCnxn)[2020-08-05 09:07:57,937] INFO
[ZooKeeperClient Kafka server] Closed.
(kafka.zookeeper.ZooKeeperClient)[2020-08-05 09:07:57,939] INFO
[ZooKeeperClient Kafka server] Initializing a new session to
node_2:2181/kafka,node_3:2181/kafka,node_1:2181/kafka.
(kafka.zookeeper.ZooKeeperClient)[2020-08-05 09:07:57,939] INFO Initiating
client connection,
connectString=node_2:2181/kafka,node_3:2181/kafka,node_1:2181/kafka
sessionTimeout=6000
watcher=kafka.zookeeper.ZooKeeperClient$ZooKeeperClientWatcher$@6ff65192
(org.apache.zookeeper.ZooKeeper)[2020-08-05 09:07:57,940] INFO [ZooKeeperClient
Kafka server] Waiting until connected.
(kafka.zookeeper.ZooKeeperClient)[2020-08-05 09:07:57,941] INFO Client will use
DIGEST-MD5 as SASL mechanism.
(org.apache.zookeeper.client.ZooKeeperSaslClient)[2020-08-05 09:07:57,943] INFO
Opening socket connection to server node_2/node_2:2181. Will attempt to
SASL-authenticate using Login Context section 'Client'
(org.apache.zookeeper.ClientCnxn)[2020-08-05 09:07:57,944] INFO Socket
connection established to node_2/node_2:2181, initiating session
(org.apache.zookeeper.ClientCnxn)[2020-08-05 09:07:57,949] INFO Session
establishment complete on server node_2/node_2:2181, sessionid =
0x373bdbbd3b00003, negotiated timeout = 6000
(org.apache.zookeeper.ClientCnxn)[2020-08-05 09:07:57,950] INFO
[ZooKeeperClient Kafka server] Connected.
(kafka.zookeeper.ZooKeeperClient)[2020-08-05 09:07:58,167] INFO Cluster ID =
5_d5S6HeQBWf0ZzwQ6TjRA (kafka.server.KafkaServer)[2020-08-05 09:07:58,253] INFO
KafkaConfig values: advertised.host.name = node_3 advertised.listeners =
SSL://node_3:9093 advertised.port = null alter.config.policy.class.name = null
alter.log.dirs.replication.quota.window.num = 11
alter.log.dirs.replication.quota.window.size.seconds = 1 authorizer.class.name
= kafka.security.auth.SimpleAclAuthorizer auto.create.topics.enable = true
auto.leader.rebalance.enable = true background.threads = 10 broker.id = -1
broker.id.generation.enable = true broker.rack = null
client.quota.callback.class = null compression.type = producer
connection.failed.authentication.delay.ms = 100 connections.max.idle.ms =
600000 connections.max.reauth.ms = 0 control.plane.listener.name = null
controlled.shutdown.enable = true controlled.shutdown.max.retries = 3
controlled.shutdown.retry.backoff.ms = 5000 controller.socket.timeout.ms =
30000 create.topic.policy.class.name = null default.replication.factor = 2
delegation.token.expiry.check.interval.ms = 3600000
delegation.token.expiry.time.ms = 86400000 delegation.token.master.key = null
delegation.token.max.lifetime.ms = 604800000
delete.records.purgatory.purge.interval.requests = 1 delete.topic.enable = true
fetch.purgatory.purge.interval.requests = 1000 group.initial.rebalance.delay.ms
= 3000 group.max.session.timeout.ms = 60000 group.max.size = 2147483647
group.min.session.timeout.ms = 10000 host.name = node_3
inter.broker.listener.name = null inter.broker.protocol.version = 2.3-IV1
kafka.metrics.polling.interval.secs = 10 kafka.metrics.reporters = []
leader.imbalance.check.interval.seconds = 300
leader.imbalance.per.broker.percentage = 10 listener.security.protocol.map =
SSL:SSL listeners = SSL://node_3:9093 log.cleaner.backoff.ms = 15000
log.cleaner.dedupe.buffer.size = 134217728 log.cleaner.delete.retention.ms =
86400000 log.cleaner.enable = true log.cleaner.io.buffer.load.factor = 0.9
log.cleaner.io.buffer.size = 524288 log.cleaner.io.max.bytes.per.second =
1.7976931348623157E308 log.cleaner.max.compaction.lag.ms = 9223372036854775807
log.cleaner.min.cleanable.ratio = 0.5 log.cleaner.min.compaction.lag.ms = 0
log.cleaner.threads = 1 log.cleanup.policy = [delete] log.dir = /tmp/kafka-logs
log.dirs = /var/lib/kafka log.flush.interval.messages = 5000
log.flush.interval.ms = 5000 log.flush.offset.checkpoint.interval.ms = 60000
log.flush.scheduler.interval.ms = 9223372036854775807
log.flush.start.offset.checkpoint.interval.ms = 60000 log.index.interval.bytes
= 4096 log.index.size.max.bytes = 10485760 log.message.downconversion.enable =
true log.message.format.version = 2.3-IV1
log.message.timestamp.difference.max.ms = 9223372036854775807
log.message.timestamp.type = CreateTime log.preallocate = false
log.retention.bytes = -1 log.retention.check.interval.ms = 300000
log.retention.hours = 336 log.retention.minutes = null log.retention.ms = null
log.roll.hours = 168 log.roll.jitter.hours = 0 log.roll.jitter.ms = null
log.roll.ms = null log.segment.bytes = 1073741824 log.segment.delete.delay.ms =
60000 max.connections = 2147483647 max.connections.per.ip = 2147483647
max.connections.per.ip.overrides = max.incremental.fetch.session.cache.slots =
1000 message.max.bytes = 2000024 metric.reporters = [] metrics.num.samples = 2
metrics.recording.level = INFO metrics.sample.window.ms = 30000
min.insync.replicas = 1 num.io.threads = 8 num.network.threads = 3
num.partitions = 4 num.recovery.threads.per.data.dir = 4
num.replica.alter.log.dirs.threads = null num.replica.fetchers = 1
offset.metadata.max.bytes = 4096 offsets.commit.required.acks = -1
offsets.commit.timeout.ms = 5000 offsets.load.buffer.size = 5242880
offsets.retention.check.interval.ms = 600000 offsets.retention.minutes = 10080
offsets.topic.compression.codec = 0 offsets.topic.num.partitions = 50
offsets.topic.replication.factor = 2 offsets.topic.segment.bytes = 104857600
password.encoder.cipher.algorithm = AES/CBC/PKCS5Padding
password.encoder.iterations = 4096 password.encoder.key.length = 128
password.encoder.keyfactory.algorithm = null password.encoder.old.secret = null
password.encoder.secret = null port = 9092 principal.builder.class = null
producer.purgatory.purge.interval.requests = 1000 queued.max.request.bytes = -1
queued.max.requests = 500 quota.consumer.default = 9223372036854775807
quota.producer.default = 9223372036854775807 quota.window.num = 11
quota.window.size.seconds = 1 replica.fetch.backoff.ms = 1000
replica.fetch.max.bytes = 1048576 replica.fetch.min.bytes = 1
replica.fetch.response.max.bytes = 10485760 replica.fetch.wait.max.ms = 500
replica.high.watermark.checkpoint.interval.ms = 5000 replica.lag.time.max.ms =
30000 replica.socket.receive.buffer.bytes = 65536 replica.socket.timeout.ms =
30000 replication.quota.window.num = 11 replication.quota.window.size.seconds =
1 request.timeout.ms = 30000 reserved.broker.max.id = 1000
sasl.client.callback.handler.class = null sasl.enabled.mechanisms = [GSSAPI]
sasl.jaas.config = null sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.principal.to.local.rules = [DEFAULT] sasl.kerberos.service.name =
null sasl.kerberos.ticket.renew.jitter = 0.05
sasl.kerberos.ticket.renew.window.factor = 0.8
sasl.login.callback.handler.class = null sasl.login.class = null
sasl.login.refresh.buffer.seconds = 300 sasl.login.refresh.min.period.seconds =
60 sasl.login.refresh.window.factor = 0.8 sasl.login.refresh.window.jitter =
0.05 sasl.mechanism.inter.broker.protocol = GSSAPI
sasl.server.callback.handler.class = null security.inter.broker.protocol = SSL
socket.receive.buffer.bytes = 102400 socket.request.max.bytes = 104857600
socket.send.buffer.bytes = 102400 ssl.cipher.suites = [] ssl.client.auth =
required ssl.enabled.protocols = [TLSv1.2]
ssl.endpoint.identification.algorithm = HTTPS ssl.key.password = [hidden]
ssl.keymanager.algorithm = SunX509 ssl.keystore.location =
/etc/ssl/kafka/kafka.keystore.jks ssl.keystore.password = [hidden]
ssl.keystore.type = JKS ssl.principal.mapping.rules = [DEFAULT] ssl.protocol =
TLS ssl.provider = null ssl.secure.random.implementation = SHA1PRNG
ssl.trustmanager.algorithm = PKIX ssl.truststore.location =
/etc/ssl/kafka/kafka.truststore.jks ssl.truststore.password = [hidden]
ssl.truststore.type = JKS
transaction.abort.timed.out.transaction.cleanup.interval.ms = 180000
transaction.max.timeout.ms = 900000
transaction.remove.expired.transaction.cleanup.interval.ms = 3600000
transaction.state.log.load.buffer.size = 5242880 transaction.state.log.min.isr
= 2 transaction.state.log.num.partitions = 50
transaction.state.log.replication.factor = 2
transaction.state.log.segment.bytes = 104857600 transactional.id.expiration.ms
= 604800000 unclean.leader.election.enable = false zookeeper.connect =
node_2:2181/kafka,node_3:2181/kafka,node_1:2181/kafka
zookeeper.connection.timeout.ms = 15000 zookeeper.max.in.flight.requests = 10
zookeeper.session.timeout.ms = 6000 zookeeper.set.acl = true
zookeeper.sync.time.ms = 2000 (kafka.server.KafkaConfig)[2020-08-05
09:07:58,272] INFO KafkaConfig values: advertised.host.name = node_3
advertised.listeners = SSL://node_3:9093 advertised.port = null
alter.config.policy.class.name = null
alter.log.dirs.replication.quota.window.num = 11
alter.log.dirs.replication.quota.window.size.seconds = 1 authorizer.class.name
= kafka.security.auth.SimpleAclAuthorizer auto.create.topics.enable = true
auto.leader.rebalance.enable = true background.threads = 10 broker.id = -1
broker.id.generation.enable = true broker.rack = null
client.quota.callback.class = null compression.type = producer
connection.failed.authentication.delay.ms = 100 connections.max.idle.ms =
600000 connections.max.reauth.ms = 0 control.plane.listener.name = null
controlled.shutdown.enable = true controlled.shutdown.max.retries = 3
controlled.shutdown.retry.backoff.ms = 5000 controller.socket.timeout.ms =
30000 create.topic.policy.class.name = null default.replication.factor = 2
delegation.token.expiry.check.interval.ms = 3600000
delegation.token.expiry.time.ms = 86400000 delegation.token.master.key = null
delegation.token.max.lifetime.ms = 604800000
delete.records.purgatory.purge.interval.requests = 1 delete.topic.enable = true
fetch.purgatory.purge.interval.requests = 1000 group.initial.rebalance.delay.ms
= 3000 group.max.session.timeout.ms = 60000 group.max.size = 2147483647
group.min.session.timeout.ms = 10000 host.name = node_3
inter.broker.listener.name = null inter.broker.protocol.version = 2.3-IV1
kafka.metrics.polling.interval.secs = 10 kafka.metrics.reporters = []
leader.imbalance.check.interval.seconds = 300
leader.imbalance.per.broker.percentage = 10 listener.security.protocol.map =
SSL:SSL listeners = SSL://node_3:9093 log.cleaner.backoff.ms = 15000
log.cleaner.dedupe.buffer.size = 134217728 log.cleaner.delete.retention.ms =
86400000 log.cleaner.enable = true log.cleaner.io.buffer.load.factor = 0.9
log.cleaner.io.buffer.size = 524288 log.cleaner.io.max.bytes.per.second =
1.7976931348623157E308 log.cleaner.max.compaction.lag.ms = 9223372036854775807
log.cleaner.min.cleanable.ratio = 0.5 log.cleaner.min.compaction.lag.ms = 0
log.cleaner.threads = 1 log.cleanup.policy = [delete] log.dir = /tmp/kafka-logs
log.dirs = /var/lib/kafka log.flush.interval.messages = 5000
log.flush.interval.ms = 5000 log.flush.offset.checkpoint.interval.ms = 60000
log.flush.scheduler.interval.ms = 9223372036854775807
log.flush.start.offset.checkpoint.interval.ms = 60000 log.index.interval.bytes
= 4096 log.index.size.max.bytes = 10485760 log.message.downconversion.enable =
true log.message.format.version = 2.3-IV1
log.message.timestamp.difference.max.ms = 9223372036854775807
log.message.timestamp.type = CreateTime log.preallocate = false
log.retention.bytes = -1 log.retention.check.interval.ms = 300000
log.retention.hours = 336 log.retention.minutes = null log.retention.ms = null
log.roll.hours = 168 log.roll.jitter.hours = 0 log.roll.jitter.ms = null
log.roll.ms = null log.segment.bytes = 1073741824 log.segment.delete.delay.ms =
60000 max.connections = 2147483647 max.connections.per.ip = 2147483647
max.connections.per.ip.overrides = max.incremental.fetch.session.cache.slots =
1000 message.max.bytes = 2000024 metric.reporters = [] metrics.num.samples = 2
metrics.recording.level = INFO metrics.sample.window.ms = 30000
min.insync.replicas = 1 num.io.threads = 8 num.network.threads = 3
num.partitions = 4 num.recovery.threads.per.data.dir = 4
num.replica.alter.log.dirs.threads = null num.replica.fetchers = 1
offset.metadata.max.bytes = 4096 offsets.commit.required.acks = -1
offsets.commit.timeout.ms = 5000 offsets.load.buffer.size = 5242880
offsets.retention.check.interval.ms = 600000 offsets.retention.minutes = 10080
offsets.topic.compression.codec = 0 offsets.topic.num.partitions = 50
offsets.topic.replication.factor = 2 offsets.topic.segment.bytes = 104857600
password.encoder.cipher.algorithm = AES/CBC/PKCS5Padding
password.encoder.iterations = 4096 password.encoder.key.length = 128
password.encoder.keyfactory.algorithm = null password.encoder.old.secret = null
password.encoder.secret = null port = 9092 principal.builder.class = null
producer.purgatory.purge.interval.requests = 1000 queued.max.request.bytes = -1
queued.max.requests = 500 quota.consumer.default = 9223372036854775807
quota.producer.default = 9223372036854775807 quota.window.num = 11
quota.window.size.seconds = 1 replica.fetch.backoff.ms = 1000
replica.fetch.max.bytes = 1048576 replica.fetch.min.bytes = 1
replica.fetch.response.max.bytes = 10485760 replica.fetch.wait.max.ms = 500
replica.high.watermark.checkpoint.interval.ms = 5000 replica.lag.time.max.ms =
30000 replica.socket.receive.buffer.bytes = 65536 replica.socket.timeout.ms =
30000 replication.quota.window.num = 11 replication.quota.window.size.seconds =
1 request.timeout.ms = 30000 reserved.broker.max.id = 1000
sasl.client.callback.handler.class = null sasl.enabled.mechanisms = [GSSAPI]
sasl.jaas.config = null sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.principal.to.local.rules = [DEFAULT] sasl.kerberos.service.name =
null sasl.kerberos.ticket.renew.jitter = 0.05
sasl.kerberos.ticket.renew.window.factor = 0.8
sasl.login.callback.handler.class = null sasl.login.class = null
sasl.login.refresh.buffer.seconds = 300 sasl.login.refresh.min.period.seconds =
60 sasl.login.refresh.window.factor = 0.8 sasl.login.refresh.window.jitter =
0.05 sasl.mechanism.inter.broker.protocol = GSSAPI
sasl.server.callback.handler.class = null security.inter.broker.protocol = SSL
socket.receive.buffer.bytes = 102400 socket.request.max.bytes = 104857600
socket.send.buffer.bytes = 102400 ssl.cipher.suites = [] ssl.client.auth =
required ssl.enabled.protocols = [TLSv1.2]
ssl.endpoint.identification.algorithm = HTTPS ssl.key.password = [hidden]
ssl.keymanager.algorithm = SunX509 ssl.keystore.location =
/etc/ssl/kafka/kafka.keystore.jks ssl.keystore.password = [hidden]
ssl.keystore.type = JKS ssl.principal.mapping.rules = [DEFAULT] ssl.protocol =
TLS ssl.provider = null ssl.secure.random.implementation = SHA1PRNG
ssl.trustmanager.algorithm = PKIX ssl.truststore.location =
/etc/ssl/kafka/kafka.truststore.jks ssl.truststore.password = [hidden]
ssl.truststore.type = JKS
transaction.abort.timed.out.transaction.cleanup.interval.ms = 180000
transaction.max.timeout.ms = 900000
transaction.remove.expired.transaction.cleanup.interval.ms = 3600000
transaction.state.log.load.buffer.size = 5242880 transaction.state.log.min.isr
= 2 transaction.state.log.num.partitions = 50
transaction.state.log.replication.factor = 2
transaction.state.log.segment.bytes = 104857600 transactional.id.expiration.ms
= 604800000 unclean.leader.election.enable = false zookeeper.connect =
node_2:2181/kafka,node_3:2181/kafka,node_1:2181/kafka
zookeeper.connection.timeout.ms = 15000 zookeeper.max.in.flight.requests = 10
zookeeper.session.timeout.ms = 6000 zookeeper.set.acl = true
zookeeper.sync.time.ms = 2000 (kafka.server.KafkaConfig)[2020-08-05
09:07:58,328] INFO [ThrottledChannelReaper-Produce]: Starting
(kafka.server.ClientQuotaManager$ThrottledChannelReaper)[2020-08-05
09:07:58,328] INFO [ThrottledChannelReaper-Fetch]: Starting
(kafka.server.ClientQuotaManager$ThrottledChannelReaper)[2020-08-05
09:07:58,331] INFO [ThrottledChannelReaper-Request]: Starting
(kafka.server.ClientQuotaManager$ThrottledChannelReaper)[2020-08-05
09:07:58,361] INFO Loading logs. (kafka.log.LogManager)[2020-08-05
09:07:58,374] INFO Logs loading complete in 13 ms.
(kafka.log.LogManager)[2020-08-05 09:07:58,403] INFO Starting log cleanup with
a period of 300000 ms. (kafka.log.LogManager)[2020-08-05 09:07:58,407] INFO
Starting log flusher with a default period of 9223372036854775807 ms.
(kafka.log.LogManager)[2020-08-05 09:07:58,805] INFO Awaiting socket
connections on node_3:9093. (kafka.network.Acceptor)[2020-08-05 09:08:00,594]
INFO [SocketServer brokerId=0] Created data-plane acceptor and processors for
endpoint : EndPoint(node_3,9093,ListenerName(SSL),SSL)
(kafka.network.SocketServer)[2020-08-05 09:08:00,596] INFO [SocketServer
brokerId=0] Started 1 acceptor threads for data-plane
(kafka.network.SocketServer)[2020-08-05 09:08:00,629] INFO
[ExpirationReaper-0-Produce]: Starting
(kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper)[2020-08-05
09:08:00,630] INFO [ExpirationReaper-0-Fetch]: Starting
(kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper)[2020-08-05
09:08:00,630] INFO [ExpirationReaper-0-DeleteRecords]: Starting
(kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper)[2020-08-05
09:08:00,631] INFO [ExpirationReaper-0-ElectPreferredLeader]: Starting
(kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper)[2020-08-05
09:08:00,660] INFO [LogDirFailureHandler]: Starting
(kafka.server.ReplicaManager$LogDirFailureHandler)[2020-08-05 09:08:00,726]
INFO Creating /brokers/ids/0 (is it secure? true)
(kafka.zk.KafkaZkClient)[2020-08-05 09:08:00,751] ERROR Error while creating
ephemeral at /brokers/ids/0, node already exists and owner '248751018843570177'
does not match current session '248751018843570179'
(kafka.zk.KafkaZkClient$CheckedEphemeral)[2020-08-05 09:08:00,757] ERROR
[KafkaServer id=0] Fatal error during KafkaServer startup. Prepare to shutdown
(kafka.server.KafkaServer)org.apache.zookeeper.KeeperException$NodeExistsException:
KeeperErrorCode = NodeExists at
org.apache.zookeeper.KeeperException.create(KeeperException.java:122) at
kafka.zk.KafkaZkClient$CheckedEphemeral.getAfterNodeExists(KafkaZkClient.scala:1784)
at kafka.zk.KafkaZkClient$CheckedEphemeral.create(KafkaZkClient.scala:1722) at
kafka.zk.KafkaZkClient.checkedEphemeralCreate(KafkaZkClient.scala:1689) at
kafka.zk.KafkaZkClient.registerBroker(KafkaZkClient.scala:97) at
kafka.server.KafkaServer.startup(KafkaServer.scala:262) at
kafka.server.KafkaServerStartable.startup(KafkaServerStartable.scala:38) at
kafka.Kafka$.main(Kafka.scala:84) at kafka.Kafka.main(Kafka.scala)[2020-08-05
09:08:00,763] INFO [KafkaServer id=0] shutting down (kafka.server.KafkaServer)
{code}
As you can see here, this broker tries to connect to Cluster with ID
_5_d5S6HeQBWf0ZzwQ6TjRA_, but other brokers are connected to Cluster with ID
_OhWuEGMeQHe66HP74rurRA_.
If we bring back the order of zookeeper servers in the config file then Kafka
broker starts normally and connects to an existing cluster.
This issue blocks us from adding new nodes to the cluster and removes the old
ones.
Cluster details:
* 3 Kafka nodes cluster running 2.3.1 (also reproduced on 2.4.0)
* 3 Zookeeper node cluster running 3.4.10
was:
We've just successfully set up a Kafka cluster consists of 3 brokers and faced
with the following issue: when we change order of zookeeper servers in
zookeeper.connect property in server.properties files and restart Kafka broker
then this Kafka broker tries to connect to a new Kafka cluster. As a result,
Kafka broker throws an error and shutdown.
For example, config server.properties on first broker:
{code:java}
broker.id=-1
...
zookeeper.connect=node_1:2181/kafka,node_2:2181/kafka,node_3:2181/kafka
{code}
We changed it to
{code:java}
broker.id=-1
...
zookeeper.connect=node_2:2181/kafka,node_3:2181/kafka,node_1:2181/kafka
{code}
and restart Kafka broker.
Logs:
{code:java}
[2020-08-05 09:07:55,658] INFO [ExpirationReaper-0-Heartbeat]: Starting
(kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper)[2020-08-05
09:07:55,658] INFO [ExpirationReaper-0-Heartbeat]: Starting
(kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper)[2020-08-05
09:07:55,658] INFO [ExpirationReaper-0-topic]: Starting
(kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper)[2020-08-05
09:07:57,070] INFO Registered kafka:type=kafka.Log4jController MBean
(kafka.utils.Log4jControllerRegistration$)[2020-08-05 09:07:57,656] INFO
Registered signal handlers for TERM, INT, HUP
(org.apache.kafka.common.utils.LoggingSignalHandler)[2020-08-05 09:07:57,657]
INFO starting (kafka.server.KafkaServer)[2020-08-05 09:07:57,658] INFO
Connecting to zookeeper on
node_2:2181/kafka,node_3:2181/kafka,node_1:2181/kafka
(kafka.server.KafkaServer)[2020-08-05 09:07:57,685] INFO [ZooKeeperClient Kafka
server] Initializing a new session to node_2:2181.
(kafka.zookeeper.ZooKeeperClient)[2020-08-05 09:07:57,690] INFO Client
environment:zookeeper.version=3.4.14-4c25d480e66aadd371de8bd2fd8da255ac140bcf,
built on 03/06/2019 16:18 GMT (org.apache.zookeeper.ZooKeeper)[2020-08-05
09:07:57,693] INFO Client environment:host.name=localhost
(org.apache.zookeeper.ZooKeeper)[2020-08-05 09:07:57,693] INFO Client
environment:java.version=11.0.8 (org.apache.zookeeper.ZooKeeper)[2020-08-05
09:07:57,696] INFO Client environment:java.vendor=Ubuntu
(org.apache.zookeeper.ZooKeeper)[2020-08-05 09:07:57,696] INFO Client
environment:java.home=/usr/lib/jvm/java-11-openjdk-amd64
(org.apache.zookeeper.ZooKeeper)[2020-08-05 09:07:57,696] INFO Client
environment:java.class.path=/opt/kafka/current/bin/../libs/activation-1.1.1.jar:/opt/kafka/current/bin/../libs/aopalliance-repackaged-2.5.0.jar:/opt/kafka/current/bin/../libs/argparse4j-0.7.0.jar:/opt/kafka/current/bin/../libs/audience-annotations-0.5.0.jar:/opt/kafka/current/bin/../libs/commons-lang3-3.8.1.jar:/opt/kafka/current/bin/../libs/connect-api-2.3.1.jar:/opt/kafka/current/bin/../libs/connect-basic-auth-extension-2.3.1.jar:/opt/kafka/current/bin/../libs/connect-file-2.3.1.jar:/opt/kafka/current/bin/../libs/connect-json-2.3.1.jar:/opt/kafka/current/bin/../libs/connect-runtime-2.3.1.jar:/opt/kafka/current/bin/../libs/connect-transforms-2.3.1.jar:/opt/kafka/current/bin/../libs/guava-20.0.jar:/opt/kafka/current/bin/../libs/hk2-api-2.5.0.jar:/opt/kafka/current/bin/../libs/hk2-locator-2.5.0.jar:/opt/kafka/current/bin/../libs/hk2-utils-2.5.0.jar:/opt/kafka/current/bin/../libs/jackson-annotations-2.10.0.jar:/opt/kafka/current/bin/../libs/jackson-core-2.10.0.jar:/opt/kafka/current/bin/../libs/jackson-databind-2.10.0.jar:/opt/kafka/current/bin/../libs/jackson-dataformat-csv-2.10.0.jar:/opt/kafka/current/bin/../libs/jackson-datatype-jdk8-2.10.0.jar:/opt/kafka/current/bin/../libs/jackson-jaxrs-base-2.10.0.jar:/opt/kafka/current/bin/../libs/jackson-jaxrs-json-provider-2.10.0.jar:/opt/kafka/current/bin/../libs/jackson-module-jaxb-annotations-2.10.0.jar:/opt/kafka/current/bin/../libs/jackson-module-paranamer-2.10.0.jar:/opt/kafka/current/bin/../libs/jackson-module-scala_2.12-2.10.0.jar:/opt/kafka/current/bin/../libs/jakarta.activation-api-1.2.1.jar:/opt/kafka/current/bin/../libs/jakarta.annotation-api-1.3.4.jar:/opt/kafka/current/bin/../libs/jakarta.inject-2.5.0.jar:/opt/kafka/current/bin/../libs/jakarta.ws.rs-api-2.1.5.jar:/opt/kafka/current/bin/../libs/jakarta.xml.bind-api-2.3.2.jar:/opt/kafka/current/bin/../libs/javassist-3.22.0-CR2.jar:/opt/kafka/current/bin/../libs/javax.servlet-api-3.1.0.jar:/opt/kafka/current/bin/../libs/javax.ws.rs-api-2.1.1.jar:/opt/kafka/current/bin/../libs/jaxb-api-2.3.0.jar:/opt/kafka/current/bin/../libs/jersey-client-2.28.jar:/opt/kafka/current/bin/../libs/jersey-common-2.28.jar:/opt/kafka/current/bin/../libs/jersey-container-servlet-2.28.jar:/opt/kafka/current/bin/../libs/jersey-container-servlet-core-2.28.jar:/opt/kafka/current/bin/../libs/jersey-hk2-2.28.jar:/opt/kafka/current/bin/../libs/jersey-media-jaxb-2.28.jar:/opt/kafka/current/bin/../libs/jersey-server-2.28.jar:/opt/kafka/current/bin/../libs/jetty-client-9.4.18.v20190429.jar:/opt/kafka/current/bin/../libs/jetty-continuation-9.4.18.v20190429.jar:/opt/kafka/current/bin/../libs/jetty-http-9.4.18.v20190429.jar:/opt/kafka/current/bin/../libs/jetty-io-9.4.18.v20190429.jar:/opt/kafka/current/bin/../libs/jetty-security-9.4.18.v20190429.jar:/opt/kafka/current/bin/../libs/jetty-server-9.4.18.v20190429.jar:/opt/kafka/current/bin/../libs/jetty-servlet-9.4.18.v20190429.jar:/opt/kafka/current/bin/../libs/jetty-servlets-9.4.18.v20190429.jar:/opt/kafka/current/bin/../libs/jetty-util-9.4.18.v20190429.jar:/opt/kafka/current/bin/../libs/jopt-simple-5.0.4.jar:/opt/kafka/current/bin/../libs/jsr305-3.0.2.jar:/opt/kafka/current/bin/../libs/kafka-clients-2.3.1.jar:/opt/kafka/current/bin/../libs/kafka-log4j-appender-2.3.1.jar:/opt/kafka/current/bin/../libs/kafka-streams-2.3.1.jar:/opt/kafka/current/bin/../libs/kafka-streams-examples-2.3.1.jar:/opt/kafka/current/bin/../libs/kafka-streams-scala_2.12-2.3.1.jar:/opt/kafka/current/bin/../libs/kafka-streams-test-utils-2.3.1.jar:/opt/kafka/current/bin/../libs/kafka-tools-2.3.1.jar:/opt/kafka/current/bin/../libs/kafka_2.12-2.3.1-sources.jar:/opt/kafka/current/bin/../libs/kafka_2.12-2.3.1.jar:/opt/kafka/current/bin/../libs/log4j-1.2.17.jar:/opt/kafka/current/bin/../libs/lz4-java-1.6.0.jar:/opt/kafka/current/bin/../libs/maven-artifact-3.6.1.jar:/opt/kafka/current/bin/../libs/metrics-core-2.2.0.jar:/opt/kafka/current/bin/../libs/osgi-resource-locator-1.0.1.jar:/opt/kafka/current/bin/../libs/paranamer-2.8.jar:/opt/kafka/current/bin/../libs/plexus-utils-3.2.0.jar:/opt/kafka/current/bin/../libs/reflections-0.9.11.jar:/opt/kafka/current/bin/../libs/rocksdbjni-5.18.3.jar:/opt/kafka/current/bin/../libs/scala-library-2.12.10.jar:/opt/kafka/current/bin/../libs/scala-library-2.12.8.jar:/opt/kafka/current/bin/../libs/scala-logging_2.12-3.9.0.jar:/opt/kafka/current/bin/../libs/scala-reflect-2.12.8.jar:/opt/kafka/current/bin/../libs/slf4j-api-1.7.26.jar:/opt/kafka/current/bin/../libs/slf4j-log4j12-1.7.26.jar:/opt/kafka/current/bin/../libs/snappy-java-1.1.7.3.jar:/opt/kafka/current/bin/../libs/spotbugs-annotations-3.1.9.jar:/opt/kafka/current/bin/../libs/validation-api-2.0.1.Final.jar:/opt/kafka/current/bin/../libs/zkclient-0.11.jar:/opt/kafka/current/bin/../libs/zookeeper-3.4.14.jar:/opt/kafka/current/bin/../libs/zstd-jni-1.4.0-1.jar
(org.apache.zookeeper.ZooKeeper)[2020-08-05 09:07:57,697] INFO Client
environment:java.library.path=/usr/java/packages/lib:/usr/lib/x86_64-linux-gnu/jni:/lib/x86_64-linux-gnu:/usr/lib/x86_64-linux-gnu:/usr/lib/jni:/lib:/usr/lib
(org.apache.zookeeper.ZooKeeper)[2020-08-05 09:07:57,697] INFO Client
environment:java.io.tmpdir=/tmp (org.apache.zookeeper.ZooKeeper)[2020-08-05
09:07:57,697] INFO Client environment:java.compiler=<NA>
(org.apache.zookeeper.ZooKeeper)[2020-08-05 09:07:57,697] INFO Client
environment:os.name=Linux (org.apache.zookeeper.ZooKeeper)[2020-08-05
09:07:57,697] INFO Client environment:os.arch=amd64
(org.apache.zookeeper.ZooKeeper)[2020-08-05 09:07:57,698] INFO Client
environment:os.version=4.15.0-66-generic
(org.apache.zookeeper.ZooKeeper)[2020-08-05 09:07:57,698] INFO Client
environment:user.name=kafka (org.apache.zookeeper.ZooKeeper)[2020-08-05
09:07:57,698] INFO Client environment:user.home=/opt/kafka
(org.apache.zookeeper.ZooKeeper)[2020-08-05 09:07:57,698] INFO Client
environment:user.dir=/ (org.apache.zookeeper.ZooKeeper)[2020-08-05
09:07:57,699] INFO Initiating client connection, connectString=node_2:2181
sessionTimeout=6000
watcher=kafka.zookeeper.ZooKeeperClient$ZooKeeperClientWatcher$@74bada02
(org.apache.zookeeper.ZooKeeper)[2020-08-05 09:07:57,718] INFO [ZooKeeperClient
Kafka server] Waiting until connected.
(kafka.zookeeper.ZooKeeperClient)[2020-08-05 09:07:57,818] INFO Client
successfully logged in. (org.apache.zookeeper.Login)[2020-08-05 09:07:57,821]
INFO Client will use DIGEST-MD5 as SASL mechanism.
(org.apache.zookeeper.client.ZooKeeperSaslClient)[2020-08-05 09:07:57,826] INFO
Opening socket connection to server node_2/node_2:2181. Will attempt to
SASL-authenticate using Login Context section 'Client'
(org.apache.zookeeper.ClientCnxn)[2020-08-05 09:07:57,832] INFO Socket
connection established to node_2/node_2:2181, initiating session
(org.apache.zookeeper.ClientCnxn)[2020-08-05 09:07:57,841] INFO Session
establishment complete on server node_2/node_2:2181, sessionid =
0x373bdbbd3b00002, negotiated timeout = 6000
(org.apache.zookeeper.ClientCnxn)[2020-08-05 09:07:57,847] INFO
[ZooKeeperClient Kafka server] Connected.
(kafka.zookeeper.ZooKeeperClient)[2020-08-05 09:07:57,925] INFO Created
zookeeper path /kafka,node_3:2181/kafka,node_1:2181/kafka
(kafka.server.KafkaServer)[2020-08-05 09:07:57,926] INFO [ZooKeeperClient Kafka
server] Closing. (kafka.zookeeper.ZooKeeperClient)[2020-08-05 09:07:57,933]
INFO Session: 0x373bdbbd3b00002 closed
(org.apache.zookeeper.ZooKeeper)[2020-08-05 09:07:57,934] INFO EventThread shut
down for session: 0x373bdbbd3b00002
(org.apache.zookeeper.ClientCnxn)[2020-08-05 09:07:57,937] INFO
[ZooKeeperClient Kafka server] Closed.
(kafka.zookeeper.ZooKeeperClient)[2020-08-05 09:07:57,939] INFO
[ZooKeeperClient Kafka server] Initializing a new session to
node_2:2181/kafka,node_3:2181/kafka,node_1:2181/kafka.
(kafka.zookeeper.ZooKeeperClient)[2020-08-05 09:07:57,939] INFO Initiating
client connection,
connectString=node_2:2181/kafka,node_3:2181/kafka,node_1:2181/kafka
sessionTimeout=6000
watcher=kafka.zookeeper.ZooKeeperClient$ZooKeeperClientWatcher$@6ff65192
(org.apache.zookeeper.ZooKeeper)[2020-08-05 09:07:57,940] INFO [ZooKeeperClient
Kafka server] Waiting until connected.
(kafka.zookeeper.ZooKeeperClient)[2020-08-05 09:07:57,941] INFO Client will use
DIGEST-MD5 as SASL mechanism.
(org.apache.zookeeper.client.ZooKeeperSaslClient)[2020-08-05 09:07:57,943] INFO
Opening socket connection to server node_2/node_2:2181. Will attempt to
SASL-authenticate using Login Context section 'Client'
(org.apache.zookeeper.ClientCnxn)[2020-08-05 09:07:57,944] INFO Socket
connection established to node_2/node_2:2181, initiating session
(org.apache.zookeeper.ClientCnxn)[2020-08-05 09:07:57,949] INFO Session
establishment complete on server node_2/node_2:2181, sessionid =
0x373bdbbd3b00003, negotiated timeout = 6000
(org.apache.zookeeper.ClientCnxn)[2020-08-05 09:07:57,950] INFO
[ZooKeeperClient Kafka server] Connected.
(kafka.zookeeper.ZooKeeperClient)[2020-08-05 09:07:58,167] INFO Cluster ID =
5_d5S6HeQBWf0ZzwQ6TjRA (kafka.server.KafkaServer)[2020-08-05 09:07:58,253] INFO
KafkaConfig values: advertised.host.name = node_3 advertised.listeners =
SSL://node_3:9093 advertised.port = null alter.config.policy.class.name = null
alter.log.dirs.replication.quota.window.num = 11
alter.log.dirs.replication.quota.window.size.seconds = 1 authorizer.class.name
= kafka.security.auth.SimpleAclAuthorizer auto.create.topics.enable = true
auto.leader.rebalance.enable = true background.threads = 10 broker.id = -1
broker.id.generation.enable = true broker.rack = null
client.quota.callback.class = null compression.type = producer
connection.failed.authentication.delay.ms = 100 connections.max.idle.ms =
600000 connections.max.reauth.ms = 0 control.plane.listener.name = null
controlled.shutdown.enable = true controlled.shutdown.max.retries = 3
controlled.shutdown.retry.backoff.ms = 5000 controller.socket.timeout.ms =
30000 create.topic.policy.class.name = null default.replication.factor = 2
delegation.token.expiry.check.interval.ms = 3600000
delegation.token.expiry.time.ms = 86400000 delegation.token.master.key = null
delegation.token.max.lifetime.ms = 604800000
delete.records.purgatory.purge.interval.requests = 1 delete.topic.enable = true
fetch.purgatory.purge.interval.requests = 1000 group.initial.rebalance.delay.ms
= 3000 group.max.session.timeout.ms = 60000 group.max.size = 2147483647
group.min.session.timeout.ms = 10000 host.name = node_3
inter.broker.listener.name = null inter.broker.protocol.version = 2.3-IV1
kafka.metrics.polling.interval.secs = 10 kafka.metrics.reporters = []
leader.imbalance.check.interval.seconds = 300
leader.imbalance.per.broker.percentage = 10 listener.security.protocol.map =
SSL:SSL listeners = SSL://node_3:9093 log.cleaner.backoff.ms = 15000
log.cleaner.dedupe.buffer.size = 134217728 log.cleaner.delete.retention.ms =
86400000 log.cleaner.enable = true log.cleaner.io.buffer.load.factor = 0.9
log.cleaner.io.buffer.size = 524288 log.cleaner.io.max.bytes.per.second =
1.7976931348623157E308 log.cleaner.max.compaction.lag.ms = 9223372036854775807
log.cleaner.min.cleanable.ratio = 0.5 log.cleaner.min.compaction.lag.ms = 0
log.cleaner.threads = 1 log.cleanup.policy = [delete] log.dir = /tmp/kafka-logs
log.dirs = /var/lib/kafka log.flush.interval.messages = 5000
log.flush.interval.ms = 5000 log.flush.offset.checkpoint.interval.ms = 60000
log.flush.scheduler.interval.ms = 9223372036854775807
log.flush.start.offset.checkpoint.interval.ms = 60000 log.index.interval.bytes
= 4096 log.index.size.max.bytes = 10485760 log.message.downconversion.enable =
true log.message.format.version = 2.3-IV1
log.message.timestamp.difference.max.ms = 9223372036854775807
log.message.timestamp.type = CreateTime log.preallocate = false
log.retention.bytes = -1 log.retention.check.interval.ms = 300000
log.retention.hours = 336 log.retention.minutes = null log.retention.ms = null
log.roll.hours = 168 log.roll.jitter.hours = 0 log.roll.jitter.ms = null
log.roll.ms = null log.segment.bytes = 1073741824 log.segment.delete.delay.ms =
60000 max.connections = 2147483647 max.connections.per.ip = 2147483647
max.connections.per.ip.overrides = max.incremental.fetch.session.cache.slots =
1000 message.max.bytes = 2000024 metric.reporters = [] metrics.num.samples = 2
metrics.recording.level = INFO metrics.sample.window.ms = 30000
min.insync.replicas = 1 num.io.threads = 8 num.network.threads = 3
num.partitions = 4 num.recovery.threads.per.data.dir = 4
num.replica.alter.log.dirs.threads = null num.replica.fetchers = 1
offset.metadata.max.bytes = 4096 offsets.commit.required.acks = -1
offsets.commit.timeout.ms = 5000 offsets.load.buffer.size = 5242880
offsets.retention.check.interval.ms = 600000 offsets.retention.minutes = 10080
offsets.topic.compression.codec = 0 offsets.topic.num.partitions = 50
offsets.topic.replication.factor = 2 offsets.topic.segment.bytes = 104857600
password.encoder.cipher.algorithm = AES/CBC/PKCS5Padding
password.encoder.iterations = 4096 password.encoder.key.length = 128
password.encoder.keyfactory.algorithm = null password.encoder.old.secret = null
password.encoder.secret = null port = 9092 principal.builder.class = null
producer.purgatory.purge.interval.requests = 1000 queued.max.request.bytes = -1
queued.max.requests = 500 quota.consumer.default = 9223372036854775807
quota.producer.default = 9223372036854775807 quota.window.num = 11
quota.window.size.seconds = 1 replica.fetch.backoff.ms = 1000
replica.fetch.max.bytes = 1048576 replica.fetch.min.bytes = 1
replica.fetch.response.max.bytes = 10485760 replica.fetch.wait.max.ms = 500
replica.high.watermark.checkpoint.interval.ms = 5000 replica.lag.time.max.ms =
30000 replica.socket.receive.buffer.bytes = 65536 replica.socket.timeout.ms =
30000 replication.quota.window.num = 11 replication.quota.window.size.seconds =
1 request.timeout.ms = 30000 reserved.broker.max.id = 1000
sasl.client.callback.handler.class = null sasl.enabled.mechanisms = [GSSAPI]
sasl.jaas.config = null sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.principal.to.local.rules = [DEFAULT] sasl.kerberos.service.name =
null sasl.kerberos.ticket.renew.jitter = 0.05
sasl.kerberos.ticket.renew.window.factor = 0.8
sasl.login.callback.handler.class = null sasl.login.class = null
sasl.login.refresh.buffer.seconds = 300 sasl.login.refresh.min.period.seconds =
60 sasl.login.refresh.window.factor = 0.8 sasl.login.refresh.window.jitter =
0.05 sasl.mechanism.inter.broker.protocol = GSSAPI
sasl.server.callback.handler.class = null security.inter.broker.protocol = SSL
socket.receive.buffer.bytes = 102400 socket.request.max.bytes = 104857600
socket.send.buffer.bytes = 102400 ssl.cipher.suites = [] ssl.client.auth =
required ssl.enabled.protocols = [TLSv1.2]
ssl.endpoint.identification.algorithm = HTTPS ssl.key.password = [hidden]
ssl.keymanager.algorithm = SunX509 ssl.keystore.location =
/etc/ssl/kafka/kafka.keystore.jks ssl.keystore.password = [hidden]
ssl.keystore.type = JKS ssl.principal.mapping.rules = [DEFAULT] ssl.protocol =
TLS ssl.provider = null ssl.secure.random.implementation = SHA1PRNG
ssl.trustmanager.algorithm = PKIX ssl.truststore.location =
/etc/ssl/kafka/kafka.truststore.jks ssl.truststore.password = [hidden]
ssl.truststore.type = JKS
transaction.abort.timed.out.transaction.cleanup.interval.ms = 180000
transaction.max.timeout.ms = 900000
transaction.remove.expired.transaction.cleanup.interval.ms = 3600000
transaction.state.log.load.buffer.size = 5242880 transaction.state.log.min.isr
= 2 transaction.state.log.num.partitions = 50
transaction.state.log.replication.factor = 2
transaction.state.log.segment.bytes = 104857600 transactional.id.expiration.ms
= 604800000 unclean.leader.election.enable = false zookeeper.connect =
node_2:2181/kafka,node_3:2181/kafka,node_1:2181/kafka
zookeeper.connection.timeout.ms = 15000 zookeeper.max.in.flight.requests = 10
zookeeper.session.timeout.ms = 6000 zookeeper.set.acl = true
zookeeper.sync.time.ms = 2000 (kafka.server.KafkaConfig)[2020-08-05
09:07:58,272] INFO KafkaConfig values: advertised.host.name = node_3
advertised.listeners = SSL://node_3:9093 advertised.port = null
alter.config.policy.class.name = null
alter.log.dirs.replication.quota.window.num = 11
alter.log.dirs.replication.quota.window.size.seconds = 1 authorizer.class.name
= kafka.security.auth.SimpleAclAuthorizer auto.create.topics.enable = true
auto.leader.rebalance.enable = true background.threads = 10 broker.id = -1
broker.id.generation.enable = true broker.rack = null
client.quota.callback.class = null compression.type = producer
connection.failed.authentication.delay.ms = 100 connections.max.idle.ms =
600000 connections.max.reauth.ms = 0 control.plane.listener.name = null
controlled.shutdown.enable = true controlled.shutdown.max.retries = 3
controlled.shutdown.retry.backoff.ms = 5000 controller.socket.timeout.ms =
30000 create.topic.policy.class.name = null default.replication.factor = 2
delegation.token.expiry.check.interval.ms = 3600000
delegation.token.expiry.time.ms = 86400000 delegation.token.master.key = null
delegation.token.max.lifetime.ms = 604800000
delete.records.purgatory.purge.interval.requests = 1 delete.topic.enable = true
fetch.purgatory.purge.interval.requests = 1000 group.initial.rebalance.delay.ms
= 3000 group.max.session.timeout.ms = 60000 group.max.size = 2147483647
group.min.session.timeout.ms = 10000 host.name = node_3
inter.broker.listener.name = null inter.broker.protocol.version = 2.3-IV1
kafka.metrics.polling.interval.secs = 10 kafka.metrics.reporters = []
leader.imbalance.check.interval.seconds = 300
leader.imbalance.per.broker.percentage = 10 listener.security.protocol.map =
SSL:SSL listeners = SSL://node_3:9093 log.cleaner.backoff.ms = 15000
log.cleaner.dedupe.buffer.size = 134217728 log.cleaner.delete.retention.ms =
86400000 log.cleaner.enable = true log.cleaner.io.buffer.load.factor = 0.9
log.cleaner.io.buffer.size = 524288 log.cleaner.io.max.bytes.per.second =
1.7976931348623157E308 log.cleaner.max.compaction.lag.ms = 9223372036854775807
log.cleaner.min.cleanable.ratio = 0.5 log.cleaner.min.compaction.lag.ms = 0
log.cleaner.threads = 1 log.cleanup.policy = [delete] log.dir = /tmp/kafka-logs
log.dirs = /var/lib/kafka log.flush.interval.messages = 5000
log.flush.interval.ms = 5000 log.flush.offset.checkpoint.interval.ms = 60000
log.flush.scheduler.interval.ms = 9223372036854775807
log.flush.start.offset.checkpoint.interval.ms = 60000 log.index.interval.bytes
= 4096 log.index.size.max.bytes = 10485760 log.message.downconversion.enable =
true log.message.format.version = 2.3-IV1
log.message.timestamp.difference.max.ms = 9223372036854775807
log.message.timestamp.type = CreateTime log.preallocate = false
log.retention.bytes = -1 log.retention.check.interval.ms = 300000
log.retention.hours = 336 log.retention.minutes = null log.retention.ms = null
log.roll.hours = 168 log.roll.jitter.hours = 0 log.roll.jitter.ms = null
log.roll.ms = null log.segment.bytes = 1073741824 log.segment.delete.delay.ms =
60000 max.connections = 2147483647 max.connections.per.ip = 2147483647
max.connections.per.ip.overrides = max.incremental.fetch.session.cache.slots =
1000 message.max.bytes = 2000024 metric.reporters = [] metrics.num.samples = 2
metrics.recording.level = INFO metrics.sample.window.ms = 30000
min.insync.replicas = 1 num.io.threads = 8 num.network.threads = 3
num.partitions = 4 num.recovery.threads.per.data.dir = 4
num.replica.alter.log.dirs.threads = null num.replica.fetchers = 1
offset.metadata.max.bytes = 4096 offsets.commit.required.acks = -1
offsets.commit.timeout.ms = 5000 offsets.load.buffer.size = 5242880
offsets.retention.check.interval.ms = 600000 offsets.retention.minutes = 10080
offsets.topic.compression.codec = 0 offsets.topic.num.partitions = 50
offsets.topic.replication.factor = 2 offsets.topic.segment.bytes = 104857600
password.encoder.cipher.algorithm = AES/CBC/PKCS5Padding
password.encoder.iterations = 4096 password.encoder.key.length = 128
password.encoder.keyfactory.algorithm = null password.encoder.old.secret = null
password.encoder.secret = null port = 9092 principal.builder.class = null
producer.purgatory.purge.interval.requests = 1000 queued.max.request.bytes = -1
queued.max.requests = 500 quota.consumer.default = 9223372036854775807
quota.producer.default = 9223372036854775807 quota.window.num = 11
quota.window.size.seconds = 1 replica.fetch.backoff.ms = 1000
replica.fetch.max.bytes = 1048576 replica.fetch.min.bytes = 1
replica.fetch.response.max.bytes = 10485760 replica.fetch.wait.max.ms = 500
replica.high.watermark.checkpoint.interval.ms = 5000 replica.lag.time.max.ms =
30000 replica.socket.receive.buffer.bytes = 65536 replica.socket.timeout.ms =
30000 replication.quota.window.num = 11 replication.quota.window.size.seconds =
1 request.timeout.ms = 30000 reserved.broker.max.id = 1000
sasl.client.callback.handler.class = null sasl.enabled.mechanisms = [GSSAPI]
sasl.jaas.config = null sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.principal.to.local.rules = [DEFAULT] sasl.kerberos.service.name =
null sasl.kerberos.ticket.renew.jitter = 0.05
sasl.kerberos.ticket.renew.window.factor = 0.8
sasl.login.callback.handler.class = null sasl.login.class = null
sasl.login.refresh.buffer.seconds = 300 sasl.login.refresh.min.period.seconds =
60 sasl.login.refresh.window.factor = 0.8 sasl.login.refresh.window.jitter =
0.05 sasl.mechanism.inter.broker.protocol = GSSAPI
sasl.server.callback.handler.class = null security.inter.broker.protocol = SSL
socket.receive.buffer.bytes = 102400 socket.request.max.bytes = 104857600
socket.send.buffer.bytes = 102400 ssl.cipher.suites = [] ssl.client.auth =
required ssl.enabled.protocols = [TLSv1.2]
ssl.endpoint.identification.algorithm = HTTPS ssl.key.password = [hidden]
ssl.keymanager.algorithm = SunX509 ssl.keystore.location =
/etc/ssl/kafka/kafka.keystore.jks ssl.keystore.password = [hidden]
ssl.keystore.type = JKS ssl.principal.mapping.rules = [DEFAULT] ssl.protocol =
TLS ssl.provider = null ssl.secure.random.implementation = SHA1PRNG
ssl.trustmanager.algorithm = PKIX ssl.truststore.location =
/etc/ssl/kafka/kafka.truststore.jks ssl.truststore.password = [hidden]
ssl.truststore.type = JKS
transaction.abort.timed.out.transaction.cleanup.interval.ms = 180000
transaction.max.timeout.ms = 900000
transaction.remove.expired.transaction.cleanup.interval.ms = 3600000
transaction.state.log.load.buffer.size = 5242880 transaction.state.log.min.isr
= 2 transaction.state.log.num.partitions = 50
transaction.state.log.replication.factor = 2
transaction.state.log.segment.bytes = 104857600 transactional.id.expiration.ms
= 604800000 unclean.leader.election.enable = false zookeeper.connect =
node_2:2181/kafka,node_3:2181/kafka,node_1:2181/kafka
zookeeper.connection.timeout.ms = 15000 zookeeper.max.in.flight.requests = 10
zookeeper.session.timeout.ms = 6000 zookeeper.set.acl = true
zookeeper.sync.time.ms = 2000 (kafka.server.KafkaConfig)[2020-08-05
09:07:58,328] INFO [ThrottledChannelReaper-Produce]: Starting
(kafka.server.ClientQuotaManager$ThrottledChannelReaper)[2020-08-05
09:07:58,328] INFO [ThrottledChannelReaper-Fetch]: Starting
(kafka.server.ClientQuotaManager$ThrottledChannelReaper)[2020-08-05
09:07:58,331] INFO [ThrottledChannelReaper-Request]: Starting
(kafka.server.ClientQuotaManager$ThrottledChannelReaper)[2020-08-05
09:07:58,361] INFO Loading logs. (kafka.log.LogManager)[2020-08-05
09:07:58,374] INFO Logs loading complete in 13 ms.
(kafka.log.LogManager)[2020-08-05 09:07:58,403] INFO Starting log cleanup with
a period of 300000 ms. (kafka.log.LogManager)[2020-08-05 09:07:58,407] INFO
Starting log flusher with a default period of 9223372036854775807 ms.
(kafka.log.LogManager)[2020-08-05 09:07:58,805] INFO Awaiting socket
connections on node_3:9093. (kafka.network.Acceptor)[2020-08-05 09:08:00,594]
INFO [SocketServer brokerId=0] Created data-plane acceptor and processors for
endpoint : EndPoint(node_3,9093,ListenerName(SSL),SSL)
(kafka.network.SocketServer)[2020-08-05 09:08:00,596] INFO [SocketServer
brokerId=0] Started 1 acceptor threads for data-plane
(kafka.network.SocketServer)[2020-08-05 09:08:00,629] INFO
[ExpirationReaper-0-Produce]: Starting
(kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper)[2020-08-05
09:08:00,630] INFO [ExpirationReaper-0-Fetch]: Starting
(kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper)[2020-08-05
09:08:00,630] INFO [ExpirationReaper-0-DeleteRecords]: Starting
(kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper)[2020-08-05
09:08:00,631] INFO [ExpirationReaper-0-ElectPreferredLeader]: Starting
(kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper)[2020-08-05
09:08:00,660] INFO [LogDirFailureHandler]: Starting
(kafka.server.ReplicaManager$LogDirFailureHandler)[2020-08-05 09:08:00,726]
INFO Creating /brokers/ids/0 (is it secure? true)
(kafka.zk.KafkaZkClient)[2020-08-05 09:08:00,751] ERROR Error while creating
ephemeral at /brokers/ids/0, node already exists and owner '248751018843570177'
does not match current session '248751018843570179'
(kafka.zk.KafkaZkClient$CheckedEphemeral)[2020-08-05 09:08:00,757] ERROR
[KafkaServer id=0] Fatal error during KafkaServer startup. Prepare to shutdown
(kafka.server.KafkaServer)org.apache.zookeeper.KeeperException$NodeExistsException:
KeeperErrorCode = NodeExists at
org.apache.zookeeper.KeeperException.create(KeeperException.java:122) at
kafka.zk.KafkaZkClient$CheckedEphemeral.getAfterNodeExists(KafkaZkClient.scala:1784)
at kafka.zk.KafkaZkClient$CheckedEphemeral.create(KafkaZkClient.scala:1722) at
kafka.zk.KafkaZkClient.checkedEphemeralCreate(KafkaZkClient.scala:1689) at
kafka.zk.KafkaZkClient.registerBroker(KafkaZkClient.scala:97) at
kafka.server.KafkaServer.startup(KafkaServer.scala:262) at
kafka.server.KafkaServerStartable.startup(KafkaServerStartable.scala:38) at
kafka.Kafka$.main(Kafka.scala:84) at kafka.Kafka.main(Kafka.scala)[2020-08-05
09:08:00,763] INFO [KafkaServer id=0] shutting down (kafka.server.KafkaServer)
{code}
As you can see here, this broker tries to connect to Cluster with ID
_5_d5S6HeQBWf0ZzwQ6TjRA_, but other brokers are connected to Cluster with ID
_OhWuEGMeQHe66HP74rurRA_.
If we bring back the order of zookeeper servers in the config file then Kafka
broker starts normally and connects to an existing cluster.
This issue blocks us from adding new nodes to the cluster and removes the old
ones.
Cluster details:
* 3 Kafka nodes cluster running 2.3.1 (also reproduced on 2.4.0)
* 3 Zookeeper node cluster running 3.4.10
> Broker try to connect to a new cluster when there are changes in
> zookeeper.connect properties
> ---------------------------------------------------------------------------------------------
>
> Key: KAFKA-10363
> URL: https://issues.apache.org/jira/browse/KAFKA-10363
> Project: Kafka
> Issue Type: Bug
> Affects Versions: 2.4.0, 2.3.1
> Environment: 3 Kafka brokers (v2.3.1, v2.4.0) with Zookeeper cluster
> (3.4.10)
> Ubuntu 18.04 LTS
> Reporter: Alexey Kornev
> Priority: Critical
>
> We've just successfully set up a Kafka cluster consists of 3 brokers and
> faced with the following issue: when we change order of zookeeper servers in
> zookeeper.connect property in server.properties files and restart Kafka
> broker then this Kafka broker tries to connect to a new Kafka cluster. As a
> result, Kafka broker throws an error and shutdown.
> For example, config server.properties on first broker:
> {code:java}
> broker.id=-1
> ...
> zookeeper.connect=node_1:2181/kafka,node_2:2181/kafka,node_3:2181/kafka
> {code}
> We changed it to
> {code:java}
> broker.id=-1
> ...
> zookeeper.connect=node_2:2181/kafka,node_3:2181/kafka,node_1:2181/kafka {code}
> and restart Kafka broker.
> Logs:
> {code:java}
> [2020-08-05 09:07:55,658] INFO [ExpirationReaper-0-Heartbeat]: Starting
> (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper)[2020-08-05
> 09:07:55,658] INFO [ExpirationReaper-0-Heartbeat]: Starting
> (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper)[2020-08-05
> 09:07:55,658] INFO [ExpirationReaper-0-topic]: Starting
> (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper)[2020-08-05
> 09:07:57,070] INFO Registered kafka:type=kafka.Log4jController MBean
> (kafka.utils.Log4jControllerRegistration$)[2020-08-05 09:07:57,656] INFO
> Registered signal handlers for TERM, INT, HUP
> (org.apache.kafka.common.utils.LoggingSignalHandler)[2020-08-05 09:07:57,657]
> INFO starting (kafka.server.KafkaServer)[2020-08-05 09:07:57,658] INFO
> Connecting to zookeeper on
> node_2:2181/kafka,node_3:2181/kafka,node_1:2181/kafka
> (kafka.server.KafkaServer)[2020-08-05 09:07:57,685] INFO [ZooKeeperClient
> Kafka server] Initializing a new session to node_2:2181.
> (kafka.zookeeper.ZooKeeperClient)[2020-08-05 09:07:57,690] INFO Client
> environment:zookeeper.version=3.4.14-4c25d480e66aadd371de8bd2fd8da255ac140bcf,
> built on 03/06/2019 16:18 GMT (org.apache.zookeeper.ZooKeeper)[2020-08-05
> 09:07:57,693] INFO Client environment:host.name=localhost
> (org.apache.zookeeper.ZooKeeper)[2020-08-05 09:07:57,693] INFO Client
> environment:java.version=11.0.8 (org.apache.zookeeper.ZooKeeper)[2020-08-05
> 09:07:57,696] INFO Client environment:java.vendor=Ubuntu
> (org.apache.zookeeper.ZooKeeper)[2020-08-05 09:07:57,696] INFO Client
> environment:java.home=/usr/lib/jvm/java-11-openjdk-amd64
> (org.apache.zookeeper.ZooKeeper)[2020-08-05 09:07:57,696] INFO Client
> environment:java.class.path=/opt/kafka/current/bin/../libs/activation-1.1.1.jar:/opt/kafka/current/bin/../libs/aopalliance-repackaged-2.5.0.jar:/opt/kafka/current/bin/../libs/argparse4j-0.7.0.jar:/opt/kafka/current/bin/../libs/audience-annotations-0.5.0.jar:/opt/kafka/current/bin/../libs/commons-lang3-3.8.1.jar:/opt/kafka/current/bin/../libs/connect-api-2.3.1.jar:/opt/kafka/current/bin/../libs/connect-basic-auth-extension-2.3.1.jar:/opt/kafka/current/bin/../libs/connect-file-2.3.1.jar:/opt/kafka/current/bin/../libs/connect-json-2.3.1.jar:/opt/kafka/current/bin/../libs/connect-runtime-2.3.1.jar:/opt/kafka/current/bin/../libs/connect-transforms-2.3.1.jar:/opt/kafka/current/bin/../libs/guava-20.0.jar:/opt/kafka/current/bin/../libs/hk2-api-2.5.0.jar:/opt/kafka/current/bin/../libs/hk2-locator-2.5.0.jar:/opt/kafka/current/bin/../libs/hk2-utils-2.5.0.jar:/opt/kafka/current/bin/../libs/jackson-annotations-2.10.0.jar:/opt/kafka/current/bin/../libs/jackson-core-2.10.0.jar:/opt/kafka/current/bin/../libs/jackson-databind-2.10.0.jar:/opt/kafka/current/bin/../libs/jackson-dataformat-csv-2.10.0.jar:/opt/kafka/current/bin/../libs/jackson-datatype-jdk8-2.10.0.jar:/opt/kafka/current/bin/../libs/jackson-jaxrs-base-2.10.0.jar:/opt/kafka/current/bin/../libs/jackson-jaxrs-json-provider-2.10.0.jar:/opt/kafka/current/bin/../libs/jackson-module-jaxb-annotations-2.10.0.jar:/opt/kafka/current/bin/../libs/jackson-module-paranamer-2.10.0.jar:/opt/kafka/current/bin/../libs/jackson-module-scala_2.12-2.10.0.jar:/opt/kafka/current/bin/../libs/jakarta.activation-api-1.2.1.jar:/opt/kafka/current/bin/../libs/jakarta.annotation-api-1.3.4.jar:/opt/kafka/current/bin/../libs/jakarta.inject-2.5.0.jar:/opt/kafka/current/bin/../libs/jakarta.ws.rs-api-2.1.5.jar:/opt/kafka/current/bin/../libs/jakarta.xml.bind-api-2.3.2.jar:/opt/kafka/current/bin/../libs/javassist-3.22.0-CR2.jar:/opt/kafka/current/bin/../libs/javax.servlet-api-3.1.0.jar:/opt/kafka/current/bin/../libs/javax.ws.rs-api-2.1.1.jar:/opt/kafka/current/bin/../libs/jaxb-api-2.3.0.jar:/opt/kafka/current/bin/../libs/jersey-client-2.28.jar:/opt/kafka/current/bin/../libs/jersey-common-2.28.jar:/opt/kafka/current/bin/../libs/jersey-container-servlet-2.28.jar:/opt/kafka/current/bin/../libs/jersey-container-servlet-core-2.28.jar:/opt/kafka/current/bin/../libs/jersey-hk2-2.28.jar:/opt/kafka/current/bin/../libs/jersey-media-jaxb-2.28.jar:/opt/kafka/current/bin/../libs/jersey-server-2.28.jar:/opt/kafka/current/bin/../libs/jetty-client-9.4.18.v20190429.jar:/opt/kafka/current/bin/../libs/jetty-continuation-9.4.18.v20190429.jar:/opt/kafka/current/bin/../libs/jetty-http-9.4.18.v20190429.jar:/opt/kafka/current/bin/../libs/jetty-io-9.4.18.v20190429.jar:/opt/kafka/current/bin/../libs/jetty-security-9.4.18.v20190429.jar:/opt/kafka/current/bin/../libs/jetty-server-9.4.18.v20190429.jar:/opt/kafka/current/bin/../libs/jetty-servlet-9.4.18.v20190429.jar:/opt/kafka/current/bin/../libs/jetty-servlets-9.4.18.v20190429.jar:/opt/kafka/current/bin/../libs/jetty-util-9.4.18.v20190429.jar:/opt/kafka/current/bin/../libs/jopt-simple-5.0.4.jar:/opt/kafka/current/bin/../libs/jsr305-3.0.2.jar:/opt/kafka/current/bin/../libs/kafka-clients-2.3.1.jar:/opt/kafka/current/bin/../libs/kafka-log4j-appender-2.3.1.jar:/opt/kafka/current/bin/../libs/kafka-streams-2.3.1.jar:/opt/kafka/current/bin/../libs/kafka-streams-examples-2.3.1.jar:/opt/kafka/current/bin/../libs/kafka-streams-scala_2.12-2.3.1.jar:/opt/kafka/current/bin/../libs/kafka-streams-test-utils-2.3.1.jar:/opt/kafka/current/bin/../libs/kafka-tools-2.3.1.jar:/opt/kafka/current/bin/../libs/kafka_2.12-2.3.1-sources.jar:/opt/kafka/current/bin/../libs/kafka_2.12-2.3.1.jar:/opt/kafka/current/bin/../libs/log4j-1.2.17.jar:/opt/kafka/current/bin/../libs/lz4-java-1.6.0.jar:/opt/kafka/current/bin/../libs/maven-artifact-3.6.1.jar:/opt/kafka/current/bin/../libs/metrics-core-2.2.0.jar:/opt/kafka/current/bin/../libs/osgi-resource-locator-1.0.1.jar:/opt/kafka/current/bin/../libs/paranamer-2.8.jar:/opt/kafka/current/bin/../libs/plexus-utils-3.2.0.jar:/opt/kafka/current/bin/../libs/reflections-0.9.11.jar:/opt/kafka/current/bin/../libs/rocksdbjni-5.18.3.jar:/opt/kafka/current/bin/../libs/scala-library-2.12.10.jar:/opt/kafka/current/bin/../libs/scala-library-2.12.8.jar:/opt/kafka/current/bin/../libs/scala-logging_2.12-3.9.0.jar:/opt/kafka/current/bin/../libs/scala-reflect-2.12.8.jar:/opt/kafka/current/bin/../libs/slf4j-api-1.7.26.jar:/opt/kafka/current/bin/../libs/slf4j-log4j12-1.7.26.jar:/opt/kafka/current/bin/../libs/snappy-java-1.1.7.3.jar:/opt/kafka/current/bin/../libs/spotbugs-annotations-3.1.9.jar:/opt/kafka/current/bin/../libs/validation-api-2.0.1.Final.jar:/opt/kafka/current/bin/../libs/zkclient-0.11.jar:/opt/kafka/current/bin/../libs/zookeeper-3.4.14.jar:/opt/kafka/current/bin/../libs/zstd-jni-1.4.0-1.jar
> (org.apache.zookeeper.ZooKeeper)[2020-08-05 09:07:57,697] INFO Client
> environment:java.library.path=/usr/java/packages/lib:/usr/lib/x86_64-linux-gnu/jni:/lib/x86_64-linux-gnu:/usr/lib/x86_64-linux-gnu:/usr/lib/jni:/lib:/usr/lib
> (org.apache.zookeeper.ZooKeeper)[2020-08-05 09:07:57,697] INFO Client
> environment:java.io.tmpdir=/tmp (org.apache.zookeeper.ZooKeeper)[2020-08-05
> 09:07:57,697] INFO Client environment:java.compiler=<NA>
> (org.apache.zookeeper.ZooKeeper)[2020-08-05 09:07:57,697] INFO Client
> environment:os.name=Linux (org.apache.zookeeper.ZooKeeper)[2020-08-05
> 09:07:57,697] INFO Client environment:os.arch=amd64
> (org.apache.zookeeper.ZooKeeper)[2020-08-05 09:07:57,698] INFO Client
> environment:os.version=4.15.0-66-generic
> (org.apache.zookeeper.ZooKeeper)[2020-08-05 09:07:57,698] INFO Client
> environment:user.name=kafka (org.apache.zookeeper.ZooKeeper)[2020-08-05
> 09:07:57,698] INFO Client environment:user.home=/opt/kafka
> (org.apache.zookeeper.ZooKeeper)[2020-08-05 09:07:57,698] INFO Client
> environment:user.dir=/ (org.apache.zookeeper.ZooKeeper)[2020-08-05
> 09:07:57,699] INFO Initiating client connection, connectString=node_2:2181
> sessionTimeout=6000
> watcher=kafka.zookeeper.ZooKeeperClient$ZooKeeperClientWatcher$@74bada02
> (org.apache.zookeeper.ZooKeeper)[2020-08-05 09:07:57,718] INFO
> [ZooKeeperClient Kafka server] Waiting until connected.
> (kafka.zookeeper.ZooKeeperClient)[2020-08-05 09:07:57,818] INFO Client
> successfully logged in. (org.apache.zookeeper.Login)[2020-08-05 09:07:57,821]
> INFO Client will use DIGEST-MD5 as SASL mechanism.
> (org.apache.zookeeper.client.ZooKeeperSaslClient)[2020-08-05 09:07:57,826]
> INFO Opening socket connection to server node_2/node_2:2181. Will attempt to
> SASL-authenticate using Login Context section 'Client'
> (org.apache.zookeeper.ClientCnxn)[2020-08-05 09:07:57,832] INFO Socket
> connection established to node_2/node_2:2181, initiating session
> (org.apache.zookeeper.ClientCnxn)[2020-08-05 09:07:57,841] INFO Session
> establishment complete on server node_2/node_2:2181, sessionid =
> 0x373bdbbd3b00002, negotiated timeout = 6000
> (org.apache.zookeeper.ClientCnxn)[2020-08-05 09:07:57,847] INFO
> [ZooKeeperClient Kafka server] Connected.
> (kafka.zookeeper.ZooKeeperClient)[2020-08-05 09:07:57,925] INFO Created
> zookeeper path /kafka,node_3:2181/kafka,node_1:2181/kafka
> (kafka.server.KafkaServer)[2020-08-05 09:07:57,926] INFO [ZooKeeperClient
> Kafka server] Closing. (kafka.zookeeper.ZooKeeperClient)[2020-08-05
> 09:07:57,933] INFO Session: 0x373bdbbd3b00002 closed
> (org.apache.zookeeper.ZooKeeper)[2020-08-05 09:07:57,934] INFO EventThread
> shut down for session: 0x373bdbbd3b00002
> (org.apache.zookeeper.ClientCnxn)[2020-08-05 09:07:57,937] INFO
> [ZooKeeperClient Kafka server] Closed.
> (kafka.zookeeper.ZooKeeperClient)[2020-08-05 09:07:57,939] INFO
> [ZooKeeperClient Kafka server] Initializing a new session to
> node_2:2181/kafka,node_3:2181/kafka,node_1:2181/kafka.
> (kafka.zookeeper.ZooKeeperClient)[2020-08-05 09:07:57,939] INFO Initiating
> client connection,
> connectString=node_2:2181/kafka,node_3:2181/kafka,node_1:2181/kafka
> sessionTimeout=6000
> watcher=kafka.zookeeper.ZooKeeperClient$ZooKeeperClientWatcher$@6ff65192
> (org.apache.zookeeper.ZooKeeper)[2020-08-05 09:07:57,940] INFO
> [ZooKeeperClient Kafka server] Waiting until connected.
> (kafka.zookeeper.ZooKeeperClient)[2020-08-05 09:07:57,941] INFO Client will
> use DIGEST-MD5 as SASL mechanism.
> (org.apache.zookeeper.client.ZooKeeperSaslClient)[2020-08-05 09:07:57,943]
> INFO Opening socket connection to server node_2/node_2:2181. Will attempt to
> SASL-authenticate using Login Context section 'Client'
> (org.apache.zookeeper.ClientCnxn)[2020-08-05 09:07:57,944] INFO Socket
> connection established to node_2/node_2:2181, initiating session
> (org.apache.zookeeper.ClientCnxn)[2020-08-05 09:07:57,949] INFO Session
> establishment complete on server node_2/node_2:2181, sessionid =
> 0x373bdbbd3b00003, negotiated timeout = 6000
> (org.apache.zookeeper.ClientCnxn)[2020-08-05 09:07:57,950] INFO
> [ZooKeeperClient Kafka server] Connected.
> (kafka.zookeeper.ZooKeeperClient)[2020-08-05 09:07:58,167] INFO Cluster ID =
> 5_d5S6HeQBWf0ZzwQ6TjRA (kafka.server.KafkaServer)[2020-08-05 09:07:58,253]
> INFO KafkaConfig values: advertised.host.name = node_3 advertised.listeners
> = SSL://node_3:9093 advertised.port = null alter.config.policy.class.name =
> null alter.log.dirs.replication.quota.window.num = 11
> alter.log.dirs.replication.quota.window.size.seconds = 1
> authorizer.class.name = kafka.security.auth.SimpleAclAuthorizer
> auto.create.topics.enable = true auto.leader.rebalance.enable = true
> background.threads = 10 broker.id = -1 broker.id.generation.enable = true
> broker.rack = null client.quota.callback.class = null compression.type =
> producer connection.failed.authentication.delay.ms = 100
> connections.max.idle.ms = 600000 connections.max.reauth.ms = 0
> control.plane.listener.name = null controlled.shutdown.enable = true
> controlled.shutdown.max.retries = 3 controlled.shutdown.retry.backoff.ms =
> 5000 controller.socket.timeout.ms = 30000 create.topic.policy.class.name =
> null default.replication.factor = 2 delegation.token.expiry.check.interval.ms
> = 3600000 delegation.token.expiry.time.ms = 86400000
> delegation.token.master.key = null delegation.token.max.lifetime.ms =
> 604800000 delete.records.purgatory.purge.interval.requests = 1
> delete.topic.enable = true fetch.purgatory.purge.interval.requests = 1000
> group.initial.rebalance.delay.ms = 3000 group.max.session.timeout.ms = 60000
> group.max.size = 2147483647 group.min.session.timeout.ms = 10000 host.name =
> node_3 inter.broker.listener.name = null inter.broker.protocol.version =
> 2.3-IV1 kafka.metrics.polling.interval.secs = 10 kafka.metrics.reporters = []
> leader.imbalance.check.interval.seconds = 300
> leader.imbalance.per.broker.percentage = 10 listener.security.protocol.map =
> SSL:SSL listeners = SSL://node_3:9093 log.cleaner.backoff.ms = 15000
> log.cleaner.dedupe.buffer.size = 134217728 log.cleaner.delete.retention.ms =
> 86400000 log.cleaner.enable = true log.cleaner.io.buffer.load.factor = 0.9
> log.cleaner.io.buffer.size = 524288 log.cleaner.io.max.bytes.per.second =
> 1.7976931348623157E308 log.cleaner.max.compaction.lag.ms =
> 9223372036854775807 log.cleaner.min.cleanable.ratio = 0.5
> log.cleaner.min.compaction.lag.ms = 0 log.cleaner.threads = 1
> log.cleanup.policy = [delete] log.dir = /tmp/kafka-logs log.dirs =
> /var/lib/kafka log.flush.interval.messages = 5000 log.flush.interval.ms =
> 5000 log.flush.offset.checkpoint.interval.ms = 60000
> log.flush.scheduler.interval.ms = 9223372036854775807
> log.flush.start.offset.checkpoint.interval.ms = 60000
> log.index.interval.bytes = 4096 log.index.size.max.bytes = 10485760
> log.message.downconversion.enable = true log.message.format.version = 2.3-IV1
> log.message.timestamp.difference.max.ms = 9223372036854775807
> log.message.timestamp.type = CreateTime log.preallocate = false
> log.retention.bytes = -1 log.retention.check.interval.ms = 300000
> log.retention.hours = 336 log.retention.minutes = null log.retention.ms =
> null log.roll.hours = 168 log.roll.jitter.hours = 0 log.roll.jitter.ms = null
> log.roll.ms = null log.segment.bytes = 1073741824 log.segment.delete.delay.ms
> = 60000 max.connections = 2147483647 max.connections.per.ip = 2147483647
> max.connections.per.ip.overrides = max.incremental.fetch.session.cache.slots
> = 1000 message.max.bytes = 2000024 metric.reporters = [] metrics.num.samples
> = 2 metrics.recording.level = INFO metrics.sample.window.ms = 30000
> min.insync.replicas = 1 num.io.threads = 8 num.network.threads = 3
> num.partitions = 4 num.recovery.threads.per.data.dir = 4
> num.replica.alter.log.dirs.threads = null num.replica.fetchers = 1
> offset.metadata.max.bytes = 4096 offsets.commit.required.acks = -1
> offsets.commit.timeout.ms = 5000 offsets.load.buffer.size = 5242880
> offsets.retention.check.interval.ms = 600000 offsets.retention.minutes =
> 10080 offsets.topic.compression.codec = 0 offsets.topic.num.partitions = 50
> offsets.topic.replication.factor = 2 offsets.topic.segment.bytes = 104857600
> password.encoder.cipher.algorithm = AES/CBC/PKCS5Padding
> password.encoder.iterations = 4096 password.encoder.key.length = 128
> password.encoder.keyfactory.algorithm = null password.encoder.old.secret =
> null password.encoder.secret = null port = 9092 principal.builder.class =
> null producer.purgatory.purge.interval.requests = 1000
> queued.max.request.bytes = -1 queued.max.requests = 500
> quota.consumer.default = 9223372036854775807 quota.producer.default =
> 9223372036854775807 quota.window.num = 11 quota.window.size.seconds = 1
> replica.fetch.backoff.ms = 1000 replica.fetch.max.bytes = 1048576
> replica.fetch.min.bytes = 1 replica.fetch.response.max.bytes = 10485760
> replica.fetch.wait.max.ms = 500 replica.high.watermark.checkpoint.interval.ms
> = 5000 replica.lag.time.max.ms = 30000 replica.socket.receive.buffer.bytes =
> 65536 replica.socket.timeout.ms = 30000 replication.quota.window.num = 11
> replication.quota.window.size.seconds = 1 request.timeout.ms = 30000
> reserved.broker.max.id = 1000 sasl.client.callback.handler.class = null
> sasl.enabled.mechanisms = [GSSAPI] sasl.jaas.config = null
> sasl.kerberos.kinit.cmd = /usr/bin/kinit
> sasl.kerberos.min.time.before.relogin = 60000
> sasl.kerberos.principal.to.local.rules = [DEFAULT] sasl.kerberos.service.name
> = null sasl.kerberos.ticket.renew.jitter = 0.05
> sasl.kerberos.ticket.renew.window.factor = 0.8
> sasl.login.callback.handler.class = null sasl.login.class = null
> sasl.login.refresh.buffer.seconds = 300 sasl.login.refresh.min.period.seconds
> = 60 sasl.login.refresh.window.factor = 0.8 sasl.login.refresh.window.jitter
> = 0.05 sasl.mechanism.inter.broker.protocol = GSSAPI
> sasl.server.callback.handler.class = null security.inter.broker.protocol =
> SSL socket.receive.buffer.bytes = 102400 socket.request.max.bytes = 104857600
> socket.send.buffer.bytes = 102400 ssl.cipher.suites = [] ssl.client.auth =
> required ssl.enabled.protocols = [TLSv1.2]
> ssl.endpoint.identification.algorithm = HTTPS ssl.key.password = [hidden]
> ssl.keymanager.algorithm = SunX509 ssl.keystore.location =
> /etc/ssl/kafka/kafka.keystore.jks ssl.keystore.password = [hidden]
> ssl.keystore.type = JKS ssl.principal.mapping.rules = [DEFAULT] ssl.protocol
> = TLS ssl.provider = null ssl.secure.random.implementation = SHA1PRNG
> ssl.trustmanager.algorithm = PKIX ssl.truststore.location =
> /etc/ssl/kafka/kafka.truststore.jks ssl.truststore.password = [hidden]
> ssl.truststore.type = JKS
> transaction.abort.timed.out.transaction.cleanup.interval.ms = 180000
> transaction.max.timeout.ms = 900000
> transaction.remove.expired.transaction.cleanup.interval.ms = 3600000
> transaction.state.log.load.buffer.size = 5242880
> transaction.state.log.min.isr = 2 transaction.state.log.num.partitions = 50
> transaction.state.log.replication.factor = 2
> transaction.state.log.segment.bytes = 104857600
> transactional.id.expiration.ms = 604800000 unclean.leader.election.enable =
> false zookeeper.connect =
> node_2:2181/kafka,node_3:2181/kafka,node_1:2181/kafka
> zookeeper.connection.timeout.ms = 15000 zookeeper.max.in.flight.requests = 10
> zookeeper.session.timeout.ms = 6000 zookeeper.set.acl = true
> zookeeper.sync.time.ms = 2000 (kafka.server.KafkaConfig)[2020-08-05
> 09:07:58,272] INFO KafkaConfig values: advertised.host.name = node_3
> advertised.listeners = SSL://node_3:9093 advertised.port = null
> alter.config.policy.class.name = null
> alter.log.dirs.replication.quota.window.num = 11
> alter.log.dirs.replication.quota.window.size.seconds = 1
> authorizer.class.name = kafka.security.auth.SimpleAclAuthorizer
> auto.create.topics.enable = true auto.leader.rebalance.enable = true
> background.threads = 10 broker.id = -1 broker.id.generation.enable = true
> broker.rack = null client.quota.callback.class = null compression.type =
> producer connection.failed.authentication.delay.ms = 100
> connections.max.idle.ms = 600000 connections.max.reauth.ms = 0
> control.plane.listener.name = null controlled.shutdown.enable = true
> controlled.shutdown.max.retries = 3 controlled.shutdown.retry.backoff.ms =
> 5000 controller.socket.timeout.ms = 30000 create.topic.policy.class.name =
> null default.replication.factor = 2 delegation.token.expiry.check.interval.ms
> = 3600000 delegation.token.expiry.time.ms = 86400000
> delegation.token.master.key = null delegation.token.max.lifetime.ms =
> 604800000 delete.records.purgatory.purge.interval.requests = 1
> delete.topic.enable = true fetch.purgatory.purge.interval.requests = 1000
> group.initial.rebalance.delay.ms = 3000 group.max.session.timeout.ms = 60000
> group.max.size = 2147483647 group.min.session.timeout.ms = 10000 host.name =
> node_3 inter.broker.listener.name = null inter.broker.protocol.version =
> 2.3-IV1 kafka.metrics.polling.interval.secs = 10 kafka.metrics.reporters = []
> leader.imbalance.check.interval.seconds = 300
> leader.imbalance.per.broker.percentage = 10 listener.security.protocol.map =
> SSL:SSL listeners = SSL://node_3:9093 log.cleaner.backoff.ms = 15000
> log.cleaner.dedupe.buffer.size = 134217728 log.cleaner.delete.retention.ms =
> 86400000 log.cleaner.enable = true log.cleaner.io.buffer.load.factor = 0.9
> log.cleaner.io.buffer.size = 524288 log.cleaner.io.max.bytes.per.second =
> 1.7976931348623157E308 log.cleaner.max.compaction.lag.ms =
> 9223372036854775807 log.cleaner.min.cleanable.ratio = 0.5
> log.cleaner.min.compaction.lag.ms = 0 log.cleaner.threads = 1
> log.cleanup.policy = [delete] log.dir = /tmp/kafka-logs log.dirs =
> /var/lib/kafka log.flush.interval.messages = 5000 log.flush.interval.ms =
> 5000 log.flush.offset.checkpoint.interval.ms = 60000
> log.flush.scheduler.interval.ms = 9223372036854775807
> log.flush.start.offset.checkpoint.interval.ms = 60000
> log.index.interval.bytes = 4096 log.index.size.max.bytes = 10485760
> log.message.downconversion.enable = true log.message.format.version = 2.3-IV1
> log.message.timestamp.difference.max.ms = 9223372036854775807
> log.message.timestamp.type = CreateTime log.preallocate = false
> log.retention.bytes = -1 log.retention.check.interval.ms = 300000
> log.retention.hours = 336 log.retention.minutes = null log.retention.ms =
> null log.roll.hours = 168 log.roll.jitter.hours = 0 log.roll.jitter.ms = null
> log.roll.ms = null log.segment.bytes = 1073741824 log.segment.delete.delay.ms
> = 60000 max.connections = 2147483647 max.connections.per.ip = 2147483647
> max.connections.per.ip.overrides = max.incremental.fetch.session.cache.slots
> = 1000 message.max.bytes = 2000024 metric.reporters = [] metrics.num.samples
> = 2 metrics.recording.level = INFO metrics.sample.window.ms = 30000
> min.insync.replicas = 1 num.io.threads = 8 num.network.threads = 3
> num.partitions = 4 num.recovery.threads.per.data.dir = 4
> num.replica.alter.log.dirs.threads = null num.replica.fetchers = 1
> offset.metadata.max.bytes = 4096 offsets.commit.required.acks = -1
> offsets.commit.timeout.ms = 5000 offsets.load.buffer.size = 5242880
> offsets.retention.check.interval.ms = 600000 offsets.retention.minutes =
> 10080 offsets.topic.compression.codec = 0 offsets.topic.num.partitions = 50
> offsets.topic.replication.factor = 2 offsets.topic.segment.bytes = 104857600
> password.encoder.cipher.algorithm = AES/CBC/PKCS5Padding
> password.encoder.iterations = 4096 password.encoder.key.length = 128
> password.encoder.keyfactory.algorithm = null password.encoder.old.secret =
> null password.encoder.secret = null port = 9092 principal.builder.class =
> null producer.purgatory.purge.interval.requests = 1000
> queued.max.request.bytes = -1 queued.max.requests = 500
> quota.consumer.default = 9223372036854775807 quota.producer.default =
> 9223372036854775807 quota.window.num = 11 quota.window.size.seconds = 1
> replica.fetch.backoff.ms = 1000 replica.fetch.max.bytes = 1048576
> replica.fetch.min.bytes = 1 replica.fetch.response.max.bytes = 10485760
> replica.fetch.wait.max.ms = 500 replica.high.watermark.checkpoint.interval.ms
> = 5000 replica.lag.time.max.ms = 30000 replica.socket.receive.buffer.bytes =
> 65536 replica.socket.timeout.ms = 30000 replication.quota.window.num = 11
> replication.quota.window.size.seconds = 1 request.timeout.ms = 30000
> reserved.broker.max.id = 1000 sasl.client.callback.handler.class = null
> sasl.enabled.mechanisms = [GSSAPI] sasl.jaas.config = null
> sasl.kerberos.kinit.cmd = /usr/bin/kinit
> sasl.kerberos.min.time.before.relogin = 60000
> sasl.kerberos.principal.to.local.rules = [DEFAULT] sasl.kerberos.service.name
> = null sasl.kerberos.ticket.renew.jitter = 0.05
> sasl.kerberos.ticket.renew.window.factor = 0.8
> sasl.login.callback.handler.class = null sasl.login.class = null
> sasl.login.refresh.buffer.seconds = 300 sasl.login.refresh.min.period.seconds
> = 60 sasl.login.refresh.window.factor = 0.8 sasl.login.refresh.window.jitter
> = 0.05 sasl.mechanism.inter.broker.protocol = GSSAPI
> sasl.server.callback.handler.class = null security.inter.broker.protocol =
> SSL socket.receive.buffer.bytes = 102400 socket.request.max.bytes = 104857600
> socket.send.buffer.bytes = 102400 ssl.cipher.suites = [] ssl.client.auth =
> required ssl.enabled.protocols = [TLSv1.2]
> ssl.endpoint.identification.algorithm = HTTPS ssl.key.password = [hidden]
> ssl.keymanager.algorithm = SunX509 ssl.keystore.location =
> /etc/ssl/kafka/kafka.keystore.jks ssl.keystore.password = [hidden]
> ssl.keystore.type = JKS ssl.principal.mapping.rules = [DEFAULT] ssl.protocol
> = TLS ssl.provider = null ssl.secure.random.implementation = SHA1PRNG
> ssl.trustmanager.algorithm = PKIX ssl.truststore.location =
> /etc/ssl/kafka/kafka.truststore.jks ssl.truststore.password = [hidden]
> ssl.truststore.type = JKS
> transaction.abort.timed.out.transaction.cleanup.interval.ms = 180000
> transaction.max.timeout.ms = 900000
> transaction.remove.expired.transaction.cleanup.interval.ms = 3600000
> transaction.state.log.load.buffer.size = 5242880
> transaction.state.log.min.isr = 2 transaction.state.log.num.partitions = 50
> transaction.state.log.replication.factor = 2
> transaction.state.log.segment.bytes = 104857600
> transactional.id.expiration.ms = 604800000 unclean.leader.election.enable =
> false zookeeper.connect =
> node_2:2181/kafka,node_3:2181/kafka,node_1:2181/kafka
> zookeeper.connection.timeout.ms = 15000 zookeeper.max.in.flight.requests = 10
> zookeeper.session.timeout.ms = 6000 zookeeper.set.acl = true
> zookeeper.sync.time.ms = 2000 (kafka.server.KafkaConfig)[2020-08-05
> 09:07:58,328] INFO [ThrottledChannelReaper-Produce]: Starting
> (kafka.server.ClientQuotaManager$ThrottledChannelReaper)[2020-08-05
> 09:07:58,328] INFO [ThrottledChannelReaper-Fetch]: Starting
> (kafka.server.ClientQuotaManager$ThrottledChannelReaper)[2020-08-05
> 09:07:58,331] INFO [ThrottledChannelReaper-Request]: Starting
> (kafka.server.ClientQuotaManager$ThrottledChannelReaper)[2020-08-05
> 09:07:58,361] INFO Loading logs. (kafka.log.LogManager)[2020-08-05
> 09:07:58,374] INFO Logs loading complete in 13 ms.
> (kafka.log.LogManager)[2020-08-05 09:07:58,403] INFO Starting log cleanup
> with a period of 300000 ms. (kafka.log.LogManager)[2020-08-05 09:07:58,407]
> INFO Starting log flusher with a default period of 9223372036854775807 ms.
> (kafka.log.LogManager)[2020-08-05 09:07:58,805] INFO Awaiting socket
> connections on node_3:9093. (kafka.network.Acceptor)[2020-08-05 09:08:00,594]
> INFO [SocketServer brokerId=0] Created data-plane acceptor and processors for
> endpoint : EndPoint(node_3,9093,ListenerName(SSL),SSL)
> (kafka.network.SocketServer)[2020-08-05 09:08:00,596] INFO [SocketServer
> brokerId=0] Started 1 acceptor threads for data-plane
> (kafka.network.SocketServer)[2020-08-05 09:08:00,629] INFO
> [ExpirationReaper-0-Produce]: Starting
> (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper)[2020-08-05
> 09:08:00,630] INFO [ExpirationReaper-0-Fetch]: Starting
> (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper)[2020-08-05
> 09:08:00,630] INFO [ExpirationReaper-0-DeleteRecords]: Starting
> (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper)[2020-08-05
> 09:08:00,631] INFO [ExpirationReaper-0-ElectPreferredLeader]: Starting
> (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper)[2020-08-05
> 09:08:00,660] INFO [LogDirFailureHandler]: Starting
> (kafka.server.ReplicaManager$LogDirFailureHandler)[2020-08-05 09:08:00,726]
> INFO Creating /brokers/ids/0 (is it secure? true)
> (kafka.zk.KafkaZkClient)[2020-08-05 09:08:00,751] ERROR Error while creating
> ephemeral at /brokers/ids/0, node already exists and owner
> '248751018843570177' does not match current session '248751018843570179'
> (kafka.zk.KafkaZkClient$CheckedEphemeral)[2020-08-05 09:08:00,757] ERROR
> [KafkaServer id=0] Fatal error during KafkaServer startup. Prepare to
> shutdown
> (kafka.server.KafkaServer)org.apache.zookeeper.KeeperException$NodeExistsException:
> KeeperErrorCode = NodeExists at
> org.apache.zookeeper.KeeperException.create(KeeperException.java:122) at
> kafka.zk.KafkaZkClient$CheckedEphemeral.getAfterNodeExists(KafkaZkClient.scala:1784)
> at kafka.zk.KafkaZkClient$CheckedEphemeral.create(KafkaZkClient.scala:1722)
> at kafka.zk.KafkaZkClient.checkedEphemeralCreate(KafkaZkClient.scala:1689) at
> kafka.zk.KafkaZkClient.registerBroker(KafkaZkClient.scala:97) at
> kafka.server.KafkaServer.startup(KafkaServer.scala:262) at
> kafka.server.KafkaServerStartable.startup(KafkaServerStartable.scala:38) at
> kafka.Kafka$.main(Kafka.scala:84) at kafka.Kafka.main(Kafka.scala)[2020-08-05
> 09:08:00,763] INFO [KafkaServer id=0] shutting down (kafka.server.KafkaServer)
> {code}
>
> As you can see here, this broker tries to connect to Cluster with ID
> _5_d5S6HeQBWf0ZzwQ6TjRA_, but other brokers are connected to Cluster with ID
> _OhWuEGMeQHe66HP74rurRA_.
> If we bring back the order of zookeeper servers in the config file then Kafka
> broker starts normally and connects to an existing cluster.
> This issue blocks us from adding new nodes to the cluster and removes the old
> ones.
> Cluster details:
> * 3 Kafka nodes cluster running 2.3.1 (also reproduced on 2.4.0)
> * 3 Zookeeper node cluster running 3.4.10
>
--
This message was sent by Atlassian Jira
(v8.3.4#803005)