cmccabe commented on a change in pull request #11417: URL: https://github.com/apache/kafka/pull/11417#discussion_r736953818
########## File path: core/src/test/scala/integration/kafka/server/QuorumTestHarness.scala ########## @@ -0,0 +1,339 @@ +/** + * Licensed to the Apache Software Foundation (ASF) under one or more + * contributor license agreements. See the NOTICE file distributed with + * this work for additional information regarding copyright ownership. + * The ASF licenses this file to You under the Apache License, Version 2.0 + * (the "License"); you may not use this file except in compliance with + * the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +package kafka.server + +import java.io.{ByteArrayOutputStream, File, PrintStream} +import java.net.InetSocketAddress +import java.util +import java.util.{Collections, Properties} +import java.util.concurrent.CompletableFuture + +import javax.security.auth.login.Configuration +import kafka.raft.KafkaRaftManager +import kafka.tools.StorageTool +import kafka.utils.{CoreUtils, Logging, TestInfoUtils, TestUtils} +import kafka.zk.{AdminZkClient, EmbeddedZookeeper, KafkaZkClient} +import org.apache.kafka.common.metrics.Metrics +import org.apache.kafka.common.{TopicPartition, Uuid} +import org.apache.kafka.common.security.JaasUtils +import org.apache.kafka.common.security.auth.SecurityProtocol +import org.apache.kafka.common.utils.Time +import org.apache.kafka.metadata.MetadataRecordSerde +import org.apache.kafka.raft.RaftConfig.{AddressSpec, InetAddressSpec} +import org.apache.kafka.server.common.ApiMessageAndVersion +import org.apache.zookeeper.client.ZKClientConfig +import org.apache.zookeeper.{WatchedEvent, Watcher, ZooKeeper} +import org.junit.jupiter.api.Assertions._ +import org.junit.jupiter.api.{AfterAll, AfterEach, BeforeAll, BeforeEach, Tag, TestInfo} + +import scala.collection.{Seq, immutable} + +trait QuorumImplementation { + def createAndStartBroker(config: KafkaConfig, + time: Time): KafkaBroker + + def shutdown(): Unit +} + +class ZooKeeperQuorumImplementation(val zookeeper: EmbeddedZookeeper, + val zkClient: KafkaZkClient, + val adminZkClient: AdminZkClient, + val log: Logging) extends QuorumImplementation { + override def createAndStartBroker(config: KafkaConfig, + time: Time): KafkaBroker = { + val server = new KafkaServer(config, time, None, false) + server.startup() + server + } + + override def shutdown(): Unit = { + CoreUtils.swallow(zkClient.close(), log) + CoreUtils.swallow(zookeeper.shutdown(), log) + } + + def shutdownZooKeeper(): Unit = { + CoreUtils.swallow(zookeeper.shutdown(), log) + } +} + +class KRaftQuorumImplementation(val raftManager: KafkaRaftManager[ApiMessageAndVersion], + val controllerServer: ControllerServer, + val metadataDir: File, + val controllerQuorumVotersFuture: CompletableFuture[util.Map[Integer, AddressSpec]], + val clusterId: String, + val log: Logging) extends QuorumImplementation { + override def createAndStartBroker(config: KafkaConfig, + time: Time): KafkaBroker = { + val broker = new BrokerServer(config = config, + metaProps = new MetaProperties(clusterId, config.nodeId), + raftManager = raftManager, + time = time, + metrics = new Metrics(), + threadNamePrefix = Some("Broker%02d_".format(config.nodeId)), + initialOfflineDirs = Seq(), + controllerQuorumVotersFuture = controllerQuorumVotersFuture, + supportedFeatures = Collections.emptyMap()) + broker.startup() + broker + } + + override def shutdown(): Unit = { + CoreUtils.swallow(raftManager.shutdown(), log) + CoreUtils.swallow(controllerServer.shutdown(), log) + } +} + +@Tag("integration") +abstract class QuorumTestHarness extends Logging { + val zkConnectionTimeout = 10000 + val zkSessionTimeout = 15000 // Allows us to avoid ZK session expiration due to GC up to 2/3 * 15000ms = 10 secs + val zkMaxInFlightRequests = Int.MaxValue + + protected def zkAclsEnabled: Option[Boolean] = None + + /** + * When in KRaft mode, the security protocol to use for the controller listener. + * Can be overridden by subclasses. + */ + protected def controllerListenerSecurityProtocol: SecurityProtocol = SecurityProtocol.PLAINTEXT + + protected def kraftControllerConfigs(): Seq[Properties] = { + Seq(new Properties()) + } + + private var implementation: QuorumImplementation = null + + def isKRaftTest(): Boolean = implementation.isInstanceOf[KRaftQuorumImplementation] + + def checkIsZKTest(): Unit = { + if (isKRaftTest()) { + throw new RuntimeException("This function can't be accessed when running the test " + + "in KRaft mode. ZooKeeper mode is required.") + } + } + + def checkIsKRaftTest(): Unit = { + if (!isKRaftTest()) { + throw new RuntimeException("This function can't be accessed when running the test " + + "in ZooKeeper mode. KRaft mode is required.") + } + } + + private def asZk(): ZooKeeperQuorumImplementation = { + checkIsZKTest() + implementation.asInstanceOf[ZooKeeperQuorumImplementation] + } + + private def asKRaft(): KRaftQuorumImplementation = { + checkIsKRaftTest() + implementation.asInstanceOf[KRaftQuorumImplementation] + } + + def zookeeper: EmbeddedZookeeper = asZk().zookeeper + + def zkClient: KafkaZkClient = asZk().zkClient + + def zkClientOrNull: KafkaZkClient = if (isKRaftTest()) null else asZk().zkClient + + def adminZkClient: AdminZkClient = asZk().adminZkClient + + def zkPort: Int = asZk().zookeeper.port + + def zkConnect: String = s"127.0.0.1:$zkPort" + + def zkConnectOrNull: String = if (isKRaftTest()) null else s"127.0.0.1:$zkPort" + + def controllerServer: ControllerServer = asKRaft().controllerServer + + @BeforeEach + def setUp(testInfo: TestInfo): Unit = { + val name = if (testInfo.getTestMethod().isPresent()) { + testInfo.getTestMethod().get().toString() + } else { + "[unspecified]" + } + if (TestInfoUtils.isKRaft(testInfo)) { + info(s"Running KRAFT test ${name}") + implementation = newKRaftQuorum(testInfo) + } else { + info(s"Running ZK test ${name}") + implementation = newZooKeeperQuorum() + } + } + + def createAndStartBroker(config: KafkaConfig, + time: Time = Time.SYSTEM): KafkaBroker = { + implementation.createAndStartBroker(config, + time) + } + + def shutdownZooKeeper(): Unit = asZk().shutdownZooKeeper() + + private def formatDirectories(directories: immutable.Seq[String], + metaProperties: MetaProperties): Unit = { + val stream = new ByteArrayOutputStream() + var out: PrintStream = null + try { + out = new PrintStream(stream) + if (StorageTool.formatCommand(out, directories, metaProperties, false) != 0) { + throw new RuntimeException(out.toString()) + } + debug(s"Formatted storage directory(ies) ${directories}") + } finally { + if (out != null) out.close() + stream.close() + } + } + + private def newKRaftQuorum(testInfo: TestInfo): KRaftQuorumImplementation = { + val clusterId = Uuid.randomUuid().toString + val metadataDir = TestUtils.tempDir() + val metaProperties = new MetaProperties(clusterId, 0) + formatDirectories(immutable.Seq(metadataDir.getAbsolutePath()), metaProperties) + val controllerMetrics = new Metrics() + val propsList = kraftControllerConfigs() + if (propsList.size != 1) { + throw new RuntimeException("Only one KRaft controller is supported for now.") + } + val props = propsList(0) + props.setProperty(KafkaConfig.ProcessRolesProp, "controller") + props.setProperty(KafkaConfig.NodeIdProp, "1000") + props.setProperty(KafkaConfig.MetadataLogDirProp, metadataDir.getAbsolutePath()) + val proto = controllerListenerSecurityProtocol.toString() + props.setProperty(KafkaConfig.ListenerSecurityProtocolMapProp, s"${proto}:${proto}") + props.setProperty(KafkaConfig.ListenersProp, s"${proto}://localhost:0") + props.setProperty(KafkaConfig.ControllerListenerNamesProp, proto) + props.setProperty(KafkaConfig.QuorumVotersProp, "1000@localhost:0") + val config = new KafkaConfig(props) + val threadNamePrefix = "Controller_" + testInfo.getDisplayName + val controllerQuorumVotersFuture = new CompletableFuture[util.Map[Integer, AddressSpec]] + val raftManager = new KafkaRaftManager( + metaProperties = metaProperties, + config = config, + recordSerde = MetadataRecordSerde.INSTANCE, + topicPartition = new TopicPartition(KafkaRaftServer.MetadataTopic, 0), + topicId = KafkaRaftServer.MetadataTopicId, + time = Time.SYSTEM, + metrics = controllerMetrics, + threadNamePrefixOpt = Option(threadNamePrefix), + controllerQuorumVotersFuture = controllerQuorumVotersFuture) + var controllerServer: ControllerServer = null + try { + controllerServer = new ControllerServer( + metaProperties = metaProperties, + config = config, + raftManager = raftManager, + time = Time.SYSTEM, + metrics = controllerMetrics, + threadNamePrefix = Option(threadNamePrefix), + controllerQuorumVotersFuture = controllerQuorumVotersFuture) + controllerServer.socketServerFirstBoundPortFuture.whenComplete((port, e) => { + if (e != null) { + error("Error completing controller socket server future", e) + controllerQuorumVotersFuture.completeExceptionally(e) + } else { + controllerQuorumVotersFuture.complete(Collections.singletonMap(1000, + new InetAddressSpec(new InetSocketAddress("localhost", port)))); Review comment: fixed -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: jira-unsubscr...@kafka.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org