[ https://issues.apache.org/jira/browse/KAFKA-8727?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Levani Kokhreidze resolved KAFKA-8727. -------------------------------------- Resolution: Duplicate Duplicate of KAFKA-6718 > Control over standby tasks host assignment > ------------------------------------------ > > Key: KAFKA-8727 > URL: https://issues.apache.org/jira/browse/KAFKA-8727 > Project: Kafka > Issue Type: New Feature > Components: streams > Reporter: Levani Kokhreidze > Priority: Minor > > *Motivation* > As of now, Kafka Streams user has no control over to which host Kafka Streams > application will create standby task. In production deployments (especially > in Kubernetes) it's quite common to have multiple instances of the same Kafka > Streams application deployed across more than one "cluster" in order to have > high availability of the system. > For example, if we have 6 Kafka Streams instances deployed across two > clusters, we'll get 3 Kafka Streams instances per cluster. With the current > implementation, Kafka Streams application may create "standby task" in the > same cluster as the active task. This is not the most optimal solution, > since, in case of cluster failure recovery time will be much bigger. This is > especially problematic for Kafka Streams application that manages large state. > > *Possible Solution* > It would be great if in the Kafka Streams configuration we could have a > possibility to inject dynamic environment variables and use that environment > variables to control over where standby task should be created. > For example, suppose I have active task *1_1* with environment variable: > *CLUSTER_ID: main01* then stnadby task for *1_1* should be created where > *CLUSTER_ID* *!=* *main01* -- This message was sent by Atlassian JIRA (v7.6.14#76016)