Shyam created SPARK-30222: ----------------------------- Summary: Still getting KafkaConsumer cache hitting max capacity of 64, removing consumer for CacheKe Key: SPARK-30222 URL: https://issues.apache.org/jira/browse/SPARK-30222 Project: Spark Issue Type: Bug Components: Structured Streaming Affects Versions: 2.4.1 Environment: {{Below are the logs.}}
2019-12-11 08:33:31,504 [Executor task launch worker for task 1050] WARN org.apache.spark.sql.kafka010.KafkaDataConsumer - KafkaConsumer cache hitting max capacity of 64, removing consumer for CacheKey(spark-kafka-source-93ee3689-79f9-42e8-b1ee-e856570205ae-1923743483-executor,COMPANY_TRANSACTIONS_INBOUND-21) 2019-12-11 08:33:32,493 [Executor task launch worker for task 1051] WARN org.apache.spark.sql.kafka010.KafkaDataConsumer - KafkaConsumer cache hitting max capacity of 64, removing consumer for CacheKey(spark-kafka-source-93ee3689-79f9-42e8-b1ee-e856570205ae-1923743483-executor,COMPANY_TRANSACTIONS_INBOUND-5) 2019-12-11 08:33:32,570 [Executor task launch worker for task 1052] WARN org.apache.spark.sql.kafka010.KafkaDataConsumer - KafkaConsumer cache hitting max capacity of 64, removing consumer for CacheKey(spark-kafka-source-93ee3689-79f9-42e8-b1ee-e856570205ae-1923743483-executor,COMPANY_TRANSACTIONS_INBOUND-9) 2019-12-11 08:33:33,441 [Executor task launch worker for task 1053] WARN org.apache.spark.sql.kafka010.KafkaDataConsumer - KafkaConsumer cache hitting max capacity of 64, removing consumer for CacheKey(spark-kafka-source-93ee3689-79f9-42e8-b1ee-e856570205ae-1923743483-executor,COMPANY_TRANSACTIONS_INBOUND-17) 2019-12-11 08:33:33,619 [Executor task launch worker for task 1054] WARN org.apache.spark.sql.kafka010.KafkaDataConsumer - KafkaConsumer cache hitting max capacity of 64, removing consumer for CacheKey(spark-kafka-source-93ee3689-79f9-42e8-b1ee-e856570205ae-1923743483-executor,COMPANY_TRANSACTIONS_INBOUND-2) 2019-12-11 08:33:34,474 [Executor task launch worker for task 1055] WARN org.apache.spark.sql.kafka010.KafkaDataConsumer - KafkaConsumer cache hitting max capacity of 64, removing consumer for CacheKey(spark-kafka-source-93ee3689-79f9-42e8-b1ee-e856570205ae-1923743483-executor,COMPANY_TRANSACTIONS_INBOUND-10) 2019-12-11 08:33:35,006 [Executor task launch worker for task 1056] WARN org.apache.spark.sql.kafka010.KafkaDataConsumer - KafkaConsumer cache hitting max capacity of 64, removing consumer for CacheKey(spark-kafka-source-93ee3689-79f9-42e8-b1ee-e856570205ae-1923743483-executor,COMPANY_TRANSACTIONS_INBOUND-6) 2019-12-11 08:33:36,326 [Executor task launch worker for task 1057] WARN org.apache.spark.sql.kafka010.KafkaDataConsumer - KafkaConsumer cache hitting max capacity of 64, removing consumer for CacheKey(spark-kafka-source-93ee3689-79f9-42e8-b1ee-e856570205ae-1923743483-executor,COMPANY_TRANSACTIONS_INBOUND-14) 2019-12-11 08:33:36,634 [Executor task launch worker for task 1058] WARN org.apache.spark.sql.kafka010.KafkaDataConsumer - KafkaConsumer cache hitting max capacity of 64, removing consumer for CacheKey(spark-kafka-source-93ee3689-79f9-42e8-b1ee-e856570205ae-1923743483-executor,COMPANY_TRANSACTIONS_INBOUND-0) 2019-12-11 08:33:37,496 [Executor task launch worker for task 1059] WARN org.apache.spark.sql.kafka010.KafkaDataConsumer - KafkaConsumer cache hitting max capacity of 64, removing consumer for CacheKey(spark-kafka-source-93ee3689-79f9-42e8-b1ee-e856570205ae-1923743483-executor,COMPANY_TRANSACTIONS_INBOUND-19) 2019-12-11 08:33:39,183 [stream execution thread for [id = b3aec196-e4f2-4ef9-973b-d5685eba917e, runId = 5c35a63a-16ad-4899-b732-1019397770bd]] WARN org.apache.spark.sql.execution.streaming.ProcessingTimeExecutor - Current batch is falling behind. The trigger interval is 15000 milliseconds, but spent 63438 milliseconds Reporter: Shyam Fix For: 2.4.1 Me using spark-sql-2.4.1 version with Kafka 0.10 v. While I try to consume data by consumer. it gives error below even after setting .option("spark.sql.kafkaConsumerCache.capacity",128) {{Dataset<Row> df = sparkSession}} {{ .readStream()}} {{ .format("kafka")}} {{ .option("kafka.bootstrap.servers", SERVERS)}} {{ .option("subscribe", TOPIC) }}{{}} {{ .option("spark.sql.kafkaConsumerCache.capacity",128) }} {{ }} {{ .load();}} {{}} {{}} {{}} -- This message was sent by Atlassian Jira (v8.3.4#803005) --------------------------------------------------------------------- To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org