Re: Number of Consumers Connected
You can specify the consumer configs in the config file pointed by "--consumer.config Consumer config properties file.", for example: auto.commit.enable=true auto.commit.interval.ms=1000 Guozhang On Tue, Feb 3, 2015 at 4:54 PM, Sa Li wrote: > Hi, Guozhang > > Thank you very much for the reply, as you mentioned, I download the latest > version > > > https://www.apache.org/dyn/closer.cgi?path=/kafka/0.8.2-beta/kafka-0.8.2-beta-src.tgz > > Untar this build and here is what I see > > root@DO-mq-dev:/home/stuser/kafka-0.8.2-beta-src/bin# > kafka-console-consumer.sh > The console consumer is a tool that reads data from Kafka and outputs it to > standard output. > Option Description > -- --- > --blacklist Blacklist of topics to exclude from > consumption. > --consumer.config Consumer config properties file. > --csv-reporter-enabled If set, the CSV metrics reporter > will > be enabled > --delete-consumer-offsets If specified, the consumer path in > zookeeper is deleted when > starting up > --formatter The name of a class to use for > formatting kafka messages for > display. (default: kafka.tools. > DefaultMessageFormatter) > --from-beginningIf the consumer does not already > have > an established offset to consume > from, start with the earliest > message present in the log rather > than the latest message. > --max-messages The maximum number of messages to > consume before exiting. If not > set, > consumption is continual. > --metrics-dir If csv-reporter-enable is set, and > this parameter isset, the csv > metrics will be outputed here > --property > --skip-message-on-error If there is an error when > processing a > message, skip it instead of halt. > --topic The topic id to consume on. > --whitelist Whitelist of topics to include for > consumption. > --zookeeper REQUIRED: The connection string for > the zookeeper connection in the > form > host:port. Multiple URLS can be > given to allow fail-over. > > Again, I am still not able to see description of --property, or I download > the wrong version? > > > thanks > > AL > > > On Tue, Feb 3, 2015 at 4:29 PM, Guozhang Wang wrote: > > > Hello Sa, > > > > Could you try the latest 0.8.2 release, whose console consumer tool has > > been polished a bit with clearer properties? > > > > Guozhang > > > > > > On Tue, Feb 3, 2015 at 10:32 AM, Sa Li wrote: > > > > > GuoZhang > > > > > > Sorry for leaving this topic for a while, I am still not clear how to > > > commit the offset to zk from commandline, I tried this > > > > > > bin/kafka-console-consumer.sh --zookeeper 10.100.71.33:2181 --topic > > > pipe-test-2 --from-beginning --property pipe > > > > > > It seems generate a console-consumer-001 in zK, but when I did that to > > > other topics, nothing in zk, (I can't read anything from consumer group > > in > > > kafka-web-console), see > > > [zk: localhost:2181(CONNECTED) 2] ls > > /consumers/web-console-consumer-38650 > > > [offsets, owners, ids] > > > [zk: localhost:2181(CONNECTED) 3] ls > > > /consumers/web-console-consumer-38650/offsets > > > [PofApiTest-1] > > > [zk: localhost:2181(CONNECTED) 4] ls > > > /consumers/web-console-consumer-38650/offsets/PofApiTest-1 > > > [3, 2, 1, 0, 7, 6, 5, 4] > > > [zk: localhost:2181(CONNECTED) 5] ls > > > /consumers/web-console-consumer-38650/offsets/PofApiTest-1/3 > > > [] > > > > > > Any ideas? > > > > > > thanks > > > > > > AL > > > > > > > > > On Tue, Jan 20, 2015 at 9:57 PM, Guozhang Wang > > wrote: > > > > > > > It seems not the latest version of Kafka, which version are you > using? > > > > > > > > On Tue, Jan 20, 2015 at 9:46 AM, Sa Li wrote: > > > > > > > > > Guozhang > > > > > > > > > > Thank you very much for reply, here I print out the > > > > > kafka-console-consumer.sh help, > > > > > > > > > > root@voluminous-mass:/srv/kafka# bin/kafka-console-consumer.sh > > > > > Missing required argument "[zookeeper]" > > > > > Option D
Re: Number of Consumers Connected
Hi, Guozhang Thank you very much for the reply, as you mentioned, I download the latest version https://www.apache.org/dyn/closer.cgi?path=/kafka/0.8.2-beta/kafka-0.8.2-beta-src.tgz Untar this build and here is what I see root@DO-mq-dev:/home/stuser/kafka-0.8.2-beta-src/bin# kafka-console-consumer.sh The console consumer is a tool that reads data from Kafka and outputs it to standard output. Option Description -- --- --blacklist Blacklist of topics to exclude from consumption. --consumer.config Consumer config properties file. --csv-reporter-enabled If set, the CSV metrics reporter will be enabled --delete-consumer-offsets If specified, the consumer path in zookeeper is deleted when starting up --formatter The name of a class to use for formatting kafka messages for display. (default: kafka.tools. DefaultMessageFormatter) --from-beginningIf the consumer does not already have an established offset to consume from, start with the earliest message present in the log rather than the latest message. --max-messages The maximum number of messages to consume before exiting. If not set, consumption is continual. --metrics-dir If csv-reporter-enable is set, and this parameter isset, the csv metrics will be outputed here --property --skip-message-on-error If there is an error when processing a message, skip it instead of halt. --topic The topic id to consume on. --whitelist Whitelist of topics to include for consumption. --zookeeper REQUIRED: The connection string for the zookeeper connection in the form host:port. Multiple URLS can be given to allow fail-over. Again, I am still not able to see description of --property, or I download the wrong version? thanks AL On Tue, Feb 3, 2015 at 4:29 PM, Guozhang Wang wrote: > Hello Sa, > > Could you try the latest 0.8.2 release, whose console consumer tool has > been polished a bit with clearer properties? > > Guozhang > > > On Tue, Feb 3, 2015 at 10:32 AM, Sa Li wrote: > > > GuoZhang > > > > Sorry for leaving this topic for a while, I am still not clear how to > > commit the offset to zk from commandline, I tried this > > > > bin/kafka-console-consumer.sh --zookeeper 10.100.71.33:2181 --topic > > pipe-test-2 --from-beginning --property pipe > > > > It seems generate a console-consumer-001 in zK, but when I did that to > > other topics, nothing in zk, (I can't read anything from consumer group > in > > kafka-web-console), see > > [zk: localhost:2181(CONNECTED) 2] ls > /consumers/web-console-consumer-38650 > > [offsets, owners, ids] > > [zk: localhost:2181(CONNECTED) 3] ls > > /consumers/web-console-consumer-38650/offsets > > [PofApiTest-1] > > [zk: localhost:2181(CONNECTED) 4] ls > > /consumers/web-console-consumer-38650/offsets/PofApiTest-1 > > [3, 2, 1, 0, 7, 6, 5, 4] > > [zk: localhost:2181(CONNECTED) 5] ls > > /consumers/web-console-consumer-38650/offsets/PofApiTest-1/3 > > [] > > > > Any ideas? > > > > thanks > > > > AL > > > > > > On Tue, Jan 20, 2015 at 9:57 PM, Guozhang Wang > wrote: > > > > > It seems not the latest version of Kafka, which version are you using? > > > > > > On Tue, Jan 20, 2015 at 9:46 AM, Sa Li wrote: > > > > > > > Guozhang > > > > > > > > Thank you very much for reply, here I print out the > > > > kafka-console-consumer.sh help, > > > > > > > > root@voluminous-mass:/srv/kafka# bin/kafka-console-consumer.sh > > > > Missing required argument "[zookeeper]" > > > > Option Description > > > > -- --- > > > > --autocommit.interval.ms The time interval at which > to > > > save > > > > the > > > > current offset in ms > > (default: > > > > 6) > > > > --blacklist Blacklist of topics to > exclude > > > from > > > > consumption. > > > > --consumer-timeout-msconsumer throws timeout > > exception > > > >
Re: Number of Consumers Connected
Hello Sa, Could you try the latest 0.8.2 release, whose console consumer tool has been polished a bit with clearer properties? Guozhang On Tue, Feb 3, 2015 at 10:32 AM, Sa Li wrote: > GuoZhang > > Sorry for leaving this topic for a while, I am still not clear how to > commit the offset to zk from commandline, I tried this > > bin/kafka-console-consumer.sh --zookeeper 10.100.71.33:2181 --topic > pipe-test-2 --from-beginning --property pipe > > It seems generate a console-consumer-001 in zK, but when I did that to > other topics, nothing in zk, (I can't read anything from consumer group in > kafka-web-console), see > [zk: localhost:2181(CONNECTED) 2] ls /consumers/web-console-consumer-38650 > [offsets, owners, ids] > [zk: localhost:2181(CONNECTED) 3] ls > /consumers/web-console-consumer-38650/offsets > [PofApiTest-1] > [zk: localhost:2181(CONNECTED) 4] ls > /consumers/web-console-consumer-38650/offsets/PofApiTest-1 > [3, 2, 1, 0, 7, 6, 5, 4] > [zk: localhost:2181(CONNECTED) 5] ls > /consumers/web-console-consumer-38650/offsets/PofApiTest-1/3 > [] > > Any ideas? > > thanks > > AL > > > On Tue, Jan 20, 2015 at 9:57 PM, Guozhang Wang wrote: > > > It seems not the latest version of Kafka, which version are you using? > > > > On Tue, Jan 20, 2015 at 9:46 AM, Sa Li wrote: > > > > > Guozhang > > > > > > Thank you very much for reply, here I print out the > > > kafka-console-consumer.sh help, > > > > > > root@voluminous-mass:/srv/kafka# bin/kafka-console-consumer.sh > > > Missing required argument "[zookeeper]" > > > Option Description > > > -- --- > > > --autocommit.interval.ms The time interval at which to > > save > > > the > > > current offset in ms > (default: > > > 6) > > > --blacklist Blacklist of topics to exclude > > from > > > consumption. > > > --consumer-timeout-msconsumer throws timeout > exception > > > after waiting this much of > time > > > without incoming messages > > > (default: > > > -1) > > > --csv-reporter-enabled If set, the CSV metrics > reporter > > > will > > > be enabled > > > --fetch-size The amount of data to fetch in > a > > > single request. (default: > > > 1048576) > > > --formatter The name of a class to use for > > > formatting kafka messages for > > > display. (default: > > > kafka.consumer. > > > DefaultMessageFormatter) > > > --from-beginningIf the consumer does not > already > > > have > > > an established offset to > > consume > > > from, start with the earliest > > > message present in the log > > rather > > > than the latest message. > > > --groupThe group id to consume on. > > > (default: > > > console-consumer-85664) > > > --max-messages The maximum number of messages > to > > > consume before exiting. If > not > > > set, > > > consumption is continual. > > > --max-wait-ms The max amount of time each > fetch > > > request waits. (default: 100) > > > --metrics-dir If csv-reporter-enable is set, > > and > > > this parameter isset, the csv > > > metrics will be outputed here > > > --min-fetch-bytes The min number of bytes each > > fetch > > > request waits for. (default: > 1) > > > --property > > > --refresh-leader-backoff-ms > > ms> metadata (default: 200) > > > --skip-message-on-error If there is an error when > > > processing a > > > message, skip it instead of > > halt. > > > --socket-buffer-size The size of the tcp RECV size. > > > (default: 2097152) > > > --socket-timeout-msThe socket timeout used for the > > > connection to the broker > > > (default: > > > 3) > > > --topic The topic id to consume on. > > > --whitelist Whitelist of topics to include > > for > > >
Re: Number of Consumers Connected
GuoZhang Sorry for leaving this topic for a while, I am still not clear how to commit the offset to zk from commandline, I tried this bin/kafka-console-consumer.sh --zookeeper 10.100.71.33:2181 --topic pipe-test-2 --from-beginning --property pipe It seems generate a console-consumer-001 in zK, but when I did that to other topics, nothing in zk, (I can't read anything from consumer group in kafka-web-console), see [zk: localhost:2181(CONNECTED) 2] ls /consumers/web-console-consumer-38650 [offsets, owners, ids] [zk: localhost:2181(CONNECTED) 3] ls /consumers/web-console-consumer-38650/offsets [PofApiTest-1] [zk: localhost:2181(CONNECTED) 4] ls /consumers/web-console-consumer-38650/offsets/PofApiTest-1 [3, 2, 1, 0, 7, 6, 5, 4] [zk: localhost:2181(CONNECTED) 5] ls /consumers/web-console-consumer-38650/offsets/PofApiTest-1/3 [] Any ideas? thanks AL On Tue, Jan 20, 2015 at 9:57 PM, Guozhang Wang wrote: > It seems not the latest version of Kafka, which version are you using? > > On Tue, Jan 20, 2015 at 9:46 AM, Sa Li wrote: > > > Guozhang > > > > Thank you very much for reply, here I print out the > > kafka-console-consumer.sh help, > > > > root@voluminous-mass:/srv/kafka# bin/kafka-console-consumer.sh > > Missing required argument "[zookeeper]" > > Option Description > > -- --- > > --autocommit.interval.ms The time interval at which to > save > > the > > current offset in ms (default: > > 6) > > --blacklist Blacklist of topics to exclude > from > > consumption. > > --consumer-timeout-msconsumer throws timeout exception > > after waiting this much of time > > without incoming messages > > (default: > > -1) > > --csv-reporter-enabled If set, the CSV metrics reporter > > will > > be enabled > > --fetch-size The amount of data to fetch in a > > single request. (default: > > 1048576) > > --formatter The name of a class to use for > > formatting kafka messages for > > display. (default: > > kafka.consumer. > > DefaultMessageFormatter) > > --from-beginningIf the consumer does not already > > have > > an established offset to > consume > > from, start with the earliest > > message present in the log > rather > > than the latest message. > > --groupThe group id to consume on. > > (default: > > console-consumer-85664) > > --max-messages The maximum number of messages to > > consume before exiting. If not > > set, > > consumption is continual. > > --max-wait-ms The max amount of time each fetch > > request waits. (default: 100) > > --metrics-dir If csv-reporter-enable is set, > and > > this parameter isset, the csv > > metrics will be outputed here > > --min-fetch-bytes The min number of bytes each > fetch > > request waits for. (default: 1) > > --property > > --refresh-leader-backoff-ms > ms> metadata (default: 200) > > --skip-message-on-error If there is an error when > > processing a > > message, skip it instead of > halt. > > --socket-buffer-size The size of the tcp RECV size. > > (default: 2097152) > > --socket-timeout-msThe socket timeout used for the > > connection to the broker > > (default: > > 3) > > --topic The topic id to consume on. > > --whitelist Whitelist of topics to include > for > > consumption. > > --zookeeper REQUIRED: The connection string > for > > the zookeeper connection in the > > form > > host:port. Multiple URLS can be > > given to allow fail-over. > > > > --property option is not provided the description, is there an exampl
Re: Number of Consumers Connected
It seems not the latest version of Kafka, which version are you using? On Tue, Jan 20, 2015 at 9:46 AM, Sa Li wrote: > Guozhang > > Thank you very much for reply, here I print out the > kafka-console-consumer.sh help, > > root@voluminous-mass:/srv/kafka# bin/kafka-console-consumer.sh > Missing required argument "[zookeeper]" > Option Description > -- --- > --autocommit.interval.ms The time interval at which to save > the > current offset in ms (default: > 6) > --blacklist Blacklist of topics to exclude from > consumption. > --consumer-timeout-msconsumer throws timeout exception > after waiting this much of time > without incoming messages > (default: > -1) > --csv-reporter-enabled If set, the CSV metrics reporter > will > be enabled > --fetch-size The amount of data to fetch in a > single request. (default: > 1048576) > --formatter The name of a class to use for > formatting kafka messages for > display. (default: > kafka.consumer. > DefaultMessageFormatter) > --from-beginningIf the consumer does not already > have > an established offset to consume > from, start with the earliest > message present in the log rather > than the latest message. > --groupThe group id to consume on. > (default: > console-consumer-85664) > --max-messages The maximum number of messages to > consume before exiting. If not > set, > consumption is continual. > --max-wait-ms The max amount of time each fetch > request waits. (default: 100) > --metrics-dir If csv-reporter-enable is set, and > this parameter isset, the csv > metrics will be outputed here > --min-fetch-bytes The min number of bytes each fetch > request waits for. (default: 1) > --property > --refresh-leader-backoff-msms> metadata (default: 200) > --skip-message-on-error If there is an error when > processing a > message, skip it instead of halt. > --socket-buffer-size The size of the tcp RECV size. > (default: 2097152) > --socket-timeout-msThe socket timeout used for the > connection to the broker > (default: > 3) > --topic The topic id to consume on. > --whitelist Whitelist of topics to include for > consumption. > --zookeeper REQUIRED: The connection string for > the zookeeper connection in the > form > host:port. Multiple URLS can be > given to allow fail-over. > > --property option is not provided the description, is there an example how > to use it? > > thanks > > AL > > On Mon, Jan 19, 2015 at 6:30 PM, Guozhang Wang wrote: > > > There is a property config you can set via bin/kafka-console-consumer.sh > to > > commit offsets to ZK, you can use bin/kafka-console-consumer.sh --help to > > list all the properties. > > > > Guozhang > > > > On Mon, Jan 19, 2015 at 5:15 PM, Sa Li wrote: > > > > > Guozhang, > > > > > > Currently we are in the stage to testing producer, our C# producer > > sending > > > data to brokers, and use > > > > > > bin/kafka-run-class.sh > org.apache.kafka.clients.tools.ProducerPerformance > > > command to produce the messages. We don't have a coded consumer to > commit > > > offset, we use > > > bin/kafka-console-consumer.sh --zookeeper command to consume, is there > a > > > command that we can use on command line to create zk path? > > > > > > thanks > > > > > > AL > > > > > > On Mon, Jan 19, 2015 at 4:14 PM, Guozhang Wang > > wrote: > > > > > > > Sa, > > > > > > > > Did your consumer ever commit offsets to Kafka? If not then no > > > > corresponding ZK path will be created. > >
Re: Number of Consumers Connected
Guozhang Thank you very much for reply, here I print out the kafka-console-consumer.sh help, root@voluminous-mass:/srv/kafka# bin/kafka-console-consumer.sh Missing required argument "[zookeeper]" Option Description -- --- --autocommit.interval.ms The time interval at which to save the current offset in ms (default: 6) --blacklist Blacklist of topics to exclude from consumption. --consumer-timeout-msconsumer throws timeout exception after waiting this much of time without incoming messages (default: -1) --csv-reporter-enabled If set, the CSV metrics reporter will be enabled --fetch-size The amount of data to fetch in a single request. (default: 1048576) --formatter The name of a class to use for formatting kafka messages for display. (default: kafka.consumer. DefaultMessageFormatter) --from-beginningIf the consumer does not already have an established offset to consume from, start with the earliest message present in the log rather than the latest message. --groupThe group id to consume on. (default: console-consumer-85664) --max-messages The maximum number of messages to consume before exiting. If not set, consumption is continual. --max-wait-ms The max amount of time each fetch request waits. (default: 100) --metrics-dir If csv-reporter-enable is set, and this parameter isset, the csv metrics will be outputed here --min-fetch-bytes The min number of bytes each fetch request waits for. (default: 1) --property --refresh-leader-backoff-ms metadata (default: 200) --skip-message-on-error If there is an error when processing a message, skip it instead of halt. --socket-buffer-size The size of the tcp RECV size. (default: 2097152) --socket-timeout-msThe socket timeout used for the connection to the broker (default: 3) --topic The topic id to consume on. --whitelist Whitelist of topics to include for consumption. --zookeeper REQUIRED: The connection string for the zookeeper connection in the form host:port. Multiple URLS can be given to allow fail-over. --property option is not provided the description, is there an example how to use it? thanks AL On Mon, Jan 19, 2015 at 6:30 PM, Guozhang Wang wrote: > There is a property config you can set via bin/kafka-console-consumer.sh to > commit offsets to ZK, you can use bin/kafka-console-consumer.sh --help to > list all the properties. > > Guozhang > > On Mon, Jan 19, 2015 at 5:15 PM, Sa Li wrote: > > > Guozhang, > > > > Currently we are in the stage to testing producer, our C# producer > sending > > data to brokers, and use > > > > bin/kafka-run-class.sh org.apache.kafka.clients.tools.ProducerPerformance > > command to produce the messages. We don't have a coded consumer to commit > > offset, we use > > bin/kafka-console-consumer.sh --zookeeper command to consume, is there a > > command that we can use on command line to create zk path? > > > > thanks > > > > AL > > > > On Mon, Jan 19, 2015 at 4:14 PM, Guozhang Wang > wrote: > > > > > Sa, > > > > > > Did your consumer ever commit offsets to Kafka? If not then no > > > corresponding ZK path will be created. > > > > > > Guozhang > > > > > > On Mon, Jan 19, 2015 at 3:58 PM, Sa Li wrote: > > > > > > > Hi, > > > > > > > > I use such tool > > > > > > > > Consumer Offset Checker > > > > > > > > Displays the: Consumer Group, Topic, Partitions, Offset, logSize, > Lag, > > > > Owner for the specified set of Topics and Consumer Group > > > > bin/kafka-run-class
Re: Number of Consumers Connected
There is a property config you can set via bin/kafka-console-consumer.sh to commit offsets to ZK, you can use bin/kafka-console-consumer.sh --help to list all the properties. Guozhang On Mon, Jan 19, 2015 at 5:15 PM, Sa Li wrote: > Guozhang, > > Currently we are in the stage to testing producer, our C# producer sending > data to brokers, and use > > bin/kafka-run-class.sh org.apache.kafka.clients.tools.ProducerPerformance > command to produce the messages. We don't have a coded consumer to commit > offset, we use > bin/kafka-console-consumer.sh --zookeeper command to consume, is there a > command that we can use on command line to create zk path? > > thanks > > AL > > On Mon, Jan 19, 2015 at 4:14 PM, Guozhang Wang wrote: > > > Sa, > > > > Did your consumer ever commit offsets to Kafka? If not then no > > corresponding ZK path will be created. > > > > Guozhang > > > > On Mon, Jan 19, 2015 at 3:58 PM, Sa Li wrote: > > > > > Hi, > > > > > > I use such tool > > > > > > Consumer Offset Checker > > > > > > Displays the: Consumer Group, Topic, Partitions, Offset, logSize, Lag, > > > Owner for the specified set of Topics and Consumer Group > > > bin/kafka-run-class.sh kafka.tools.ConsumerOffsetChecker > > > > > > To be able to know the consumer group, in zkCli.sh > > > > > > [zk: localhost:2181(CONNECTED) 3] ls / > > > [transactional, admin, zookeeper, consumers, config, controller, storm, > > > brokers, controller_epoch] > > > [zk: localhost:2181(CONNECTED) 4] ls /consumers > > > [web-console-consumer-99295, web-console-consumer-37853, > > > web-console-consumer-30841, perf-consumer-92283, perf-consumer-21631, > > > perf-consumer-95281, perf-consumer-59296, web-console-consumer-52126, > > > web-console-consumer-89137, perf-consumer-72484, perf-consumer-80363, > > > web-console-consumer-47543, web-console-consumer-22509, > > > perf-consumer-16954, perf-consumer-53957, perf-consumer-39448, > > > web-console-consumer-17021, perf-consumer-88693, > > > web-console-consumer-48744, web-console-consumer-82543, > > > perf-consumer-89565, web-console-consumer-97959, perf-consumer-40427, > > > web-console-consumer-95350, web-console-consumer-26473, > > > web-console-consumer-79384, web-console-consumer-8, > > > perf-consumer-91681, web-console-consumer-36136, > > > web-console-consumer-86924, perf-consumer-24510, perf-consumer-5888, > > > perf-consumer-73534, perf-consumer-92985, perf-consumer-7675, > > > perf-consumer-52306, perf-consumer-87352, web-console-consumer-30400] > > > [zk: localhost:2181(CONNECTED) 5] > > > > > > I then run > > > > > > root@exemplary-birds:/srv/kafka# bin/kafka-run-class.sh > > > kafka.tools.ConsumerOffsetChecker --topic PofApiTest-1 --group > > > web-console-consumer-48744 > > > Group Topic Pid Offset > > > logSize Lag Owner > > > Exception in thread "main" > > org.I0Itec.zkclient.exception.ZkNoNodeException: > > > org.apache.zookeeper.KeeperException$NoNodeException: KeeperErrorCode = > > > NoNode for /consumers/web-console-consumer-48744/offsets/PofApiTest-1/0 > > > at > > > org.I0Itec.zkclient.exception.ZkException.create(ZkException.java:47) > > > at > > > org.I0Itec.zkclient.ZkClient.retryUntilConnected(ZkClient.java:685) > > > at org.I0Itec.zkclient.ZkClient.readData(ZkClient.java:766) > > > at org.I0Itec.zkclient.ZkClient.readData(ZkClient.java:761) > > > at kafka.utils.ZkUtils$.readData(ZkUtils.scala:461) > > > at > > > > > > > > > kafka.tools.ConsumerOffsetChecker$.kafka$tools$ConsumerOffsetChecker$$processPartition(ConsumerOffsetChecker.scala:59) > > > at > > > > > > > > > kafka.tools.ConsumerOffsetChecker$$anonfun$kafka$tools$ConsumerOffsetChecker$$processTopic$1.apply$mcVI$sp(ConsumerOffsetChecker.scala:89) > > > at > > > > > > > > > kafka.tools.ConsumerOffsetChecker$$anonfun$kafka$tools$ConsumerOffsetChecker$$processTopic$1.apply(ConsumerOffsetChecker.scala:89) > > > at > > > > > > > > > kafka.tools.ConsumerOffsetChecker$$anonfun$kafka$tools$ConsumerOffsetChecker$$processTopic$1.apply(ConsumerOffsetChecker.scala:89) > > > at > > > > > > > > > scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59) > > > at > > > scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47) > > > at > > > > > > > > > kafka.tools.ConsumerOffsetChecker$.kafka$tools$ConsumerOffsetChecker$$processTopic(ConsumerOffsetChecker.scala:88) > > > at > > > > > > > > > kafka.tools.ConsumerOffsetChecker$$anonfun$main$3.apply(ConsumerOffsetChecker.scala:153) > > > at > > > > > > > > > kafka.tools.ConsumerOffsetChecker$$anonfun$main$3.apply(ConsumerOffsetChecker.scala:153) > > > at scala.collection.immutable.List.foreach(List.scala:318) > > > at > > > > kafka.tools.ConsumerOffsetChecker$.main(ConsumerOffsetChecker.scala:152) > > > at > > > kafka.tools.ConsumerOffsetChecker.main(ConsumerOff
Re: Number of Consumers Connected
Guozhang, Currently we are in the stage to testing producer, our C# producer sending data to brokers, and use bin/kafka-run-class.sh org.apache.kafka.clients.tools.ProducerPerformance command to produce the messages. We don't have a coded consumer to commit offset, we use bin/kafka-console-consumer.sh --zookeeper command to consume, is there a command that we can use on command line to create zk path? thanks AL On Mon, Jan 19, 2015 at 4:14 PM, Guozhang Wang wrote: > Sa, > > Did your consumer ever commit offsets to Kafka? If not then no > corresponding ZK path will be created. > > Guozhang > > On Mon, Jan 19, 2015 at 3:58 PM, Sa Li wrote: > > > Hi, > > > > I use such tool > > > > Consumer Offset Checker > > > > Displays the: Consumer Group, Topic, Partitions, Offset, logSize, Lag, > > Owner for the specified set of Topics and Consumer Group > > bin/kafka-run-class.sh kafka.tools.ConsumerOffsetChecker > > > > To be able to know the consumer group, in zkCli.sh > > > > [zk: localhost:2181(CONNECTED) 3] ls / > > [transactional, admin, zookeeper, consumers, config, controller, storm, > > brokers, controller_epoch] > > [zk: localhost:2181(CONNECTED) 4] ls /consumers > > [web-console-consumer-99295, web-console-consumer-37853, > > web-console-consumer-30841, perf-consumer-92283, perf-consumer-21631, > > perf-consumer-95281, perf-consumer-59296, web-console-consumer-52126, > > web-console-consumer-89137, perf-consumer-72484, perf-consumer-80363, > > web-console-consumer-47543, web-console-consumer-22509, > > perf-consumer-16954, perf-consumer-53957, perf-consumer-39448, > > web-console-consumer-17021, perf-consumer-88693, > > web-console-consumer-48744, web-console-consumer-82543, > > perf-consumer-89565, web-console-consumer-97959, perf-consumer-40427, > > web-console-consumer-95350, web-console-consumer-26473, > > web-console-consumer-79384, web-console-consumer-8, > > perf-consumer-91681, web-console-consumer-36136, > > web-console-consumer-86924, perf-consumer-24510, perf-consumer-5888, > > perf-consumer-73534, perf-consumer-92985, perf-consumer-7675, > > perf-consumer-52306, perf-consumer-87352, web-console-consumer-30400] > > [zk: localhost:2181(CONNECTED) 5] > > > > I then run > > > > root@exemplary-birds:/srv/kafka# bin/kafka-run-class.sh > > kafka.tools.ConsumerOffsetChecker --topic PofApiTest-1 --group > > web-console-consumer-48744 > > Group Topic Pid Offset > > logSize Lag Owner > > Exception in thread "main" > org.I0Itec.zkclient.exception.ZkNoNodeException: > > org.apache.zookeeper.KeeperException$NoNodeException: KeeperErrorCode = > > NoNode for /consumers/web-console-consumer-48744/offsets/PofApiTest-1/0 > > at > > org.I0Itec.zkclient.exception.ZkException.create(ZkException.java:47) > > at > > org.I0Itec.zkclient.ZkClient.retryUntilConnected(ZkClient.java:685) > > at org.I0Itec.zkclient.ZkClient.readData(ZkClient.java:766) > > at org.I0Itec.zkclient.ZkClient.readData(ZkClient.java:761) > > at kafka.utils.ZkUtils$.readData(ZkUtils.scala:461) > > at > > > > > kafka.tools.ConsumerOffsetChecker$.kafka$tools$ConsumerOffsetChecker$$processPartition(ConsumerOffsetChecker.scala:59) > > at > > > > > kafka.tools.ConsumerOffsetChecker$$anonfun$kafka$tools$ConsumerOffsetChecker$$processTopic$1.apply$mcVI$sp(ConsumerOffsetChecker.scala:89) > > at > > > > > kafka.tools.ConsumerOffsetChecker$$anonfun$kafka$tools$ConsumerOffsetChecker$$processTopic$1.apply(ConsumerOffsetChecker.scala:89) > > at > > > > > kafka.tools.ConsumerOffsetChecker$$anonfun$kafka$tools$ConsumerOffsetChecker$$processTopic$1.apply(ConsumerOffsetChecker.scala:89) > > at > > > > > scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59) > > at > > scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47) > > at > > > > > kafka.tools.ConsumerOffsetChecker$.kafka$tools$ConsumerOffsetChecker$$processTopic(ConsumerOffsetChecker.scala:88) > > at > > > > > kafka.tools.ConsumerOffsetChecker$$anonfun$main$3.apply(ConsumerOffsetChecker.scala:153) > > at > > > > > kafka.tools.ConsumerOffsetChecker$$anonfun$main$3.apply(ConsumerOffsetChecker.scala:153) > > at scala.collection.immutable.List.foreach(List.scala:318) > > at > > kafka.tools.ConsumerOffsetChecker$.main(ConsumerOffsetChecker.scala:152) > > at > > kafka.tools.ConsumerOffsetChecker.main(ConsumerOffsetChecker.scala) > > Caused by: org.apache.zookeeper.KeeperException$NoNodeException: > > KeeperErrorCode = NoNode for > > /consumers/web-console-consumer-48744/offsets/PofApiTest-1/0 > > at > > org.apache.zookeeper.KeeperException.create(KeeperException.java:102) > > at > > org.apache.zookeeper.KeeperException.create(KeeperException.java:42) > > at org.apache.zookeeper.ZooKeeper.getData(ZooKeeper.java:927) > > at org.apache.zookeeper.ZooKee
Re: Number of Consumers Connected
Sa, Did your consumer ever commit offsets to Kafka? If not then no corresponding ZK path will be created. Guozhang On Mon, Jan 19, 2015 at 3:58 PM, Sa Li wrote: > Hi, > > I use such tool > > Consumer Offset Checker > > Displays the: Consumer Group, Topic, Partitions, Offset, logSize, Lag, > Owner for the specified set of Topics and Consumer Group > bin/kafka-run-class.sh kafka.tools.ConsumerOffsetChecker > > To be able to know the consumer group, in zkCli.sh > > [zk: localhost:2181(CONNECTED) 3] ls / > [transactional, admin, zookeeper, consumers, config, controller, storm, > brokers, controller_epoch] > [zk: localhost:2181(CONNECTED) 4] ls /consumers > [web-console-consumer-99295, web-console-consumer-37853, > web-console-consumer-30841, perf-consumer-92283, perf-consumer-21631, > perf-consumer-95281, perf-consumer-59296, web-console-consumer-52126, > web-console-consumer-89137, perf-consumer-72484, perf-consumer-80363, > web-console-consumer-47543, web-console-consumer-22509, > perf-consumer-16954, perf-consumer-53957, perf-consumer-39448, > web-console-consumer-17021, perf-consumer-88693, > web-console-consumer-48744, web-console-consumer-82543, > perf-consumer-89565, web-console-consumer-97959, perf-consumer-40427, > web-console-consumer-95350, web-console-consumer-26473, > web-console-consumer-79384, web-console-consumer-8, > perf-consumer-91681, web-console-consumer-36136, > web-console-consumer-86924, perf-consumer-24510, perf-consumer-5888, > perf-consumer-73534, perf-consumer-92985, perf-consumer-7675, > perf-consumer-52306, perf-consumer-87352, web-console-consumer-30400] > [zk: localhost:2181(CONNECTED) 5] > > I then run > > root@exemplary-birds:/srv/kafka# bin/kafka-run-class.sh > kafka.tools.ConsumerOffsetChecker --topic PofApiTest-1 --group > web-console-consumer-48744 > Group Topic Pid Offset > logSize Lag Owner > Exception in thread "main" org.I0Itec.zkclient.exception.ZkNoNodeException: > org.apache.zookeeper.KeeperException$NoNodeException: KeeperErrorCode = > NoNode for /consumers/web-console-consumer-48744/offsets/PofApiTest-1/0 > at > org.I0Itec.zkclient.exception.ZkException.create(ZkException.java:47) > at > org.I0Itec.zkclient.ZkClient.retryUntilConnected(ZkClient.java:685) > at org.I0Itec.zkclient.ZkClient.readData(ZkClient.java:766) > at org.I0Itec.zkclient.ZkClient.readData(ZkClient.java:761) > at kafka.utils.ZkUtils$.readData(ZkUtils.scala:461) > at > > kafka.tools.ConsumerOffsetChecker$.kafka$tools$ConsumerOffsetChecker$$processPartition(ConsumerOffsetChecker.scala:59) > at > > kafka.tools.ConsumerOffsetChecker$$anonfun$kafka$tools$ConsumerOffsetChecker$$processTopic$1.apply$mcVI$sp(ConsumerOffsetChecker.scala:89) > at > > kafka.tools.ConsumerOffsetChecker$$anonfun$kafka$tools$ConsumerOffsetChecker$$processTopic$1.apply(ConsumerOffsetChecker.scala:89) > at > > kafka.tools.ConsumerOffsetChecker$$anonfun$kafka$tools$ConsumerOffsetChecker$$processTopic$1.apply(ConsumerOffsetChecker.scala:89) > at > > scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59) > at > scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47) > at > > kafka.tools.ConsumerOffsetChecker$.kafka$tools$ConsumerOffsetChecker$$processTopic(ConsumerOffsetChecker.scala:88) > at > > kafka.tools.ConsumerOffsetChecker$$anonfun$main$3.apply(ConsumerOffsetChecker.scala:153) > at > > kafka.tools.ConsumerOffsetChecker$$anonfun$main$3.apply(ConsumerOffsetChecker.scala:153) > at scala.collection.immutable.List.foreach(List.scala:318) > at > kafka.tools.ConsumerOffsetChecker$.main(ConsumerOffsetChecker.scala:152) > at > kafka.tools.ConsumerOffsetChecker.main(ConsumerOffsetChecker.scala) > Caused by: org.apache.zookeeper.KeeperException$NoNodeException: > KeeperErrorCode = NoNode for > /consumers/web-console-consumer-48744/offsets/PofApiTest-1/0 > at > org.apache.zookeeper.KeeperException.create(KeeperException.java:102) > at > org.apache.zookeeper.KeeperException.create(KeeperException.java:42) > at org.apache.zookeeper.ZooKeeper.getData(ZooKeeper.java:927) > at org.apache.zookeeper.ZooKeeper.getData(ZooKeeper.java:956) > at org.I0Itec.zkclient.ZkConnection.readData(ZkConnection.java:103) > at org.I0Itec.zkclient.ZkClient$9.call(ZkClient.java:770) > at org.I0Itec.zkclient.ZkClient$9.call(ZkClient.java:766) > at > org.I0Itec.zkclient.ZkClient.retryUntilConnected(ZkClient.java:675) > ... 15 more > > So consumer groups make confusing, I didn't specify the consumer-group-id > in producer, the only place I know to config group is consumer.properties > > > #consumer group id > group.id=test-consumer-group > > > Any hints? Thanks > > > AL > > > > > > > On Mon, Dec 15, 2014 at 6:46 PM, nitin sharma > > wrote: > > > got it ... thank
Re: Number of Consumers Connected
Hi, I use such tool Consumer Offset Checker Displays the: Consumer Group, Topic, Partitions, Offset, logSize, Lag, Owner for the specified set of Topics and Consumer Group bin/kafka-run-class.sh kafka.tools.ConsumerOffsetChecker To be able to know the consumer group, in zkCli.sh [zk: localhost:2181(CONNECTED) 3] ls / [transactional, admin, zookeeper, consumers, config, controller, storm, brokers, controller_epoch] [zk: localhost:2181(CONNECTED) 4] ls /consumers [web-console-consumer-99295, web-console-consumer-37853, web-console-consumer-30841, perf-consumer-92283, perf-consumer-21631, perf-consumer-95281, perf-consumer-59296, web-console-consumer-52126, web-console-consumer-89137, perf-consumer-72484, perf-consumer-80363, web-console-consumer-47543, web-console-consumer-22509, perf-consumer-16954, perf-consumer-53957, perf-consumer-39448, web-console-consumer-17021, perf-consumer-88693, web-console-consumer-48744, web-console-consumer-82543, perf-consumer-89565, web-console-consumer-97959, perf-consumer-40427, web-console-consumer-95350, web-console-consumer-26473, web-console-consumer-79384, web-console-consumer-8, perf-consumer-91681, web-console-consumer-36136, web-console-consumer-86924, perf-consumer-24510, perf-consumer-5888, perf-consumer-73534, perf-consumer-92985, perf-consumer-7675, perf-consumer-52306, perf-consumer-87352, web-console-consumer-30400] [zk: localhost:2181(CONNECTED) 5] I then run root@exemplary-birds:/srv/kafka# bin/kafka-run-class.sh kafka.tools.ConsumerOffsetChecker --topic PofApiTest-1 --group web-console-consumer-48744 Group Topic Pid Offset logSize Lag Owner Exception in thread "main" org.I0Itec.zkclient.exception.ZkNoNodeException: org.apache.zookeeper.KeeperException$NoNodeException: KeeperErrorCode = NoNode for /consumers/web-console-consumer-48744/offsets/PofApiTest-1/0 at org.I0Itec.zkclient.exception.ZkException.create(ZkException.java:47) at org.I0Itec.zkclient.ZkClient.retryUntilConnected(ZkClient.java:685) at org.I0Itec.zkclient.ZkClient.readData(ZkClient.java:766) at org.I0Itec.zkclient.ZkClient.readData(ZkClient.java:761) at kafka.utils.ZkUtils$.readData(ZkUtils.scala:461) at kafka.tools.ConsumerOffsetChecker$.kafka$tools$ConsumerOffsetChecker$$processPartition(ConsumerOffsetChecker.scala:59) at kafka.tools.ConsumerOffsetChecker$$anonfun$kafka$tools$ConsumerOffsetChecker$$processTopic$1.apply$mcVI$sp(ConsumerOffsetChecker.scala:89) at kafka.tools.ConsumerOffsetChecker$$anonfun$kafka$tools$ConsumerOffsetChecker$$processTopic$1.apply(ConsumerOffsetChecker.scala:89) at kafka.tools.ConsumerOffsetChecker$$anonfun$kafka$tools$ConsumerOffsetChecker$$processTopic$1.apply(ConsumerOffsetChecker.scala:89) at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59) at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47) at kafka.tools.ConsumerOffsetChecker$.kafka$tools$ConsumerOffsetChecker$$processTopic(ConsumerOffsetChecker.scala:88) at kafka.tools.ConsumerOffsetChecker$$anonfun$main$3.apply(ConsumerOffsetChecker.scala:153) at kafka.tools.ConsumerOffsetChecker$$anonfun$main$3.apply(ConsumerOffsetChecker.scala:153) at scala.collection.immutable.List.foreach(List.scala:318) at kafka.tools.ConsumerOffsetChecker$.main(ConsumerOffsetChecker.scala:152) at kafka.tools.ConsumerOffsetChecker.main(ConsumerOffsetChecker.scala) Caused by: org.apache.zookeeper.KeeperException$NoNodeException: KeeperErrorCode = NoNode for /consumers/web-console-consumer-48744/offsets/PofApiTest-1/0 at org.apache.zookeeper.KeeperException.create(KeeperException.java:102) at org.apache.zookeeper.KeeperException.create(KeeperException.java:42) at org.apache.zookeeper.ZooKeeper.getData(ZooKeeper.java:927) at org.apache.zookeeper.ZooKeeper.getData(ZooKeeper.java:956) at org.I0Itec.zkclient.ZkConnection.readData(ZkConnection.java:103) at org.I0Itec.zkclient.ZkClient$9.call(ZkClient.java:770) at org.I0Itec.zkclient.ZkClient$9.call(ZkClient.java:766) at org.I0Itec.zkclient.ZkClient.retryUntilConnected(ZkClient.java:675) ... 15 more So consumer groups make confusing, I didn't specify the consumer-group-id in producer, the only place I know to config group is consumer.properties #consumer group id group.id=test-consumer-group Any hints? Thanks AL On Mon, Dec 15, 2014 at 6:46 PM, nitin sharma wrote: > got it ... thanks a lot. > > Regards, > Nitin Kumar Sharma. > > > On Mon, Dec 15, 2014 at 9:26 PM, Gwen Shapira > wrote: > > > > Hi Nitin, > > > > Go to where you installed zookeeper and run: > > > > bin/zkCli.sh -server 127.0.0.1:2181 > > > > On Mon, Dec 15, 2014 at 6:09 PM, nitin sharma > > wrote: > > > Thanks Neha and Gwen for your responses.. > > > > > > @Gwen -- Kindly explain how to perform
Re: Number of Consumers Connected
got it ... thanks a lot. Regards, Nitin Kumar Sharma. On Mon, Dec 15, 2014 at 9:26 PM, Gwen Shapira wrote: > > Hi Nitin, > > Go to where you installed zookeeper and run: > > bin/zkCli.sh -server 127.0.0.1:2181 > > On Mon, Dec 15, 2014 at 6:09 PM, nitin sharma > wrote: > > Thanks Neha and Gwen for your responses.. > > > > @Gwen -- Kindly explain how to perform the steps you have mentioned. how > > should i connect to a zookeeper..? > > > > Regards, > > Nitin Kumar Sharma. > > > > > > On Mon, Dec 15, 2014 at 6:36 PM, Neha Narkhede > wrote: > >> > >> In addition to Gwen's suggestion, we actually don't have jmx metrics > that > >> give you a list of actively consuming processes. > >> > >> On Mon, Dec 15, 2014 at 12:59 PM, Gwen Shapira > >> wrote: > >> > > >> > Currently you can find the number of consumer groups through > ZooKeeper: > >> > > >> > connect to ZK and run > >> > ls /consumers > >> > > >> > and count the number of results > >> > > >> > On Mon, Dec 15, 2014 at 11:34 AM, nitin sharma > >> > wrote: > >> > > Hi Team, > >> > > > >> > > Is it possible to know how many Consumer Group connected to kafka > >> broker > >> > Ids > >> > > and as well as how many Instances within a Group are fetching > messages > >> > from > >> > > Kafka Brokers > >> > > > >> > > Regards, > >> > > Nitin Kumar Sharma. > >> > > >> > >> > >> -- > >> Thanks, > >> Neha > >> >
Re: Number of Consumers Connected
Hi Nitin, Go to where you installed zookeeper and run: bin/zkCli.sh -server 127.0.0.1:2181 On Mon, Dec 15, 2014 at 6:09 PM, nitin sharma wrote: > Thanks Neha and Gwen for your responses.. > > @Gwen -- Kindly explain how to perform the steps you have mentioned. how > should i connect to a zookeeper..? > > Regards, > Nitin Kumar Sharma. > > > On Mon, Dec 15, 2014 at 6:36 PM, Neha Narkhede wrote: >> >> In addition to Gwen's suggestion, we actually don't have jmx metrics that >> give you a list of actively consuming processes. >> >> On Mon, Dec 15, 2014 at 12:59 PM, Gwen Shapira >> wrote: >> > >> > Currently you can find the number of consumer groups through ZooKeeper: >> > >> > connect to ZK and run >> > ls /consumers >> > >> > and count the number of results >> > >> > On Mon, Dec 15, 2014 at 11:34 AM, nitin sharma >> > wrote: >> > > Hi Team, >> > > >> > > Is it possible to know how many Consumer Group connected to kafka >> broker >> > Ids >> > > and as well as how many Instances within a Group are fetching messages >> > from >> > > Kafka Brokers >> > > >> > > Regards, >> > > Nitin Kumar Sharma. >> > >> >> >> -- >> Thanks, >> Neha >>
Re: Number of Consumers Connected
Thanks Neha and Gwen for your responses.. @Gwen -- Kindly explain how to perform the steps you have mentioned. how should i connect to a zookeeper..? Regards, Nitin Kumar Sharma. On Mon, Dec 15, 2014 at 6:36 PM, Neha Narkhede wrote: > > In addition to Gwen's suggestion, we actually don't have jmx metrics that > give you a list of actively consuming processes. > > On Mon, Dec 15, 2014 at 12:59 PM, Gwen Shapira > wrote: > > > > Currently you can find the number of consumer groups through ZooKeeper: > > > > connect to ZK and run > > ls /consumers > > > > and count the number of results > > > > On Mon, Dec 15, 2014 at 11:34 AM, nitin sharma > > wrote: > > > Hi Team, > > > > > > Is it possible to know how many Consumer Group connected to kafka > broker > > Ids > > > and as well as how many Instances within a Group are fetching messages > > from > > > Kafka Brokers > > > > > > Regards, > > > Nitin Kumar Sharma. > > > > > -- > Thanks, > Neha >
Re: Number of Consumers Connected
In addition to Gwen's suggestion, we actually don't have jmx metrics that give you a list of actively consuming processes. On Mon, Dec 15, 2014 at 12:59 PM, Gwen Shapira wrote: > > Currently you can find the number of consumer groups through ZooKeeper: > > connect to ZK and run > ls /consumers > > and count the number of results > > On Mon, Dec 15, 2014 at 11:34 AM, nitin sharma > wrote: > > Hi Team, > > > > Is it possible to know how many Consumer Group connected to kafka broker > Ids > > and as well as how many Instances within a Group are fetching messages > from > > Kafka Brokers > > > > Regards, > > Nitin Kumar Sharma. > -- Thanks, Neha
Re: Number of Consumers Connected
Currently you can find the number of consumer groups through ZooKeeper: connect to ZK and run ls /consumers and count the number of results On Mon, Dec 15, 2014 at 11:34 AM, nitin sharma wrote: > Hi Team, > > Is it possible to know how many Consumer Group connected to kafka broker Ids > and as well as how many Instances within a Group are fetching messages from > Kafka Brokers > > Regards, > Nitin Kumar Sharma.