Re: using the new logstash-kafka plugin
Scott you will have to do just one topic per input right now but multiple topics per group, whitelisting and blacklisting just got merged into jruby-kafka and I'm working them up the chain to my logstash-kafka and then pass it to the logstash-input/output/-kafka plugin. From: Scott Chapman sc...@woofplanet.com Sent: Wednesday, January 21, 2015 8:32 PM To: users@kafka.apache.org Subject: using the new logstash-kafka plugin We are starting to use the new logstash-kafka plugin, and I am wondering if it is possible to read multiple topics? Or do you need to create separate logstashes for each topic to parse? We are consuming multi-line logs from a service running on a bunch of different hosts, so we address that by creating single partition topics for our producers. We then want to have logstash consume them for ELK. Thanks in advance!
delete log file
Hi all, The kafka broker's disk is almost full, and the log file is still opened by kafka through lsof command. I know to change log.retention.hours in server.properties, but I don't want to restart the kafka server. Is there any utility to delete log file without impacting kafka? Thanks Regards YuanJia Li
Re: clarification of the per-topic retention.bytes setting
Hi David, The per-topic configs will just override the global configs for that specific topic; for the retention.bytes config it will be applied to all partitions of that topic. So if you have two topics each with two partitions and replication factor 1 with retention.bytes valued A then the total limit will be 2 (topics) * 2 (partitions per topic) * A (bytes) And if you specify per-topic-retention for one of the topics to value B the total limit will be changed to 2 (partitions per topic) * A (bytes) + 2 (partitions per topic) * B (bytes) Guozhang On Thu, Jan 22, 2015 at 4:38 AM, David Watzke wat...@avast.com wrote: Hi list, please help me understand the per-topic retention.* setting (in kafka 0.8.1.1) done by: bin/kafka-topics.sh --zookeeper $ZK --alter --topic $TOPIC --config retention.bytes=VALUE I understand from this: http://search-hadoop.com/m/4TaT4E9f78/retention.bytes; subj=estimating+log+retention+bytes that log.retention.bytes limit applies to a single partition, so basically each partition directory (in all directories listed in log.dirs) can take up at most log.retention.bytes bytes. But what happens if I increase this limit for a single topic? Is this limit per-that-topic's-partition? So the overall topic size limit would be LIMIT = retention.bytes * TOPIC'S_PARTITION_COUNT and it could take up to LIMIT * TOPIC'S_REPLICATION_FACTOR of disk space? Or is this setting per-topic, meaning that retention.bytes property sets the upper overall topic size limit directly? Thanks in advance! -- David Watzke -- -- Guozhang
Re: How to setup inter DC replication in Kafka 0.8.1.1
For now (0.8.2) if you have a hard crash then you will likely have duplicates or even data loss unfortunately, I do not have a good solution to this on top of my head. On Wed, Jan 21, 2015 at 10:14 PM, Madhukar Bharti bhartimadhu...@gmail.com wrote: Thanks Guozhang for your reply. Checked the details as mentioned above. Issue KAFKA-1650 has been fixed but it will come with 0.8.3. :-( By MM Killed, it is hard-failed(kill -9 )/crashed due to system reboot or any other reason. In that case some of the messages get duplicated. Can you please suggest a way to do inter DC replication for time being without loss/duplication of messages? On Thu, Jan 22, 2015 at 6:39 AM, Guozhang Wang wangg...@gmail.com wrote: Hi Madhukar, We are currently working on improving MM to avoid data loss / duplicates, details can be found in KAFKA-1650 / 1839 / 1840. For your concern, by saying MM killed do you mean it is shutdown or hard-failed / crashed that can cause duplicates? Guozhang On Mon, Jan 19, 2015 at 10:51 PM, Madhukar Bharti bhartimadhu...@gmail.com wrote: Hi, I want to setup inter DC replication b/w Kafka clusters. Is there any inbuilt tool to do this? I already have tried MirrorMaker tool but the problem is, if MM killed then some messages get duplicated. I don't want to duplicate the messages. Please suggest a way to do this. Please share your experience how you have setup this. Thanks in Advance, Madhukar -- -- Guozhang -- Thanks and Regards, Madhukar Bharti -- -- Guozhang
clarification of the per-topic retention.bytes setting
Hi list, please help me understand the per-topic retention.* setting (in kafka 0.8.1.1) done by: bin/kafka-topics.sh --zookeeper $ZK --alter --topic $TOPIC --config retention.bytes=VALUE I understand from this: http://search-hadoop.com/m/4TaT4E9f78/retention.bytessubj=estimating+log+retention+bytes that log.retention.bytes limit applies to a single partition, so basically each partition directory (in all directories listed in log.dirs) can take up at most log.retention.bytes bytes. But what happens if I increase this limit for a single topic? Is this limit per-that-topic's-partition? So the overall topic size limit would be LIMIT = retention.bytes * TOPIC'S_PARTITION_COUNT and it could take up to LIMIT * TOPIC'S_REPLICATION_FACTOR of disk space? Or is this setting per-topic, meaning that retention.bytes property sets the upper overall topic size limit directly? Thanks in advance! -- David Watzke
Help: Kafka LeaderNotAvailableException
Hi, Let me overview on the issue that I am facing on producing message in Kafka: I have horthonworks HDP-2.1 installed, along that we have Kafka on other node. * On kafka node: Start Zookeepeer Start Kafka Broker service Send message/producer Consume message - Works (Note: here we start Zookeeper locally on kafka01 node) * Issue side: Now in HDP-2.1 we have Zookeeper inbuild we have Zookeeper service running on master node I go to Kafka server Started Kafka Broker (In config\server.properties file I have added zookeeper.connect with maasternode:2181) Then I start producer Send message... after that we got error like kafka.common.LeaderNotAvailableException [2015-01-17 05:54:09,465] WARN Error while fetching metadata [{TopicMetadata for topic fred - No partition metadata for topic fred due to kafka.common.LeaderNotAvailableException}] for topic [fred]: class kafka.common.LeaderNotAvailableException (kafka.producer.BrokerPartitionInfo) [2015-01-17 05:54:09,659] WARN Error while fetching metadata [{TopicMetadata for topic fred - No partition metadata for topic fred due to kafka.common.LeaderNotAvailableException}] for topic [fred]: class kafka.common.LeaderNotAvailableException (kafka.producer.BrokerPartitionInfo) [2015-01-17 05:54:09,659] ERROR Failed to collate messages by topic, partition due to: Failed to fetch topic metadata for topic: fred (kafka.producer.async.DefaultEventHandler) [2015-01-17 05:54:09,802] WARN Error while fetching metadata [{TopicMetadata for topic fred - No partition metadata for topic fred due to kafka.common.LeaderNotAvailableException}] for topic [fred]: class kafka.common.LeaderNotAvailableException (kafka.producer.BrokerPartitionInfo) [2015-01-17 05:54:09,820] ERROR Failed to send requests for topics fred with correlation ids in [0,8] (kafka.producer.async.DefaultEventHandler) [2015-01-17 05:54:09,822] ERROR Error in handling batch of 1 events (kafka.producer.async.ProducerSendThread) kafka.common.FailedToSendMessageException: Failed to send messages after 3 tries. at kafka.producer.async.DefaultEventHandler.handle(DefaultEventHandler.scala:90) at kafka.producer.async.ProducerSendThread.tryToHandle(ProducerSendThread.scala:104) at kafka.producer.async.ProducerSendThread$$anonfun$processEvents$3.apply(ProducerSendThread.scala:87) Can someone suggest what is going wrong. Thanks. -Vishal Regards, Vishal Software Dev Staff Engineer Dell | Bangalore Ext : 79268
Re: using the new logstash-kafka plugin
Hey Joe, with other input types (like file) one can reference things like the path in the filter section. Is it possible to refer to the topic_id in the filter section? I tried and nothing obvious worked. We are encoding a few things (like host name, and type) in the name of the topic, and would like to grok those values out. Let me know if anything comes to mind. Thanks! -Scott On Thu Jan 22 2015 at 10:00:21 AM Joseph Lawson jlaw...@roomkey.com wrote: Just trying to get everything in prior to the 1.5 release. From: Scott Chapman sc...@woofplanet.com Sent: Thursday, January 22, 2015 9:32 AM To: users@kafka.apache.org Subject: Re: using the new logstash-kafka plugin Awesome, what release are you targeting? Or are you able to make updates to the plugin outside of kafka? On Thu Jan 22 2015 at 9:31:26 AM Joseph Lawson jlaw...@roomkey.com wrote: Scott you will have to do just one topic per input right now but multiple topics per group, whitelisting and blacklisting just got merged into jruby-kafka and I'm working them up the chain to my logstash-kafka and then pass it to the logstash-input/output/-kafka plugin. From: Scott Chapman sc...@woofplanet.com Sent: Wednesday, January 21, 2015 8:32 PM To: users@kafka.apache.org Subject: using the new logstash-kafka plugin We are starting to use the new logstash-kafka plugin, and I am wondering if it is possible to read multiple topics? Or do you need to create separate logstashes for each topic to parse? We are consuming multi-line logs from a service running on a bunch of different hosts, so we address that by creating single partition topics for our producers. We then want to have logstash consume them for ELK. Thanks in advance!
Re: using the new logstash-kafka plugin
Just trying to get everything in prior to the 1.5 release. From: Scott Chapman sc...@woofplanet.com Sent: Thursday, January 22, 2015 9:32 AM To: users@kafka.apache.org Subject: Re: using the new logstash-kafka plugin Awesome, what release are you targeting? Or are you able to make updates to the plugin outside of kafka? On Thu Jan 22 2015 at 9:31:26 AM Joseph Lawson jlaw...@roomkey.com wrote: Scott you will have to do just one topic per input right now but multiple topics per group, whitelisting and blacklisting just got merged into jruby-kafka and I'm working them up the chain to my logstash-kafka and then pass it to the logstash-input/output/-kafka plugin. From: Scott Chapman sc...@woofplanet.com Sent: Wednesday, January 21, 2015 8:32 PM To: users@kafka.apache.org Subject: using the new logstash-kafka plugin We are starting to use the new logstash-kafka plugin, and I am wondering if it is possible to read multiple topics? Or do you need to create separate logstashes for each topic to parse? We are consuming multi-line logs from a service running on a bunch of different hosts, so we address that by creating single partition topics for our producers. We then want to have logstash consume them for ELK. Thanks in advance!
Re: kafka production server test
Hi, Guozhang Can I run this package remotely test another server? which mean I run this package on dev but testing kafka system on production? thanks AL On Thu, Jan 22, 2015 at 2:55 PM, Sa Li sal...@gmail.com wrote: Hi, Guozhang, Good to know such package, will try it now. :-) thanks On Thu, Jan 22, 2015 at 2:40 PM, Guozhang Wang wangg...@gmail.com wrote: Hi Sa, Have you looked into the system test package? It contains a suite of tests on different failure modes of Kafka brokers. Guozhang On Thu, Jan 22, 2015 at 12:00 PM, Sa Li sal...@gmail.com wrote: Hi, All We are about to deliver kafka production server, I have been working on different test, like performance test from linkedin. This is a 3-node cluster, with 5 nodes zkEnsemble. I assume there are lots of tests I need to do, like network, node failure, flush time, etc. Is there is completed guide to instruct the tests for kafka production servers? thanks -- Alec Li -- -- Guozhang -- Alec Li -- Alec Li
Re: warning on startup of consumer app with 0.8.2 rc2
Hmm, kafka-console-consumer in 0.8.2 rc2 is running fine. Do you have multiple kafka jars in your classpath? Thanks, Jun On Thu, Jan 22, 2015 at 4:58 PM, Jason Rosenberg j...@squareup.com wrote: 2015-01-23 00:55:25,273 WARN [async-message-sender-0] common.AppInfo$ - Can't read Kafka version from MANIFEST.MF. Possible cause: java.lang.NullPointerException
Re: Partition reassignment doesn't work on few partitions
Perhaps you can upgrade all brokers and then try? Thanks, Jun On Wed, Jan 21, 2015 at 9:53 PM, Raghu Udiyar ra...@helpshift.com wrote: No errors in the state-change log or the controller. Its as if the controller never got the request for that partition. Regarding the upgrade, we did upgrade one of the nodes, and initiate the replication. Here, the controller at 0.8.0 and this node at 0.8.1.1. In this case, when we initiated the reassignment, the following error was logged on the destination broker : [2015-01-21 13:36:18,101] WARN Broker 7 ignoring LeaderAndIsr request with correlation id 837 from controller 1 epoch 44 as broker is not in assigned replica list 5,6 for partition [test-topic,0] (state.change.logger) 5,6 is from where we were moving to topic to 7. 7 is the new broker at 0.8.1.1 I'm guessing the controller is in a weird way. -- Raghu On Thu, Jan 22, 2015 at 5:57 AM, Jun Rao j...@confluent.io wrote: Any error in the controller and state-change log? Also, you may want to upgrade to 0.8.1, which fixed some reassignment issues. Thanks, Jun On Wed, Jan 21, 2015 at 12:38 PM, Raghu Udiyar ra...@helpshift.com wrote: Hello, I have a 6 node kafka cluster (0.8.0) where partition reassignment doesn't seem to work on a few partitions. This happens within the same, as well as across other topics. Following is the behavior observed : 1. For a successful reassignment, the kafka-reassign-partitions.sh returns success, I see the controller initiating the reassignment, and the destination brokers start replica fetcher threads. 2. For the unsuccessful reassignment, the tool returns success, but there is nothing in the controller logs nor the destination brokers. Also, for the ones that are successful, some don’t finish replication correctly. I can see that the destination brokers get stuck after a few thousand offsets (checked in JMX), and doesn’t move after that. The controller keeps on waiting for the fetchers to complete, but never gets there. Anyone seen this issue before? Is there a way to reset the state of the controller? or re-elect a new one? Thanks, Raghu
Re: delete log file
In general, you shouldn't delete any files while the broker is up. If you have no other choices, you can try deleting the older log segments and hope that no consumer or the log cleaner are using them. Thanks, Jun On Thu, Jan 22, 2015 at 12:40 AM, YuanJia Li yuanjia8...@163.com wrote: Hi all, The kafka broker's disk is almost full, and the log file is still opened by kafka through lsof command. I know to change log.retention.hours in server.properties, but I don't want to restart the kafka server. Is there any utility to delete log file without impacting kafka? Thanks Regards YuanJia Li
Re: Help: Kafka LeaderNotAvailableException
Any error in the controller and the broker log? Thanks, Jun On Thu, Jan 22, 2015 at 1:33 AM, vishal_ja...@dell.com wrote: Hi, Let me overview on the issue that I am facing on producing message in Kafka: I have horthonworks HDP-2.1 installed, along that we have Kafka on other node. * On kafka node: Start Zookeepeer Start Kafka Broker service Send message/producer Consume message - Works (Note: here we start Zookeeper locally on kafka01 node) * Issue side: Now in HDP-2.1 we have Zookeeper inbuild we have Zookeeper service running on master node I go to Kafka server Started Kafka Broker (In config\server.properties file I have added zookeeper.connect with maasternode:2181) Then I start producer Send message... after that we got error like kafka.common.LeaderNotAvailableException [2015-01-17 05:54:09,465] WARN Error while fetching metadata [{TopicMetadata for topic fred - No partition metadata for topic fred due to kafka.common.LeaderNotAvailableException}] for topic [fred]: class kafka.common.LeaderNotAvailableException (kafka.producer.BrokerPartitionInfo) [2015-01-17 05:54:09,659] WARN Error while fetching metadata [{TopicMetadata for topic fred - No partition metadata for topic fred due to kafka.common.LeaderNotAvailableException}] for topic [fred]: class kafka.common.LeaderNotAvailableException (kafka.producer.BrokerPartitionInfo) [2015-01-17 05:54:09,659] ERROR Failed to collate messages by topic, partition due to: Failed to fetch topic metadata for topic: fred (kafka.producer.async.DefaultEventHandler) [2015-01-17 05:54:09,802] WARN Error while fetching metadata [{TopicMetadata for topic fred - No partition metadata for topic fred due to kafka.common.LeaderNotAvailableException}] for topic [fred]: class kafka.common.LeaderNotAvailableException (kafka.producer.BrokerPartitionInfo) [2015-01-17 05:54:09,820] ERROR Failed to send requests for topics fred with correlation ids in [0,8] (kafka.producer.async.DefaultEventHandler) [2015-01-17 05:54:09,822] ERROR Error in handling batch of 1 events (kafka.producer.async.ProducerSendThread) kafka.common.FailedToSendMessageException: Failed to send messages after 3 tries. at kafka.producer.async.DefaultEventHandler.handle(DefaultEventHandler.scala:90) at kafka.producer.async.ProducerSendThread.tryToHandle(ProducerSendThread.scala:104) at kafka.producer.async.ProducerSendThread$$anonfun$processEvents$3.apply(ProducerSendThread.scala:87) Can someone suggest what is going wrong. Thanks. -Vishal Regards, Vishal Software Dev Staff Engineer Dell | Bangalore Ext : 79268
Re: Help: Kafka LeaderNotAvailableException
It sounds like you have two zookeepers, one for HDP and one for Kafka. Did you move Kafka from one zookeeper to another? Perhaps Kafka finds the topics (logs) on disk, but they do not exist in ZK because you are using a different zookeeper now. Gwen On Thu, Jan 22, 2015 at 6:38 PM, Jun Rao j...@confluent.io wrote: Any error in the controller and the broker log? Thanks, Jun On Thu, Jan 22, 2015 at 1:33 AM, vishal_ja...@dell.com wrote: Hi, Let me overview on the issue that I am facing on producing message in Kafka: I have horthonworks HDP-2.1 installed, along that we have Kafka on other node. * On kafka node: Start Zookeepeer Start Kafka Broker service Send message/producer Consume message - Works (Note: here we start Zookeeper locally on kafka01 node) * Issue side: Now in HDP-2.1 we have Zookeeper inbuild we have Zookeeper service running on master node I go to Kafka server Started Kafka Broker (In config\server.properties file I have added zookeeper.connect with maasternode:2181) Then I start producer Send message... after that we got error like kafka.common.LeaderNotAvailableException [2015-01-17 05:54:09,465] WARN Error while fetching metadata [{TopicMetadata for topic fred - No partition metadata for topic fred due to kafka.common.LeaderNotAvailableException}] for topic [fred]: class kafka.common.LeaderNotAvailableException (kafka.producer.BrokerPartitionInfo) [2015-01-17 05:54:09,659] WARN Error while fetching metadata [{TopicMetadata for topic fred - No partition metadata for topic fred due to kafka.common.LeaderNotAvailableException}] for topic [fred]: class kafka.common.LeaderNotAvailableException (kafka.producer.BrokerPartitionInfo) [2015-01-17 05:54:09,659] ERROR Failed to collate messages by topic, partition due to: Failed to fetch topic metadata for topic: fred (kafka.producer.async.DefaultEventHandler) [2015-01-17 05:54:09,802] WARN Error while fetching metadata [{TopicMetadata for topic fred - No partition metadata for topic fred due to kafka.common.LeaderNotAvailableException}] for topic [fred]: class kafka.common.LeaderNotAvailableException (kafka.producer.BrokerPartitionInfo) [2015-01-17 05:54:09,820] ERROR Failed to send requests for topics fred with correlation ids in [0,8] (kafka.producer.async.DefaultEventHandler) [2015-01-17 05:54:09,822] ERROR Error in handling batch of 1 events (kafka.producer.async.ProducerSendThread) kafka.common.FailedToSendMessageException: Failed to send messages after 3 tries. at kafka.producer.async.DefaultEventHandler.handle(DefaultEventHandler.scala:90) at kafka.producer.async.ProducerSendThread.tryToHandle(ProducerSendThread.scala:104) at kafka.producer.async.ProducerSendThread$$anonfun$processEvents$3.apply(ProducerSendThread.scala:87) Can someone suggest what is going wrong. Thanks. -Vishal Regards, Vishal Software Dev Staff Engineer Dell | Bangalore Ext : 79268
Re: kafka production server test
Hi, Guozhang, Good to know such package, will try it now. :-) thanks On Thu, Jan 22, 2015 at 2:40 PM, Guozhang Wang wangg...@gmail.com wrote: Hi Sa, Have you looked into the system test package? It contains a suite of tests on different failure modes of Kafka brokers. Guozhang On Thu, Jan 22, 2015 at 12:00 PM, Sa Li sal...@gmail.com wrote: Hi, All We are about to deliver kafka production server, I have been working on different test, like performance test from linkedin. This is a 3-node cluster, with 5 nodes zkEnsemble. I assume there are lots of tests I need to do, like network, node failure, flush time, etc. Is there is completed guide to instruct the tests for kafka production servers? thanks -- Alec Li -- -- Guozhang -- Alec Li
Re: kafka production server test
Hi Sa, Have you looked into the system test package? It contains a suite of tests on different failure modes of Kafka brokers. Guozhang On Thu, Jan 22, 2015 at 12:00 PM, Sa Li sal...@gmail.com wrote: Hi, All We are about to deliver kafka production server, I have been working on different test, like performance test from linkedin. This is a 3-node cluster, with 5 nodes zkEnsemble. I assume there are lots of tests I need to do, like network, node failure, flush time, etc. Is there is completed guide to instruct the tests for kafka production servers? thanks -- Alec Li -- -- Guozhang
warning on startup of consumer app with 0.8.2 rc2
2015-01-23 00:55:25,273 WARN [async-message-sender-0] common.AppInfo$ - Can't read Kafka version from MANIFEST.MF. Possible cause: java.lang.NullPointerException
Re: Re: delete log file
Hi Yang, I think my problem is not the same with yours. My production environment is fine, log.retention.hours is 720. My disk is almost full just beacause of too much data. I want to have a utility or command to delete manually, instead of waiting until it expires. Thanks for your reply. Thanks Regards YuanJia Li From: Yang Fang Date: 2015-01-23 12:15 To: users Subject: Re: delete log file YuanJia, I want to know why your broker's disk is almost full . Maybe your issue is the same with mine. Is it one broker out of service for a long time, then it fails back, at first it looks good, but after some hours, its disk is almost full. Other brokers are all fine, avg remaining capacity is about 25%. In my production environment, log.retention.hours is 24, broker's data volume is about 2TB. The reason it that when broker decide to delete logs is based on its start time and log.retention.hours. On Thu, Jan 22, 2015 at 4:40 PM, YuanJia Li yuanjia8...@163.com wrote: Hi all, The kafka broker's disk is almost full, and the log file is still opened by kafka through lsof command. I know to change log.retention.hours in server.properties, but I don't want to restart the kafka server. Is there any utility to delete log file without impacting kafka? Thanks Regards YuanJia Li
kafka topic not exist after creation with api
Hi all , My application creates kafka topic at runtime with AdminUtils.createTopic but that topic is not available when I try to produce to it. If I run bin/kafka-console-consumer.sh --topic $topic --zookeeper $zkStr --from-beginning it would throw UnknownTopicOrPartitionException. Also, I found the following error in controller log. [2015-01-23 12:44:41,825] WARN [Controller-0-to-broker-0-send-thread], Controller 0 fails to send a request to broker id:0,host:node10,port:9092 (kafka.controller.RequestSendThread) java.io.EOFException: Received -1 when reading from channel, socket has likely been closed. at kafka.utils.Utils$.read(Utils.scala:381) at kafka.network.BoundedByteBufferReceive.readFrom(BoundedByteBufferReceive.scala:54) at kafka.network.Receive$class.readCompletely(Transmission.scala:56) at kafka.network.BoundedByteBufferReceive.readCompletely(BoundedByteBufferReceive.scala:29) at kafka.network.BlockingChannel.receive(BlockingChannel.scala:108) at kafka.controller.RequestSendThread.doWork(ControllerChannelManager.scala:146) at kafka.utils.ShutdownableThread.run(ShutdownableThread.scala:60) [2015-01-23 12:44:41,827] ERROR [Controller-0-to-broker-0-send-thread], Controller 0 epoch 3 failed to send request Name:UpdateMetadataRequest;Version:0;Controller:0;ControllerEpoch:3;CorrelationId:80;ClientId:id_0-host_null-port_9092;AliveBrokers:id:0,host:node10,port:9092;PartitionState:[app1_test_0_testB,0] - (LeaderAndIsrInfo:(Leader:0,ISR:0,LeaderEpoch:0,ControllerEpoch:3),ReplicationFactor:1),AllReplicas:0) to broker id:0,host:node10,port:9092. Reconnecting to broker. (kafka.controller.RequestSendThread) java.nio.channels.ClosedChannelException at kafka.network.BlockingChannel.send(BlockingChannel.scala:97) at kafka.controller.RequestSendThread.liftedTree1$1(ControllerChannelManager.scala:132) at kafka.controller.RequestSendThread.doWork(ControllerChannelManager.scala:131) at kafka.utils.ShutdownableThread.run(ShutdownableThread.scala:60) I've tried fetching metadata after creating topic to ensure its availability but the fetch failed after several round of retries. My kafka is built from 0.8.2 trunk running on local node. Any thoughts would be appreciated Thanks, Manu
New mirror maker consumer.config question
Hi, I discovered that the new mirror maker implementation in trunk now only accept one consumer.config property instead of a list of them which means we can only supply one source per mirror maker process. Is it a reason for it? If I have multiple source kafka clusters do I need to setup multiple mirror maker processes? -- Regards, Tao
Apache Spark broadcast error: Error sending message as driverActor is null [message = UpdateBlockInfo(BlockManagerId
HiI'm using Apache Spark 1.1.0 and I'm currently having issue with broadcast method. So when I call broadcast function on a small dataset to a 5 nodes cluster, I experiencing the Error sending message as driverActor is null after broadcast the variables several times (apps running under jboss). Any help would be appreciate.ThanksEdwin
Re: Apache Spark broadcast error: Error sending message as driverActor is null [message = UpdateBlockInfo(BlockManagerId
Yep, sorry, had a rough day.. On Thursday, January 22, 2015 2:25 PM, Guozhang Wang wangg...@gmail.com wrote: Hi Zijing, Sounds like you sent to the wrong mailing list :P Guozhang On Thu, Jan 22, 2015 at 11:12 AM, Zijing Guo alter...@yahoo.com.invalid wrote: HiI'm using Apache Spark 1.1.0 and I'm currently having issue with broadcast method. So when I call broadcast function on a small dataset to a 5 nodes cluster, I experiencing the Error sending message as driverActor is null after broadcast the variables several times (apps running under jboss). Any help would be appreciate.ThanksEdwin -- -- Guozhang
Re: Apache Spark broadcast error: Error sending message as driverActor is null [message = UpdateBlockInfo(BlockManagerId
Hi Zijing, Sounds like you sent to the wrong mailing list :P Guozhang On Thu, Jan 22, 2015 at 11:12 AM, Zijing Guo alter...@yahoo.com.invalid wrote: HiI'm using Apache Spark 1.1.0 and I'm currently having issue with broadcast method. So when I call broadcast function on a small dataset to a 5 nodes cluster, I experiencing the Error sending message as driverActor is null after broadcast the variables several times (apps running under jboss). Any help would be appreciate.ThanksEdwin -- -- Guozhang
kafka production server test
Hi, All We are about to deliver kafka production server, I have been working on different test, like performance test from linkedin. This is a 3-node cluster, with 5 nodes zkEnsemble. I assume there are lots of tests I need to do, like network, node failure, flush time, etc. Is there is completed guide to instruct the tests for kafka production servers? thanks -- Alec Li
Re: Help: Kafka LeaderNotAvailableException
Vishal, Does this error happen every time you are sending? Or just the first time? Joe Stein On Thu, Jan 22, 2015 at 4:33 AM, vishal_ja...@dell.com wrote: Hi, Let me overview on the issue that I am facing on producing message in Kafka: I have horthonworks HDP-2.1 installed, along that we have Kafka on other node. * On kafka node: Start Zookeepeer Start Kafka Broker service Send message/producer Consume message - Works (Note: here we start Zookeeper locally on kafka01 node) * Issue side: Now in HDP-2.1 we have Zookeeper inbuild we have Zookeeper service running on master node I go to Kafka server Started Kafka Broker (In config\server.properties file I have added zookeeper.connect with maasternode:2181) Then I start producer Send message... after that we got error like kafka.common.LeaderNotAvailableException [2015-01-17 05:54:09,465] WARN Error while fetching metadata [{TopicMetadata for topic fred - No partition metadata for topic fred due to kafka.common.LeaderNotAvailableException}] for topic [fred]: class kafka.common.LeaderNotAvailableException (kafka.producer.BrokerPartitionInfo) [2015-01-17 05:54:09,659] WARN Error while fetching metadata [{TopicMetadata for topic fred - No partition metadata for topic fred due to kafka.common.LeaderNotAvailableException}] for topic [fred]: class kafka.common.LeaderNotAvailableException (kafka.producer.BrokerPartitionInfo) [2015-01-17 05:54:09,659] ERROR Failed to collate messages by topic, partition due to: Failed to fetch topic metadata for topic: fred (kafka.producer.async.DefaultEventHandler) [2015-01-17 05:54:09,802] WARN Error while fetching metadata [{TopicMetadata for topic fred - No partition metadata for topic fred due to kafka.common.LeaderNotAvailableException}] for topic [fred]: class kafka.common.LeaderNotAvailableException (kafka.producer.BrokerPartitionInfo) [2015-01-17 05:54:09,820] ERROR Failed to send requests for topics fred with correlation ids in [0,8] (kafka.producer.async.DefaultEventHandler) [2015-01-17 05:54:09,822] ERROR Error in handling batch of 1 events (kafka.producer.async.ProducerSendThread) kafka.common.FailedToSendMessageException: Failed to send messages after 3 tries. at kafka.producer.async.DefaultEventHandler.handle(DefaultEventHandler.scala:90) at kafka.producer.async.ProducerSendThread.tryToHandle(ProducerSendThread.scala:104) at kafka.producer.async.ProducerSendThread$$anonfun$processEvents$3.apply(ProducerSendThread.scala:87) Can someone suggest what is going wrong. Thanks. -Vishal Regards, Vishal Software Dev Staff Engineer Dell | Bangalore Ext : 79268
Counting # of Message Brokers Receive Per Minute Per Topic
Hi Kafka Team, I need to count message received by entire Kafka Broker Cluster for a particular topic. I have 3 brokers, so do I need to sum the COUNT metric or just one server count reflect all server count. It seems that count is always increasing (although metric name is *MessagesInPerSec* so count does not get reset) kafka.server:type=BrokerTopicMetrics,name=topic-MessagesInPerSec Please clarify. Thanks, Bhavesh
kafka sending duplicate content to consumer
Hi, i want know in which situation does kafka send same event multiple times to consumer. Is there a consumer side configuration to tell kafka to send only once and stop retries? -- Regards Tousif Khazi