Re: using the new logstash-kafka plugin

2015-01-22 Thread Joseph Lawson
Scott you will have to do just one topic per input right now but multiple 
topics per group, whitelisting and blacklisting just got merged into 
jruby-kafka and I'm working them up the chain to my logstash-kafka and then 
pass it to the logstash-input/output/-kafka plugin.


From: Scott Chapman sc...@woofplanet.com
Sent: Wednesday, January 21, 2015 8:32 PM
To: users@kafka.apache.org
Subject: using the new logstash-kafka plugin

We are starting to use the new logstash-kafka plugin, and I am wondering if
it is possible to read multiple topics? Or do you need to create separate
logstashes for each topic to parse?

We are consuming multi-line logs from a service running on a bunch of
different hosts, so we address that by creating single partition topics for
our producers.

We then want to have logstash consume them for ELK.

Thanks in advance!


delete log file

2015-01-22 Thread YuanJia Li
Hi all,
The kafka broker's disk is almost full, and the log file is still opened by 
kafka through lsof command.
I know to change log.retention.hours in server.properties, but I don't want to 
restart the kafka server.
Is there any utility to delete log file without impacting kafka?

Thanks  Regards
YuanJia Li

Re: clarification of the per-topic retention.bytes setting

2015-01-22 Thread Guozhang Wang
Hi David,

The per-topic configs will just override the global configs for that
specific topic; for the retention.bytes config it will be applied to all
partitions of that topic.

So if you have two topics each with two partitions and replication factor 1
with retention.bytes valued A then the total limit will be

2 (topics) * 2 (partitions per topic) * A (bytes)

And if you specify per-topic-retention for one of the topics to value B
the total limit will be changed to

2 (partitions per topic) * A (bytes) + 2 (partitions per topic) * B (bytes)

Guozhang

On Thu, Jan 22, 2015 at 4:38 AM, David Watzke wat...@avast.com wrote:

 Hi list,

 please help me understand the per-topic retention.* setting (in kafka
 0.8.1.1) done by:

 bin/kafka-topics.sh --zookeeper $ZK --alter --topic $TOPIC --config
 retention.bytes=VALUE

 I understand from this:
 http://search-hadoop.com/m/4TaT4E9f78/retention.bytes;
 subj=estimating+log+retention+bytes
 that log.retention.bytes limit applies to a single partition, so basically
 each partition directory (in all directories listed in log.dirs) can take
 up at most log.retention.bytes bytes.

 But what happens if I increase this limit for a single topic? Is this
 limit per-that-topic's-partition? So the overall topic size limit would be

 LIMIT = retention.bytes * TOPIC'S_PARTITION_COUNT

 and it could take up to LIMIT * TOPIC'S_REPLICATION_FACTOR of disk space?

 Or is this setting per-topic, meaning that retention.bytes property sets
 the upper overall topic size limit directly?

 Thanks in advance!

 --
 David Watzke




-- 
-- Guozhang


Re: How to setup inter DC replication in Kafka 0.8.1.1

2015-01-22 Thread Guozhang Wang
For now (0.8.2) if you have a hard crash then you will likely have
duplicates or even data loss unfortunately, I do not have a good solution
to this on top of my head.

On Wed, Jan 21, 2015 at 10:14 PM, Madhukar Bharti bhartimadhu...@gmail.com
wrote:

 Thanks Guozhang for your reply.

 Checked the details as mentioned above. Issue KAFKA-1650 has been fixed but
 it will come with 0.8.3. :-(

 By MM Killed, it is hard-failed(kill -9 )/crashed due to system reboot or
 any other reason. In that case some of the messages get duplicated.

 Can you please suggest a way to do inter DC replication for time being
 without loss/duplication of messages?



 On Thu, Jan 22, 2015 at 6:39 AM, Guozhang Wang wangg...@gmail.com wrote:

  Hi Madhukar,
 
  We are currently working on improving MM to avoid data loss / duplicates,
  details can be found in KAFKA-1650 / 1839 / 1840.
 
  For your concern, by saying MM killed do you mean it is shutdown or
  hard-failed / crashed that can cause duplicates?
 
  Guozhang
 
  On Mon, Jan 19, 2015 at 10:51 PM, Madhukar Bharti 
  bhartimadhu...@gmail.com
  wrote:
 
   Hi,
  
   I want to setup inter DC replication b/w Kafka clusters. Is there any
   inbuilt tool to do this?
  
   I already have tried MirrorMaker tool but the problem is, if MM killed
  then
   some messages get duplicated. I don't want to duplicate the messages.
  
   Please suggest a way to do this. Please share your experience how you
  have
   setup this.
  
  
   Thanks in Advance,
   Madhukar
  
 
 
 
  --
  -- Guozhang
 



 --
 Thanks and Regards,
 Madhukar Bharti




-- 
-- Guozhang


clarification of the per-topic retention.bytes setting

2015-01-22 Thread David Watzke

Hi list,

please help me understand the per-topic retention.* setting (in kafka 
0.8.1.1) done by:


bin/kafka-topics.sh --zookeeper $ZK --alter --topic $TOPIC --config 
retention.bytes=VALUE


I understand from this:
http://search-hadoop.com/m/4TaT4E9f78/retention.bytessubj=estimating+log+retention+bytes
that log.retention.bytes limit applies to a single partition, so 
basically each partition directory (in all directories listed in 
log.dirs) can take up at most log.retention.bytes bytes.


But what happens if I increase this limit for a single topic? Is this 
limit per-that-topic's-partition? So the overall topic size limit would be


LIMIT = retention.bytes * TOPIC'S_PARTITION_COUNT

and it could take up to LIMIT * TOPIC'S_REPLICATION_FACTOR of disk space?

Or is this setting per-topic, meaning that retention.bytes property 
sets the upper overall topic size limit directly?


Thanks in advance!

--
David Watzke



Help: Kafka LeaderNotAvailableException

2015-01-22 Thread Vishal_Jain2
Hi,
Let me overview on the issue that I am facing on producing message in Kafka:
I have horthonworks HDP-2.1 installed, along that we have Kafka on other node.

* On kafka node:
Start Zookeepeer
Start Kafka Broker service
Send message/producer
Consume message - Works (Note: here we start Zookeeper locally on kafka01 node)

* Issue side:
Now in HDP-2.1 we have Zookeeper inbuild  we have Zookeeper service running 
on master node
I go to Kafka server  Started Kafka Broker
(In config\server.properties file I have added zookeeper.connect with 
maasternode:2181)
Then I start producer  Send message... after that we got error like 
kafka.common.LeaderNotAvailableException

[2015-01-17 05:54:09,465] WARN Error while fetching metadata [{TopicMetadata 
for topic fred -
No partition metadata for topic fred due to 
kafka.common.LeaderNotAvailableException}] for topic [fred]: class 
kafka.common.LeaderNotAvailableException (kafka.producer.BrokerPartitionInfo)
[2015-01-17 05:54:09,659] WARN Error while fetching metadata [{TopicMetadata 
for topic fred -
No partition metadata for topic fred due to 
kafka.common.LeaderNotAvailableException}] for topic [fred]: class 
kafka.common.LeaderNotAvailableException (kafka.producer.BrokerPartitionInfo)
[2015-01-17 05:54:09,659] ERROR Failed to collate messages by topic, partition 
due to: Failed to fetch topic metadata for topic: fred 
(kafka.producer.async.DefaultEventHandler)
[2015-01-17 05:54:09,802] WARN Error while fetching metadata [{TopicMetadata 
for topic fred -
No partition metadata for topic fred due to 
kafka.common.LeaderNotAvailableException}] for topic [fred]: class 
kafka.common.LeaderNotAvailableException (kafka.producer.BrokerPartitionInfo)
[2015-01-17 05:54:09,820] ERROR Failed to send requests for topics fred with 
correlation ids in [0,8] (kafka.producer.async.DefaultEventHandler)
[2015-01-17 05:54:09,822] ERROR Error in handling batch of 1 events 
(kafka.producer.async.ProducerSendThread)
kafka.common.FailedToSendMessageException: Failed to send messages after 3 
tries.
at kafka.producer.async.DefaultEventHandler.handle(DefaultEventHandler.scala:90)
at 
kafka.producer.async.ProducerSendThread.tryToHandle(ProducerSendThread.scala:104)
at 
kafka.producer.async.ProducerSendThread$$anonfun$processEvents$3.apply(ProducerSendThread.scala:87)

Can someone suggest what is going wrong.
Thanks.





-Vishal


Regards,
Vishal
Software Dev Staff Engineer
Dell | Bangalore
Ext : 79268



Re: using the new logstash-kafka plugin

2015-01-22 Thread Scott Chapman
Hey Joe, with other input types (like file) one can reference things like
the path in the filter section.

Is it possible to refer to the topic_id in the filter section? I tried and
nothing obvious worked.

We are encoding a few things (like host name, and type) in the name of the
topic, and would like to grok those values out.

Let me know if anything comes to mind.

Thanks!

-Scott

On Thu Jan 22 2015 at 10:00:21 AM Joseph Lawson jlaw...@roomkey.com wrote:

 Just trying to get everything in prior to the 1.5 release.

 
 From: Scott Chapman sc...@woofplanet.com
 Sent: Thursday, January 22, 2015 9:32 AM
 To: users@kafka.apache.org
 Subject: Re: using the new logstash-kafka plugin

 Awesome, what release are you targeting? Or are you able to make updates to
 the plugin outside of kafka?

 On Thu Jan 22 2015 at 9:31:26 AM Joseph Lawson jlaw...@roomkey.com
 wrote:

  Scott you will have to do just one topic per input right now but multiple
  topics per group, whitelisting and blacklisting just got merged into
  jruby-kafka and I'm working them up the chain to my logstash-kafka and
 then
  pass it to the logstash-input/output/-kafka plugin.
 
  
  From: Scott Chapman sc...@woofplanet.com
  Sent: Wednesday, January 21, 2015 8:32 PM
  To: users@kafka.apache.org
  Subject: using the new logstash-kafka plugin
 
  We are starting to use the new logstash-kafka plugin, and I am wondering
 if
  it is possible to read multiple topics? Or do you need to create separate
  logstashes for each topic to parse?
 
  We are consuming multi-line logs from a service running on a bunch of
  different hosts, so we address that by creating single partition topics
 for
  our producers.
 
  We then want to have logstash consume them for ELK.
 
  Thanks in advance!
 



Re: using the new logstash-kafka plugin

2015-01-22 Thread Joseph Lawson
Just trying to get everything in prior to the 1.5 release.


From: Scott Chapman sc...@woofplanet.com
Sent: Thursday, January 22, 2015 9:32 AM
To: users@kafka.apache.org
Subject: Re: using the new logstash-kafka plugin

Awesome, what release are you targeting? Or are you able to make updates to
the plugin outside of kafka?

On Thu Jan 22 2015 at 9:31:26 AM Joseph Lawson jlaw...@roomkey.com wrote:

 Scott you will have to do just one topic per input right now but multiple
 topics per group, whitelisting and blacklisting just got merged into
 jruby-kafka and I'm working them up the chain to my logstash-kafka and then
 pass it to the logstash-input/output/-kafka plugin.

 
 From: Scott Chapman sc...@woofplanet.com
 Sent: Wednesday, January 21, 2015 8:32 PM
 To: users@kafka.apache.org
 Subject: using the new logstash-kafka plugin

 We are starting to use the new logstash-kafka plugin, and I am wondering if
 it is possible to read multiple topics? Or do you need to create separate
 logstashes for each topic to parse?

 We are consuming multi-line logs from a service running on a bunch of
 different hosts, so we address that by creating single partition topics for
 our producers.

 We then want to have logstash consume them for ELK.

 Thanks in advance!



Re: kafka production server test

2015-01-22 Thread Sa Li
Hi, Guozhang

Can I run this package remotely test another server? which mean I run this
package on dev but testing kafka system on production?

thanks

AL

On Thu, Jan 22, 2015 at 2:55 PM, Sa Li sal...@gmail.com wrote:

 Hi, Guozhang,

 Good to know such package, will try it now. :-)

 thanks

 On Thu, Jan 22, 2015 at 2:40 PM, Guozhang Wang wangg...@gmail.com wrote:

 Hi Sa,

 Have you looked into the system test package? It contains a suite of tests
 on different failure modes of Kafka brokers.

 Guozhang


 On Thu, Jan 22, 2015 at 12:00 PM, Sa Li sal...@gmail.com wrote:

  Hi, All
 
  We are about to deliver kafka production server, I have been working on
  different test, like performance test from linkedin. This is a 3-node
  cluster, with 5 nodes zkEnsemble. I assume there are lots of tests I
 need
  to do, like network, node failure, flush time, etc. Is there is
 completed
  guide to instruct the tests for kafka production servers?
 
  thanks
 
  --
 
  Alec Li
 



 --
 -- Guozhang




 --

 Alec Li




-- 

Alec Li


Re: warning on startup of consumer app with 0.8.2 rc2

2015-01-22 Thread Jun Rao
Hmm, kafka-console-consumer in 0.8.2 rc2 is running fine. Do you have
multiple kafka jars in your classpath?

Thanks,

Jun

On Thu, Jan 22, 2015 at 4:58 PM, Jason Rosenberg j...@squareup.com wrote:

 2015-01-23 00:55:25,273  WARN [async-message-sender-0] common.AppInfo$ -
 Can't read Kafka version from MANIFEST.MF. Possible cause:
 java.lang.NullPointerException



Re: Partition reassignment doesn't work on few partitions

2015-01-22 Thread Jun Rao
Perhaps you can upgrade all brokers and then try?

Thanks,

Jun

On Wed, Jan 21, 2015 at 9:53 PM, Raghu Udiyar ra...@helpshift.com wrote:

 No errors in the state-change log or the controller. Its as if the
 controller never got the request for that partition.

 Regarding the upgrade, we did upgrade one of the nodes, and initiate the
 replication. Here, the controller at 0.8.0 and this node at 0.8.1.1. In
 this case, when we initiated the reassignment, the following error was
 logged on the destination broker :

 [2015-01-21 13:36:18,101] WARN Broker 7 ignoring LeaderAndIsr request with
 correlation id 837 from controller 1 epoch 44 as broker is not in assigned
 replica list 5,6 for partition [test-topic,0] (state.change.logger)

 5,6 is from where we were moving to topic to 7. 7 is the new broker at
 0.8.1.1

 I'm guessing the controller is in a weird way.

 -- Raghu


 On Thu, Jan 22, 2015 at 5:57 AM, Jun Rao j...@confluent.io wrote:

  Any error in the controller and state-change log? Also, you may want to
  upgrade to 0.8.1, which fixed some reassignment issues.
 
  Thanks,
 
  Jun
 
  On Wed, Jan 21, 2015 at 12:38 PM, Raghu Udiyar ra...@helpshift.com
  wrote:
 
   Hello,
  
   I have a 6 node kafka cluster (0.8.0) where partition reassignment
  doesn't
   seem to work on a few partitions. This happens within the same, as well
  as
   across other topics. Following is the behavior observed :
  
   1. For a successful reassignment, the kafka-reassign-partitions.sh
  returns
   success, I see the controller initiating the reassignment, and
   the destination brokers start replica fetcher threads.
   2. For the unsuccessful reassignment, the tool returns success, but
 there
   is nothing in the controller logs nor the destination brokers.
  
   Also, for the ones that are successful, some don’t finish replication
   correctly. I can see that the destination brokers get stuck after a few
   thousand offsets (checked in JMX), and doesn’t move after that. The
   controller keeps on waiting for the fetchers to complete, but never
 gets
   there.
  
   Anyone seen this issue before? Is there a way to reset the state of the
   controller? or re-elect a new one?
  
   Thanks,
   Raghu
  
 



Re: delete log file

2015-01-22 Thread Jun Rao
In general, you shouldn't delete any files while the broker is up. If you
have no other choices, you can try deleting the older log segments and hope
that no consumer or the log cleaner are using them.

Thanks,

Jun

On Thu, Jan 22, 2015 at 12:40 AM, YuanJia Li yuanjia8...@163.com wrote:

 Hi all,
 The kafka broker's disk is almost full, and the log file is still opened
 by kafka through lsof command.
 I know to change log.retention.hours in server.properties, but I don't
 want to restart the kafka server.
 Is there any utility to delete log file without impacting kafka?

 Thanks  Regards
 YuanJia Li


Re: Help: Kafka LeaderNotAvailableException

2015-01-22 Thread Jun Rao
Any error in the controller and the broker log?

Thanks,

Jun

On Thu, Jan 22, 2015 at 1:33 AM, vishal_ja...@dell.com wrote:

 Hi,
 Let me overview on the issue that I am facing on producing message in
 Kafka:
 I have horthonworks HDP-2.1 installed, along that we have Kafka on other
 node.

 * On kafka node:
 Start Zookeepeer
 Start Kafka Broker service
 Send message/producer
 Consume message - Works (Note: here we start Zookeeper locally on kafka01
 node)

 * Issue side:
 Now in HDP-2.1 we have Zookeeper inbuild  we have Zookeeper service
 running on master node
 I go to Kafka server  Started Kafka Broker
 (In config\server.properties file I have added zookeeper.connect with
 maasternode:2181)
 Then I start producer  Send message... after that we got error like
 kafka.common.LeaderNotAvailableException

 [2015-01-17 05:54:09,465] WARN Error while fetching metadata
 [{TopicMetadata for topic fred -
 No partition metadata for topic fred due to
 kafka.common.LeaderNotAvailableException}] for topic [fred]: class
 kafka.common.LeaderNotAvailableException
 (kafka.producer.BrokerPartitionInfo)
 [2015-01-17 05:54:09,659] WARN Error while fetching metadata
 [{TopicMetadata for topic fred -
 No partition metadata for topic fred due to
 kafka.common.LeaderNotAvailableException}] for topic [fred]: class
 kafka.common.LeaderNotAvailableException
 (kafka.producer.BrokerPartitionInfo)
 [2015-01-17 05:54:09,659] ERROR Failed to collate messages by topic,
 partition due to: Failed to fetch topic metadata for topic: fred
 (kafka.producer.async.DefaultEventHandler)
 [2015-01-17 05:54:09,802] WARN Error while fetching metadata
 [{TopicMetadata for topic fred -
 No partition metadata for topic fred due to
 kafka.common.LeaderNotAvailableException}] for topic [fred]: class
 kafka.common.LeaderNotAvailableException
 (kafka.producer.BrokerPartitionInfo)
 [2015-01-17 05:54:09,820] ERROR Failed to send requests for topics fred
 with correlation ids in [0,8] (kafka.producer.async.DefaultEventHandler)
 [2015-01-17 05:54:09,822] ERROR Error in handling batch of 1 events
 (kafka.producer.async.ProducerSendThread)
 kafka.common.FailedToSendMessageException: Failed to send messages after 3
 tries.
 at
 kafka.producer.async.DefaultEventHandler.handle(DefaultEventHandler.scala:90)
 at
 kafka.producer.async.ProducerSendThread.tryToHandle(ProducerSendThread.scala:104)
 at
 kafka.producer.async.ProducerSendThread$$anonfun$processEvents$3.apply(ProducerSendThread.scala:87)

 Can someone suggest what is going wrong.
 Thanks.





 -Vishal


 Regards,
 Vishal
 Software Dev Staff Engineer
 Dell | Bangalore
 Ext : 79268




Re: Help: Kafka LeaderNotAvailableException

2015-01-22 Thread Gwen Shapira
It sounds like you have two zookeepers, one for HDP and one for Kafka.
Did you move Kafka from one zookeeper to another?

Perhaps Kafka finds the topics (logs) on disk, but they do not exist
in ZK because you are using a different zookeeper now.

Gwen

On Thu, Jan 22, 2015 at 6:38 PM, Jun Rao j...@confluent.io wrote:
 Any error in the controller and the broker log?

 Thanks,

 Jun

 On Thu, Jan 22, 2015 at 1:33 AM, vishal_ja...@dell.com wrote:

 Hi,
 Let me overview on the issue that I am facing on producing message in
 Kafka:
 I have horthonworks HDP-2.1 installed, along that we have Kafka on other
 node.

 * On kafka node:
 Start Zookeepeer
 Start Kafka Broker service
 Send message/producer
 Consume message - Works (Note: here we start Zookeeper locally on kafka01
 node)

 * Issue side:
 Now in HDP-2.1 we have Zookeeper inbuild  we have Zookeeper service
 running on master node
 I go to Kafka server  Started Kafka Broker
 (In config\server.properties file I have added zookeeper.connect with
 maasternode:2181)
 Then I start producer  Send message... after that we got error like
 kafka.common.LeaderNotAvailableException

 [2015-01-17 05:54:09,465] WARN Error while fetching metadata
 [{TopicMetadata for topic fred -
 No partition metadata for topic fred due to
 kafka.common.LeaderNotAvailableException}] for topic [fred]: class
 kafka.common.LeaderNotAvailableException
 (kafka.producer.BrokerPartitionInfo)
 [2015-01-17 05:54:09,659] WARN Error while fetching metadata
 [{TopicMetadata for topic fred -
 No partition metadata for topic fred due to
 kafka.common.LeaderNotAvailableException}] for topic [fred]: class
 kafka.common.LeaderNotAvailableException
 (kafka.producer.BrokerPartitionInfo)
 [2015-01-17 05:54:09,659] ERROR Failed to collate messages by topic,
 partition due to: Failed to fetch topic metadata for topic: fred
 (kafka.producer.async.DefaultEventHandler)
 [2015-01-17 05:54:09,802] WARN Error while fetching metadata
 [{TopicMetadata for topic fred -
 No partition metadata for topic fred due to
 kafka.common.LeaderNotAvailableException}] for topic [fred]: class
 kafka.common.LeaderNotAvailableException
 (kafka.producer.BrokerPartitionInfo)
 [2015-01-17 05:54:09,820] ERROR Failed to send requests for topics fred
 with correlation ids in [0,8] (kafka.producer.async.DefaultEventHandler)
 [2015-01-17 05:54:09,822] ERROR Error in handling batch of 1 events
 (kafka.producer.async.ProducerSendThread)
 kafka.common.FailedToSendMessageException: Failed to send messages after 3
 tries.
 at
 kafka.producer.async.DefaultEventHandler.handle(DefaultEventHandler.scala:90)
 at
 kafka.producer.async.ProducerSendThread.tryToHandle(ProducerSendThread.scala:104)
 at
 kafka.producer.async.ProducerSendThread$$anonfun$processEvents$3.apply(ProducerSendThread.scala:87)

 Can someone suggest what is going wrong.
 Thanks.





 -Vishal


 Regards,
 Vishal
 Software Dev Staff Engineer
 Dell | Bangalore
 Ext : 79268




Re: kafka production server test

2015-01-22 Thread Sa Li
Hi, Guozhang,

Good to know such package, will try it now. :-)

thanks

On Thu, Jan 22, 2015 at 2:40 PM, Guozhang Wang wangg...@gmail.com wrote:

 Hi Sa,

 Have you looked into the system test package? It contains a suite of tests
 on different failure modes of Kafka brokers.

 Guozhang


 On Thu, Jan 22, 2015 at 12:00 PM, Sa Li sal...@gmail.com wrote:

  Hi, All
 
  We are about to deliver kafka production server, I have been working on
  different test, like performance test from linkedin. This is a 3-node
  cluster, with 5 nodes zkEnsemble. I assume there are lots of tests I need
  to do, like network, node failure, flush time, etc. Is there is completed
  guide to instruct the tests for kafka production servers?
 
  thanks
 
  --
 
  Alec Li
 



 --
 -- Guozhang




-- 

Alec Li


Re: kafka production server test

2015-01-22 Thread Guozhang Wang
Hi Sa,

Have you looked into the system test package? It contains a suite of tests
on different failure modes of Kafka brokers.

Guozhang


On Thu, Jan 22, 2015 at 12:00 PM, Sa Li sal...@gmail.com wrote:

 Hi, All

 We are about to deliver kafka production server, I have been working on
 different test, like performance test from linkedin. This is a 3-node
 cluster, with 5 nodes zkEnsemble. I assume there are lots of tests I need
 to do, like network, node failure, flush time, etc. Is there is completed
 guide to instruct the tests for kafka production servers?

 thanks

 --

 Alec Li




-- 
-- Guozhang


warning on startup of consumer app with 0.8.2 rc2

2015-01-22 Thread Jason Rosenberg
2015-01-23 00:55:25,273  WARN [async-message-sender-0] common.AppInfo$ -
Can't read Kafka version from MANIFEST.MF. Possible cause:
java.lang.NullPointerException


Re: Re: delete log file

2015-01-22 Thread YuanJia Li
Hi Yang,
I think my problem is not the same with yours.
My production environment is fine, log.retention.hours is 720. My disk is 
almost full just beacause of too much data.
I want to have a utility or command to delete manually, instead of waiting 
until it expires.
Thanks for your reply.



Thanks  Regards
YuanJia Li

From: Yang Fang
Date: 2015-01-23 12:15
To: users
Subject: Re: delete log file
YuanJia,
I want to know why your broker's disk is almost full . Maybe your issue is
the same with mine. Is it one broker out of service for a long time, then
it fails back,  at first it looks good, but after some hours, its disk is
almost full. Other brokers are all fine, avg remaining capacity is about
25%. In my production environment, log.retention.hours is 24, broker's data
volume is about 2TB. The reason it that when broker decide to delete logs
is based on its start time and log.retention.hours.

On Thu, Jan 22, 2015 at 4:40 PM, YuanJia Li yuanjia8...@163.com wrote:

 Hi all,
 The kafka broker's disk is almost full, and the log file is still opened
 by kafka through lsof command.
 I know to change log.retention.hours in server.properties, but I don't
 want to restart the kafka server.
 Is there any utility to delete log file without impacting kafka?

 Thanks  Regards
 YuanJia Li

kafka topic not exist after creation with api

2015-01-22 Thread Manu Zhang
Hi all ,
My application creates kafka topic at runtime with AdminUtils.createTopic
but that topic is not available when I try to produce to it.

If I run bin/kafka-console-consumer.sh --topic $topic --zookeeper $zkStr
--from-beginning it would throw UnknownTopicOrPartitionException.

 Also, I found the following error in controller log.

[2015-01-23 12:44:41,825] WARN [Controller-0-to-broker-0-send-thread],
Controller 0 fails to send a request to broker id:0,host:node10,port:9092
(kafka.controller.RequestSendThread)
java.io.EOFException: Received -1 when reading from channel, socket has
likely been closed.
at kafka.utils.Utils$.read(Utils.scala:381)
at
kafka.network.BoundedByteBufferReceive.readFrom(BoundedByteBufferReceive.scala:54)
at kafka.network.Receive$class.readCompletely(Transmission.scala:56)
at
kafka.network.BoundedByteBufferReceive.readCompletely(BoundedByteBufferReceive.scala:29)
at kafka.network.BlockingChannel.receive(BlockingChannel.scala:108)
at
kafka.controller.RequestSendThread.doWork(ControllerChannelManager.scala:146)
at kafka.utils.ShutdownableThread.run(ShutdownableThread.scala:60)
[2015-01-23 12:44:41,827] ERROR [Controller-0-to-broker-0-send-thread],
Controller 0 epoch 3 failed to send request
Name:UpdateMetadataRequest;Version:0;Controller:0;ControllerEpoch:3;CorrelationId:80;ClientId:id_0-host_null-port_9092;AliveBrokers:id:0,host:node10,port:9092;PartitionState:[app1_test_0_testB,0]
-
(LeaderAndIsrInfo:(Leader:0,ISR:0,LeaderEpoch:0,ControllerEpoch:3),ReplicationFactor:1),AllReplicas:0)
to broker id:0,host:node10,port:9092. Reconnecting to broker.
(kafka.controller.RequestSendThread)
java.nio.channels.ClosedChannelException
at kafka.network.BlockingChannel.send(BlockingChannel.scala:97)
at
kafka.controller.RequestSendThread.liftedTree1$1(ControllerChannelManager.scala:132)
at
kafka.controller.RequestSendThread.doWork(ControllerChannelManager.scala:131)
at kafka.utils.ShutdownableThread.run(ShutdownableThread.scala:60)

I've tried fetching metadata after creating topic to ensure its
availability but the fetch failed after several round of retries.

My kafka is built from 0.8.2 trunk running on local node.

Any thoughts would be appreciated

Thanks,
Manu


New mirror maker consumer.config question

2015-01-22 Thread tao xiao
Hi,

I discovered that the new mirror maker implementation in trunk now only
accept one consumer.config property instead of a list of them which means
we can only supply one source per mirror maker process. Is it a reason for
it? If I have multiple source kafka clusters do I need to setup multiple
mirror maker processes?

-- 
Regards,
Tao


Apache Spark broadcast error: Error sending message as driverActor is null [message = UpdateBlockInfo(BlockManagerId

2015-01-22 Thread Zijing Guo
HiI'm using Apache Spark 1.1.0 and I'm currently having issue with broadcast 
method. So when I call broadcast function on a small dataset to a 5 nodes 
cluster, I experiencing the Error sending message as driverActor is null 
after broadcast the variables several times (apps running under jboss). 

Any help would be appreciate.ThanksEdwin


Re: Apache Spark broadcast error: Error sending message as driverActor is null [message = UpdateBlockInfo(BlockManagerId

2015-01-22 Thread Zijing Guo
Yep, sorry, had a rough day.. 

 On Thursday, January 22, 2015 2:25 PM, Guozhang Wang wangg...@gmail.com 
wrote:
   

 Hi Zijing,

Sounds like you sent to the wrong mailing list :P

Guozhang

On Thu, Jan 22, 2015 at 11:12 AM, Zijing Guo alter...@yahoo.com.invalid
wrote:

 HiI'm using Apache Spark 1.1.0 and I'm currently having issue with
 broadcast method. So when I call broadcast function on a small dataset to a
 5 nodes cluster, I experiencing the
 Error sending message as driverActor is null after broadcast the
 variables several times (apps running under jboss).

 Any help would be appreciate.ThanksEdwin




-- 
-- Guozhang




Re: Apache Spark broadcast error: Error sending message as driverActor is null [message = UpdateBlockInfo(BlockManagerId

2015-01-22 Thread Guozhang Wang
Hi Zijing,

Sounds like you sent to the wrong mailing list :P

Guozhang

On Thu, Jan 22, 2015 at 11:12 AM, Zijing Guo alter...@yahoo.com.invalid
wrote:

 HiI'm using Apache Spark 1.1.0 and I'm currently having issue with
 broadcast method. So when I call broadcast function on a small dataset to a
 5 nodes cluster, I experiencing the
 Error sending message as driverActor is null after broadcast the
 variables several times (apps running under jboss).

 Any help would be appreciate.ThanksEdwin




-- 
-- Guozhang


kafka production server test

2015-01-22 Thread Sa Li
Hi, All

We are about to deliver kafka production server, I have been working on
different test, like performance test from linkedin. This is a 3-node
cluster, with 5 nodes zkEnsemble. I assume there are lots of tests I need
to do, like network, node failure, flush time, etc. Is there is completed
guide to instruct the tests for kafka production servers?

thanks

-- 

Alec Li


Re: Help: Kafka LeaderNotAvailableException

2015-01-22 Thread Joe Stein
Vishal,

Does this error happen every time you are sending? Or just the first time?

Joe Stein

On Thu, Jan 22, 2015 at 4:33 AM, vishal_ja...@dell.com wrote:

 Hi,
 Let me overview on the issue that I am facing on producing message in
 Kafka:
 I have horthonworks HDP-2.1 installed, along that we have Kafka on other
 node.

 * On kafka node:
 Start Zookeepeer
 Start Kafka Broker service
 Send message/producer
 Consume message - Works (Note: here we start Zookeeper locally on kafka01
 node)

 * Issue side:
 Now in HDP-2.1 we have Zookeeper inbuild  we have Zookeeper service
 running on master node
 I go to Kafka server  Started Kafka Broker
 (In config\server.properties file I have added zookeeper.connect with
 maasternode:2181)
 Then I start producer  Send message... after that we got error like
 kafka.common.LeaderNotAvailableException

 [2015-01-17 05:54:09,465] WARN Error while fetching metadata
 [{TopicMetadata for topic fred -
 No partition metadata for topic fred due to
 kafka.common.LeaderNotAvailableException}] for topic [fred]: class
 kafka.common.LeaderNotAvailableException
 (kafka.producer.BrokerPartitionInfo)
 [2015-01-17 05:54:09,659] WARN Error while fetching metadata
 [{TopicMetadata for topic fred -
 No partition metadata for topic fred due to
 kafka.common.LeaderNotAvailableException}] for topic [fred]: class
 kafka.common.LeaderNotAvailableException
 (kafka.producer.BrokerPartitionInfo)
 [2015-01-17 05:54:09,659] ERROR Failed to collate messages by topic,
 partition due to: Failed to fetch topic metadata for topic: fred
 (kafka.producer.async.DefaultEventHandler)
 [2015-01-17 05:54:09,802] WARN Error while fetching metadata
 [{TopicMetadata for topic fred -
 No partition metadata for topic fred due to
 kafka.common.LeaderNotAvailableException}] for topic [fred]: class
 kafka.common.LeaderNotAvailableException
 (kafka.producer.BrokerPartitionInfo)
 [2015-01-17 05:54:09,820] ERROR Failed to send requests for topics fred
 with correlation ids in [0,8] (kafka.producer.async.DefaultEventHandler)
 [2015-01-17 05:54:09,822] ERROR Error in handling batch of 1 events
 (kafka.producer.async.ProducerSendThread)
 kafka.common.FailedToSendMessageException: Failed to send messages after 3
 tries.
 at
 kafka.producer.async.DefaultEventHandler.handle(DefaultEventHandler.scala:90)
 at
 kafka.producer.async.ProducerSendThread.tryToHandle(ProducerSendThread.scala:104)
 at
 kafka.producer.async.ProducerSendThread$$anonfun$processEvents$3.apply(ProducerSendThread.scala:87)

 Can someone suggest what is going wrong.
 Thanks.





 -Vishal


 Regards,
 Vishal
 Software Dev Staff Engineer
 Dell | Bangalore
 Ext : 79268




Counting # of Message Brokers Receive Per Minute Per Topic

2015-01-22 Thread Bhavesh Mistry
Hi Kafka Team,

I need to count message received by entire Kafka Broker Cluster for a
particular topic.

I have 3 brokers, so do I need to sum the COUNT metric or just one server
count reflect all server count.  It seems that count is always increasing
(although metric name is *MessagesInPerSec* so count does not get reset)

kafka.server:type=BrokerTopicMetrics,name=topic-MessagesInPerSec

Please clarify.

Thanks,
Bhavesh


kafka sending duplicate content to consumer

2015-01-22 Thread Tousif
Hi,

i want know in which situation does kafka send same event  multiple times
to consumer. Is there a consumer side configuration to tell kafka to send
only once and stop retries?

-- 


Regards
Tousif Khazi