Re: SimpleConsumer cannot read KeyedMessage.

2013-11-01 Thread Lu Xuechao
The consumer starts from offset 0. Yes, in the log dir. On Fri, Nov 1, 2013 at 4:06 PM, Jun Rao wrote: > Which offset did you use for fetching? Is there data in the kafka log dir? > > Thanks, > > Jun > > > On Fri, Nov 1, 2013 at 11:48 AM, Lu Xuechao wrote: > > &g

Re: SimpleConsumer cannot read KeyedMessage.

2013-11-01 Thread Lu Xuechao
checked fetchResponse.hasError() but has no error. On Fri, Nov 1, 2013 at 7:45 AM, Jun Rao wrote: > Did you check the error code associated with each partition in the fetch > response? > > Thanks, > > Jun > > > On Thu, Oct 31, 2013 at 9:59 PM, Lu Xuechao wrote: &g

Re: SimpleConsumer cannot read KeyedMessage.

2013-10-31 Thread Lu Xuechao
I enabled gzip compression. Each topic has 10 partitions and each partition is handled by 1 simple consumer thread. All consumers stop to iterate after iterate first several responses. The responses still return with bytes, but cannot iterate. On Thu, Oct 31, 2013 at 9:59 PM, Lu Xuechao wrote

Re: SimpleConsumer cannot read KeyedMessage.

2013-10-31 Thread Lu Xuechao
A/FAQ#FAQ-Whybrokersdonotreceiveproducersentmessages%3F > ? > > Thanks, > > Jun > > > On Thu, Oct 31, 2013 at 2:23 PM, Lu Xuechao wrote: > > > It seems the reason is I enabled gzip compression. > > > > what the code would like to consume compressed messages? > > > > thank

Re: SimpleConsumer cannot read KeyedMessage.

2013-10-31 Thread Lu Xuechao
It seems the reason is I enabled gzip compression. what the code would like to consume compressed messages? thanks. On Thu, Oct 31, 2013 at 11:26 AM, Lu Xuechao wrote: > Hi, > > I am following the > https://cwiki.apache.org/confluence/display/KAFKA/0.8.0+SimpleConsumer+Exampl

SimpleConsumer cannot read KeyedMessage.

2013-10-31 Thread Lu Xuechao
Hi, I am following the https://cwiki.apache.org/confluence/display/KAFKA/0.8.0+SimpleConsumer+Example When I send KeyedMessage with StringEncoder, I can get the messages sent: for (MessageAndOffset messageAndOffset : fetchResponse.messageSet(m_topic, m_partition)) { //handle messages } But whe

Re: Please help find the reason behind performance test results.

2013-10-21 Thread Lu Xuechao
? > > Thanks, > Neha > > > On Mon, Oct 21, 2013 at 12:01 PM, Lu Xuechao wrote: > > > Hi, > > > > We observed below correlation between kafka configuration and performance > > data: > > > > 1. producer insertion rate drops as compression enabled,

Please help find the reason behind performance test results.

2013-10-21 Thread Lu Xuechao
Hi, We observed below correlation between kafka configuration and performance data: 1. producer insertion rate drops as compression enabled, especially gzip; 2. when the producer batch size is below 200, we got high CPU/network IO on brokers, got high network IO on producers/consumers; What's th

Only enable compressed storage on broker?

2013-10-21 Thread Lu Xuechao
Hi, I wonder if it is supported to enable compression for storage on brokers but the producer sends messages uncompressed and consumer receives messages uncompressed? thanks.

Re: Question on Kafka partitions.

2013-09-12 Thread Lu Xuechao
t; > On Wed, Sep 11, 2013 at 9:59 PM, Lu Xuechao wrote: > > > Hi Team, > > > > I have some questions regarding Kafka partitions: > > > > 1. Based on my understanding, the partitions of the same broker have > > contention on disk IO. Say If I have 10 hard d

Question on Kafka partitions.

2013-09-11 Thread Lu Xuechao
Hi Team, I have some questions regarding Kafka partitions: 1. Based on my understanding, the partitions of the same broker have contention on disk IO. Say If I have 10 hard drives, can I specify all the partitions spread evenly on those drives? 2. If I configure default.replication.factor=2, the

Re: Unable to send and consume compressed events.

2013-08-30 Thread Lu Xuechao
Hi, Joe. wiki updated. Hope it helps. On Fri, Aug 30, 2013 at 3:22 PM, Joe Stein wrote: > I feel like this is maybe a usual case as we have heard it before now a few > bits > > Lu Xuechao would you mind updating the FAQ > https://cwiki.apache.org/confluence/display/KAFKA/FA

Re: Unable to send and consume compressed events.

2013-08-29 Thread Lu Xuechao
Hi Jun, Thanks for you help. Finally, I found the reason by enabling producer side DEBUG info output. The snappy jar is not included in the classpath. Added it and it worked. Thanks again. On Fri, Aug 30, 2013 at 12:53 PM, Lu Xuechao wrote: > No. > > > On Fri, Aug 30, 2013 at 1

Re: Unable to send and consume compressed events.

2013-08-29 Thread Lu Xuechao
No. On Fri, Aug 30, 2013 at 11:57 AM, Jun Rao wrote: > These are the metadata requests. Do you see Producer requests from your > client? > > Thanks, > > Jun > > > On Thu, Aug 29, 2013 at 5:40 PM, Lu Xuechao wrote: > > > After I sent 1,000 compressed even

Re: Unable to send and consume compressed events.

2013-08-29 Thread Lu Xuechao
Update: Sending compressed events with console producer works: kafka-console-producer.bat --broker-list localhost:9092 --sync --topic topic1 --compress I am working on Windows 7. On Fri, Aug 30, 2013 at 8:40 AM, Lu Xuechao wrote: > After I sent 1,000 compressed events, I saw these messa

Re: Unable to send and consume compressed events.

2013-08-29 Thread Lu Xuechao
; Jun > > > On Thu, Aug 29, 2013 at 6:29 AM, Lu Xuechao wrote: > > > Let me post my test code here. I could see producer.send(data); returned > > with no error. > > > > public class TestProducer extends Thread { > > private final Producer producer;

Re: Unable to send and consume compressed events.

2013-08-29 Thread Lu Xuechao
_request_required_acks = "1"; final static String p_producer_type = "async"; final static String p_batch_num = "100"; final static String p_compression_codec = "1"; final static String p_message_send_retries = "3"; final static

Re: Unable to send and consume compressed events.

2013-08-29 Thread Lu Xuechao
ponding errors in > the broker logs? With the configuration below, I don't think any errors > will be reported back to the producer. > > You could also try setting erquest.required.acks=1 to see if errors are > reported back to the client. > > On 8/29/13 4:40 AM, "Lu Xuecha

Unable to send and consume compressed events.

2013-08-29 Thread Lu Xuechao
Hi , I am trying to enable gzip compression for my events. But after I switched compression.codec to "1" I found the produced events were even not be persisted to disk log file. Of course, the consumer could not receive any compressed events. I sent 10,000 or more events but the broker's log file