RE: Looks like consumer fetchers get stopped we are not getting any data

2014-01-11 Thread Seshadri, Balaji
Currently its set to default,whats the timeout should we increase to.

From: Guozhang Wang [wangg...@gmail.com]
Sent: Saturday, January 11, 2014 3:53 PM
To: users@kafka.apache.org
Subject: Re: Looks like consumer fetchers get stopped we are not getting any 
data

>From the logs it seems the consumer's ZK registry log has lost, while
KAFKA-693 is mainly due to server side issue. Could you check if there is a
session timeout from the consumer on ZK log.

Guozhang


On Sat, Jan 11, 2014 at 2:33 PM, Seshadri, Balaji
wrote:

> We found the below bug in Kafka JIRA.
>
> https://issues.apache.org/jira/browse/KAFKA-693
>
> Could you guys please let us know if this bug is causing the fetchers
> getting stooped because of rebalance failure,we are using 0.8-beta and this
> fix seems to be in 0.8 version.
>
> The issue happens only when we run it in tomcat,when we use JUnits with
> same group.id to test we receive the messages.It would be great if you
> could help us confirm if it is a bug.
>
> Also we would like to know if we need to upgrade the server/client to 0.8
> or just server side code.
>
>
> 
> From: Rob Withers [robert.w.with...@gmail.com]
> Sent: Friday, January 10, 2014 10:17 PM
> To: users@kafka.apache.org
> Subject: Re: Looks like consumer fetchers get stopped we are not getting
> any data
>
> That was an interesting section too.  Which GC settings would you suggest?
>
> Thank you,
> - charlie
>
> > On Jan 10, 2014, at 10:11 PM, Jun Rao  wrote:
> >
> > Have you looked at our FAQ, especially
> >
> https://cwiki.apache.org/confluence/display/KAFKA/FAQ#FAQ-Whyaretheremanyrebalancesinmyconsumerlog
> > ?
> >
> > Thanks,
> >
> > Jun
> >
> >
> > On Fri, Jan 10, 2014 at 2:25 PM, Seshadri, Balaji
> > wrote:
> >
> >> Any clue would be helpful.
> >>
> >> -Original Message-
> >> From: Seshadri, Balaji [mailto:balaji.sesha...@dish.com]
> >> Sent: Friday, January 10, 2014 12:46 PM
> >> To: users@kafka.apache.org
> >> Subject: RE: Looks like consumer fetchers get stopped we are not getting
> >> any data
> >>
> >> Yes rebalance begins and exceptions occurs.
> >>
> >>
> >> {2014-01-10 00:58:11,293} INFO
> >>
> [account-info-updated-hadoop-consumer_tm1mwdpl04-1389222553159-ad59660b_watcher_executor]
> >> (?:?) - [account-i
> >> nfo-updated-hadoop-consumer_tm1mwdpl04-1389222553159-ad59660b], Cleared
> the
> >> data chunks in all the consumer message iterators
> >> {2014-01-10 00:58:11,293} INFO
> >>
> [account-info-updated-hadoop-consumer_tm1mwdpl04-1389222553159-ad59660b_watcher_executor]
> >> (?:?) - [account-i
> >> nfo-updated-hadoop-consumer_tm1mwdpl04-1389222553159-ad59660b],
> Committing
> >> all offsets after clearing the fetcher queues
> >> {2014-01-10 00:58:11,298} DEBUG [catalina-exec-12-SendThread(
> >> tvip-m1-mw-zookeeper.dish.com:2181)] (ClientCnxn.java:759) - Got ping
> >> response for sessionid: 0x1437b2879870005 af ter 0ms
> >> {2014-01-10 00:58:11,313} INFO
> >>
> [account-activated-hadoop-consumer_tm1mwdpl04-1389222557906-562b6738_watcher_executor]
> >> (?:?) -
> >> [account-activated-hadoop-consumer_tm1mwdpl04-1389222557906-562b6738],
> >> begin rebalancing consumer
> >> account-activated-hadoop-consumer_tm1mwdpl04-1389222557906-562b6738 try
> #1
> >> {2014-01-10 00:58:11,314} DEBUG [catalina-exec-12-SendThread(
> >> tvip-m1-mw-zookeeper.dish.com:2181)] (ClientCnxn.java:839) - Reading
> >> reply sessionid:0x1437b2879870005, packet:: clientPath:null
> serverPath:null
> >> finished:false header:: 627,8  replyHeader:: 627,51539619966,0
>  request::
> >> '/brokers/ids,F  response:: v{'1}
> >> {2014-01-10 00:58:11,315} DEBUG [catalina-exec-12-SendThread(
> >> tvip-m1-mw-zookeeper.dish.com:2181)] (ClientCnxn.java:839) - Reading
> >> reply sessionid:0x1437b2879870005, packet:: clientPath:null
> serverPath:null
> >> finished:false header:: 628,4  replyHeader:: 628,51539619966,0
>  request::
> >> '/brokers/ids/1,F  response::
> >>
> #7b2022686f7374223a22746d312d6b61666b6162726f6b6572313031222c20226a6d785f706f7274223a393939392c2022706f7274223a393039322c202276657273696f6e223a31207d,s{47244644685,47244644685,1388537628753,1388537628753,0,0,0,163056791896588316,74,0,47244644685}
> >> {2014-01-10 00:58:11,316} DEBUG [catalina-exec-12-SendThread(
> >> tvip-m1-mw-zookeeper.dish.com:2181)] (ClientCnxn.java:839) - Reading
> >> reply sessionid:0x1437b2879870005, packet:: clientPath:null
> serverPath:null
> >> finished:false header:: 629,4  replyHeader:: 629,51539619966,-101
> >> request::
> >>
> '/consumers/account-activated-hadoop-consumer/ids/account-activated-hadoop-consumer_tm1mwdpl04-1389222557906-562b6738,F
> >> response::
> >> {2014-01-10 00:58:11,316} INFO
> >>
> [account-activated-hadoop-consumer_tm1mwdpl04-1389222557906-562b6738_watcher_executor]
> >> (?:?) -
> >> [account-activated-hadoop-consumer_tm1mwdpl04-1389222557906-562b6738],
> >> exception during rebalance
> >> org.I0Itec.zkclient.exception.ZkNoNodeException:
> >> org

restoring broker node / zookeeper question

2014-01-11 Thread Ran RanUser
I have two questions:

1.

We plan to initially run Kafka without partitioning or replication, simply
having two single node Kafka (broker) servers in two geographically diverse
locations.  Using periodic backups/snapshots across locations for disaster
recovery.

Our producer / consumer applications will process to/from each location.

For such configurations, is it possible to run the Kafka broker without
Zookeeper?  If not, any plans for such a simplified configuration?  We'd
like to remove the extra app from the stack (Zookeeper) since it's not
being utilized.

2.

If we need to restore a Kafka node (simplified design above), I was hoping
Kafka and Zookeeper could bootstrap itself as long as we can restore a copy
of the /tmp/kafka-logs data. This dir contains the topics, log data, and
high watermark.

Unfortunately, it seems without the Zookeeper data, we cannot restore the
node.

Is there a way to restore a Kafka node if the Zookeeper data cannot be
restored?  What are the specific requirements to restore a Kafka broker
from backups, if the /tmp/zookeeper must also be restored, are there any
inconsistent state issues (i.e. backup of /tmp/kafka-logs and
/tmp/zookeeper are not synchronized)?


Thank you!


RE: Looks like consumer fetchers get stopped we are not getting any data

2014-01-11 Thread Nanjegowda, Mithunraj

We have attached more logs where the fetchers are stopped.
https://issues.apache.org/jira/browse/KAFKA-693 issue mentions about this 
issue, and do we need to upgrade to 0.8.0 release for this fix?


Regards,
RJ

-Original Message-
From: Seshadri, Balaji
Sent: Saturday, January 11, 2014 3:34 PM
To: users@kafka.apache.org
Cc: Gulia, Vikram; Nanjegowda, Mithunraj; Behera, Ajit; Withers, Robert; 
ShankenpurMayanna, Diwakar
Subject: RE: Looks like consumer fetchers get stopped we are not getting any 
data

We found the below bug in Kafka JIRA.

https://issues.apache.org/jira/browse/KAFKA-693

Could you guys please let us know if this bug is causing the fetchers getting 
stooped because of rebalance failure,we are using 0.8-beta and this fix seems 
to be in 0.8 version.

The issue happens only when we run it in tomcat,when we use JUnits with same 
group.id to test we receive the messages.It would be great if you could help us 
confirm if it is a bug.

Also we would like to know if we need to upgrade the server/client to 0.8 or 
just server side code.



From: Rob Withers [robert.w.with...@gmail.com]
Sent: Friday, January 10, 2014 10:17 PM
To: users@kafka.apache.org
Subject: Re: Looks like consumer fetchers get stopped we are not getting any 
data

That was an interesting section too.  Which GC settings would you suggest?

Thank you,
- charlie

> On Jan 10, 2014, at 10:11 PM, Jun Rao  wrote:
>
> Have you looked at our FAQ, especially
> https://cwiki.apache.org/confluence/display/KAFKA/FAQ#FAQ-Whyaretheremanyrebalancesinmyconsumerlog
> ?
>
> Thanks,
>
> Jun
>
>
> On Fri, Jan 10, 2014 at 2:25 PM, Seshadri, Balaji
> wrote:
>
>> Any clue would be helpful.
>>
>> -Original Message-
>> From: Seshadri, Balaji [mailto:balaji.sesha...@dish.com]
>> Sent: Friday, January 10, 2014 12:46 PM
>> To: users@kafka.apache.org
>> Subject: RE: Looks like consumer fetchers get stopped we are not getting
>> any data
>>
>> Yes rebalance begins and exceptions occurs.
>>
>>
>> {2014-01-10 00:58:11,293} INFO
>> [account-info-updated-hadoop-consumer_tm1mwdpl04-1389222553159-ad59660b_watcher_executor]
>> (?:?) - [account-i
>> nfo-updated-hadoop-consumer_tm1mwdpl04-1389222553159-ad59660b], Cleared the
>> data chunks in all the consumer message iterators
>> {2014-01-10 00:58:11,293} INFO
>> [account-info-updated-hadoop-consumer_tm1mwdpl04-1389222553159-ad59660b_watcher_executor]
>> (?:?) - [account-i
>> nfo-updated-hadoop-consumer_tm1mwdpl04-1389222553159-ad59660b], Committing
>> all offsets after clearing the fetcher queues
>> {2014-01-10 00:58:11,298} DEBUG [catalina-exec-12-SendThread(
>> tvip-m1-mw-zookeeper.dish.com:2181)] (ClientCnxn.java:759) - Got ping
>> response for sessionid: 0x1437b2879870005 af ter 0ms
>> {2014-01-10 00:58:11,313} INFO
>> [account-activated-hadoop-consumer_tm1mwdpl04-1389222557906-562b6738_watcher_executor]
>> (?:?) -
>> [account-activated-hadoop-consumer_tm1mwdpl04-1389222557906-562b6738],
>> begin rebalancing consumer
>> account-activated-hadoop-consumer_tm1mwdpl04-1389222557906-562b6738 try #1
>> {2014-01-10 00:58:11,314} DEBUG [catalina-exec-12-SendThread(
>> tvip-m1-mw-zookeeper.dish.com:2181)] (ClientCnxn.java:839) - Reading
>> reply sessionid:0x1437b2879870005, packet:: clientPath:null serverPath:null
>> finished:false header:: 627,8  replyHeader:: 627,51539619966,0  request::
>> '/brokers/ids,F  response:: v{'1}
>> {2014-01-10 00:58:11,315} DEBUG [catalina-exec-12-SendThread(
>> tvip-m1-mw-zookeeper.dish.com:2181)] (ClientCnxn.java:839) - Reading
>> reply sessionid:0x1437b2879870005, packet:: clientPath:null serverPath:null
>> finished:false header:: 628,4  replyHeader:: 628,51539619966,0  request::
>> '/brokers/ids/1,F  response::
>> #7b2022686f7374223a22746d312d6b61666b6162726f6b6572313031222c20226a6d785f706f7274223a393939392c2022706f7274223a393039322c202276657273696f6e223a31207d,s{47244644685,47244644685,1388537628753,1388537628753,0,0,0,163056791896588316,74,0,47244644685}
>> {2014-01-10 00:58:11,316} DEBUG [catalina-exec-12-SendThread(
>> tvip-m1-mw-zookeeper.dish.com:2181)] (ClientCnxn.java:839) - Reading
>> reply sessionid:0x1437b2879870005, packet:: clientPath:null serverPath:null
>> finished:false header:: 629,4  replyHeader:: 629,51539619966,-101
>> request::
>> '/consumers/account-activated-hadoop-consumer/ids/account-activated-hadoop-consumer_tm1mwdpl04-1389222557906-562b6738,F
>> response::
>> {2014-01-10 00:58:11,316} INFO
>> [account-activated-hadoop-consumer_tm1mwdpl04-1389222557906-562b6738_watcher_executor]
>> (?:?) -
>> [account-activated-hadoop-consumer_tm1mwdpl04-1389222557906-562b6738],
>> exception during rebalance
>> org.I0Itec.zkclient.exception.ZkNoNodeException:
>> org.apache.zookeeper.KeeperException$NoNodeException: KeeperErrorCode =
>> NoNode for
>> /consumers/account-activated-hadoop-consumer/ids/account-activated-hadoop-consumer_tm1mwdpl04-1389222557906-562b6738
>>at
>> org.I0Itec.zkclient.exception.ZkException

Re: Looks like consumer fetchers get stopped we are not getting any data

2014-01-11 Thread Guozhang Wang
>From the logs it seems the consumer's ZK registry log has lost, while
KAFKA-693 is mainly due to server side issue. Could you check if there is a
session timeout from the consumer on ZK log.

Guozhang


On Sat, Jan 11, 2014 at 2:33 PM, Seshadri, Balaji
wrote:

> We found the below bug in Kafka JIRA.
>
> https://issues.apache.org/jira/browse/KAFKA-693
>
> Could you guys please let us know if this bug is causing the fetchers
> getting stooped because of rebalance failure,we are using 0.8-beta and this
> fix seems to be in 0.8 version.
>
> The issue happens only when we run it in tomcat,when we use JUnits with
> same group.id to test we receive the messages.It would be great if you
> could help us confirm if it is a bug.
>
> Also we would like to know if we need to upgrade the server/client to 0.8
> or just server side code.
>
>
> 
> From: Rob Withers [robert.w.with...@gmail.com]
> Sent: Friday, January 10, 2014 10:17 PM
> To: users@kafka.apache.org
> Subject: Re: Looks like consumer fetchers get stopped we are not getting
> any data
>
> That was an interesting section too.  Which GC settings would you suggest?
>
> Thank you,
> - charlie
>
> > On Jan 10, 2014, at 10:11 PM, Jun Rao  wrote:
> >
> > Have you looked at our FAQ, especially
> >
> https://cwiki.apache.org/confluence/display/KAFKA/FAQ#FAQ-Whyaretheremanyrebalancesinmyconsumerlog
> > ?
> >
> > Thanks,
> >
> > Jun
> >
> >
> > On Fri, Jan 10, 2014 at 2:25 PM, Seshadri, Balaji
> > wrote:
> >
> >> Any clue would be helpful.
> >>
> >> -Original Message-
> >> From: Seshadri, Balaji [mailto:balaji.sesha...@dish.com]
> >> Sent: Friday, January 10, 2014 12:46 PM
> >> To: users@kafka.apache.org
> >> Subject: RE: Looks like consumer fetchers get stopped we are not getting
> >> any data
> >>
> >> Yes rebalance begins and exceptions occurs.
> >>
> >>
> >> {2014-01-10 00:58:11,293} INFO
> >>
> [account-info-updated-hadoop-consumer_tm1mwdpl04-1389222553159-ad59660b_watcher_executor]
> >> (?:?) - [account-i
> >> nfo-updated-hadoop-consumer_tm1mwdpl04-1389222553159-ad59660b], Cleared
> the
> >> data chunks in all the consumer message iterators
> >> {2014-01-10 00:58:11,293} INFO
> >>
> [account-info-updated-hadoop-consumer_tm1mwdpl04-1389222553159-ad59660b_watcher_executor]
> >> (?:?) - [account-i
> >> nfo-updated-hadoop-consumer_tm1mwdpl04-1389222553159-ad59660b],
> Committing
> >> all offsets after clearing the fetcher queues
> >> {2014-01-10 00:58:11,298} DEBUG [catalina-exec-12-SendThread(
> >> tvip-m1-mw-zookeeper.dish.com:2181)] (ClientCnxn.java:759) - Got ping
> >> response for sessionid: 0x1437b2879870005 af ter 0ms
> >> {2014-01-10 00:58:11,313} INFO
> >>
> [account-activated-hadoop-consumer_tm1mwdpl04-1389222557906-562b6738_watcher_executor]
> >> (?:?) -
> >> [account-activated-hadoop-consumer_tm1mwdpl04-1389222557906-562b6738],
> >> begin rebalancing consumer
> >> account-activated-hadoop-consumer_tm1mwdpl04-1389222557906-562b6738 try
> #1
> >> {2014-01-10 00:58:11,314} DEBUG [catalina-exec-12-SendThread(
> >> tvip-m1-mw-zookeeper.dish.com:2181)] (ClientCnxn.java:839) - Reading
> >> reply sessionid:0x1437b2879870005, packet:: clientPath:null
> serverPath:null
> >> finished:false header:: 627,8  replyHeader:: 627,51539619966,0
>  request::
> >> '/brokers/ids,F  response:: v{'1}
> >> {2014-01-10 00:58:11,315} DEBUG [catalina-exec-12-SendThread(
> >> tvip-m1-mw-zookeeper.dish.com:2181)] (ClientCnxn.java:839) - Reading
> >> reply sessionid:0x1437b2879870005, packet:: clientPath:null
> serverPath:null
> >> finished:false header:: 628,4  replyHeader:: 628,51539619966,0
>  request::
> >> '/brokers/ids/1,F  response::
> >>
> #7b2022686f7374223a22746d312d6b61666b6162726f6b6572313031222c20226a6d785f706f7274223a393939392c2022706f7274223a393039322c202276657273696f6e223a31207d,s{47244644685,47244644685,1388537628753,1388537628753,0,0,0,163056791896588316,74,0,47244644685}
> >> {2014-01-10 00:58:11,316} DEBUG [catalina-exec-12-SendThread(
> >> tvip-m1-mw-zookeeper.dish.com:2181)] (ClientCnxn.java:839) - Reading
> >> reply sessionid:0x1437b2879870005, packet:: clientPath:null
> serverPath:null
> >> finished:false header:: 629,4  replyHeader:: 629,51539619966,-101
> >> request::
> >>
> '/consumers/account-activated-hadoop-consumer/ids/account-activated-hadoop-consumer_tm1mwdpl04-1389222557906-562b6738,F
> >> response::
> >> {2014-01-10 00:58:11,316} INFO
> >>
> [account-activated-hadoop-consumer_tm1mwdpl04-1389222557906-562b6738_watcher_executor]
> >> (?:?) -
> >> [account-activated-hadoop-consumer_tm1mwdpl04-1389222557906-562b6738],
> >> exception during rebalance
> >> org.I0Itec.zkclient.exception.ZkNoNodeException:
> >> org.apache.zookeeper.KeeperException$NoNodeException: KeeperErrorCode =
> >> NoNode for
> >>
> /consumers/account-activated-hadoop-consumer/ids/account-activated-hadoop-consumer_tm1mwdpl04-1389222557906-562b6738
> >>at
> >> org.I0Itec.zkclient.exception.ZkException.create(ZkException.java:47)
> >> 

RE: Looks like consumer fetchers get stopped we are not getting any data

2014-01-11 Thread Seshadri, Balaji
We found the below bug in Kafka JIRA.

https://issues.apache.org/jira/browse/KAFKA-693

Could you guys please let us know if this bug is causing the fetchers getting 
stooped because of rebalance failure,we are using 0.8-beta and this fix seems 
to be in 0.8 version.

The issue happens only when we run it in tomcat,when we use JUnits with same 
group.id to test we receive the messages.It would be great if you could help us 
confirm if it is a bug.

Also we would like to know if we need to upgrade the server/client to 0.8 or 
just server side code.



From: Rob Withers [robert.w.with...@gmail.com]
Sent: Friday, January 10, 2014 10:17 PM
To: users@kafka.apache.org
Subject: Re: Looks like consumer fetchers get stopped we are not getting any 
data

That was an interesting section too.  Which GC settings would you suggest?

Thank you,
- charlie

> On Jan 10, 2014, at 10:11 PM, Jun Rao  wrote:
>
> Have you looked at our FAQ, especially
> https://cwiki.apache.org/confluence/display/KAFKA/FAQ#FAQ-Whyaretheremanyrebalancesinmyconsumerlog
> ?
>
> Thanks,
>
> Jun
>
>
> On Fri, Jan 10, 2014 at 2:25 PM, Seshadri, Balaji
> wrote:
>
>> Any clue would be helpful.
>>
>> -Original Message-
>> From: Seshadri, Balaji [mailto:balaji.sesha...@dish.com]
>> Sent: Friday, January 10, 2014 12:46 PM
>> To: users@kafka.apache.org
>> Subject: RE: Looks like consumer fetchers get stopped we are not getting
>> any data
>>
>> Yes rebalance begins and exceptions occurs.
>>
>>
>> {2014-01-10 00:58:11,293} INFO
>> [account-info-updated-hadoop-consumer_tm1mwdpl04-1389222553159-ad59660b_watcher_executor]
>> (?:?) - [account-i
>> nfo-updated-hadoop-consumer_tm1mwdpl04-1389222553159-ad59660b], Cleared the
>> data chunks in all the consumer message iterators
>> {2014-01-10 00:58:11,293} INFO
>> [account-info-updated-hadoop-consumer_tm1mwdpl04-1389222553159-ad59660b_watcher_executor]
>> (?:?) - [account-i
>> nfo-updated-hadoop-consumer_tm1mwdpl04-1389222553159-ad59660b], Committing
>> all offsets after clearing the fetcher queues
>> {2014-01-10 00:58:11,298} DEBUG [catalina-exec-12-SendThread(
>> tvip-m1-mw-zookeeper.dish.com:2181)] (ClientCnxn.java:759) - Got ping
>> response for sessionid: 0x1437b2879870005 af ter 0ms
>> {2014-01-10 00:58:11,313} INFO
>> [account-activated-hadoop-consumer_tm1mwdpl04-1389222557906-562b6738_watcher_executor]
>> (?:?) -
>> [account-activated-hadoop-consumer_tm1mwdpl04-1389222557906-562b6738],
>> begin rebalancing consumer
>> account-activated-hadoop-consumer_tm1mwdpl04-1389222557906-562b6738 try #1
>> {2014-01-10 00:58:11,314} DEBUG [catalina-exec-12-SendThread(
>> tvip-m1-mw-zookeeper.dish.com:2181)] (ClientCnxn.java:839) - Reading
>> reply sessionid:0x1437b2879870005, packet:: clientPath:null serverPath:null
>> finished:false header:: 627,8  replyHeader:: 627,51539619966,0  request::
>> '/brokers/ids,F  response:: v{'1}
>> {2014-01-10 00:58:11,315} DEBUG [catalina-exec-12-SendThread(
>> tvip-m1-mw-zookeeper.dish.com:2181)] (ClientCnxn.java:839) - Reading
>> reply sessionid:0x1437b2879870005, packet:: clientPath:null serverPath:null
>> finished:false header:: 628,4  replyHeader:: 628,51539619966,0  request::
>> '/brokers/ids/1,F  response::
>> #7b2022686f7374223a22746d312d6b61666b6162726f6b6572313031222c20226a6d785f706f7274223a393939392c2022706f7274223a393039322c202276657273696f6e223a31207d,s{47244644685,47244644685,1388537628753,1388537628753,0,0,0,163056791896588316,74,0,47244644685}
>> {2014-01-10 00:58:11,316} DEBUG [catalina-exec-12-SendThread(
>> tvip-m1-mw-zookeeper.dish.com:2181)] (ClientCnxn.java:839) - Reading
>> reply sessionid:0x1437b2879870005, packet:: clientPath:null serverPath:null
>> finished:false header:: 629,4  replyHeader:: 629,51539619966,-101
>> request::
>> '/consumers/account-activated-hadoop-consumer/ids/account-activated-hadoop-consumer_tm1mwdpl04-1389222557906-562b6738,F
>> response::
>> {2014-01-10 00:58:11,316} INFO
>> [account-activated-hadoop-consumer_tm1mwdpl04-1389222557906-562b6738_watcher_executor]
>> (?:?) -
>> [account-activated-hadoop-consumer_tm1mwdpl04-1389222557906-562b6738],
>> exception during rebalance
>> org.I0Itec.zkclient.exception.ZkNoNodeException:
>> org.apache.zookeeper.KeeperException$NoNodeException: KeeperErrorCode =
>> NoNode for
>> /consumers/account-activated-hadoop-consumer/ids/account-activated-hadoop-consumer_tm1mwdpl04-1389222557906-562b6738
>>at
>> org.I0Itec.zkclient.exception.ZkException.create(ZkException.java:47)
>>at
>> org.I0Itec.zkclient.ZkClient.retryUntilConnected(ZkClient.java:685)
>>at org.I0Itec.zkclient.ZkClient.readData(ZkClient.java:766)
>>at org.I0Itec.zkclient.ZkClient.readData(ZkClient.java:761)
>>at kafka.utils.ZkUtils$.readData(Unknown Source)
>>at kafka.consumer.TopicCount$.constructTopicCount(Unknown Source)
>>at
>> kafka.consumer.ZookeeperConsumerConnector$ZKRebalancerListener.kafka$consumer$ZookeeperConsumerConnector$ZKRebala

Re: custom kafka consumer - strangeness

2014-01-11 Thread Gerrit Jansen van Vuuren
I'm also seeing the following.

I consume the data in the queue.
Then after 10 seconds send another fetch request (with the incremented
offset), and never receives a response from the broker, my code eventually
times out (after 30seconds).

The broker writes Expiring fetch request Name: FetchRequest; Version: 0;
CorrelationId: 1389443537; ClientId: 1; ReplicaId: -1; MaxWait: 1000 ms;
MinBytes: 1 bytes; RequestInfo: [ping,0] ->
PartitionFetchInfo(187,1048576).

This corresponds with the timed out fetch request.






On Sat, Jan 11, 2014 at 12:19 PM, Gerrit Jansen van Vuuren <
gerrit...@gmail.com> wrote:

> Hi,
>
>
> No the offsets are not the same. I've printed out the values to see this,
> and its not the case.
>
>
>
> On Fri, Jan 10, 2014 at 5:02 PM, Jun Rao  wrote:
>
>> Are the offset used in the 2 fetch requests the same? If so, you will get
>> the same messages twice. You consumer is responsible for advancing the
>> offsets after consumption.
>>
>> Thanks,
>>
>> Jun
>>
>>
>> On Thu, Jan 9, 2014 at 1:00 PM, Gerrit Jansen van Vuuren <
>> gerrit...@gmail.com> wrote:
>>
>> > Hi,
>> >
>> > I'm writing a custom consumer for kafka 0.8.
>> > Everything works except for the following:
>> >
>> > a. connect, send fetch, read all results
>> > b. send fetch
>> > c. send fetch
>> > d. send fetch
>> > e. via the console publisher, publish 2 messages
>> > f. send fetch :corr-id 1
>> > g. read 2 messages published :offsets [10 11] :corr-id 1
>> > h. send fetch :corr-id 2
>> > i. read 2 messages published :offsets [10 11] :corr-id 2
>> > j.  send fetch ...
>> >
>> > The problem is I get the messages sent twice as a response to two
>> separate
>> > fetch requests. The correlation id is distinct so it cannot be that I
>> read
>> > the response twice. The offsets of the 2 messages are are the same so
>> they
>> > are duplicates, and its not the producer sending the messages twice.
>> >
>> > Note: the same connection is kept open the whole time, and I send
>> > block,receive then send again, after the first 2 messages are read, the
>> > offsets are incremented and the next fetch will ask kafka to give it
>> > messages from the new offsets.
>> >
>> > any ideas of why kafka would be sending the messages again on the second
>> > fetch request?
>> >
>> > Regards,
>> >  Gerrit
>> >
>>
>
>


Re: custom kafka consumer - strangeness

2014-01-11 Thread Gerrit Jansen van Vuuren
Hi,


No the offsets are not the same. I've printed out the values to see this,
and its not the case.



On Fri, Jan 10, 2014 at 5:02 PM, Jun Rao  wrote:

> Are the offset used in the 2 fetch requests the same? If so, you will get
> the same messages twice. You consumer is responsible for advancing the
> offsets after consumption.
>
> Thanks,
>
> Jun
>
>
> On Thu, Jan 9, 2014 at 1:00 PM, Gerrit Jansen van Vuuren <
> gerrit...@gmail.com> wrote:
>
> > Hi,
> >
> > I'm writing a custom consumer for kafka 0.8.
> > Everything works except for the following:
> >
> > a. connect, send fetch, read all results
> > b. send fetch
> > c. send fetch
> > d. send fetch
> > e. via the console publisher, publish 2 messages
> > f. send fetch :corr-id 1
> > g. read 2 messages published :offsets [10 11] :corr-id 1
> > h. send fetch :corr-id 2
> > i. read 2 messages published :offsets [10 11] :corr-id 2
> > j.  send fetch ...
> >
> > The problem is I get the messages sent twice as a response to two
> separate
> > fetch requests. The correlation id is distinct so it cannot be that I
> read
> > the response twice. The offsets of the 2 messages are are the same so
> they
> > are duplicates, and its not the producer sending the messages twice.
> >
> > Note: the same connection is kept open the whole time, and I send
> > block,receive then send again, after the first 2 messages are read, the
> > offsets are incremented and the next fetch will ask kafka to give it
> > messages from the new offsets.
> >
> > any ideas of why kafka would be sending the messages again on the second
> > fetch request?
> >
> > Regards,
> >  Gerrit
> >
>


Re: Kafka with Docker - producer disconnecting

2014-01-11 Thread Thomas

Hi,

you can have a look at https://github.com/wurstmeister/kafka-docker
maybe that helps

On 10/01/2014 19:30, Alex Artigues wrote:

Hi everyone, I am attempting to run Zookeeper and Kafka in dockers.

Both startup normally and Kafka connects ok.  I think my containers are
linked fine because I am able to list topics, and create topics.

The producer however never delivers messages.  It connects using the
provided shell scripts, and it seems I am able to type messages, but I
cannot ctl+c to stop it.  in the logs there is an INFO message that the
client disconnects, but no good reason why.

This is in the ZK logs:
2014-01-10 17:55:39,893 - WARN  [NIOServerCxn.Factory:
0.0.0.0/0.0.0.0:2181:NIOServerCnxn@349] - caught end of stream exception
EndOfStreamException: Unable to read additional data from client sessionid
0x1437d4304fe, likely client has closed socket


If anyone has any clues I would really appreciate it, thanks.