Jay, This makes perfect sense and it is very helpful.
thanks, Jason On Thu, Apr 18, 2013 at 12:20 PM, Jay Kreps <jay.kr...@gmail.com> wrote: > I think there is a misunderstanding here. The log file division is > completely transparent to the consumer. The consumer consumes by > offset (0,1,2,3,...) and these offsets span log files (i.e. there is a > single, monotonically increasing set of offsets over all log files in > a partition). In other words consumers ALWAYS consume across log files > based on the position (offset) in the set of all logs in the > partition. > > If you want fewer log files change the settings so that the log files > are bigger. However the only impact of this is fewer file handles, it > won't impact the consumer in any way at all. > > Not sure if that helps... > -Jay > > On Thu, Apr 18, 2013 at 8:36 AM, Jason Huang <jason.hu...@icare.com> > wrote: > > Number of open file handles is one thing. But I am thinking more of a use > > case where you may want to fetch information from two separate log files > of > > same topic and same partition. Could we use one high level consumer API > > call to achieve that? One naive idea is that if we can merge those two > log > > files (since they are of same topic and in partition I thought we might > be > > able to merge them) into one, then using one high level consumer API will > > do the job. > > > > > > On Thu, Apr 18, 2013 at 11:26 AM, Neha Narkhede <neha.narkh...@gmail.com > >wrote: > > > >> There is no way to merge those log files in 0.8. But why would you want > to > >> do that ? To reduce the number of open file handles ? > >> > >> Thanks, > >> Neha > >> > >> On Thursday, April 18, 2013, Jason Huang wrote: > >> > >> > Just curious - given the situation above, if you have two sets of .log > >> and > >> > .index files under the same topic and same partition, is there any > way to > >> > merge these two sets of log files and index files? > >> > > >> > > >> > On Thu, Apr 18, 2013 at 5:42 AM, Swapnil Ghike <sgh...@linkedin.com> > >> > wrote: > >> > > >> > > Otherwise if you are using the high level consumer > >> > > (ZookeeperConsumerConnector) in code, you will have to start a new > >> > > consumer with a different group id. > >> > > > >> > > Thanks, > >> > > Swapnil > >> > > > >> > > On 4/18/13 2:09 AM, "Swapnil Ghike" <sgh...@linkedin.com> wrote: > >> > > > >> > > >Snehalata, > >> > > > > >> > > >Are you using the console consumer? You can pass a --from-beginning > >> flag > >> > > >to the console consumer to get all the messages posted to that > topic. > >> > > > > >> > > >Thanks, > >> > > >Swapnil > >> > > > > >> > > >On 4/18/13 1:32 AM, "Snehalata Nagaje" > >> > > ><snehalata.nag...@harbingergroup.com> wrote: > >> > > > > >> > > >>Hi Neha, > >> > > >> > >> > > >>I was able to figure this out, this is happening due to time based > >> > > >>rolling. > >> > > >> > >> > > >>But now problem, when I fetch the messages for particular topic, > >> server > >> > > >>is > >> > > >>reading only one log file, returns those messages only. > >> > > >> > >> > > >>Is there any way I can combine all log files which are already > >> produced > >> > > >>into > >> > > >>one, so that I can get all messages posted in that topic. > >> > > >> > >> > > >>Thanks, > >> > > >>Snehalata > >> > > >> > >> > > >> > >> > > >>-----Original Message----- > >> > > >>From: Snehalata Nagaje [mailto: > snehalata.nag...@harbingergroup.com] > >> > > >>Sent: Thursday, April 18, 2013 11:41 AM > >> > > >>To: users@kafka.apache.org > >> > > >>Subject: RE: producer creating multiple log files after message > >> sending > >> > > >> > >> > > >>Thanks Neha, > >> > > >> > >> > > >>-----Original Message----- > >> > > >>From: Neha Narkhede [mailto:neha.narkh...@gmail.com] > >> > > >>Sent: Wednesday, April 17, 2013 9:08 PM > >> > > >>To: users@kafka.apache.org > >> > > >>Subject: Re: producer creating multiple log files after message > >> sending > >> > > >> > >> > > >>Hi Snehalata, > >> > > >> > >> > > >>Would you grep through the log4j files for the topic under > >> > > >>consideration and look for "Rolling.*due to". This will tell us > why > >> > > >>the log is getting rolled over. > >> > > >> > >> > > >>Thanks, > >> > > >>Neha > >> > > >> > >> > > >>On Wed, Apr 17, 2013 at 12:05 AM, Snehalata Nagaje > >> > > >><snehalata.nag...@harbingergroup.com> wrote: > >> > > >>> Hi, > >> > > >>> > >> > > >>> > >> > > >>> > >> > > >>> I am using kafak 0.8 version and sending messages to kafka > server > >> > using > >> > > >>> below code, > >> > > >>> > >> > > >>> > >> > > >>> > >> > > >>> Producer<Integer, String> producer = new Producer<Integer, > >> > String>(new > >> > > >>> ProducerConfig(props)); > >> > > >>> > >> > > >>> producer.send(new > KeyedMessage(topicName,partitionKey,jsonString)); > >> > > >>> > >> > > >>> producer.close(); > >> > > >>> > >> > > >>> > >> > > >>> > >> > > >>> I am able to post messages with, but after certain days, kafka > >> > creating > >> > > >>> multiple log files, messages are posted in new log file, > although > >> old > >> > > >>>file > >> > > >>> size is very less. > >> > > >>> > >> > > >>> > >> > > >>> > >> > > >>> What can be the reason for creating multiple log files? Does > >> message > >> > > >>format > >> > > >>> matters while sending message to kafka? > >> > > >>> > >> > > >>> > >> > > >>> > >> > > >>> Thanks, > >> > > >>> > >> > > >>> Snehalata > >> > > >>> > >> > > >>> > >> > > >>> Disclaimer: > >> > > >>> This e-mail may contain Privileged/Confidential information and > is > >> > > >>intended on > >> >