>> couldn't this be written in a way that it doesn't block?

This can't be done since file channels are blocking in nature.

Thanks,
Neha

On Wed, May 9, 2012 at 9:35 PM, S Ahmed <sahmed1...@gmail.com> wrote:

> couldn't this be written in a way that it doesn't block? or limits the
> time, like it makes a copy of the reference, then replaces it when a newed
> up channel (synchronized).
>
> Oh, this isn't possible because the var is mapped to a file at the o/s
> level?
>
> On Tue, May 8, 2012 at 5:38 PM, Jay Kreps <jay.kr...@gmail.com> wrote:
>
> > filechannel.force() always fully syncs the file to disk. This is done
> > irrespective of message boundaries. The file is locked during this
> > time so other appends are blocked.
> >
> > -Jay
> >
> > On Tue, May 8, 2012 at 1:44 PM, S Ahmed <sahmed1...@gmail.com> wrote:
> > > Slowly trying to understand it, have to wramp up on my scala.
> > >
> > > When the flush/sink occurrs, does it pull items of the collection 1 by
> 1
> > or
> > > does it do this in bulk somehow while locking the collection?
> > >
> > > On Mon, May 7, 2012 at 3:14 PM, Neha Narkhede <neha.narkh...@gmail.com
> > >wrote:
> > >
> > >> Ahmed,
> > >>
> > >> The related code is in kafka.log.*. The message to file persistence is
> > >> inside FileMessageSet.scala.
> > >>
> > >> Thanks,
> > >> Neha
> > >>
> > >> On Mon, May 7, 2012 at 12:12 PM, S Ahmed <sahmed1...@gmail.com>
> wrote:
> > >>
> > >> > I can barely read scala, but I'm curious where the applications
> > performs
> > >> > the operation of taking the in-memory log and persisting it to the
> > >> > database, all the while accepting new log messages and removing the
> > keys
> > >> of
> > >> > the messages that have been persisted to disk.
> > >> >
> > >> > I'm guessing you have used the concurrenthashmap where the key is a
> > >> topic,
> > >> > and once the flush timeout has been reached a background thread will
> > >> >  somehow persist and remove the keys.
> > >> >
> > >>
> >
>

Reply via email to