Uh Oh! Its showing the reporter of the problem as me but it wasn't me in
reality :) I am not able to modify it. Please feel free to change it :)

Cheers,
Akash A

On Sun, Oct 2, 2011 at 10:38 AM, Akash Ashok <thehellma...@gmail.com> wrote:

> I've opened up a JIRA for this
> https://issues.apache.org/jira/browse/HBASE-4529
>
> Cheers,
> Akash A
>
>
> On Sun, Oct 2, 2011 at 6:04 AM, karthik tunga <karthik.tu...@gmail.com>wrote:
>
>> Hey Stack,
>>
>> Along with the log replaying part, logic is also needed for log roll over.
>> This, I think, easier compared to the merging of the logs. Any edits less
>> than the last sequence number  on the file system can be removed from all
>> the WALs.
>>
>> Cheers,
>> Karthik
>>
>> On 1 October 2011 18:05, Jesse Yates <jesse.k.ya...@gmail.com> wrote:
>>
>> > I think adding the abstraction layer and making it not only pluggable,
>> but
>> > configurable would be great.
>> >
>> >  It would be nice to be able to tie into a service that logs directly to
>> > disk, rather than go through HDFS giving some potentially awesome
>> speedup
>> > at
>> > the cost of having to write a logging service that handles replication,
>> > etc.
>> > Side note, Accumulo is using their own service to storing the WAL,
>> rather
>> > than HDFS and I suspect that plays a big role in people's claim of its
>> > ability to do 'outperform' HBase.
>> >
>> > -Jesse Yates
>> >
>> > On Sat, Oct 1, 2011 at 2:04 PM, Stack <st...@duboce.net> wrote:
>> >
>> > > Yes.  For sure.  Would need to check that the split can deal w/
>> > > multiple logs written by the one server concurrently (sort by sequence
>> > > edit id after sorting on all the rest that makes up a wal log key).
>> > >
>> > > St.Ack
>> > >
>> > > On Sat, Oct 1, 2011 at 1:36 PM, karthik tunga <
>> karthik.tu...@gmail.com>
>> > > wrote:
>> > > > Hey,
>> > > >
>> > > > Doesn't multiple WALs need some kind of merging when recovering from
>> a
>> > > crash
>> > > > ?
>> > > >
>> > > > Cheers,
>> > > > Karthik
>> > > >
>> > > >
>> > > > On 1 October 2011 15:17, Stack <st...@duboce.net> wrote:
>> > > >
>> > > >> +1 on making WAL pluggable so we can experiment.  Being able to
>> write
>> > > >> multiple WALs at once should be easy enough to do (the WAL split
>> code
>> > > >> should be able to handle it). Also a suggestion made a while back
>> was
>> > > >> making it so hbase could be configured to write two filesystems --
>> > > >> there'd be hbase.rootdir as now -- and then we'd allow specifying
>> > > >> another fs to use for writing WALs (If not specified, we'd just use
>> > > >> hbase.rootdir for all filesystem interactions as now).
>> > > >>
>> > > >> St.Ack
>> > > >>
>> > > >> On Sat, Oct 1, 2011 at 10:56 AM, Dhruba Borthakur <
>> dhr...@gmail.com>
>> > > >> wrote:
>> > > >> > I have been experimenting with the WAL settings too. It is
>> obvious
>> > > that
>> > > >> > turning off the wal makes ur transactions go faster, HDFS
>> write/sync
>> > > are
>> > > >> not
>> > > >> > yet very optimized for high throughput small writes.
>> > > >> >
>> > > >> > However, irrespective of whether I have one wal or two, I have
>> > seeing
>> > > the
>> > > >> > same throughput. I have experimented with an HDFS setting that
>> > allows
>> > > >> > writing/sync to multiple replicas in parallel, and that has
>> > increased
>> > > >> > performance for my test workload, see
>> > > >> > https://issues.apache.org/jira/browse/HDFS-1783.
>> > > >> >
>> > > >> > About using one wal or two, it will be nice if we can separate
>> out
>> > the
>> > > >> wal
>> > > >> > API elegantly and make it pluggable. In that case, we can
>> experiment
>> > > >> HBase
>> > > >> > with multiple systems. Once we have it pluggable, we can make the
>> > > habse
>> > > >> wal
>> > > >> > go to a separate HDFS (pure SSD based maybe?).
>> > > >> >
>> > > >> > -dhruba
>> > > >> >
>> > > >> >
>> > > >> > On Sat, Oct 1, 2011 at 8:09 AM, Akash Ashok <
>> thehellma...@gmail.com
>> > >
>> > > >> wrote:
>> > > >> >
>> > > >> >> Hey,
>> > > >> >> I've see that setting writeToWAL(false) boosts up the writes
>> like
>> > > crazy.
>> > > >> I
>> > > >> >> was just thinking having MuiltipleWAL on HBase. I understand
>> that
>> > > this
>> > > >> is a
>> > > >> >> consideration in BigTable paper that a WAL per region is not
>> used
>> > > >> because
>> > > >> >> it
>> > > >> >> might result in a lot of disk seeks when there are large number
>> of
>> > > >> reasons.
>> > > >> >> But how about having as many WALs as the number of HardDrives in
>> > the
>> > > >> >> system.
>> > > >> >> I see that the recommended configs for HBase are 4 - 12 hard
>> drives
>> > > per
>> > > >> >> node. This might kick the writes up a notch.
>> > > >> >>
>> > > >> >> Would like to know the general opinion on this one?
>> > > >> >>
>> > > >> >> Cheers,
>> > > >> >> Akash A
>> > > >> >>
>> > > >> >
>> > > >> >
>> > > >> >
>> > > >> > --
>> > > >> > Connect to me at http://www.facebook.com/dhruba
>> > > >> >
>> > > >>
>> > > >
>> > >
>> >
>>
>
>

Reply via email to