[ https://issues.apache.org/jira/browse/HBASE-5699?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
stack updated HBASE-5699: ------------------------- Release Note: HBase's write-ahead-log (WAL) can now be configured to use multiple HDFS pipelines in parallel to provide better write throughput for clusters by using additional disks. By default, HBase will still use only a single HDFS-based WAL. To run with multiple WALs, alter the hbase-site.xml property "hbase.wal.provider" to have the value "multiwal". To return to having HBase determine what kind of WAL implementation to use either remove the property all together or set it to "defaultProvider". Altering the WAL provider used by a particular RegionServer requires restarting that instance. RegionServers using the original WAL implementation and those using the "multiwal" implementation can each handle recovery of either set of WALs, so a zero-downtime configuration update is possible through a rolling restart. This issue introduces the following configurations: hbase.wal.regiongrouping.numgroups is how many provider instances the 'multiwal' should create. Default is two. hbase.wal.regiongrouping.strategy is the strategy to use figuring which provider instance an edit should go to. Default is 'identity'. Identity is the only current built-in option. hbase.wal.regiongrouping.delegate is the type of the provider the multiwal creates. Default is default from WALFactory. was: HBase's write-ahead-log (WAL) can now be configured to use multiple HDFS pipelines in parallel to provide better write throughput for clusters by using additional disks. By default, HBase will still use only a single HDFS-based WAL. To run with multiple WALs, alter the hbase-site.xml property "hbase.wal.provider" to have the value "multiwal". To return to having HBase determine what kind of WAL implementation to use either remove the property all together or set it to "defaultProvider". Altering the WAL provider used by a particular RegionServer requires restarting that instance. RegionServers using the original WAL implementation and those using the "multiwal" implementation can each handle recovery of either set of WALs, so a zero-downtime configuration update is possible through a rolling restart. > Run with > 1 WAL in HRegionServer > --------------------------------- > > Key: HBASE-5699 > URL: https://issues.apache.org/jira/browse/HBASE-5699 > Project: HBase > Issue Type: Improvement > Components: Performance, wal > Reporter: binlijin > Assignee: Sean Busbey > Priority: Critical > Fix For: 1.0.0, 1.1.0 > > Attachments: HBASE-5699.3.patch.txt, HBASE-5699.4.patch.txt, > HBASE-5699_#workers_vs_MiB_per_s_1x1col_512Bval_wal_count_1,2,4.tiff, > HBASE-5699_disabled_and_regular_#workers_vs_MiB_per_s_1x1col_512Bval_wal_count_1,2,4.tiff, > HBASE-5699_write_iops_multiwal-1_1_to_200_threads.tiff, > HBASE-5699_write_iops_multiwal-2_10,50,120,190,260,330,400_threads.tiff, > HBASE-5699_write_iops_multiwal-4_10,50,120,190,260,330,400_threads.tiff, > HBASE-5699_write_iops_multiwal-6_10,50,120,190,260,330,400_threads.tiff, > HBASE-5699_write_iops_upstream_1_to_200_threads.tiff, PerfHbase.txt, > hbase-5699_multiwal_400-threads_stats_sync_heavy.txt, > hbase-5699_total_throughput_sync_heavy.txt, > results-hbase5699-upstream.txt.bz2, results-hbase5699-wals-1.txt.bz2, > results-updated-hbase5699-wals-2.txt.bz2, > results-updated-hbase5699-wals-4.txt.bz2, > results-updated-hbase5699-wals-6.txt.bz2 > > -- This message was sent by Atlassian JIRA (v7.6.3#76005)