[jira] [Updated] (HBASE-14326) HBase book: fix definition of max min size to compact

2015-10-21 Thread Misty Stanley-Jones (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-14326?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Misty Stanley-Jones updated HBASE-14326:

  Resolution: Fixed
Hadoop Flags: Reviewed
  Status: Resolved  (was: Patch Available)

Pushed to master. Thanks stack!

> HBase book: fix definition of max min size to compact
> -
>
> Key: HBASE-14326
> URL: https://issues.apache.org/jira/browse/HBASE-14326
> Project: HBase
>  Issue Type: Bug
>  Components: documentation
>Reporter: Vladimir Rodionov
>Assignee: Misty Stanley-Jones
> Fix For: 2.0.0
>
> Attachments: HBASE-14326.patch
>
>
> I think we need to change wording/definition of these config parameters in 
> HBase book, they are misleading:
> {quote}
> hbase.hstore.compaction.min.size
> Description
> A StoreFile smaller than this size will always be eligible for minor 
> compaction. HFiles this size or larger are evaluated by 
> hbase.hstore.compaction.ratio to determine if they are eligible. Because this 
> limit represents the "automatic include"limit for all StoreFiles smaller than 
> this value, this value may need to be reduced in write-heavy environments 
> where many StoreFiles in the 1-2 MB range are being flushed, because every 
> StoreFile will be targeted for compaction and the resulting StoreFiles may 
> still be under the minimum size and require further compaction. If this 
> parameter is lowered, the ratio check is triggered more quickly. This 
> addressed some issues seen in earlier versions of HBase but changing this 
> parameter is no longer necessary in most situations. Default: 128 MB 
> expressed in bytes.
> Default
> 134217728
> hbase.hstore.compaction.max.size
> Description
> A StoreFile larger than this size will be excluded from compaction. The 
> effect of raising hbase.hstore.compaction.max.size is fewer, larger 
> StoreFiles that do not get compacted often. If you feel that compaction is 
> happening too often without much benefit, you can try raising this value. 
> Default: the value of LONG.MAX_VALUE, expressed in bytes.
> hbase.hstore.compaction.ratio
> Description
> For minor compaction, this ratio is used to determine whether a given 
> StoreFile which is larger than hbase.hstore.compaction.min.size is eligible 
> for compaction. Its effect is to limit compaction of large StoreFiles. The 
> value of hbase.hstore.compaction.ratio is expressed as a floating-point 
> decimal. A large ratio, such as 10, will produce a single giant StoreFile. 
> Conversely, a low value, such as .25, will produce behavior similar to the 
> BigTable compaction algorithm, producing four StoreFiles. A moderate value of 
> between 1.0 and 1.4 is recommended. When tuning this value, you are balancing 
> write costs with read costs. Raising the value (to something like 1.4) will 
> have more write costs, because you will compact larger StoreFiles. However, 
> during reads, HBase will need to seek through fewer StoreFiles to accomplish 
> the read. Consider this approach if you cannot take advantage of Bloom 
> filters. Otherwise, you can lower this value to something like 1.0 to reduce 
> the background cost of writes, and use Bloom filters to control the number of 
> StoreFiles touched during reads. For most cases, the default value is 
> appropriate.
> Default
> 1.2F
> {quote}
> For details, see HBASE-14263.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HBASE-14326) HBase book: fix definition of max min size to compact

2015-10-11 Thread Misty Stanley-Jones (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-14326?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Misty Stanley-Jones updated HBASE-14326:

Attachment: HBASE-14326.patch

Hopefully this patch clarifies the docs.

> HBase book: fix definition of max min size to compact
> -
>
> Key: HBASE-14326
> URL: https://issues.apache.org/jira/browse/HBASE-14326
> Project: HBase
>  Issue Type: Bug
>  Components: documentation
>Reporter: Vladimir Rodionov
>Assignee: Misty Stanley-Jones
> Fix For: 2.0.0
>
> Attachments: HBASE-14326.patch
>
>
> I think we need to change wording/definition of these config parameters in 
> HBase book, they are misleading:
> {quote}
> hbase.hstore.compaction.min.size
> Description
> A StoreFile smaller than this size will always be eligible for minor 
> compaction. HFiles this size or larger are evaluated by 
> hbase.hstore.compaction.ratio to determine if they are eligible. Because this 
> limit represents the "automatic include"limit for all StoreFiles smaller than 
> this value, this value may need to be reduced in write-heavy environments 
> where many StoreFiles in the 1-2 MB range are being flushed, because every 
> StoreFile will be targeted for compaction and the resulting StoreFiles may 
> still be under the minimum size and require further compaction. If this 
> parameter is lowered, the ratio check is triggered more quickly. This 
> addressed some issues seen in earlier versions of HBase but changing this 
> parameter is no longer necessary in most situations. Default: 128 MB 
> expressed in bytes.
> Default
> 134217728
> hbase.hstore.compaction.max.size
> Description
> A StoreFile larger than this size will be excluded from compaction. The 
> effect of raising hbase.hstore.compaction.max.size is fewer, larger 
> StoreFiles that do not get compacted often. If you feel that compaction is 
> happening too often without much benefit, you can try raising this value. 
> Default: the value of LONG.MAX_VALUE, expressed in bytes.
> hbase.hstore.compaction.ratio
> Description
> For minor compaction, this ratio is used to determine whether a given 
> StoreFile which is larger than hbase.hstore.compaction.min.size is eligible 
> for compaction. Its effect is to limit compaction of large StoreFiles. The 
> value of hbase.hstore.compaction.ratio is expressed as a floating-point 
> decimal. A large ratio, such as 10, will produce a single giant StoreFile. 
> Conversely, a low value, such as .25, will produce behavior similar to the 
> BigTable compaction algorithm, producing four StoreFiles. A moderate value of 
> between 1.0 and 1.4 is recommended. When tuning this value, you are balancing 
> write costs with read costs. Raising the value (to something like 1.4) will 
> have more write costs, because you will compact larger StoreFiles. However, 
> during reads, HBase will need to seek through fewer StoreFiles to accomplish 
> the read. Consider this approach if you cannot take advantage of Bloom 
> filters. Otherwise, you can lower this value to something like 1.0 to reduce 
> the background cost of writes, and use Bloom filters to control the number of 
> StoreFiles touched during reads. For most cases, the default value is 
> appropriate.
> Default
> 1.2F
> {quote}
> For details, see HBASE-14263.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HBASE-14326) HBase book: fix definition of max min size to compact

2015-10-11 Thread Misty Stanley-Jones (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-14326?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Misty Stanley-Jones updated HBASE-14326:

Status: Patch Available  (was: Open)

> HBase book: fix definition of max min size to compact
> -
>
> Key: HBASE-14326
> URL: https://issues.apache.org/jira/browse/HBASE-14326
> Project: HBase
>  Issue Type: Bug
>  Components: documentation
>Reporter: Vladimir Rodionov
>Assignee: Misty Stanley-Jones
> Fix For: 2.0.0
>
> Attachments: HBASE-14326.patch
>
>
> I think we need to change wording/definition of these config parameters in 
> HBase book, they are misleading:
> {quote}
> hbase.hstore.compaction.min.size
> Description
> A StoreFile smaller than this size will always be eligible for minor 
> compaction. HFiles this size or larger are evaluated by 
> hbase.hstore.compaction.ratio to determine if they are eligible. Because this 
> limit represents the "automatic include"limit for all StoreFiles smaller than 
> this value, this value may need to be reduced in write-heavy environments 
> where many StoreFiles in the 1-2 MB range are being flushed, because every 
> StoreFile will be targeted for compaction and the resulting StoreFiles may 
> still be under the minimum size and require further compaction. If this 
> parameter is lowered, the ratio check is triggered more quickly. This 
> addressed some issues seen in earlier versions of HBase but changing this 
> parameter is no longer necessary in most situations. Default: 128 MB 
> expressed in bytes.
> Default
> 134217728
> hbase.hstore.compaction.max.size
> Description
> A StoreFile larger than this size will be excluded from compaction. The 
> effect of raising hbase.hstore.compaction.max.size is fewer, larger 
> StoreFiles that do not get compacted often. If you feel that compaction is 
> happening too often without much benefit, you can try raising this value. 
> Default: the value of LONG.MAX_VALUE, expressed in bytes.
> hbase.hstore.compaction.ratio
> Description
> For minor compaction, this ratio is used to determine whether a given 
> StoreFile which is larger than hbase.hstore.compaction.min.size is eligible 
> for compaction. Its effect is to limit compaction of large StoreFiles. The 
> value of hbase.hstore.compaction.ratio is expressed as a floating-point 
> decimal. A large ratio, such as 10, will produce a single giant StoreFile. 
> Conversely, a low value, such as .25, will produce behavior similar to the 
> BigTable compaction algorithm, producing four StoreFiles. A moderate value of 
> between 1.0 and 1.4 is recommended. When tuning this value, you are balancing 
> write costs with read costs. Raising the value (to something like 1.4) will 
> have more write costs, because you will compact larger StoreFiles. However, 
> during reads, HBase will need to seek through fewer StoreFiles to accomplish 
> the read. Consider this approach if you cannot take advantage of Bloom 
> filters. Otherwise, you can lower this value to something like 1.0 to reduce 
> the background cost of writes, and use Bloom filters to control the number of 
> StoreFiles touched during reads. For most cases, the default value is 
> appropriate.
> Default
> 1.2F
> {quote}
> For details, see HBASE-14263.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HBASE-14326) HBase book: fix definition of max min size to compact

2015-08-27 Thread Vladimir Rodionov (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-14326?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vladimir Rodionov updated HBASE-14326:
--
Description: 
I think we need to change wording/definition of these config parameters in 
HBase book, they are misleading:
{quote}
hbase.hstore.compaction.min.size
Description
A StoreFile smaller than this size will always be eligible for minor 
compaction. HFiles this size or larger are evaluated by 
hbase.hstore.compaction.ratio to determine if they are eligible. Because this 
limit represents the automatic includelimit for all StoreFiles smaller than 
this value, this value may need to be reduced in write-heavy environments where 
many StoreFiles in the 1-2 MB range are being flushed, because every StoreFile 
will be targeted for compaction and the resulting StoreFiles may still be under 
the minimum size and require further compaction. If this parameter is lowered, 
the ratio check is triggered more quickly. This addressed some issues seen in 
earlier versions of HBase but changing this parameter is no longer necessary in 
most situations. Default: 128 MB expressed in bytes.
Default
134217728
hbase.hstore.compaction.max.size
Description
A StoreFile larger than this size will be excluded from compaction. The effect 
of raising hbase.hstore.compaction.max.size is fewer, larger StoreFiles that do 
not get compacted often. If you feel that compaction is happening too often 
without much benefit, you can try raising this value. Default: the value of 
LONG.MAX_VALUE, expressed in bytes.
hbase.hstore.compaction.ratio
Description
For minor compaction, this ratio is used to determine whether a given StoreFile 
which is larger than hbase.hstore.compaction.min.size is eligible for 
compaction. Its effect is to limit compaction of large StoreFiles. The value of 
hbase.hstore.compaction.ratio is expressed as a floating-point decimal. A large 
ratio, such as 10, will produce a single giant StoreFile. Conversely, a low 
value, such as .25, will produce behavior similar to the BigTable compaction 
algorithm, producing four StoreFiles. A moderate value of between 1.0 and 1.4 
is recommended. When tuning this value, you are balancing write costs with read 
costs. Raising the value (to something like 1.4) will have more write costs, 
because you will compact larger StoreFiles. However, during reads, HBase will 
need to seek through fewer StoreFiles to accomplish the read. Consider this 
approach if you cannot take advantage of Bloom filters. Otherwise, you can 
lower this value to something like 1.0 to reduce the background cost of writes, 
and use Bloom filters to control the number of StoreFiles touched during reads. 
For most cases, the default value is appropriate.

Default
1.2F
{quote}

For details, see HBASE-14263.

  was:
I think we need to change wording/definition of these config parameters in 
HBase book, they are misleading:
{quote}
hbase.hstore.compaction.min.size
Description
A StoreFile smaller than this size will always be eligible for minor 
compaction. HFiles this size or larger are evaluated by 
hbase.hstore.compaction.ratio to determine if they are eligible. Because this 
limit represents the automatic includelimit for all StoreFiles smaller than 
this value, this value may need to be reduced in write-heavy environments where 
many StoreFiles in the 1-2 MB range are being flushed, because every StoreFile 
will be targeted for compaction and the resulting StoreFiles may still be under 
the minimum size and require further compaction. If this parameter is lowered, 
the ratio check is triggered more quickly. This addressed some issues seen in 
earlier versions of HBase but changing this parameter is no longer necessary in 
most situations. Default: 128 MB expressed in bytes.
Default
134217728
hbase.hstore.compaction.max.size
Description
A StoreFile larger than this size will be excluded from compaction. The effect 
of raising hbase.hstore.compaction.max.size is fewer, larger StoreFiles that do 
not get compacted often. If you feel that compaction is happening too often 
without much benefit, you can try raising this value. Default: the value of 
LONG.MAX_VALUE, expressed in bytes.
{quote}

For details, see HBASE-14263.


 HBase book: fix definition of max min size to compact
 -

 Key: HBASE-14326
 URL: https://issues.apache.org/jira/browse/HBASE-14326
 Project: HBase
  Issue Type: Bug
  Components: documentation
Reporter: Vladimir Rodionov
 Fix For: 2.0.0


 I think we need to change wording/definition of these config parameters in 
 HBase book, they are misleading:
 {quote}
 hbase.hstore.compaction.min.size
 Description
 A StoreFile smaller than this size will always be eligible for minor 
 compaction. HFiles this size or larger are evaluated by