Older files will be purged by default HBase compactor if all data inside
expired (you have TTL for data?)

As for custom compaction policy you can refer to FIFOCompactionPolicy class
to get the idea how custom compaction works.

-Vlad


On Wed, Mar 22, 2017 at 12:29 PM, jeff saremi <jeffsar...@hotmail.com>
wrote:

> I mentioned some of this in another thread. We have a readonly database
> which get bulk loaded using HFiles.
> We want to keep only two versions/generations of data. Since the size of
> data is massive we need to delete the older generation.
>
> Since we write one single HBase for each region for each CF for each
> generation, could we just yank the older files using a separate standalone
> process? (sounds a little scary)
>
> If not, could we write a custom compactor? what's involved (some pointers
> please)? thanks
>
> Jeff
>

Reply via email to