[ 
https://issues.apache.org/jira/browse/HBASE-9208?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13739883#comment-13739883
 ] 

Dave Latham commented on HBASE-9208:
------------------------------------

The CleanerChore is shared code between cleaning up HLogs in .oldlogs (a single 
directory with many files) and cleaning up HFiles in archive (which can have 
many nested subdirectories).  The proposal here assumes that as long as we get 
file listings from HDFS first and then check for references to those files 
second that we're safe to delete the file if there are no such references (of 
course the existing code has the same assumption).  So that leaves a choice:

(a) We can do this batch for each directory at a time only which should still 
solve the issue for HLogs (so long as they stay in a flat directory) but 
wouldn't allow the HFile archive to use as much of the optimization. Or:

(b) We can first do a full recursive load of all files in the base directory, 
and call a batch filter across all those files.  Then, in order to remove now 
empty subdirectories we need to do some bookkeeping to associate which ones we 
believe may now be empty (of course some subdirectories may have new entries 
created in the mean time).

Since I'm focused for the moment on HLogs I'm planning on sticking with the 
simpler (a) for now, unless I hear some clamoring for the fuller solution (b).

One other question occurred when doing some work on CleanerChore for this.  If 
an IOException occurs during deleting an entry in the top level directory the 
chore continues attempting to delete the other entries.  However, if an 
IOException is thrown deleting an entry in a subdirectory it aborts trying 
other entries in the subdirectory.  I'd prefer to see consistency (and it would 
make it easier to share the code) if we either give up the entire current chore 
iteration on an unexpected file system IOException (i.e. we still tolerate 
IOException when deleting a non-empty directory) or alternately attempt to 
continue crawling all subentries.  Does anyone have an opinion on which is 
better (or disagree about making the behavior consistent between top level 
entries and sublevel entries)?
                
> ReplicationLogCleaner slow at large scale
> -----------------------------------------
>
>                 Key: HBASE-9208
>                 URL: https://issues.apache.org/jira/browse/HBASE-9208
>             Project: HBase
>          Issue Type: Improvement
>            Reporter: Dave Latham
>            Assignee: Dave Latham
>             Fix For: 0.94.12
>
>
> At a large scale the ReplicationLogCleaner fails to clean up .oldlogs as fast 
> as the cluster is producing them.  For each old HLog file that has been 
> replicated and should be deleted the ReplicationLogCleaner checks every 
> replication queue in ZooKeeper before removing it.  This means that as a 
> cluster scales up the number of files to delete scales as well as the time to 
> delete each file so the cleanup chore scales quadratically.  In our case it 
> reached the point where the oldlogs were growing faster than they were being 
> cleaned up.
> We're now running with a patch that allows the ReplicationLogCleaner to 
> refresh its list of files in the replication queues from ZooKeeper just once 
> for each batch of files the CleanerChore wants to evaluate.
> I'd propose updating FileCleanerDelegate to take a List<FileStatus> rather 
> than a single one at a time.  This would allow file cleaners that check an 
> external resource for references such as ZooKeeper (for 
> ReplicationLogCleaner) or HDFS (for SnapshotLogCleaner which looks like it 
> may also have similar trouble at scale) to load those references once per 
> batch rather than for every log.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

Reply via email to