[ https://issues.apache.org/jira/browse/HBASE-9208?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13739883#comment-13739883 ]
Dave Latham commented on HBASE-9208: ------------------------------------ The CleanerChore is shared code between cleaning up HLogs in .oldlogs (a single directory with many files) and cleaning up HFiles in archive (which can have many nested subdirectories). The proposal here assumes that as long as we get file listings from HDFS first and then check for references to those files second that we're safe to delete the file if there are no such references (of course the existing code has the same assumption). So that leaves a choice: (a) We can do this batch for each directory at a time only which should still solve the issue for HLogs (so long as they stay in a flat directory) but wouldn't allow the HFile archive to use as much of the optimization. Or: (b) We can first do a full recursive load of all files in the base directory, and call a batch filter across all those files. Then, in order to remove now empty subdirectories we need to do some bookkeeping to associate which ones we believe may now be empty (of course some subdirectories may have new entries created in the mean time). Since I'm focused for the moment on HLogs I'm planning on sticking with the simpler (a) for now, unless I hear some clamoring for the fuller solution (b). One other question occurred when doing some work on CleanerChore for this. If an IOException occurs during deleting an entry in the top level directory the chore continues attempting to delete the other entries. However, if an IOException is thrown deleting an entry in a subdirectory it aborts trying other entries in the subdirectory. I'd prefer to see consistency (and it would make it easier to share the code) if we either give up the entire current chore iteration on an unexpected file system IOException (i.e. we still tolerate IOException when deleting a non-empty directory) or alternately attempt to continue crawling all subentries. Does anyone have an opinion on which is better (or disagree about making the behavior consistent between top level entries and sublevel entries)? > ReplicationLogCleaner slow at large scale > ----------------------------------------- > > Key: HBASE-9208 > URL: https://issues.apache.org/jira/browse/HBASE-9208 > Project: HBase > Issue Type: Improvement > Reporter: Dave Latham > Assignee: Dave Latham > Fix For: 0.94.12 > > > At a large scale the ReplicationLogCleaner fails to clean up .oldlogs as fast > as the cluster is producing them. For each old HLog file that has been > replicated and should be deleted the ReplicationLogCleaner checks every > replication queue in ZooKeeper before removing it. This means that as a > cluster scales up the number of files to delete scales as well as the time to > delete each file so the cleanup chore scales quadratically. In our case it > reached the point where the oldlogs were growing faster than they were being > cleaned up. > We're now running with a patch that allows the ReplicationLogCleaner to > refresh its list of files in the replication queues from ZooKeeper just once > for each batch of files the CleanerChore wants to evaluate. > I'd propose updating FileCleanerDelegate to take a List<FileStatus> rather > than a single one at a time. This would allow file cleaners that check an > external resource for references such as ZooKeeper (for > ReplicationLogCleaner) or HDFS (for SnapshotLogCleaner which looks like it > may also have similar trouble at scale) to load those references once per > batch rather than for every log. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira