[ https://issues.apache.org/jira/browse/HBASE-22072?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16799144#comment-16799144 ]
Anoop Sam John edited comment on HBASE-22072 at 3/22/19 4:06 PM: ----------------------------------------------------------------- bq.It surprise me a little, because some files in a list compacted 3 days ago, marked as "compactedAway", so new scanners should not read them, and existing scanners should release reference as far as hbase.client.scanner.timeout.period achived. I am going to inspect source of hbase client, maybe there is a bug inside and hbase.client.scanner.timeout.period setting ignored if scaner.close() was not called. That is a real surprise that file is not deleted by the Chore service even after long time !! I just did a code read.. Seems we dont consider whether a file is compactedAway one or not before opening a new scanner on it when scan req comes for this Store. Seems this is causing the issue. {code} public List<KeyValueScanner> getScanners(boolean cacheBlocks, boolean usePread, boolean isCompaction, ScanQueryMatcher matcher, byte[] startRow, boolean includeStartRow, byte[] stopRow, boolean includeStopRow, long readPt) throws IOException { Collection<HStoreFile> storeFilesToScan; List<KeyValueScanner> memStoreScanners; this.lock.readLock().lock(); try { storeFilesToScan = this.storeEngine.getStoreFileManager().getFilesForScan(startRow, includeStartRow, stopRow, includeStopRow); memStoreScanners = this.memstore.getScanners(readPt); } finally { this.lock.readLock().unlock(); } .... {code} {code} public final Collection<HStoreFile> getFilesForScan(byte[] startRow, boolean includeStartRow, byte[] stopRow, boolean includeStopRow) { // We cannot provide any useful input and already have the files sorted by seqNum. return getStorefiles(); } public final Collection<HStoreFile> getStorefiles() { return storefiles; } {code} was (Author: anoop.hbase): bq.It surprise me a little, because some files in a list compacted 3 days ago, marked as "compactedAway", so new scanners should not read them, and existing scanners should release reference as far as {{hbase.client.scanner.timeout.period }}achived. I am going to inspect source of hbase client, maybe there is a bug inside and {{hbase.client.scanner.timeout.period}} setting ignored if scaner.close() was not called.{{}} That is a real surprise that file is not deleted by the Chore service even after long time !! I just did a code read.. Seems we dont consider whether a file is compactedAway one or not before opening a new scanner on it when scan req comes for this Store. Seems this is causing the issue. {code} public List<KeyValueScanner> getScanners(boolean cacheBlocks, boolean usePread, boolean isCompaction, ScanQueryMatcher matcher, byte[] startRow, boolean includeStartRow, byte[] stopRow, boolean includeStopRow, long readPt) throws IOException { Collection<HStoreFile> storeFilesToScan; List<KeyValueScanner> memStoreScanners; this.lock.readLock().lock(); try { storeFilesToScan = this.storeEngine.getStoreFileManager().getFilesForScan(startRow, includeStartRow, stopRow, includeStopRow); memStoreScanners = this.memstore.getScanners(readPt); } finally { this.lock.readLock().unlock(); } .... {code} {code} public final Collection<HStoreFile> getFilesForScan(byte[] startRow, boolean includeStartRow, byte[] stopRow, boolean includeStopRow) { // We cannot provide any useful input and already have the files sorted by seqNum. return getStorefiles(); } public final Collection<HStoreFile> getStorefiles() { return storefiles; } {code} > High read/write intensive regions may cause long crash recovery > --------------------------------------------------------------- > > Key: HBASE-22072 > URL: https://issues.apache.org/jira/browse/HBASE-22072 > Project: HBase > Issue Type: Bug > Components: Performance, Recovery > Affects Versions: 2.1.2 > Reporter: Pavel > Priority: Major > > Compaction of high read loaded region may leave compacted files undeleted > because of existing scan references: > INFO org.apache.hadoop.hbase.regionserver.HStore - Can't archive compacted > file hdfs://hdfs-ha/hbase... because of either isCompactedAway=true or file > has reference, isReferencedInReads=true, refCount=1, skipping for now > If region is either high write loaded this happens quite often and region may > have few storefiles and tons of undeleted compacted hdfs files. > Region keeps all that files (in my case thousands) untill graceful region > closing procedure, which ignores existing references and drop obsolete files. > It works fine unless consuming some extra hdfs space, but only in case of > normal region closing. If region server crashes than new region server, > responsible for that overfiling region, reads hdfs folder and try to deal > with all undeleted files, producing tons of storefiles, compaction tasks and > consuming abnormal amount of memory, wich may lead to OutOfMemory Exception > and further region servers crash. This stops writing to region because number > of storefiles reach *hbase.hstore.blockingStoreFiles* limit, forces high GC > duty and may take hours to compact all files into working set of files. > Workaround is a periodically check hdfs folders files count and force region > assign for ones with too many files. > It could be nice if regionserver had a setting similar to > hbase.hstore.blockingStoreFiles and invoke attempt to drop undeleted > compacted files if number of files reaches this setting. -- This message was sent by Atlassian JIRA (v7.6.3#76005)