Hello,

I have met StackOverflowError in region server.
Detail Error log here...

HBase version is 1.2.6

DAYS:36,787 DEBUG [regionserver/longCompactions] 
regionserver.CompactSplitThread: Not compacting xxx. because compaction request 
was cancelled
DAYS:36,787 DEBUG [regionserver/shortCompactions] 
compactions.ExploringCompactionPolicy: Exploring compaction algori
thm has selected 0 files of size 0 starting at candidate #-1 after considering 
3 permutations with 0 in ratio
DAYS:36,787 DEBUG [regionserver/shortCompactions] 
compactions.RatioBasedCompactionPolicy: Not compacting files becau
se we only have 0 files ready for compaction. Need 3 to initiate.
DAYS:36,787 DEBUG [regionserver/shortCompactions] 
regionserver.CompactSplitThread: Not compacting xxx. because compaction request 
was cancelled
DAYS:38,028 ERROR [B.defaultRpcServer.handler=x,queue=x,port=x] ipc.RpcServer: 
Unexpected throwable object
java.lang.StackOverflowError
        at 
org.apache.hadoop.hdfs.shortcircuit.ShortCircuitCache.unref(ShortCircuitCache.java:525)
        at 
org.apache.hadoop.hdfs.shortcircuit.ShortCircuitReplica.unref(ShortCircuitReplica.java:141)
        at 
org.apache.hadoop.hdfs.BlockReaderLocal.close(BlockReaderLocal.java:644)
        at 
org.apache.hadoop.hdfs.DFSInputStream.closeCurrentBlockReader(DFSInputStream.java:1682)
        at 
org.apache.hadoop.hdfs.DFSInputStream.blockSeekTo(DFSInputStream.java:616)
        at 
org.apache.hadoop.hdfs.DFSInputStream.readWithStrategy(DFSInputStream.java:882)
        at org.apache.hadoop.hdfs.DFSInputStream.read(DFSInputStream.java:934)
        at java.io.DataInputStream.read(DataInputStream.java:149)
        at 
org.apache.hadoop.hbase.io.hfile.HFileBlock.readWithExtra(HFileBlock.java:709)
        at 
org.apache.hadoop.hbase.io.hfile.HFileBlock$AbstractFSReader.readAtOffset(HFileBlock.java:1440)
        at 
org.apache.hadoop.hbase.io.hfile.HFileBlock$FSReaderImpl.readBlockDataInternal(HFileBlock.java:1648)
        at 
org.apache.hadoop.hbase.io.hfile.HFileBlock$FSReaderImpl.readBlockData(HFileBlock.java:1532)
        at 
org.apache.hadoop.hbase.io.hfile.HFileReaderV2.readBlock(HFileReaderV2.java:452)
        at 
org.apache.hadoop.hbase.io.hfile.HFileBlockIndex$BlockIndexReader.loadDataBlockWithScanInfo(HFileBlockIndex.java:271)
        at 
org.apache.hadoop.hbase.io.hfile.HFileReaderV2$AbstractScannerV2.seekTo(HFileReaderV2.java:649)
        at 
org.apache.hadoop.hbase.io.hfile.HFileReaderV2$AbstractScannerV2.seekTo(HFileReaderV2.java:599)
        at 
org.apache.hadoop.hbase.regionserver.StoreFileScanner.seekAtOrAfter(StoreFileScanner.java:268)
        at 
org.apache.hadoop.hbase.regionserver.StoreFileScanner.seekToPreviousRow(StoreFileScanner.java:461)
        at 
org.apache.hadoop.hbase.regionserver.StoreFileScanner.seekToPreviousRow(StoreFileScanner.java:476)
        at 
org.apache.hadoop.hbase.regionserver.StoreFileScanner.seekToPreviousRow(StoreFileScanner.java:476)
        at 
org.apache.hadoop.hbase.regionserver.StoreFileScanner.seekToPreviousRow(StoreFileScanner.java:476)
        at 
org.apache.hadoop.hbase.regionserver.StoreFileScanner.seekToPreviousRow(StoreFileScanner.java:476)
        at 
org.apache.hadoop.hbase.regionserver.StoreFileScanner.seekToPreviousRow(StoreFileScanner.java:476)
        at 
org.apache.hadoop.hbase.regionserver.StoreFileScanner.seekToPreviousRow(StoreFileScanner.java:476)
        at 
org.apache.hadoop.hbase.regionserver.StoreFileScanner.seekToPreviousRow(StoreFileScanner.java:476)
        at 
org.apache.hadoop.hbase.regionserver.StoreFileScanner.seekToPreviousRow(StoreFileScanner.java:476)

Has anyone experienced similar problems?

Best regards,
Minwoo Kang

Reply via email to