Apache9 commented on PR #4596:
URL: https://github.com/apache/hbase/pull/4596#issuecomment-1175618564

   > I think we have some other issues with the server side changes, I see 
multiple compaction failures:
   > 
   > ```
   > 2022-07-05 19:00:06,828 WARN  [20-longCompactions-0] hdfs.DFSClient - 
Failed to connect to dn1/10.118.172.92:50010 for block 
BP-1698803826-10.118.165.103-1656451448670:blk_1074099523_359025, add to 
deadNodes and continue. 
   > java.net.ConnectException: Connection refused
   >    at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
   >    at 
sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:716)
   >    at 
org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206)
   >    at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:539)
   >    at 
org.apache.hadoop.hdfs.DFSClient.newConnectedPeer(DFSClient.java:2913)
   >    at 
org.apache.hadoop.hdfs.client.impl.BlockReaderFactory.nextTcpPeer(BlockReaderFactory.java:851)
   >    at 
org.apache.hadoop.hdfs.client.impl.BlockReaderFactory.getRemoteBlockReaderFromTcp(BlockReaderFactory.java:753)
   >    at 
org.apache.hadoop.hdfs.client.impl.BlockReaderFactory.build(BlockReaderFactory.java:387)
   >    at 
org.apache.hadoop.hdfs.DFSInputStream.getBlockReader(DFSInputStream.java:843)
   >    at 
org.apache.hadoop.hdfs.DFSInputStream.blockSeekTo(DFSInputStream.java:772)
   >    at 
org.apache.hadoop.hdfs.DFSInputStream.seekToBlockSource(DFSInputStream.java:1825)
   >    at 
org.apache.hadoop.hdfs.DFSInputStream.readBuffer(DFSInputStream.java:1035)
   >    at 
org.apache.hadoop.hdfs.DFSInputStream.readWithStrategy(DFSInputStream.java:1074)
   >    at org.apache.hadoop.hdfs.DFSInputStream.read(DFSInputStream.java:1138)
   >    at 
org.apache.hadoop.fs.FSDataInputStream.read(FSDataInputStream.java:148)
   >    at 
org.apache.hadoop.hbase.io.util.BlockIOUtils.readWithExtra(BlockIOUtils.java:180)
   >    at 
org.apache.hadoop.hbase.io.hfile.HFileBlock$FSReaderImpl.readAtOffset(HFileBlock.java:1450)
   >    at 
org.apache.hadoop.hbase.io.hfile.HFileBlock$FSReaderImpl.readBlockDataInternal(HFileBlock.java:1681)
   >    at 
org.apache.hadoop.hbase.io.hfile.HFileBlock$FSReaderImpl.readBlockData(HFileBlock.java:1492)
   >    at 
org.apache.hadoop.hbase.io.hfile.HFileReaderImpl.readBlock(HFileReaderImpl.java:1308)
   >    at 
org.apache.hadoop.hbase.io.hfile.HFileReaderImpl$HFileScannerImpl.readNextDataBlock(HFileReaderImpl.java:739)
   >    at 
org.apache.hadoop.hbase.io.hfile.HFileReaderImpl$EncodedScanner.next(HFileReaderImpl.java:1480)
   >    at 
org.apache.hadoop.hbase.io.HalfStoreFileReader$1.next(HalfStoreFileReader.java:140)
   >    at 
org.apache.hadoop.hbase.regionserver.StoreFileScanner.next(StoreFileScanner.java:194)
   >    at 
org.apache.hadoop.hbase.regionserver.KeyValueHeap.next(KeyValueHeap.java:112)
   >    at 
org.apache.hadoop.hbase.regionserver.StoreScanner.next(StoreScanner.java:677)
   >    at 
org.apache.hadoop.hbase.regionserver.compactions.Compactor.performCompaction(Compactor.java:417)
   >    at 
org.apache.hadoop.hbase.regionserver.compactions.Compactor.compact(Compactor.java:345)
   >    at 
org.apache.hadoop.hbase.regionserver.compactions.DefaultCompactor.compact(DefaultCompactor.java:65)
   >    at 
org.apache.hadoop.hbase.regionserver.DefaultStoreEngine$DefaultCompactionContext.compact(DefaultStoreEngine.java:122)
   >    at org.apache.hadoop.hbase.regionserver.HStore.compact(HStore.java:1464)
   >    at 
org.apache.hadoop.hbase.regionserver.HRegion.compact(HRegion.java:2286)
   >    at 
org.apache.hadoop.hbase.regionserver.CompactSplit$CompactionRunner.doCompaction(CompactSplit.java:618)
   >    at 
org.apache.hadoop.hbase.regionserver.CompactSplit$CompactionRunner.run(CompactSplit.java:666)
   >    at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
   >    at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
   >    at java.lang.Thread.run(Thread.java:750)
   > ```
   > 
   > ```
   > 2022-07-05 18:59:43,010 WARN  [20-longCompactions-0] 
impl.BlockReaderFactory - I/O error constructing remote block reader.
   > java.net.ConnectException: Connection refused
   >    at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
   >    at 
sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:716)
   >    at 
org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206)
   >    at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:539)
   >    at 
org.apache.hadoop.hdfs.DFSClient.newConnectedPeer(DFSClient.java:2913)
   >    at 
org.apache.hadoop.hdfs.client.impl.BlockReaderFactory.nextTcpPeer(BlockReaderFactory.java:851)
   >    at 
org.apache.hadoop.hdfs.client.impl.BlockReaderFactory.getRemoteBlockReaderFromTcp(BlockReaderFactory.java:753)
   >    at 
org.apache.hadoop.hdfs.client.impl.BlockReaderFactory.build(BlockReaderFactory.java:387)
   >    at 
org.apache.hadoop.hdfs.DFSInputStream.getBlockReader(DFSInputStream.java:843)
   >    at 
org.apache.hadoop.hdfs.DFSInputStream.blockSeekTo(DFSInputStream.java:772)
   >    at 
org.apache.hadoop.hdfs.DFSInputStream.seekToBlockSource(DFSInputStream.java:1825)
   >    at 
org.apache.hadoop.hdfs.DFSInputStream.readBuffer(DFSInputStream.java:1035)
   >    at 
org.apache.hadoop.hdfs.DFSInputStream.readWithStrategy(DFSInputStream.java:1074)
   >    at org.apache.hadoop.hdfs.DFSInputStream.read(DFSInputStream.java:1138)
   >    at 
org.apache.hadoop.fs.FSDataInputStream.read(FSDataInputStream.java:148)
   >    at 
org.apache.hadoop.hbase.io.util.BlockIOUtils.readWithExtra(BlockIOUtils.java:180)
   >    at 
org.apache.hadoop.hbase.io.hfile.HFileBlock$FSReaderImpl.readAtOffset(HFileBlock.java:1450)
   >    at 
org.apache.hadoop.hbase.io.hfile.HFileBlock$FSReaderImpl.readBlockDataInternal(HFileBlock.java:1681)
   >    at 
org.apache.hadoop.hbase.io.hfile.HFileBlock$FSReaderImpl.readBlockData(HFileBlock.java:1492)
   >    at 
org.apache.hadoop.hbase.io.hfile.HFileReaderImpl.readBlock(HFileReaderImpl.java:1308)
   >    at 
org.apache.hadoop.hbase.io.hfile.HFileReaderImpl$HFileScannerImpl.readNextDataBlock(HFileReaderImpl.java:739)
   >    at 
org.apache.hadoop.hbase.io.hfile.HFileReaderImpl$EncodedScanner.next(HFileReaderImpl.java:1480)
   >    at 
org.apache.hadoop.hbase.regionserver.StoreFileScanner.next(StoreFileScanner.java:194)
   >    at 
org.apache.hadoop.hbase.regionserver.KeyValueHeap.next(KeyValueHeap.java:112)
   >    at 
org.apache.hadoop.hbase.regionserver.StoreScanner.next(StoreScanner.java:677)
   >    at 
org.apache.hadoop.hbase.regionserver.compactions.Compactor.performCompaction(Compactor.java:417)
   >    at 
org.apache.hadoop.hbase.regionserver.compactions.Compactor.compact(Compactor.java:345)
   >    at 
org.apache.hadoop.hbase.regionserver.compactions.DefaultCompactor.compact(DefaultCompactor.java:65)
   >    at 
org.apache.hadoop.hbase.regionserver.DefaultStoreEngine$DefaultCompactionContext.compact(DefaultStoreEngine.java:122)
   >    at org.apache.hadoop.hbase.regionserver.HStore.compact(HStore.java:1464)
   >    at 
org.apache.hadoop.hbase.regionserver.HRegion.compact(HRegion.java:2286)
   >    at 
org.apache.hadoop.hbase.regionserver.CompactSplit$CompactionRunner.doCompaction(CompactSplit.java:618)
   >    at 
org.apache.hadoop.hbase.regionserver.CompactSplit$CompactionRunner.run(CompactSplit.java:666)
   >    at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
   >    at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
   >    at java.lang.Thread.run(Thread.java:750)
   > ```
   > 
   > Multiple BP abandon errors as well:
   > 
   > ```
   > 2022-07-05 18:59:44,696 WARN  [Thread-1126] hdfs.DataStreamer - Abandoning 
BP-1698803826-10.118.165.103-1656451448670:blk_1074101424_360934
   > ```
   > 
   > Basically HDFS RPC calls are impacted.
   
   The changes here do not affect  the DFSClient part. Especially, the above 
failures are in the compaction thread pool, not in the rpc handler threads, 
which are not affected by the changes here too. Here we just changed the 
implementation of the decode part in our rpc server implementation when SASL is 
enabled.
   
   So I guess it is just some other problems happen at the same time...


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@hbase.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org

Reply via email to