yuangjiang opened a new issue, #6265:
URL: https://github.com/apache/iceberg/issues/6265

   ### Apache Iceberg version
   
   _No response_
   
   ### Query engine
   
   _No response_
   
   ### Please describe the bug 🐞
   
   I created the iceberg table using spark 3.3.1 iceberg hadoop catalog. The 
following exception message appears when I delete. 
   
   22/11/24 17:22:29 WARN DFSClient: Failed to connect to /192.168.x.x:1004 for 
block, add to deadNodes and continue. 
java.nio.channels.ClosedByInterruptException
        java.nio.channels.ClosedByInterruptException
                at 
java.nio.channels.spi.AbstractInterruptibleChannel.end(AbstractInterruptibleChannel.java:202)
                at 
sun.nio.ch.SocketChannelImpl.connect(SocketChannelImpl.java:659)
                at 
org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:192)
                at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:531)
                at 
org.apache.hadoop.hdfs.DFSClient.newConnectedPeer(DFSClient.java:3441)
                at 
org.apache.hadoop.hdfs.BlockReaderFactory.nextTcpPeer(BlockReaderFactory.java:777)
                at 
org.apache.hadoop.hdfs.BlockReaderFactory.getRemoteBlockReaderFromTcp(BlockReaderFactory.java:694)
                at 
org.apache.hadoop.hdfs.BlockReaderFactory.build(BlockReaderFactory.java:355)
                at 
org.apache.hadoop.hdfs.DFSInputStream.blockSeekTo(DFSInputStream.java:665)
                at 
org.apache.hadoop.hdfs.DFSInputStream.seekToBlockSource(DFSInputStream.java:1567)
                at 
org.apache.hadoop.hdfs.DFSInputStream.readBuffer(DFSInputStream.java:847)
                at 
org.apache.hadoop.hdfs.DFSInputStream.readWithStrategy(DFSInputStream.java:883)
                at 
org.apache.hadoop.hdfs.DFSInputStream.read(DFSInputStream.java:926)
                at java.io.DataInputStream.read(DataInputStream.java:149)
                at 
org.apache.iceberg.hadoop.HadoopStreams$HadoopSeekableInputStream.read(HadoopStreams.java:112)
                at 
org.apache.iceberg.avro.AvroIO$AvroInputStreamAdapter.read(AvroIO.java:117)
                at 
org.apache.iceberg.shaded.org.apache.avro.file.DataFileReader$SeekableInputStream.read(DataFileReader.java:292)
                at 
org.apache.iceberg.shaded.org.apache.avro.io.BinaryDecoder$InputStreamByteSource.tryReadRaw(BinaryDecoder.java:895)
                at 
org.apache.iceberg.shaded.org.apache.avro.io.BinaryDecoder.isEnd(BinaryDecoder.java:518)
                at 
org.apache.iceberg.shaded.org.apache.avro.file.DataFileStream.hasNextBlock(DataFileStream.java:287)
                at 
org.apache.iceberg.shaded.org.apache.avro.file.DataFileStream.hasNext(DataFileStream.java:212)
                at 
org.apache.iceberg.avro.AvroIterable$AvroReuseIterator.hasNext(AvroIterable.java:191)
                at 
org.apache.iceberg.io.CloseableIterable$4$1.hasNext(CloseableIterable.java:110)
                at 
org.apache.iceberg.io.FilterIterator.advance(FilterIterator.java:64)
                at 
org.apache.iceberg.io.FilterIterator.hasNext(FilterIterator.java:49)
                at 
org.apache.iceberg.io.FilterIterator.advance(FilterIterator.java:64)
                at 
org.apache.iceberg.io.FilterIterator.hasNext(FilterIterator.java:49)
                at 
org.apache.iceberg.io.FilterIterator.advance(FilterIterator.java:64)
                at 
org.apache.iceberg.io.FilterIterator.hasNext(FilterIterator.java:49)
                at 
org.apache.iceberg.io.CloseableIterable$4$1.hasNext(CloseableIterable.java:110)
                at 
org.apache.iceberg.util.ParallelIterable$ParallelIterator.lambda$null$1(ParallelIterable.java:69)
                at 
java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
                at java.util.concurrent.FutureTask.run(FutureTask.java:266)
                at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
                at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
                at java.lang.Thread.run(Thread.java:748)
        22/11/24 17:22:29 WARN BlockReaderFactory: I/O error constructing 
remote block reader.
        java.nio.channels.ClosedByInterruptException
                at 
java.nio.channels.spi.AbstractInterruptibleChannel.end(AbstractInterruptibleChannel.java:202)
                at 
sun.nio.ch.SocketChannelImpl.connect(SocketChannelImpl.java:659)
                at 
org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:192)
                at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:531)
                at 
org.apache.hadoop.hdfs.DFSClient.newConnectedPeer(DFSClient.java:3441)
                at 
org.apache.hadoop.hdfs.BlockReaderFactory.nextTcpPeer(BlockReaderFactory.java:777)
                at 
org.apache.hadoop.hdfs.BlockReaderFactory.getRemoteBlockReaderFromTcp(BlockReaderFactory.java:694)
                at 
org.apache.hadoop.hdfs.BlockReaderFactory.build(BlockReaderFactory.java:355)
                at 
org.apache.hadoop.hdfs.DFSInputStream.blockSeekTo(DFSInputStream.java:665)
                at 
org.apache.hadoop.hdfs.DFSInputStream.seekToBlockSource(DFSInputStream.java:1567)
                at 
org.apache.hadoop.hdfs.DFSInputStream.readBuffer(DFSInputStream.java:847)
                at 
org.apache.hadoop.hdfs.DFSInputStream.readWithStrategy(DFSInputStream.java:883)
                at 
org.apache.hadoop.hdfs.DFSInputStream.read(DFSInputStream.java:926)
                at java.io.DataInputStream.read(DataInputStream.java:149)
                at 
org.apache.iceberg.hadoop.HadoopStreams$HadoopSeekableInputStream.read(HadoopStreams.java:112)
                at 
org.apache.iceberg.avro.AvroIO$AvroInputStreamAdapter.read(AvroIO.java:117)
                at 
org.apache.iceberg.shaded.org.apache.avro.file.DataFileReader$SeekableInputStream.read(DataFileReader.java:292)
                at 
org.apache.iceberg.shaded.org.apache.avro.io.BinaryDecoder$InputStreamByteSource.tryReadRaw(BinaryDecoder.java:895)
                at 
org.apache.iceberg.shaded.org.apache.avro.io.BinaryDecoder.isEnd(BinaryDecoder.java:518)
                at 
org.apache.iceberg.shaded.org.apache.avro.file.DataFileStream.hasNextBlock(DataFileStream.java:287)
                at 
org.apache.iceberg.shaded.org.apache.avro.file.DataFileStream.hasNext(DataFileStream.java:212)
                at 
org.apache.iceberg.avro.AvroIterable$AvroReuseIterator.hasNext(AvroIterable.java:191)
                at 
org.apache.iceberg.io.CloseableIterable$4$1.hasNext(CloseableIterable.java:110)
                at 
org.apache.iceberg.io.FilterIterator.advance(FilterIterator.java:64)
                at 
org.apache.iceberg.io.FilterIterator.hasNext(FilterIterator.java:49)
                at 
org.apache.iceberg.io.FilterIterator.advance(FilterIterator.java:64)
                at 
org.apache.iceberg.io.FilterIterator.hasNext(FilterIterator.java:49)
                at 
org.apache.iceberg.io.FilterIterator.advance(FilterIterator.java:64)
                at 
org.apache.iceberg.io.FilterIterator.hasNext(FilterIterator.java:49)
                at 
org.apache.iceberg.io.CloseableIterable$4$1.hasNext(CloseableIterable.java:110)
                at 
org.apache.iceberg.util.ParallelIterable$ParallelIterator.lambda$null$1(ParallelIterable.java:69)
                at 
java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
                at java.util.concurrent.FutureTask.run(FutureTask.java:266)
                at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
                at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
                at java.lang.Thread.run(Thread.java:748)


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]


---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to