hudi-bot opened a new issue, #15030:
URL: https://github.com/apache/hudi/issues/15030

   When switching DEFAULT_METADATA_ENABLE_FOR_READERS to true and setting 
"hoodie.metadata.enable.full.scan.log.files" to false, running presto queries 
with hudi-presto-bundle on HDFS in docker demo throws 
UnsupportedOperationException during HFile log merging, because 
HadoopExtendedFileSystem does not implement getScheme().
   {code:java}
   2022-04-23T07:26:13.085Z INFO hive-hive-0 
org.apache.hudi.common.table.log.AbstractHoodieLogRecordReader Reading a data 
block from file 
hdfs://namenode:8020/user/hive/warehouse/stock_ticks_cow/.hoodie/metadata/files/.files-0000_00000000000000.log.1_0-10-10
 at instant 202204230723040192022-04-23T07:26:13.086Z INFO hive-hive-0 
org.apache.hudi.common.table.log.AbstractHoodieLogRecordReader Merging the 
final data blocks2022-04-23T07:26:13.086Z INFO hive-hive-0 
org.apache.hudi.common.table.log.AbstractHoodieLogRecordReader Number of 
remaining logblocks to merge 32022-04-23T07:26:13.185Z INFO hive-hive-0 
org.apache.hudi.common.table.log.AbstractHoodieLogRecordReader Number of 
remaining logblocks to merge 22022-04-23T07:26:13.190Z ERROR hive-hive-0 
org.apache.hudi.common.table.log.AbstractHoodieLogRecordReader Got exception 
when reading log filejava.lang.UnsupportedOperationException: Not implemented 
by the HadoopExtendedFileSystem FileSystem implementationat 
org.apache.hadoop.fs.FileSys
 tem.getScheme(FileSystem.java:219)at 
org.apache.hudi.common.table.log.block.HoodieHFileDataBlock.lookupRecords(HoodieHFileDataBlock.java:205)at
 
org.apache.hudi.common.table.log.block.HoodieDataBlock.getRecordIterator(HoodieDataBlock.java:168)at
 
org.apache.hudi.common.table.log.AbstractHoodieLogRecordReader.getRecordsIterator(AbstractHoodieLogRecordReader.java:488)at
 
org.apache.hudi.common.table.log.AbstractHoodieLogRecordReader.processDataBlock(AbstractHoodieLogRecordReader.java:378)at
 
org.apache.hudi.common.table.log.AbstractHoodieLogRecordReader.processQueuedBlocksForInstant(AbstractHoodieLogRecordReader.java:466)at
 
org.apache.hudi.common.table.log.AbstractHoodieLogRecordReader.scanInternal(AbstractHoodieLogRecordReader.java:342)at
 
org.apache.hudi.common.table.log.AbstractHoodieLogRecordReader.scan(AbstractHoodieLogRecordReader.java:195)at
 
org.apache.hudi.metadata.HoodieMetadataMergedLogRecordReader.getRecordsByKeys(HoodieMetadataMergedLogRecordReader.java:124)at
 org.apache.hudi.m
 
etadata.HoodieBackedTableMetadata.readLogRecords(HoodieBackedTableMetadata.java:257)at
 
org.apache.hudi.metadata.HoodieBackedTableMetadata.lambda$getRecordsByKeys$0(HoodieBackedTableMetadata.java:213)at
 java.util.HashMap.forEach(HashMap.java:1289)at 
org.apache.hudi.metadata.HoodieBackedTableMetadata.getRecordsByKeys(HoodieBackedTableMetadata.java:200)at
 
org.apache.hudi.metadata.HoodieBackedTableMetadata.getRecordByKey(HoodieBackedTableMetadata.java:140)at
 
org.apache.hudi.metadata.BaseTableMetadata.fetchAllFilesInPartition(BaseTableMetadata.java:312)at
 
org.apache.hudi.metadata.BaseTableMetadata.getAllFilesInPartition(BaseTableMetadata.java:135)at
 
org.apache.hudi.metadata.HoodieMetadataFileSystemView.listPartition(HoodieMetadataFileSystemView.java:65)at
 
org.apache.hudi.common.table.view.AbstractTableFileSystemView.lambda$ensurePartitionLoadedCorrectly$9(AbstractTableFileSystemView.java:304)at
 
java.util.concurrent.ConcurrentHashMap.computeIfAbsent(ConcurrentHashMap.java:1660)at
 org.apac
 
he.hudi.common.table.view.AbstractTableFileSystemView.ensurePartitionLoadedCorrectly(AbstractTableFileSystemView.java:295)at
 
org.apache.hudi.common.table.view.AbstractTableFileSystemView.getLatestBaseFiles(AbstractTableFileSystemView.java:478)at
 
org.apache.hudi.hadoop.HoodieROTablePathFilter.accept(HoodieROTablePathFilter.java:189)at
 
com.facebook.presto.hive.util.HiveFileIterator.lambda$getLocatedFileStatusRemoteIterator$0(HiveFileIterator.java:103)at
 com.google.common.collect.Iterators$5.computeNext(Iterators.java:639)at 
com.google.common.collect.AbstractIterator.tryToComputeNext(AbstractIterator.java:141)at
 
com.google.common.collect.AbstractIterator.hasNext(AbstractIterator.java:136)at 
com.facebook.presto.hive.util.HiveFileIterator.computeNext(HiveFileIterator.java:69)at
 
com.facebook.presto.hive.util.HiveFileIterator.computeNext(HiveFileIterator.java:40)at
 
com.google.common.collect.AbstractIterator.tryToComputeNext(AbstractIterator.java:141)at
 com.google.common.collect.AbstractIte
 rator.hasNext(AbstractIterator.java:136)at 
java.util.Spliterators$IteratorSpliterator.tryAdvance(Spliterators.java:1811)at 
java.util.stream.StreamSpliterators$WrappingSpliterator.lambda$initPartialTraversalState$0(StreamSpliterators.java:294)at
 
java.util.stream.StreamSpliterators$AbstractWrappingSpliterator.fillBuffer(StreamSpliterators.java:206)at
 
java.util.stream.StreamSpliterators$AbstractWrappingSpliterator.doAdvance(StreamSpliterators.java:161)at
 
java.util.stream.StreamSpliterators$WrappingSpliterator.tryAdvance(StreamSpliterators.java:300)at
 java.util.Spliterators$1Adapter.hasNext(Spliterators.java:681)at 
com.facebook.presto.hive.BackgroundHiveSplitLoader.loadSplits(BackgroundHiveSplitLoader.java:195)at
 
com.facebook.presto.hive.BackgroundHiveSplitLoader.access$300(BackgroundHiveSplitLoader.java:40)at
 
com.facebook.presto.hive.BackgroundHiveSplitLoader$HiveSplitLoaderTask.process(BackgroundHiveSplitLoader.java:121)at
 com.facebook.presto.hive.util.ResumableTasks.safeProcessTask(R
 esumableTasks.java:47)at 
com.facebook.presto.hive.util.ResumableTasks.access$000(ResumableTasks.java:20)at
 com.facebook.presto.hive.util.ResumableTasks$1.run(ResumableTasks.java:35)at 
com.facebook.airlift.concurrent.BoundedExecutor.drainQueue(BoundedExecutor.java:78)at
 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)at
 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)at
 java.lang.Thread.run(Thread.java:748)2022-04-23T07:26:13.215Z ERROR 
hive-hive-0 org.apache.hudi.hadoop.HoodieROTablePathFilter Error checking path 
:hdfs://namenode:8020/user/hive/warehouse/stock_ticks_cow/2018/08/31/.hoodie_partition_metadata,
 under folder: 
hdfs://namenode:8020/user/hive/warehouse/stock_ticks_cow/2018/08/31org.apache.hudi.exception.HoodieMetadataException:
 Failed to retrieve files in partition 
hdfs://namenode:8020/user/hive/warehouse/stock_ticks_cow/2018/08/31 from 
metadataat org.apache.hudi.metadata.BaseTableMetadata.getAllFilesInPartition
 (BaseTableMetadata.java:137)at 
org.apache.hudi.metadata.HoodieMetadataFileSystemView.listPartition(HoodieMetadataFileSystemView.java:65)at
 
org.apache.hudi.common.table.view.AbstractTableFileSystemView.lambda$ensurePartitionLoadedCorrectly$9(AbstractTableFileSystemView.java:304)at
 
java.util.concurrent.ConcurrentHashMap.computeIfAbsent(ConcurrentHashMap.java:1660)at
 
org.apache.hudi.common.table.view.AbstractTableFileSystemView.ensurePartitionLoadedCorrectly(AbstractTableFileSystemView.java:295)at
 
org.apache.hudi.common.table.view.AbstractTableFileSystemView.getLatestBaseFiles(AbstractTableFileSystemView.java:478)at
 
org.apache.hudi.hadoop.HoodieROTablePathFilter.accept(HoodieROTablePathFilter.java:189)at
 
com.facebook.presto.hive.util.HiveFileIterator.lambda$getLocatedFileStatusRemoteIterator$0(HiveFileIterator.java:103)at
 com.google.common.collect.Iterators$5.computeNext(Iterators.java:639)at 
com.google.common.collect.AbstractIterator.tryToComputeNext(AbstractIterator.java:141)at
 com.g
 oogle.common.collect.AbstractIterator.hasNext(AbstractIterator.java:136)at 
com.facebook.presto.hive.util.HiveFileIterator.computeNext(HiveFileIterator.java:69)at
 
com.facebook.presto.hive.util.HiveFileIterator.computeNext(HiveFileIterator.java:40)at
 
com.google.common.collect.AbstractIterator.tryToComputeNext(AbstractIterator.java:141)at
 
com.google.common.collect.AbstractIterator.hasNext(AbstractIterator.java:136)at 
java.util.Spliterators$IteratorSpliterator.tryAdvance(Spliterators.java:1811)at 
java.util.stream.StreamSpliterators$WrappingSpliterator.lambda$initPartialTraversalState$0(StreamSpliterators.java:294)at
 
java.util.stream.StreamSpliterators$AbstractWrappingSpliterator.fillBuffer(StreamSpliterators.java:206)at
 
java.util.stream.StreamSpliterators$AbstractWrappingSpliterator.doAdvance(StreamSpliterators.java:161)at
 
java.util.stream.StreamSpliterators$WrappingSpliterator.tryAdvance(StreamSpliterators.java:300)at
 java.util.Spliterators$1Adapter.hasNext(Spliterators.java:681)at com
 
.facebook.presto.hive.BackgroundHiveSplitLoader.loadSplits(BackgroundHiveSplitLoader.java:195)at
 
com.facebook.presto.hive.BackgroundHiveSplitLoader.access$300(BackgroundHiveSplitLoader.java:40)at
 
com.facebook.presto.hive.BackgroundHiveSplitLoader$HiveSplitLoaderTask.process(BackgroundHiveSplitLoader.java:121)at
 
com.facebook.presto.hive.util.ResumableTasks.safeProcessTask(ResumableTasks.java:47)at
 
com.facebook.presto.hive.util.ResumableTasks.access$000(ResumableTasks.java:20)at
 com.facebook.presto.hive.util.ResumableTasks$1.run(ResumableTasks.java:35)at 
com.facebook.airlift.concurrent.BoundedExecutor.drainQueue(BoundedExecutor.java:78)at
 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)at
 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)at
 java.lang.Thread.run(Thread.java:748)Caused by: 
org.apache.hudi.exception.HoodieException: Exception when reading log fileĀ at 
org.apache.hudi.common.table.log.AbstractHoodieLogRecordReade
 r.scanInternal(AbstractHoodieLogRecordReader.java:351)at 
org.apache.hudi.common.table.log.AbstractHoodieLogRecordReader.scan(AbstractHoodieLogRecordReader.java:195)at
 
org.apache.hudi.metadata.HoodieMetadataMergedLogRecordReader.getRecordsByKeys(HoodieMetadataMergedLogRecordReader.java:124)at
 
org.apache.hudi.metadata.HoodieBackedTableMetadata.readLogRecords(HoodieBackedTableMetadata.java:257)at
 
org.apache.hudi.metadata.HoodieBackedTableMetadata.lambda$getRecordsByKeys$0(HoodieBackedTableMetadata.java:213)at
 java.util.HashMap.forEach(HashMap.java:1289)at 
org.apache.hudi.metadata.HoodieBackedTableMetadata.getRecordsByKeys(HoodieBackedTableMetadata.java:200)at
 
org.apache.hudi.metadata.HoodieBackedTableMetadata.getRecordByKey(HoodieBackedTableMetadata.java:140)at
 
org.apache.hudi.metadata.BaseTableMetadata.fetchAllFilesInPartition(BaseTableMetadata.java:312)at
 
org.apache.hudi.metadata.BaseTableMetadata.getAllFilesInPartition(BaseTableMetadata.java:135)...
 30 moreCaused by: java.lang.Unsup
 portedOperationException: Not implemented by the HadoopExtendedFileSystem 
FileSystem implementationat 
org.apache.hadoop.fs.FileSystem.getScheme(FileSystem.java:219)at 
org.apache.hudi.common.table.log.block.HoodieHFileDataBlock.lookupRecords(HoodieHFileDataBlock.java:205)at
 
org.apache.hudi.common.table.log.block.HoodieDataBlock.getRecordIterator(HoodieDataBlock.java:168)at
 
org.apache.hudi.common.table.log.AbstractHoodieLogRecordReader.getRecordsIterator(AbstractHoodieLogRecordReader.java:488)at
 
org.apache.hudi.common.table.log.AbstractHoodieLogRecordReader.processDataBlock(AbstractHoodieLogRecordReader.java:378)at
 
org.apache.hudi.common.table.log.AbstractHoodieLogRecordReader.processQueuedBlocksForInstant(AbstractHoodieLogRecordReader.java:466)at
 
org.apache.hudi.common.table.log.AbstractHoodieLogRecordReader.scanInternal(AbstractHoodieLogRecordReader.java:342)...
 39 more {code}
   
   ## JIRA info
   
   - Link: https://issues.apache.org/jira/browse/HUDI-3952
   - Type: Task
   - Epic: https://issues.apache.org/jira/browse/HUDI-3303
   
   
   ---
   
   
   ## Comments
   
   23/Apr/22 08:36;codope;Should be fixed by 
[https://github.com/prestodb/presto/pull/17477]
   
   Updated the title to reflect the follow up task after the above patch is 
released in presto.;;;


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]

Reply via email to