xuekaiqi created KYLIN-4453: ------------------------------- Summary: Query on refreshed cube failed with FileNotFoundException Key: KYLIN-4453 URL: https://issues.apache.org/jira/browse/KYLIN-4453 Project: Kylin Issue Type: New Feature Components: Storage - Parquet Reporter: xuekaiqi Assignee: nichunen Fix For: v4.0.0
Steps to reproduce: # Build a segment of any cube # Refresh the segment # Query the cube, get error message like {code:java} java.io.FileNotFoundException: File file:/Users/kyligence/Downloads/localmeta_n/vvv/parquet/gg/20200401000000_20200403000000/4/part-00000-cdaa5f21-34dd-432d-865e-92089a7ffa03-c000.snappy.parquet does not exist It is possible the underlying files have been updated. You can explicitly invalidate the cache in Spark by running 'REFRESH TABLE tableName' command in SQL or by recreating the Dataset/DataFrame involved. at org.apache.spark.sql.execution.datasources.FileScanRDD$$anon$1.org$apache$spark$sql$execution$datasources$FileScanRDD$$anon$$readCurrentFile(FileScanRDD.scala:127) at org.apache.spark.sql.execution.datasources.FileScanRDD$$anon$1.nextIterator(FileScanRDD.scala:177) at org.apache.spark.sql.execution.datasources.FileScanRDD$$anon$1.hasNext(FileScanRDD.scala:101) at org.apache.spark.sql.catalyst.expressions.GeneratedClass$GeneratedIteratorForCodegenStage1.scan_nextBatch_0$(Unknown Source) at org.apache.spark.sql.catalyst.expressions.GeneratedClass$GeneratedIteratorForCodegenStage1.agg_doAggregateWithKeys_0$(Unknown Source) at org.apache.spark.sql.catalyst.expressions.GeneratedClass$GeneratedIteratorForCodegenStage1.processNext(Unknown Source) at org.apache.spark.sql.execution.BufferedRowIterator.hasNext(BufferedRowIterator.java:43) at {code} -- This message was sent by Atlassian Jira (v8.3.4#803005)