sunchao commented on a change in pull request #32753: URL: https://github.com/apache/spark/pull/32753#discussion_r659984532
########## File path: sql/core/src/main/java/org/apache/spark/sql/execution/datasources/parquet/ParquetReadState.java ########## @@ -33,31 +51,104 @@ /** The remaining number of values to read in the current batch */ int valuesToReadInBatch; - ParquetReadState(int maxDefinitionLevel) { + ParquetReadState(int maxDefinitionLevel, PrimitiveIterator.OfLong rowIndexes) { this.maxDefinitionLevel = maxDefinitionLevel; + this.rowRanges = rowIndexes == null ? null : constructRanges(rowIndexes); + nextRange(); } /** - * Called at the beginning of reading a new batch. + * Construct a list of row ranges from the given `rowIndexes`. For example, suppose the + * `rowIndexes` are `[0, 1, 2, 4, 5, 7, 8, 9]`, it will be converted into 3 row ranges: Review comment: It gives you an iterator so yeah generating them on the fly: https://github.com/apache/parquet-mr/blob/master/parquet-column/src/main/java/org/apache/parquet/internal/filter2/columnindex/RowRanges.java#L253. The indexes are generated from `Range` which is very similar to what we defined here. I'm planning to file a JIRA in parquet-mr to just return the original `Range`s so we don't have to do this step in Spark. -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org --------------------------------------------------------------------- To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org For additional commands, e-mail: reviews-h...@spark.apache.org