parthchandra commented on code in PR #2032:
URL: https://github.com/apache/datafusion-comet/pull/2032#discussion_r2217038689
##########
common/src/main/java/org/apache/comet/parquet/BatchReader.java:
##########
@@ -183,9 +183,7 @@ public BatchReader(
this.taskContext = TaskContext$.MODULE$.get();
}
- public BatchReader(AbstractColumnReader[] columnReaders) {
- // Todo: set useDecimal128 and useLazyMaterialization
- int numColumns = columnReaders.length;
+ public BatchReader(int numColumns) {
Review Comment:
> passed to Batch Reader at
[here](https://github.com/apache/iceberg/blob/main/spark/v3.5/spark/src/main/java/org/apache/iceberg/spark/data/vectorized/VectorizedSparkParquetReaders.java#L86)
This does not sound correct. BatchReader has no API to set a column reader
other than the constructor changed by this PR. The only other way to set column
readers is by calling `init` which will then create the appropriate column
readers.
Also, I notice that the current version of the constructor ignores the
column readers passed in.
Any BatchReader created with this constructor (either this PR or the current
version) will have an array of nulls as the column readers.
How is such a BatchReader usable (or useful)?
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To unsubscribe, e-mail: [email protected]
For queries about this service, please contact Infrastructure at:
[email protected]
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]