HonahX commented on code in PR #902:
URL: https://github.com/apache/iceberg-python/pull/902#discussion_r1673398145
##########
pyiceberg/table/__init__.py:
##########
@@ -1884,8 +1884,9 @@ def to_arrow_batch_reader(self) -> pa.RecordBatchReader:
from pyiceberg.io.pyarrow import project_batches, schema_to_pyarrow
+ target_schema = schema_to_pyarrow(self.projection())
Review Comment:
> I think the only time we are casting the types is on write, where we may
want to downcast it for forward compatibility.
+1 Currently, we use "large_*" types during write. I think it could be
better if we can write file based on the input pyarrow dataframe schema: if the
dataframe is `string`, we also write with `string`
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To unsubscribe, e-mail: [email protected]
For queries about this service, please contact Infrastructure at:
[email protected]
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]