[ 
https://issues.apache.org/jira/browse/ARROW-10168?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Neville Dipale resolved ARROW-10168.
------------------------------------
    Fix Version/s: 2.0.0
       Resolution: Fixed

Issue resolved by pull request 8354
[https://github.com/apache/arrow/pull/8354]

> [Rust] [Parquet] Extend arrow schema conversion to projected fields
> -------------------------------------------------------------------
>
>                 Key: ARROW-10168
>                 URL: https://issues.apache.org/jira/browse/ARROW-10168
>             Project: Apache Arrow
>          Issue Type: Sub-task
>          Components: Rust
>    Affects Versions: 1.0.1
>            Reporter: Neville Dipale
>            Priority: Major
>              Labels: pull-request-available
>             Fix For: 2.0.0
>
>          Time Spent: 1h
>  Remaining Estimate: 0h
>
> When writing Arrow data to Parquet, we serialise the schema's IPC 
> representation. This schema is then read back by the Parquet reader, and used 
> to preserve the array type information from the original Arrow data.
> We however do not rely on the above mechanism when reading projected columns 
> from a Parquet file; i.e. if we have a file with 3 columns, but we only read 
> 2 columns, we do not yet rely on the serialised arrow schema; and can thus 
> lose type information.
> This behaviour was deliberately left out, as the function 
> *parquet_to_arrow_schema_by_columns* does not check for the existence of 
> arrow schema in the metadata.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

Reply via email to