Roman Karlstetter created ARROW-13024:
-----------------------------------------

             Summary: Decoding byte stream split encoded parquet columns fails 
when file has nulls
                 Key: ARROW-13024
                 URL: https://issues.apache.org/jira/browse/ARROW-13024
             Project: Apache Arrow
          Issue Type: Bug
          Components: C++, Parquet
    Affects Versions: 4.0.0, 3.0.0, 2.0.0
            Reporter: Roman Karlstetter


Reading from a parquet file fails with the following error

{{Data size too small for number of values (corrupted file?)}}.

This happens for the case when there is a {{BYTE_STREAM_SPLIT}}-encoded column 
which has less values stored than number of rows, which is the case when the 
column has null values (definition levels are present).

The problematic part is the condition checked in 
{{ByteStreamSplitDecoder<DType>::SetData}}, which raises the error if the 
number of values does not match the size of the data array.

I'm unsure whether I have enough experience with the internals of the 
encoding/decoding part of this implementation to fix this issue, but my 
suggestion would be to initialize {{num_values_in_buffer_}} with 
{{len/static_cast<int64_t>(sizeof(T))}}.




--
This message was sent by Atlassian Jira
(v8.3.4#803005)

Reply via email to