[ https://issues.apache.org/jira/browse/DRILL-4349?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Khurram Faraaz closed DRILL-4349. --------------------------------- > parquet reader returns wrong results when reading a nullable column that > starts with a large number of nulls (>30k) > ------------------------------------------------------------------------------------------------------------------- > > Key: DRILL-4349 > URL: https://issues.apache.org/jira/browse/DRILL-4349 > Project: Apache Drill > Issue Type: Bug > Components: Storage - Parquet > Affects Versions: 1.4.0 > Reporter: Deneche A. Hakim > Assignee: Deneche A. Hakim > Priority: Critical > Fix For: 1.6.0 > > Attachments: drill4349.tar.gz > > > While reading a nullable column, if in a single pass we only read null > values, the parquet reader resets the value of pageReader.readPosInBytes > which will lead to wrong data read from the file. > To reproduce the issue, create a csv file (repro.csv) with 2 columns (id, > val) with 50100 rows, where id equals to the row number and val is empty for > the first 50k rows, and equal to id for the remaining rows. > create a parquet table from the csv file: > {noformat} > CREATE TABLE `repro_parquet` AS SELECT CAST(columns[0] AS INT) AS id, > CAST(NULLIF(columns[1], '') AS DOUBLE) AS val from `repro.csv`; > {noformat} > Now if you query any of the non null values you will get wrong results: > {noformat} > 0: jdbc:drill:zk=local> select * from `repro_parquet` where id>=50000 limit > 10; > +--------+---------------------------+ > | id | val | > +--------+---------------------------+ > | 50000 | 9.11337776337441E-309 | > | 50001 | 3.26044E-319 | > | 50002 | 1.4916681476489723E-154 | > | 50003 | 2.0000000018890676 | > | 50004 | 2.681561588521345E154 | > | 50005 | -2.1016574E-317 | > | 50006 | -1.4916681476489723E-154 | > | 50007 | -2.0000000018890676 | > | 50008 | -2.681561588521345E154 | > | 50009 | 2.1016574E-317 | > +--------+---------------------------+ > 10 rows selected (0.238 seconds) > {noformat} > and here are the expected values: > {noformat} > 0: jdbc:drill:zk=local> select * from `repro.csv` where cast(columns[0] as > int)>=50000 limit 10; > +--------------------+ > | columns | > +--------------------+ > | ["50000","50000"] | > | ["50001","50001"] | > | ["50002","50002"] | > | ["50003","50003"] | > | ["50004","50004"] | > | ["50005","50005"] | > | ["50006","50006"] | > | ["50007","50007"] | > | ["50008","50008"] | > | ["50009","50009"] | > +--------------------+ > {noformat} > I confirmed that the file is written correctly and the issue is in the > parquet reader (already have a fix for it) -- This message was sent by Atlassian JIRA (v6.3.4#6332)