yli1994 commented on issue #14726:
URL: https://github.com/apache/arrow/issues/14726#issuecomment-1326961628

   > Parquet files are written with compression turned on by default, which 
means that usually the size on disk is much (depending on the data several 
times!) smaller than the actual in-memory size of the data.
   > 
   > Can you confirm if the file is written with compression?
   > 
   > cc @jorisvandenbossche
   
   Hi @assignUser ,
   
   I wrote parquet both in "snappy" and "zstd" compression format, and the 
sizes are 202G and 158G respectively, but what I expect is that "memory map" 
reading method should not increase the memory occupied.


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]

Reply via email to