danny0405 commented on issue #8164:
URL: https://github.com/apache/hudi/issues/8164#issuecomment-1473313206

   
   1. Did you enable the ckp yet? Flink sink relies the ckp success event for 
Hudi trasanction commiting;
   2. Both bulk_insert and append_write use the `BulkInsertWriterHelper` to 
write the parquet files direcly, there is no UPSERTs, if `FLINK_STATE` is used, 
things are very diffrent, the `StreamWriteFunction` would kick in;
   3. You can just set up the compress options within the Flink SQL options, or 
the `HoodiePipeline#options` you have used:
   
   e.g.
   
   ```sql
   create table xxx(
   ) with (
     'connector' = 'hudi',
     'hoodie.parquet.compression.codec' = 'gzip'
   );
   ```
   
   ```java
   HoodiePipeline.builder("xxx")
       .option("hoodie.parquet.compression.codec", "gzip")
   ```
   
   The default codec is already gzip, probably that is the reason you do not 
perceive any difference


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: commits-unsubscr...@hudi.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org

Reply via email to