danny0405 commented on issue #8164: URL: https://github.com/apache/hudi/issues/8164#issuecomment-1473313206
1. Did you enable the ckp yet? Flink sink relies the ckp success event for Hudi trasanction commiting; 2. Both bulk_insert and append_write use the `BulkInsertWriterHelper` to write the parquet files direcly, there is no UPSERTs, if `FLINK_STATE` is used, things are very diffrent, the `StreamWriteFunction` would kick in; 3. You can just set up the compress options within the Flink SQL options, or the `HoodiePipeline#options` you have used: e.g. ```sql create table xxx( ) with ( 'connector' = 'hudi', 'hoodie.parquet.compression.codec' = 'gzip' ); ``` ```java HoodiePipeline.builder("xxx") .option("hoodie.parquet.compression.codec", "gzip") ``` The default codec is already gzip, probably that is the reason you do not perceive any difference -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: commits-unsubscr...@hudi.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org