WangCHX opened a new issue, #6487: URL: https://github.com/apache/hudi/issues/6487
**Describe the problem you faced** we accidentally configure wrong primary key in the spark write config, it cause duplicate data. wondering if there is a way to avoid it. **To Reproduce** change the primary config in write config and run the spark job. **Expected behavior** maybe should block the spark job to write data if the primary key config is different from the primary key in the original table. **Environment Description** * Hudi version : 0.11.0 * Spark version : 3.2.1 * Storage (HDFS/S3/GCS..) : GCS * Running on Docker? (yes/no) : yes. on k8s. -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: commits-unsubscr...@hudi.apache.org.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org