WangCHX opened a new issue, #6487:
URL: https://github.com/apache/hudi/issues/6487

   **Describe the problem you faced**
   
   we accidentally configure wrong primary key in the spark write config, it 
cause duplicate data. wondering if there is a way to avoid it. 
   
   **To Reproduce**
   change the primary config in write config and run the spark job.
   
   **Expected behavior**
   maybe should block the spark job to write data if the primary key config is 
different from the primary key in the original table.
   
   
   **Environment Description**
   
   * Hudi version : 0.11.0
   
   * Spark version : 3.2.1
   
   * Storage (HDFS/S3/GCS..) : GCS
   
   * Running on Docker? (yes/no) : yes. on k8s.
   
   
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: commits-unsubscr...@hudi.apache.org.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org

Reply via email to