stream2000 opened a new pull request, #9887:
URL: https://github.com/apache/hudi/pull/9887

   ### Change Logs
   
   TestInsertTable#Test Bulk Insert Into Consistent Hashing Bucket Index Table 
failed continuously, I enabled it again and reproduced it locally. 
   
   My test will cause the write job abort, and when 
org.apache.hudi.spark3.internal.HoodieDataSourceInternalBatchWrite#abort is 
called, all the subtasks may not have already been canceled. So if we roll back 
the current instant immediately, new files may still be written after rollback 
scheduled, which will cause dirty data.
   
   We don't need to rollback the failed instant immediately, so to avoid this 
problem, I remove the line that rollback the instant in 
`org.apache.hudi.spark3.internal.HoodieDataSourceInternalBatchWrite#abort `. 
   
   ### Impact
   
   Bulk insert as row do not rollback failed instant on abort
   
   ### Risk level (write none, low medium or high below)
   
   none
   
   ### Documentation Update
   
   none
   
   ### Contributor's checklist
   
   - [x] Read through [contributor's 
guide](https://hudi.apache.org/contribute/how-to-contribute)
   - [x] Change Logs and Impact were stated clearly
   - [x] Adequate tests were added if applicable
   - [ ] CI passed
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: commits-unsubscr...@hudi.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org

Reply via email to