stream2000 opened a new pull request, #9887: URL: https://github.com/apache/hudi/pull/9887
### Change Logs TestInsertTable#Test Bulk Insert Into Consistent Hashing Bucket Index Table failed continuously, I enabled it again and reproduced it locally. My test will cause the write job abort, and when org.apache.hudi.spark3.internal.HoodieDataSourceInternalBatchWrite#abort is called, all the subtasks may not have already been canceled. So if we roll back the current instant immediately, new files may still be written after rollback scheduled, which will cause dirty data. We don't need to rollback the failed instant immediately, so to avoid this problem, I remove the line that rollback the instant in `org.apache.hudi.spark3.internal.HoodieDataSourceInternalBatchWrite#abort `. ### Impact Bulk insert as row do not rollback failed instant on abort ### Risk level (write none, low medium or high below) none ### Documentation Update none ### Contributor's checklist - [x] Read through [contributor's guide](https://hudi.apache.org/contribute/how-to-contribute) - [x] Change Logs and Impact were stated clearly - [x] Adequate tests were added if applicable - [ ] CI passed -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: commits-unsubscr...@hudi.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org