[ https://issues.apache.org/jira/browse/HBASE-29146?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Duo Zhang resolved HBASE-29146. ------------------------------- Fix Version/s: 2.7.0 3.0.0-beta-2 2.6.3 Hadoop Flags: Reviewed Resolution: Fixed > Incremental backups can fail due to not cleaning up the MR bulkload output > directory > ------------------------------------------------------------------------------------ > > Key: HBASE-29146 > URL: https://issues.apache.org/jira/browse/HBASE-29146 > Project: HBase > Issue Type: Bug > Components: backup&restore > Reporter: Hernan Gelaf-Romer > Assignee: Hernan Gelaf-Romer > Priority: Major > Labels: pull-request-available > Fix For: 2.7.0, 3.0.0-beta-2, 2.6.3 > > > At the moment, if a bulkload file is archived as we're backing it up, or if > there are archive files that need to be loaded, we run into the following > exception > > Unexpected Exception : org.apache.hadoop.mapred.FileAlreadyExistsException > > The happens because we're running MR jobs to re-split these HFiles, and we > cannot run subsequent MR jobs against the same target bulk output directory > without first cleaning that directory up. We should make sure we properly > clean up the bulkload output directory in between runs of > MapReduceHFileSplitterJob -- This message was sent by Atlassian Jira (v8.20.10#820010)