HFiles are missing from an incremental load -------------------------------------------
Key: HBASE-5210 URL: https://issues.apache.org/jira/browse/HBASE-5210 Project: HBase Issue Type: Bug Components: mapreduce Affects Versions: 0.90.2 Environment: HBase 0.90.2 with Hadoop-0.20.2 (with durable sync). RHEL 2.6.18-164.15.1.el5. 4 node cluster (1 master, 3 slaves) Reporter: Lawrence Simpson We run an overnight map/reduce job that loads data from an external source and adds that data to an existing HBase table. The input files have been loaded into hdfs. The map/reduce job uses the HFileOutputFormat (and the TotalOrderPartitioner) to create HFiles which are subsequently added to the HBase table. On at least two separate occasions (that we know of), a range of output would be missing for a given day. The range of keys for the missing values corresponded to those of a particular region. This implied that a complete HFile somehow went missing from the job. Further investigation revealed the following: * Two different reducers (running in separate JVMs and thus separate class loaders) * in the same server can end up using the same file names for their * HFiles. The scenario is as follows: * 1. Both reducers start near the same time. * 2. The first reducer reaches the point where it wants to write its first file. * 3. It uses the StoreFile class which contains a static Random object * which is initialized by default using a timestamp. * 4. The file name is generated using the random number generator. * 5. The file name is checked against other existing files. * 6. The file is written into temporary files in a directory named * after the reducer attempt. * 7. The second reduce task reaches the same point, but its StoreClass * (which is now in the file system's cache) gets loaded within the * time resolution of the OS and thus initializes its Random() * object with the same seed as the first task. * 8. The second task also checks for an existing file with the name * generated by the random number generator and finds no conflict * because each task is writing files in its own temporary folder. * 9. The first task finishes and gets its temporary files committed * to the "real" folder specified for output of the HFiles. * 10. The second task then reaches its own conclusion and commits its * files (moveTaskOutputs). The released Hadoop code just overwrites * any files with the same name. No warning messages or anything. * The first task's HFiles just go missing. * * Note: The reducers here are NOT different attempts at the same * reduce task. They are different reduce tasks so data is * really lost. I am currently testing a fix in which I have added code to the Hadoop FileOutputCommitter.moveTaskOutputs method to check for a conflict with an existing file in the final output folder and to rename the HFile if needed. This may not be appropriate for all uses of FileOutputFormat. So I have put this into a new class which is then used by a subclass of HFileOutputFormat. Subclassing of FileOutputCommitter itself was a bit more of a problem due to private declarations. I don't know if my approach is the best fix for the problem. If someone more knowledgeable than myself deems that it is, I will be happy to share what I have done and by that time I may have some information on the results. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira