[ https://issues.apache.org/jira/browse/HBASE-8935?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13709735#comment-13709735 ]
Hudson commented on HBASE-8935: ------------------------------- FAILURE: Integrated in HBase-TRUNK-on-Hadoop-2.0.0 #618 (See [https://builds.apache.org/job/HBase-TRUNK-on-Hadoop-2.0.0/618/]) HBASE-8935 IntegrationTestBigLinkedList fails under load on 0.94 due to some scan issues - add logging (sershe: rev 1503524) * /hbase/trunk/hbase-it/src/test/java/org/apache/hadoop/hbase/test/IntegrationTestBigLinkedList.java * /hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/TableRecordReaderImpl.java * /hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegionServer.java > IntegrationTestBigLinkedList fails under load on 0.94 due to some scan issues > ----------------------------------------------------------------------------- > > Key: HBASE-8935 > URL: https://issues.apache.org/jira/browse/HBASE-8935 > Project: HBase > Issue Type: Bug > Reporter: Sergey Shelukhin > Assignee: Sergey Shelukhin > Attachments: HBASE-8935-v0-94-logging.patch, HBASE-8935-v0.patch > > > We observe occasional failures (4-5k undefined and unreferenced nodes in the > list, running) when the cluster is stressed while running > IntegrationTestBigLinkedList on 0.94. Unfortunately debug logging is disabled. > If Verify is rerun after the fact, the data is there, so it's not data loss. > All the missing keys come from very narrow range from the very beginning of > one region; most of this region is not affected. > In the case I'm looking at, region range is > {code}\xD0\x0C`=%\xA2\xD3\xA8\x0A\xF8\x917\x05\x94\x1D> .. > \xD4\x0Bx\xAF\x88\x0C\xF4"7\x9A\x9F{\xCE\x0E\x8A{code} > and problematic renage > {code}\xD0\x0C`QLD\xED2\xD5c\x8D\xDB5\x01\xD2H] ... > \xD0\x0D\x89)\x11\x0E8\xC5\x08o\xD7\xE3$\xB3\xAAu{code} > There are no splits and compactions on the region during MR job; there are no > compactions after that could have affected the data, although there is one > flush that happened long after, and region was moved and reopened (before I > ran verify manually that showed that data is in HBase). > One thing that happened was that the scanners used by all the map tasks had > lease expirations during the MR job that had missing data, some of them twice. > First scanner expiration for each task I looked at was exactly 1 minute after > "Processing split" log message, when the scanner is opened. > Only the tasks where the scanners have expired twice (2 of them) logged any > errors; presumably one expiration in each task happened after the reading was > finished, because everything was slow and scanner wasn't closed in time - no > errors or exceptions are logged in the tasks where scanner only expired once. > The job I ran afterwards had no scanner expirations so it does close them > correctly in normal case... > The task that was reading the problematic range had one scanner expiration > and didn't log any errors. > It's a little bit special (or it may be a total red herring) in that in other > tasks, scanner expired while the task was splitting partial outputs (which > may mean end of input reading?), whereas in the task that lost data spilling > started long (~40s) after the scanner expired. > However there's one another such task (spilling started 25s after scanner > expired) and it didn't log any errors and didn't lose data. > At this point I am not sure about the root cause but I suspect it might be > related to scanner expiration handling, given the patterns. > One thing for sure is that there isn't enough logging... so I will start with > adding that. > -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira