[ https://issues.apache.org/jira/browse/HBASE-8105?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13603949#comment-13603949 ]
philo vivero commented on HBASE-8105: ------------------------------------- I do have copious logs for the time of the outage, on the order of megabytes. Do you want all of them, or should I grep for particular regexp's that would be more useful to you? In general, it starts with this: 2013-03-12 17:11:43,241 FATAL org.apache.hadoop.hbase.regionserver.HRegionServer: ABORTING region server rs9,60020,1363131519016: IOE in log roller Then a bunch of things like this: 2013-03-12 17:11:43,241 FATAL org.apache.hadoop.hbase.regionserver.wal.HLog: Could not append. Requesting close of hlog 2013-03-12 17:11:43,242 FATAL org.apache.hadoop.hbase.regionserver.wal.HLog: Could not append. Requesting close of hlog 2013-03-12 17:11:43,242 FATAL org.apache.hadoop.hbase.regionserver.wal.HLog: Could not append. Requesting close of hlog 2013-03-12 17:11:45,245 FATAL org.apache.hadoop.hbase.regionserver.wal.HLog: Could not sync. Requesting close of hlog 2013-03-12 17:11:45,246 ERROR org.apache.hadoop.hbase.regionserver.wal.HLog: Error while syncing, requesting close of hlog 2013-03-12 17:11:47,254 FATAL org.apache.hadoop.hbase.regionserver.wal.HLog: Could not sync. Requesting close of hlog 2013-03-12 17:11:47,254 ERROR org.apache.hadoop.hbase.regionserver.wal.HLog: Error while syncing, requesting close of hlog Those last two errors repeat alternately for a long time, then eventually: 2013-03-12 17:26:13,888 ERROR org.apache.hadoop.hbase.regionserver.HRegionServer: Close and delete failed 2013-03-12 17:27:25,892 ERROR org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper: ZooKeeper delete failed after 3 retries Once we allow RS back into the network and bounce the RS, these ERRORs appear: 2013-03-12 17:51:15,908 ERROR org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper: Node /hbase/replication/peers already exists and this is not a retry 2013-03-12 17:51:15,914 ERROR org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper: Node /hbase/replication/rs already exists and this is not a retry 2013-03-12 17:51:15,925 ERROR org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper: Node /hbase/replication/state already exists and this is not a retry The full logs are 23MB, and "egrep 'ERROR|FATAL'" for the affected time period is 112KB. Still want it attached? > RegionServer Doesn't Rejoin Cluster after Netsplit > -------------------------------------------------- > > Key: HBASE-8105 > URL: https://issues.apache.org/jira/browse/HBASE-8105 > Project: HBase > Issue Type: Bug > Components: regionserver > Affects Versions: 0.92.1 > Environment: Linux Ubuntu 10.04 LTS > Reporter: philo vivero > > Running a 15-node HBase cluster. Testing various failure scenarios. Segregate > one RegionServer from the cluster by firewalling off every port except SSH > (because we need to be able to re-enable the node later). > After the RS is automatically removed from the cluster, we re-enable all > ports again, but RS never rejoins the cluster. > I suspect the possibility this is desired behaviour, but haven't found proof > so far. The code doesn't have any comment indicating this is the behaviour > desired: > http://grepcode.com/file/repo1.maven.org/maven2/org.apache.hbase/hbase/0.92.2/org/apache/hadoop/hbase/regionserver/HRegionServer.java/ > See lines starting at 624, public void run(). It makes it through the first > try/catch block, but then loops inside the second try/catch block. Our > hypothesis is that it never gets out naturally. > If we bounce the RegionServer process, then it rejoins the cluster. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira