hbase git commit: HBASE-21101 Remove the waitUntilAllRegionsAssigned call after split in TestTruncateTableProcedure

2018-08-23 Thread zhangduo
Repository: hbase
Updated Branches:
  refs/heads/master 2d911fdc2 -> 6a5b4f2a5


HBASE-21101 Remove the waitUntilAllRegionsAssigned call after split in 
TestTruncateTableProcedure


Project: http://git-wip-us.apache.org/repos/asf/hbase/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/6a5b4f2a
Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/6a5b4f2a
Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/6a5b4f2a

Branch: refs/heads/master
Commit: 6a5b4f2a5c188f8eef4f2250b8b7db7dd1e750e4
Parents: 2d911fd
Author: Duo Zhang 
Authored: Thu Aug 23 18:04:01 2018 +0800
Committer: Duo Zhang 
Committed: Thu Aug 23 18:06:03 2018 +0800

--
 .../hadoop/hbase/master/procedure/TestTruncateTableProcedure.java   | 1 -
 1 file changed, 1 deletion(-)
--


http://git-wip-us.apache.org/repos/asf/hbase/blob/6a5b4f2a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/procedure/TestTruncateTableProcedure.java
--
diff --git 
a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/procedure/TestTruncateTableProcedure.java
 
b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/procedure/TestTruncateTableProcedure.java
index 5b38b17..b4a86d7 100644
--- 
a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/procedure/TestTruncateTableProcedure.java
+++ 
b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/procedure/TestTruncateTableProcedure.java
@@ -353,7 +353,6 @@ public class TestTruncateTableProcedure extends 
TestTableDDLProcedureBase {
   throws IOException, InterruptedException {
 // split a region
 UTIL.getAdmin().split(tableName, new byte[] { '0' });
-UTIL.waitUntilAllRegionsAssigned(tableName);
 
 // wait until split really happens
 UTIL.waitFor(6,



[12/51] [partial] hbase-site git commit: Published site at 6a5b4f2a5c188f8eef4f2250b8b7db7dd1e750e4.

2018-08-23 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/handler/OpenRegionHandler.PostOpenDeployTasksThread.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/handler/OpenRegionHandler.PostOpenDeployTasksThread.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/handler/OpenRegionHandler.PostOpenDeployTasksThread.html
index 2709ea3..4a11f27 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/handler/OpenRegionHandler.PostOpenDeployTasksThread.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/handler/OpenRegionHandler.PostOpenDeployTasksThread.html
@@ -37,309 +37,299 @@
 029import 
org.apache.hadoop.hbase.executor.EventType;
 030import 
org.apache.hadoop.hbase.regionserver.HRegion;
 031import 
org.apache.hadoop.hbase.regionserver.Region;
-032import 
org.apache.hadoop.hbase.regionserver.RegionServerAccounting;
-033import 
org.apache.hadoop.hbase.regionserver.RegionServerServices;
-034import 
org.apache.hadoop.hbase.regionserver.RegionServerServices.PostOpenDeployContext;
-035import 
org.apache.hadoop.hbase.regionserver.RegionServerServices.RegionStateTransitionContext;
-036import 
org.apache.hadoop.hbase.util.CancelableProgressable;
-037import 
org.apache.yetus.audience.InterfaceAudience;
-038import org.slf4j.Logger;
-039import org.slf4j.LoggerFactory;
-040import 
org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionStateTransition.TransitionCode;
-041/**
-042 * Handles opening of a region on a 
region server.
-043 * 

-044 * This is executed after receiving an OPEN RPC from the master or client. -045 */ -046@InterfaceAudience.Private -047public class OpenRegionHandler extends EventHandler { -048 private static final Logger LOG = LoggerFactory.getLogger(OpenRegionHandler.class); -049 -050 protected final RegionServerServices rsServices; -051 -052 private final RegionInfo regionInfo; -053 private final TableDescriptor htd; -054 private final long masterSystemTime; -055 -056 public OpenRegionHandler(final Server server, -057 final RegionServerServices rsServices, RegionInfo regionInfo, -058 TableDescriptor htd, long masterSystemTime) { -059this(server, rsServices, regionInfo, htd, masterSystemTime, EventType.M_RS_OPEN_REGION); -060 } -061 -062 protected OpenRegionHandler(final Server server, -063 final RegionServerServices rsServices, final RegionInfo regionInfo, -064 final TableDescriptor htd, long masterSystemTime, EventType eventType) { -065super(server, eventType); -066this.rsServices = rsServices; -067this.regionInfo = regionInfo; -068this.htd = htd; -069this.masterSystemTime = masterSystemTime; -070 } -071 -072 public RegionInfo getRegionInfo() { -073return regionInfo; -074 } -075 -076 @Override -077 public void process() throws IOException { -078boolean openSuccessful = false; -079final String regionName = regionInfo.getRegionNameAsString(); -080HRegion region = null; -081 -082try { -083 if (this.server.isStopped() || this.rsServices.isStopping()) { -084return; -085 } -086 final String encodedName = regionInfo.getEncodedName(); -087 -088 // 2 different difficult situations can occur -089 // 1) The opening was cancelled. This is an expected situation -090 // 2) The region is now marked as online while we're suppose to open. This would be a bug. -091 -092 // Check that this region is not already online -093 if (this.rsServices.getRegion(encodedName) != null) { -094LOG.error("Region " + encodedName + -095" was already online when we started processing the opening. " + -096"Marking this new attempt as failed"); -097return; -098 } -099 -100 // Check that we're still supposed to open the region. -101 // If fails, just return. Someone stole the region from under us. -102 if (!isRegionStillOpening()){ -103LOG.error("Region " + encodedName + " opening cancelled"); -104return; -105 } -106 -107 // Open region. After a successful open, failures in subsequent -108 // processing needs to do a close as part of cleanup. -109 region = openRegion(); -110 if (region == null) { -111return; -112 } -113 -114 if (!updateMeta(region, masterSystemTime) || this.server.isStopped() || -115 this.rsServices.isStopping()) { -116return; -117 } -118 -119 if (!isRegionStillOpening()) { -120return; -121 } -122 -123 // Successful region open, and add it to MutableOnlineRegions -124 this.rsServices.addRegion(region); -125 openSuccessful = true; -126 -127 // Done! Successful region open -128 LOG.debug("Opened " + reg


[07/51] [partial] hbase-site git commit: Published site at 6a5b4f2a5c188f8eef4f2250b8b7db7dd1e750e4.

2018-08-23 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/testdevapidocs/org/apache/hadoop/hbase/class-use/HBaseTestingUtility.html
--
diff --git 
a/testdevapidocs/org/apache/hadoop/hbase/class-use/HBaseTestingUtility.html 
b/testdevapidocs/org/apache/hadoop/hbase/class-use/HBaseTestingUtility.html
index 522d316..2dd2c14 100644
--- a/testdevapidocs/org/apache/hadoop/hbase/class-use/HBaseTestingUtility.html
+++ b/testdevapidocs/org/apache/hadoop/hbase/class-use/HBaseTestingUtility.html
@@ -2380,54 +2380,58 @@
 TestServerCrashProcedure.util 
 
 
+private static HBaseTestingUtility
+TestCreateTableProcedureMuitipleRegions.UTIL 
+
+
 protected static HBaseTestingUtility
 TestTableDDLProcedureBase.UTIL 
 
-
+
 protected static HBaseTestingUtility
 MasterProcedureSchedulerPerformanceEvaluation.UTIL 
 
-
+
 protected static HBaseTestingUtility
 TestCreateNamespaceProcedure.UTIL 
 
-
+
 protected static HBaseTestingUtility
 TestModifyNamespaceProcedure.UTIL 
 
-
+
 private static HBaseTestingUtility
 TestProcedurePriority.UTIL 
 
-
+
 protected static HBaseTestingUtility
 TestMasterProcedureEvents.UTIL 
 
-
+
 protected static HBaseTestingUtility
 TestDeleteNamespaceProcedure.UTIL 
 
-
+
 protected static HBaseTestingUtility
 TestMasterObserverPostCalls.UTIL 
 
-
+
 protected static HBaseTestingUtility
 TestWALProcedureStoreOnHDFS.UTIL 
 
-
+
 protected static HBaseTestingUtility
 TestSafemodeBringsDownMaster.UTIL 
 
-
+
 protected static HBaseTestingUtility
 TestMasterFailoverWithProcedures.UTIL 
 
-
+
 protected static HBaseTestingUtility
 TestMasterProcedureWalLease.UTIL 
 
-
+
 protected static HBaseTestingUtility
 TestProcedureAdmin.UTIL 
 
@@ -2836,274 +2840,278 @@
 TestNewVersionBehaviorFromClientSide.TEST_UTIL 
 
 
+protected static HBaseTestingUtility
+TestRecoveredEditsReplayAndAbort.TEST_UTIL 
+
+
 private static HBaseTestingUtility
 TestCompactSplitThread.TEST_UTIL 
 
-
+
 (package private) static HBaseTestingUtility
 TestRegionMergeTransactionOnCluster.TEST_UTIL 
 
-
+
 protected static HBaseTestingUtility
 TestCompactionPolicy.TEST_UTIL 
 
-
+
 private static HBaseTestingUtility
 TestWalAndCompactingMemStoreFlush.TEST_UTIL 
 
-
+
 private static HBaseTestingUtility
 TestWALMonotonicallyIncreasingSeqId.TEST_UTIL 
 
-
+
 private static HBaseTestingUtility
 TestDeleteMobTable.TEST_UTIL 
 
-
+
 private static HBaseTestingUtility
 TestBlocksScanned.TEST_UTIL 
 
-
+
 private static HBaseTestingUtility
 TestCompoundBloomFilter.TEST_UTIL 
 
-
+
 private static HBaseTestingUtility
 TestMultiColumnScanner.TEST_UTIL 
 
-
+
 protected static HBaseTestingUtility
 TestHRegion.TEST_UTIL 
 
-
+
 private static HBaseTestingUtility
 TestScannerWithBulkload.TEST_UTIL 
 
-
+
 private static HBaseTestingUtility
 TestHMobStore.TEST_UTIL 
 
-
+
 private static HBaseTestingUtility
 TestStoreFileScannerWithTagCompression.TEST_UTIL 
 
-
+
 private static HBaseTestingUtility
 TestScannerWithCorruptHFile.TEST_UTIL 
 
-
+
 private static HBaseTestingUtility
 TestEndToEndSplitTransaction.TEST_UTIL 
 
-
+
 private static HBaseTestingUtility
 TestScannerHeartbeatMessages.TEST_UTIL 
 
-
+
 private static HBaseTestingUtility
 TestIsDeleteFailure.TEST_UTIL 
 
-
+
 private HBaseTestingUtility
 TestStoreFileRefresherChore.TEST_UTIL 
 
-
+
 private static HBaseTestingUtility
 TestRegionMove.TEST_UTIL 
 
-
+
 private static HBaseTestingUtility
 TestRegionServerMetrics.TEST_UTIL 
 
-
+
 private static HBaseTestingUtility
 TestCacheOnWriteInSchema.TEST_UTIL 
 
-
+
 private static HBaseTestingUtility
 TestSettingTimeoutOnBlockingPoint.TEST_UTIL 
 
-
+
 private static HBaseTestingUtility
 TestHRegionReplayEvents.TEST_UTIL 
 
-
+
 private static HBaseTestingUtility
 TestRemoveRegionMetrics.TEST_UTIL 
 
-
+
 private static HBaseTestingUtility
 TestRecoveredEdits.TEST_UTIL 
 
-
+
 private static HBaseTestingUtility
 TestHRegionOnCluster.TEST_UTIL 
 
-
+
 private static HBaseTestingUtility
 TestHStore.TEST_UTIL 
 
-
+
 private static HBaseTestingUtility
 TestSplitLogWorker.TEST_UTIL 
 
-
+
 private HBaseTestingUtility
 TestRegionServerHostname.TEST_UTIL 
 
-
+
 private static HBaseTestingUtility
 TestEncryptionRandomKeying.TEST_UTIL 
 
-
+
 private static HBaseTestingUtility
 TestPerColumnFamilyFlush.TEST_UTIL 
 
-
+
 private static HBaseTestingUtility
 TestSeekOptimizations.TEST_UTIL 
 
-
+
 (package private) HBaseTestingUtility
 TestReversibleScanners.TEST_UTIL 
 
-
+
 private static HBaseTestingUtility
 TestMasterAddressTracker.TEST_UTIL 
 
-
+
 private HBaseTestingUtility
 TestAtomicOperation.TEST_UTIL 
 
-
+
 private static HBaseTestingUtility
 TestHdfsSnapshotHRegion.TEST_UTIL 
 
-
+
 private static HBaseTestingUtility
 TestRegionServerReadRequestMetrics.TEST_UTIL 
 
-
+
 private static HBaseTestingUtility
 TestTimestampFilterSeekHint.TEST_UTIL 
 
-
+
 private static HBaseTestingUtility
 TestScanWithBloomError.TEST_UTIL 
 
-
+
 private HBaseTestingUtility
 TestCluster

[05/51] [partial] hbase-site git commit: Published site at 6a5b4f2a5c188f8eef4f2250b8b7db7dd1e750e4.

2018-08-23 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/testdevapidocs/org/apache/hadoop/hbase/regionserver/TestHRegion.FlushThread.html
--
diff --git 
a/testdevapidocs/org/apache/hadoop/hbase/regionserver/TestHRegion.FlushThread.html
 
b/testdevapidocs/org/apache/hadoop/hbase/regionserver/TestHRegion.FlushThread.html
index 6cf1758..c19b494 100644
--- 
a/testdevapidocs/org/apache/hadoop/hbase/regionserver/TestHRegion.FlushThread.html
+++ 
b/testdevapidocs/org/apache/hadoop/hbase/regionserver/TestHRegion.FlushThread.html
@@ -122,7 +122,7 @@ var activeTableTab = "activeTableTab";
 
 
 
-protected class TestHRegion.FlushThread
+protected class TestHRegion.FlushThread
 extends https://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html?is-external=true";
 title="class or interface in java.lang">Thread
 
 
@@ -255,7 +255,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html
 
 
 done
-private volatile boolean done
+private volatile boolean done
 
 
 
@@ -264,7 +264,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html
 
 
 error
-private https://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true";
 title="class or interface in java.lang">Throwable error
+private https://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true";
 title="class or interface in java.lang">Throwable error
 
 
 
@@ -281,7 +281,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html
 
 
 FlushThread
-FlushThread()
+FlushThread()
 
 
 
@@ -298,7 +298,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html
 
 
 done
-public void done()
+public void done()
 
 
 
@@ -307,7 +307,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html
 
 
 checkNoError
-public void checkNoError()
+public void checkNoError()
 
 
 
@@ -316,7 +316,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html
 
 
 run
-public void run()
+public void run()
 
 Specified by:
 https://docs.oracle.com/javase/8/docs/api/java/lang/Runnable.html?is-external=true#run--";
 title="class or interface in java.lang">run in 
interface https://docs.oracle.com/javase/8/docs/api/java/lang/Runnable.html?is-external=true";
 title="class or interface in java.lang">Runnable
@@ -331,7 +331,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html
 
 
 flush
-public void flush()
+public void flush()
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/testdevapidocs/org/apache/hadoop/hbase/regionserver/TestHRegion.GetTillDoneOrException.html
--
diff --git 
a/testdevapidocs/org/apache/hadoop/hbase/regionserver/TestHRegion.GetTillDoneOrException.html
 
b/testdevapidocs/org/apache/hadoop/hbase/regionserver/TestHRegion.GetTillDoneOrException.html
index dfcc2f4..ed2474e 100644
--- 
a/testdevapidocs/org/apache/hadoop/hbase/regionserver/TestHRegion.GetTillDoneOrException.html
+++ 
b/testdevapidocs/org/apache/hadoop/hbase/regionserver/TestHRegion.GetTillDoneOrException.html
@@ -122,7 +122,7 @@ var activeTableTab = "activeTableTab";
 
 
 
-class TestHRegion.GetTillDoneOrException
+class TestHRegion.GetTillDoneOrException
 extends https://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html?is-external=true";
 title="class or interface in java.lang">Thread
 
 
@@ -254,7 +254,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html
 
 
 g
-private final org.apache.hadoop.hbase.client.Get g
+private final org.apache.hadoop.hbase.client.Get g
 
 
 
@@ -263,7 +263,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html
 
 
 done
-private final https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/atomic/AtomicBoolean.html?is-external=true";
 title="class or interface in java.util.concurrent.atomic">AtomicBoolean done
+private final https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/atomic/AtomicBoolean.html?is-external=true";
 title="class or interface in java.util.concurrent.atomic">AtomicBoolean done
 
 
 
@@ -272,7 +272,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html
 
 
 count
-private final https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/atomic/AtomicInteger.html?is-external=true";
 title="class or interface in java.util.concurrent.atomic">AtomicInteger count
+private final https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/atomic/AtomicInteger.html?is-external=true";
 title="class or interface in java.util.concurrent.atomic">AtomicInteger count
 
 
 
@@ -281,7 +281,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html
 
 
 e
-private https://docs.oracle.com/javase/8/docs/api/java/lang/Exception.html?is-external=true";
 title="class or interface in java.lang">Exception e
+private https://docs.oracle.com/ja

[10/51] [partial] hbase-site git commit: Published site at 6a5b4f2a5c188f8eef4f2250b8b7db7dd1e750e4.

2018-08-23 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/license.html
--
diff --git a/license.html b/license.html
index 4e3807b..3c955ea 100644
--- a/license.html
+++ b/license.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase – Project Licenses
 
@@ -491,7 +491,7 @@
 https://www.apache.org/";>The Apache Software 
Foundation.
 All rights reserved.  
 
-  Last Published: 
2018-08-22
+  Last Published: 
2018-08-23
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/mail-lists.html
--
diff --git a/mail-lists.html b/mail-lists.html
index 08614e5..0662a3b 100644
--- a/mail-lists.html
+++ b/mail-lists.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase – Project Mailing Lists
 
@@ -341,7 +341,7 @@
 https://www.apache.org/";>The Apache Software 
Foundation.
 All rights reserved.  
 
-  Last Published: 
2018-08-22
+  Last Published: 
2018-08-23
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/metrics.html
--
diff --git a/metrics.html b/metrics.html
index b827866..dbd21b0 100644
--- a/metrics.html
+++ b/metrics.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase –  
   Apache HBase (TM) Metrics
@@ -459,7 +459,7 @@ export HBASE_REGIONSERVER_OPTS="$HBASE_JMX_OPTS 
-Dcom.sun.management.jmxrem
 https://www.apache.org/";>The Apache Software 
Foundation.
 All rights reserved.  
 
-  Last Published: 
2018-08-22
+  Last Published: 
2018-08-23
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/old_news.html
--
diff --git a/old_news.html b/old_news.html
index 3d1136e..06bb751 100644
--- a/old_news.html
+++ b/old_news.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase – 
   Old Apache HBase (TM) News
@@ -440,7 +440,7 @@ under the License. -->
 https://www.apache.org/";>The Apache Software 
Foundation.
 All rights reserved.  
 
-  Last Published: 
2018-08-22
+  Last Published: 
2018-08-23
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/plugin-management.html
--
diff --git a/plugin-management.html b/plugin-management.html
index 16121d6..bd1c48c 100644
--- a/plugin-management.html
+++ b/plugin-management.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase – Project Plugin Management
 
@@ -440,7 +440,7 @@
 https://www.apache.org/";>The Apache Software 
Foundation.
 All rights reserved.  
 
-  Last Published: 
2018-08-22
+  Last Published: 
2018-08-23
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/plugins.html
--
diff --git a/plugins.html b/plugins.html
index bdb9cf8..03aa14b 100644
--- a/plugins.html
+++ b/plugins.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase – Project Plugins
 
@@ -375,7 +375,7 @@
 https://www.apache.org/";>The Apache Software 
Foundation.
 All rights reserved.  
 
-  Last Published: 
2018-08-22
+  Last Published: 
2018-08-23
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/poweredbyhbase.html
--
diff --git a/poweredbyhbase.html b/poweredbyhbase.html
index f434aca..f1edaaa 100644
--- a/poweredbyhbase.html
+++ b/poweredbyhbase.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase – Powered By Apache HBase™
 
@@ -769,7 +769,7 @@ under the License. -->
 https://www.apache.org/";>The Apache Software 
Foundation.
 All rights reserved.  
 
-  Last Published: 
2018-08-22
+  Last Published: 
2018-08-23
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/project-info.html
--
diff --git a/project-info.html b/project-info.html
index 

[03/51] [partial] hbase-site git commit: Published site at 6a5b4f2a5c188f8eef4f2250b8b7db7dd1e750e4.

2018-08-23 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/testdevapidocs/org/apache/hadoop/hbase/regionserver/TestHRegionWithInMemoryFlush.html
--
diff --git 
a/testdevapidocs/org/apache/hadoop/hbase/regionserver/TestHRegionWithInMemoryFlush.html
 
b/testdevapidocs/org/apache/hadoop/hbase/regionserver/TestHRegionWithInMemoryFlush.html
index cfc19fe..db6ff4d 100644
--- 
a/testdevapidocs/org/apache/hadoop/hbase/regionserver/TestHRegionWithInMemoryFlush.html
+++ 
b/testdevapidocs/org/apache/hadoop/hbase/regionserver/TestHRegionWithInMemoryFlush.html
@@ -18,7 +18,7 @@
 catch(err) {
 }
 //-->
-var methods = {"i0":10};
+var methods = {"i0":10,"i1":10};
 var tabs = {65535:["t0","All Methods"],2:["t2","Instance 
Methods"],8:["t4","Concrete Methods"]};
 var altColor = "altColor";
 var rowColor = "rowColor";
@@ -114,7 +114,7 @@ var activeTableTab = "activeTableTab";
 
 
 
-public class TestHRegionWithInMemoryFlush
+public class TestHRegionWithInMemoryFlush
 extends TestHRegion
 A test similar to TestHRegion, but with in-memory flush 
families.
  Also checks wal truncation after in-memory compaction.
@@ -204,6 +204,12 @@ extends 
+void
+testFlushAndMemstoreSizeCounting()
+A test case of HBASE-21041
+
+
 
 
 
@@ -239,7 +245,7 @@ extends 
 
 CLASS_RULE
-public static final HBaseClassTestRule CLASS_RULE
+public static final HBaseClassTestRule CLASS_RULE
 
 
 
@@ -256,7 +262,7 @@ extends 
 
 TestHRegionWithInMemoryFlush
-public TestHRegionWithInMemoryFlush()
+public TestHRegionWithInMemoryFlush()
 
 
 
@@ -270,10 +276,10 @@ extends 
 
 
-
+
 
 initHRegion
-public org.apache.hadoop.hbase.regionserver.HRegion initHRegion(org.apache.hadoop.hbase.TableName tableName,
+public org.apache.hadoop.hbase.regionserver.HRegion initHRegion(org.apache.hadoop.hbase.TableName tableName,
 
byte[] startKey,
 
byte[] stopKey,
 
boolean isReadOnly,
@@ -292,6 +298,23 @@ extends 
+
+
+
+
+testFlushAndMemstoreSizeCounting
+public void testFlushAndMemstoreSizeCounting()
+  throws https://docs.oracle.com/javase/8/docs/api/java/lang/Exception.html?is-external=true";
 title="class or interface in java.lang">Exception
+A test case of HBASE-21041
+
+Overrides:
+testFlushAndMemstoreSizeCounting in
 class TestHRegion
+Throws:
+https://docs.oracle.com/javase/8/docs/api/java/lang/Exception.html?is-external=true";
 title="class or interface in java.lang">Exception - Exception
+
+
+
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/testdevapidocs/org/apache/hadoop/hbase/regionserver/TestHStore.MyCompactingMemStore.html
--
diff --git 
a/testdevapidocs/org/apache/hadoop/hbase/regionserver/TestHStore.MyCompactingMemStore.html
 
b/testdevapidocs/org/apache/hadoop/hbase/regionserver/TestHStore.MyCompactingMemStore.html
index 4ae96ed..ee5fdef 100644
--- 
a/testdevapidocs/org/apache/hadoop/hbase/regionserver/TestHStore.MyCompactingMemStore.html
+++ 
b/testdevapidocs/org/apache/hadoop/hbase/regionserver/TestHStore.MyCompactingMemStore.html
@@ -187,7 +187,7 @@ extends 
org.apache.hadoop.hbase.regionserver.CompactingMemStore
 
 
 Fields inherited from 
class org.apache.hadoop.hbase.regionserver.AbstractMemStore
-FIXED_OVERHEAD, snapshot, snapshotId
+FIXED_OVERHEAD, regionServices, snapshot, snapshotId
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/testdevapidocs/org/apache/hadoop/hbase/regionserver/TestHStore.MyCompactingMemStoreWithCustomCompactor.html
--
diff --git 
a/testdevapidocs/org/apache/hadoop/hbase/regionserver/TestHStore.MyCompactingMemStoreWithCustomCompactor.html
 
b/testdevapidocs/org/apache/hadoop/hbase/regionserver/TestHStore.MyCompactingMemStoreWithCustomCompactor.html
index a19a39b..80b8746 100644
--- 
a/testdevapidocs/org/apache/hadoop/hbase/regionserver/TestHStore.MyCompactingMemStoreWithCustomCompactor.html
+++ 
b/testdevapidocs/org/apache/hadoop/hbase/regionserver/TestHStore.MyCompactingMemStoreWithCustomCompactor.html
@@ -179,7 +179,7 @@ extends 
org.apache.hadoop.hbase.regionserver.CompactingMemStore
 
 
 Fields inherited from 
class org.apache.hadoop.hbase.regionserver.AbstractMemStore
-FIXED_OVERHEAD, snapshot, snapshotId
+FIXED_OVERHEAD, regionServices, snapshot, snapshotId
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/testdevapidocs/org/apache/hadoop/hbase/regionserver/TestHeapMemoryManager.RegionServerAccountingStub.html
--
diff --git 
a/testdevapidocs/org/apache/hadoop/hbase/regionserver/TestHeapMemoryManager.RegionServerAccountingStub.html
 
b/testdevapidocs/org

[24/51] [partial] hbase-site git commit: Published site at 6a5b4f2a5c188f8eef4f2250b8b7db7dd1e750e4.

2018-08-23 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.MutationBatchOperation.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.MutationBatchOperation.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.MutationBatchOperation.html
index db8431b..a8cb7c4 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.MutationBatchOperation.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.MutationBatchOperation.html
@@ -885,7766 +885,7797 @@
 877   * @return What the next sequence 
(edit) id should be.
 878   * @throws IOException e
 879   */
-880  private long initialize(final 
CancelableProgressable reporter) throws IOException {
-881
-882//Refuse to open the region if there 
is no column family in the table
-883if 
(htableDescriptor.getColumnFamilyCount() == 0) {
-884  throw new 
DoNotRetryIOException("Table " + 
htableDescriptor.getTableName().getNameAsString()+
-885  " should have at least one 
column family.");
-886}
-887
-888MonitoredTask status = 
TaskMonitor.get().createStatus("Initializing region " + this);
-889long nextSeqId = -1;
-890try {
-891  nextSeqId = 
initializeRegionInternals(reporter, status);
-892  return nextSeqId;
-893} finally {
-894  // nextSeqid will be -1 if the 
initialization fails.
-895  // At least it will be 0 
otherwise.
-896  if (nextSeqId == -1) {
-897status.abort("Exception during 
region " + getRegionInfo().getRegionNameAsString() +
-898  " initialization.");
-899  }
-900}
-901  }
-902
-903  private long 
initializeRegionInternals(final CancelableProgressable reporter,
-904  final MonitoredTask status) throws 
IOException {
-905if (coprocessorHost != null) {
-906  status.setStatus("Running 
coprocessor pre-open hook");
-907  coprocessorHost.preOpen();
-908}
-909
-910// Write HRI to a file in case we 
need to recover hbase:meta
-911// Only the primary replica should 
write .regioninfo
-912if 
(this.getRegionInfo().getReplicaId() == RegionInfo.DEFAULT_REPLICA_ID) {
-913  status.setStatus("Writing region 
info on filesystem");
-914  fs.checkRegionInfoOnFilesystem();
-915}
-916
-917// Initialize all the HStores
-918status.setStatus("Initializing all 
the Stores");
-919long maxSeqId = 
initializeStores(reporter, status);
-920this.mvcc.advanceTo(maxSeqId);
-921if 
(ServerRegionReplicaUtil.shouldReplayRecoveredEdits(this)) {
-922  Collection stores = 
this.stores.values();
-923  try {
-924// update the stores that we are 
replaying
-925LOG.debug("replaying wal for " + 
this.getRegionInfo().getEncodedName());
-926
stores.forEach(HStore::startReplayingFromWAL);
-927// Recover any edits if 
available.
-928maxSeqId = Math.max(maxSeqId,
-929  
replayRecoveredEditsIfAny(this.fs.getRegionDir(), maxSeqIdInStores, reporter, 
status));
-930// Make sure mvcc is up to max.
-931this.mvcc.advanceTo(maxSeqId);
-932  } finally {
-933LOG.debug("stopping wal replay 
for " + this.getRegionInfo().getEncodedName());
-934// update the stores that we are 
done replaying
-935
stores.forEach(HStore::stopReplayingFromWAL);
-936  }
-937}
-938this.lastReplayedOpenRegionSeqId = 
maxSeqId;
-939
-940
this.writestate.setReadOnly(ServerRegionReplicaUtil.isReadOnly(this));
-941this.writestate.flushRequested = 
false;
-942this.writestate.compacting.set(0);
-943
-944if (this.writestate.writesEnabled) 
{
-945  LOG.debug("Cleaning up temporary 
data for " + this.getRegionInfo().getEncodedName());
-946  // Remove temporary data left over 
from old regions
-947  status.setStatus("Cleaning up 
temporary data from old regions");
-948  fs.cleanupTempDir();
-949}
-950
-951if (this.writestate.writesEnabled) 
{
-952  status.setStatus("Cleaning up 
detritus from prior splits");
-953  // Get rid of any splits or merges 
that were lost in-progress.  Clean out
-954  // these directories here on open.  
We may be opening a region that was
-955  // being split but we crashed in 
the middle of it all.
-956  LOG.debug("Cleaning up detritus for 
" + this.getRegionInfo().getEncodedName());
-957  fs.cleanupAnySplitDetritus();
-958  fs.cleanupMergesDir();
-959}
+880  @VisibleForTesting
+881  long initialize(final 
CancelableProgressable reporter) throws IOException {
+882
+883//Refuse to open the region if there 
is no column family in the table
+884if 
(htableDescriptor.getColumnFamilyCount() == 0) {
+885  throw new 
DoNotRetryIOException("Table " + 
htableDescriptor.getTableName().getNameAsString()+
+886  " should have at least one 
column 

[06/51] [partial] hbase-site git commit: Published site at 6a5b4f2a5c188f8eef4f2250b8b7db7dd1e750e4.

2018-08-23 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/testdevapidocs/org/apache/hadoop/hbase/master/procedure/TestCreateTableProcedure.html
--
diff --git 
a/testdevapidocs/org/apache/hadoop/hbase/master/procedure/TestCreateTableProcedure.html
 
b/testdevapidocs/org/apache/hadoop/hbase/master/procedure/TestCreateTableProcedure.html
index 5d2c1df..7959886 100644
--- 
a/testdevapidocs/org/apache/hadoop/hbase/master/procedure/TestCreateTableProcedure.html
+++ 
b/testdevapidocs/org/apache/hadoop/hbase/master/procedure/TestCreateTableProcedure.html
@@ -18,7 +18,7 @@
 catch(err) {
 }
 //-->
-var methods = 
{"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10};
+var methods = 
{"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10};
 var tabs = {65535:["t0","All Methods"],2:["t2","Instance 
Methods"],8:["t4","Concrete Methods"]};
 var altColor = "altColor";
 var rowColor = "rowColor";
@@ -114,7 +114,7 @@ var activeTableTab = "activeTableTab";
 
 
 
-public class TestCreateTableProcedure
+public class TestCreateTableProcedure
 extends TestTableDDLProcedureBase
 
 
@@ -166,10 +166,6 @@ extends F2 
 
 
-private static org.slf4j.Logger
-LOG 
-
-
 org.junit.rules.TestName
 name 
 
@@ -222,38 +218,34 @@ extends 
 void
-testMRegions() 
-
-
-void
 testOnHDFSFailure() 
 
-
+
 void
 testRecoveryAndDoubleExecution() 
 
-
+
 void
 testRollbackAndDoubleExecution() 
 
-
+
 private void
 testRollbackAndDoubleExecution(org.apache.hadoop.hbase.client.TableDescriptorBuilder builder) 
 
-
+
 void
 testRollbackAndDoubleExecutionOnMobTable() 
 
-
+
 void
 testSimpleCreate() 
 
-
+
 private void
 testSimpleCreate(org.apache.hadoop.hbase.TableName tableName,
 byte[][] splitKeys) 
 
-
+
 void
 testSimpleCreateWithSplits() 
 
@@ -292,16 +284,7 @@ extends 
 
 CLASS_RULE
-public static final HBaseClassTestRule CLASS_RULE
-
-
-
-
-
-
-
-LOG
-private static final org.slf4j.Logger LOG
+public static final HBaseClassTestRule CLASS_RULE
 
 
 
@@ -310,7 +293,7 @@ extends 
 
 F1
-private static final https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String F1
+private static final https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String F1
 
 See Also:
 Constant
 Field Values
@@ -323,7 +306,7 @@ extends 
 
 F2
-private static final https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String F2
+private static final https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String F2
 
 See Also:
 Constant
 Field Values
@@ -336,7 +319,7 @@ extends 
 
 name
-public org.junit.rules.TestName name
+public org.junit.rules.TestName name
 
 
 
@@ -353,7 +336,7 @@ extends 
 
 TestCreateTableProcedure
-public TestCreateTableProcedure()
+public TestCreateTableProcedure()
 
 
 
@@ -370,7 +353,7 @@ extends 
 
 testSimpleCreate
-public void testSimpleCreate()
+public void testSimpleCreate()
   throws https://docs.oracle.com/javase/8/docs/api/java/lang/Exception.html?is-external=true";
 title="class or interface in java.lang">Exception
 
 Throws:
@@ -384,7 +367,7 @@ extends 
 
 testSimpleCreateWithSplits
-public void testSimpleCreateWithSplits()
+public void testSimpleCreateWithSplits()
 throws https://docs.oracle.com/javase/8/docs/api/java/lang/Exception.html?is-external=true";
 title="class or interface in java.lang">Exception
 
 Throws:
@@ -398,7 +381,7 @@ extends 
 
 testSimpleCreate
-private void testSimpleCreate(org.apache.hadoop.hbase.TableName tableName,
+private void testSimpleCreate(org.apache.hadoop.hbase.TableName tableName,
   byte[][] splitKeys)
throws https://docs.oracle.com/javase/8/docs/api/java/lang/Exception.html?is-external=true";
 title="class or interface in java.lang">Exception
 
@@ -413,7 +396,7 @@ extends 
 
 testCreateWithoutColumnFamily
-public void testCreateWithoutColumnFamily()
+public void testCreateWithoutColumnFamily()
throws https://docs.oracle.com/javase/8/docs/api/java/lang/Exception.html?is-external=true";
 title="class or interface in java.lang">Exception
 
 Throws:
@@ -427,7 +410,7 @@ extends 
 
 testCreateExisting
-public void testCreateExisting()
+public void testCreateExisting()
 throws https://docs.oracle.com/javase/8/docs/api/java/lang/Exception.html?is-external=true";
 title="class or interface in java.lang">Exception
 
 Throws:
@@ -441,7 +424,7 @@ extends 
 
 testRecoveryAndDoubleExecution
-public void testRecoveryAndDoubleExecution()
+public void testRecoveryAndDoubleExecution()

[13/51] [partial] hbase-site git commit: Published site at 6a5b4f2a5c188f8eef4f2250b8b7db7dd1e750e4.

2018-08-23 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/MutableSegment.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/MutableSegment.html 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/MutableSegment.html
index 598cf81..26b0752 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/MutableSegment.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/MutableSegment.html
@@ -55,88 +55,92 @@
 047  + ClassSize.REFERENCE
 048  + ClassSize.ATOMIC_BOOLEAN);
 049
-050  protected MutableSegment(CellSet 
cellSet, CellComparator comparator, MemStoreLAB memStoreLAB) {
-051super(cellSet, comparator, 
memStoreLAB, TimeRangeTracker.create(TimeRangeTracker.Type.SYNC));
-052incMemStoreSize(0, DEEP_OVERHEAD, 0); 
// update the mutable segment metadata
-053  }
-054
-055  /**
-056   * Adds the given cell into the 
segment
-057   * @param cell the cell to add
-058   * @param mslabUsed whether using 
MSLAB
-059   */
-060  public void add(Cell cell, boolean 
mslabUsed, MemStoreSizing memStoreSizing,
-061  boolean sizeAddedPreOperation) {
-062internalAdd(cell, mslabUsed, 
memStoreSizing, sizeAddedPreOperation);
-063  }
-064
-065  public void upsert(Cell cell, long 
readpoint, MemStoreSizing memStoreSizing,
-066  boolean sizeAddedPreOperation) {
-067internalAdd(cell, false, 
memStoreSizing, sizeAddedPreOperation);
+050  protected MutableSegment(CellSet 
cellSet, CellComparator comparator,
+051  MemStoreLAB memStoreLAB, 
MemStoreSizing memstoreSizing) {
+052super(cellSet, comparator, 
memStoreLAB, TimeRangeTracker.create(TimeRangeTracker.Type.SYNC));
+053incMemStoreSize(0, DEEP_OVERHEAD, 0); 
// update the mutable segment metadata
+054if (memstoreSizing != null) {
+055  memstoreSizing.incMemStoreSize(0, 
DEEP_OVERHEAD, 0);
+056}
+057  }
+058
+059  /**
+060   * Adds the given cell into the 
segment
+061   * @param cell the cell to add
+062   * @param mslabUsed whether using 
MSLAB
+063   */
+064  public void add(Cell cell, boolean 
mslabUsed, MemStoreSizing memStoreSizing,
+065  boolean sizeAddedPreOperation) {
+066internalAdd(cell, mslabUsed, 
memStoreSizing, sizeAddedPreOperation);
+067  }
 068
-069// Get the Cells for the 
row/family/qualifier regardless of timestamp.
-070// For this case we want to clean up 
any other puts
-071Cell firstCell = 
PrivateCellUtil.createFirstOnRowColTS(cell, HConstants.LATEST_TIMESTAMP);
-072SortedSet ss = 
this.tailSet(firstCell);
-073Iterator it = 
ss.iterator();
-074// versions visible to oldest 
scanner
-075int versionsVisible = 0;
-076while (it.hasNext()) {
-077  Cell cur = it.next();
-078
-079  if (cell == cur) {
-080// ignore the one just put in
-081continue;
-082  }
-083  // check that this is the row and 
column we are interested in, otherwise bail
-084  if (CellUtil.matchingRows(cell, 
cur) && CellUtil.matchingQualifier(cell, cur)) {
-085// only remove Puts that 
concurrent scanners cannot possibly see
-086if (cur.getTypeByte() == 
KeyValue.Type.Put.getCode() && cur.getSequenceId() <= readpoint) {
-087  if (versionsVisible >= 1) 
{
-088// if we get here we have 
seen at least one version visible to the oldest scanner,
-089// which means we can prove 
that no scanner will see this version
-090
-091// false means there was a 
change, so give us the size.
-092// TODO when the removed cell 
ie.'cur' having its data in MSLAB, we can not release that
-093// area. Only the Cell object 
as such going way. We need to consider cellLen to be
-094// decreased there as 0 only. 
Just keeping it as existing code now. We need to know the
-095// removed cell is from MSLAB 
or not. Will do once HBASE-16438 is in
-096int cellLen = 
getCellLength(cur);
-097long heapSize = 
heapSizeChange(cur, true);
-098long offHeapSize = 
offHeapSizeChange(cur, true);
-099incMemStoreSize(-cellLen, 
-heapSize, -offHeapSize);
-100if (memStoreSizing != null) 
{
-101  
memStoreSizing.decMemStoreSize(cellLen, heapSize, offHeapSize);
-102}
-103it.remove();
-104  } else {
-105versionsVisible++;
-106  }
-107}
-108  } else {
-109// past the row or column, done
-110break;
-111  }
-112}
-113  }
-114
-115  public boolean setInMemoryFlushed() {
-116return flushed.compareAndSet(false, 
true);
+069  public void upsert(Cell cell, long 
readpoint, MemStoreSizing memStoreSizing,
+070  boolean sizeAddedPreOperation) {
+071internalAdd(cell, false, 
memStoreSizing, sizeAddedPreOperation);
+072
+073// Get the Cells for th

[19/51] [partial] hbase-site git commit: Published site at 6a5b4f2a5c188f8eef4f2250b8b7db7dd1e750e4.

2018-08-23 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RowLockContext.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RowLockContext.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RowLockContext.html
index db8431b..a8cb7c4 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RowLockContext.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RowLockContext.html
@@ -885,7766 +885,7797 @@
 877   * @return What the next sequence 
(edit) id should be.
 878   * @throws IOException e
 879   */
-880  private long initialize(final 
CancelableProgressable reporter) throws IOException {
-881
-882//Refuse to open the region if there 
is no column family in the table
-883if 
(htableDescriptor.getColumnFamilyCount() == 0) {
-884  throw new 
DoNotRetryIOException("Table " + 
htableDescriptor.getTableName().getNameAsString()+
-885  " should have at least one 
column family.");
-886}
-887
-888MonitoredTask status = 
TaskMonitor.get().createStatus("Initializing region " + this);
-889long nextSeqId = -1;
-890try {
-891  nextSeqId = 
initializeRegionInternals(reporter, status);
-892  return nextSeqId;
-893} finally {
-894  // nextSeqid will be -1 if the 
initialization fails.
-895  // At least it will be 0 
otherwise.
-896  if (nextSeqId == -1) {
-897status.abort("Exception during 
region " + getRegionInfo().getRegionNameAsString() +
-898  " initialization.");
-899  }
-900}
-901  }
-902
-903  private long 
initializeRegionInternals(final CancelableProgressable reporter,
-904  final MonitoredTask status) throws 
IOException {
-905if (coprocessorHost != null) {
-906  status.setStatus("Running 
coprocessor pre-open hook");
-907  coprocessorHost.preOpen();
-908}
-909
-910// Write HRI to a file in case we 
need to recover hbase:meta
-911// Only the primary replica should 
write .regioninfo
-912if 
(this.getRegionInfo().getReplicaId() == RegionInfo.DEFAULT_REPLICA_ID) {
-913  status.setStatus("Writing region 
info on filesystem");
-914  fs.checkRegionInfoOnFilesystem();
-915}
-916
-917// Initialize all the HStores
-918status.setStatus("Initializing all 
the Stores");
-919long maxSeqId = 
initializeStores(reporter, status);
-920this.mvcc.advanceTo(maxSeqId);
-921if 
(ServerRegionReplicaUtil.shouldReplayRecoveredEdits(this)) {
-922  Collection stores = 
this.stores.values();
-923  try {
-924// update the stores that we are 
replaying
-925LOG.debug("replaying wal for " + 
this.getRegionInfo().getEncodedName());
-926
stores.forEach(HStore::startReplayingFromWAL);
-927// Recover any edits if 
available.
-928maxSeqId = Math.max(maxSeqId,
-929  
replayRecoveredEditsIfAny(this.fs.getRegionDir(), maxSeqIdInStores, reporter, 
status));
-930// Make sure mvcc is up to max.
-931this.mvcc.advanceTo(maxSeqId);
-932  } finally {
-933LOG.debug("stopping wal replay 
for " + this.getRegionInfo().getEncodedName());
-934// update the stores that we are 
done replaying
-935
stores.forEach(HStore::stopReplayingFromWAL);
-936  }
-937}
-938this.lastReplayedOpenRegionSeqId = 
maxSeqId;
-939
-940
this.writestate.setReadOnly(ServerRegionReplicaUtil.isReadOnly(this));
-941this.writestate.flushRequested = 
false;
-942this.writestate.compacting.set(0);
-943
-944if (this.writestate.writesEnabled) 
{
-945  LOG.debug("Cleaning up temporary 
data for " + this.getRegionInfo().getEncodedName());
-946  // Remove temporary data left over 
from old regions
-947  status.setStatus("Cleaning up 
temporary data from old regions");
-948  fs.cleanupTempDir();
-949}
-950
-951if (this.writestate.writesEnabled) 
{
-952  status.setStatus("Cleaning up 
detritus from prior splits");
-953  // Get rid of any splits or merges 
that were lost in-progress.  Clean out
-954  // these directories here on open.  
We may be opening a region that was
-955  // being split but we crashed in 
the middle of it all.
-956  LOG.debug("Cleaning up detritus for 
" + this.getRegionInfo().getEncodedName());
-957  fs.cleanupAnySplitDetritus();
-958  fs.cleanupMergesDir();
-959}
+880  @VisibleForTesting
+881  long initialize(final 
CancelableProgressable reporter) throws IOException {
+882
+883//Refuse to open the region if there 
is no column family in the table
+884if 
(htableDescriptor.getColumnFamilyCount() == 0) {
+885  throw new 
DoNotRetryIOException("Table " + 
htableDescriptor.getTableName().getNameAsString()+
+886  " should have at least one 
column family.");
+887}
+888
+889Monito

[38/51] [partial] hbase-site git commit: Published site at 6a5b4f2a5c188f8eef4f2250b8b7db7dd1e750e4.

2018-08-23 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/MemStoreSize.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/MemStoreSize.html 
b/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/MemStoreSize.html
index 4012248..e3d2db2 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/MemStoreSize.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/MemStoreSize.html
@@ -134,6 +134,12 @@
long currentSeqId) 
 
 
+MemStoreSize
+HRegion.dropMemStoreContents()
+Be careful, this method will drop all data in the memstore 
of this region.
+
+
+
 private MemStoreSize
 HRegion.dropMemStoreContentsForSeqId(long seqId,
 HStore store)
@@ -141,105 +147,105 @@
  if the memstore edits have seqNums smaller than the given seq id
 
 
-
+
 MemStoreSize
 HStore.getFlushableSize() 
 
-
+
 MemStoreSize
 MemStore.getFlushableSize()
 Flush will first clear out the data in snapshot if any (It 
will take a second flush
  invocation to clear the current Cell set).
 
 
-
+
 MemStoreSize
 CompactingMemStore.getFlushableSize() 
 
-
+
 MemStoreSize
 Store.getFlushableSize() 
 
-
+
 MemStoreSize
 DefaultMemStore.getFlushableSize() 
 
-
+
 MemStoreSize
 HStore.getMemStoreSize() 
 
-
+
 MemStoreSize
 NonThreadSafeMemStoreSizing.getMemStoreSize() 
 
-
+
 MemStoreSize
 Segment.getMemStoreSize() 
 
-
+
 MemStoreSize
 ThreadSafeMemStoreSizing.getMemStoreSize() 
 
-
+
 MemStoreSize
 MemStoreSizing.getMemStoreSize() 
 
-
+
 MemStoreSize
 Store.getMemStoreSize() 
 
-
+
 MemStoreSize
 MemStoreSnapshot.getMemStoreSize() 
 
-
+
 MemStoreSize
 CompactionPipeline.getPipelineSize() 
 
-
+
 MemStoreSize
 AbstractMemStore.getSnapshotSize() 
 
-
+
 MemStoreSize
 HStore.getSnapshotSize() 
 
-
+
 MemStoreSize
 MemStore.getSnapshotSize()
 Return the size of the snapshot(s) if any
 
 
-
+
 MemStoreSize
 Store.getSnapshotSize() 
 
-
+
 MemStoreSize
 CompactionPipeline.getTailSize() 
 
-
+
 MemStoreSize
 StoreFlushContext.prepare()
 Prepare for a store flush (create snapshot)
  Requires pausing writes.
 
 
-
+
 MemStoreSize
 HStore.StoreFlusherImpl.prepare()
 This is not thread safe.
 
 
-
+
 MemStoreSize
 MemStore.size() 
 
-
+
 MemStoreSize
 CompactingMemStore.size() 
 
-
+
 MemStoreSize
 DefaultMemStore.size() 
 
@@ -253,36 +259,29 @@
 
 
 
-void
-RegionServerAccounting.addRegionReplayEditsSize(byte[] regionName,
-MemStoreSize memStoreSize)
-Add memStoreSize to replayEditsPerRegion.
-
-
-
 default long
 MemStoreSizing.decMemStoreSize(MemStoreSize delta) 
 
-
+
 (package private) void
 HRegion.decrMemStoreSize(MemStoreSize mss) 
 
-
+
 (package private) void
 RegionServerAccounting.incGlobalMemStoreSize(MemStoreSize mss) 
 
-
+
 (package private) void
 HRegion.incMemStoreSize(MemStoreSize mss)
 Increase the size of mem store in this region and the size 
of global mem
  store
 
 
-
+
 default long
 MemStoreSizing.incMemStoreSize(MemStoreSize delta) 
 
-
+
 private boolean
 HRegion.isFlushSize(MemStoreSize size) 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/MemStoreSizing.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/MemStoreSizing.html 
b/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/MemStoreSizing.html
index d6eaaf7..1029e86 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/MemStoreSizing.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/MemStoreSizing.html
@@ -189,19 +189,6 @@
 
 
 
-
-Fields in org.apache.hadoop.hbase.regionserver
 with type parameters of type MemStoreSizing 
-
-Modifier and Type
-Field and Description
-
-
-
-private https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/ConcurrentMap.html?is-external=true";
 title="class or interface in 
java.util.concurrent">ConcurrentMap
-RegionServerAccounting.replayEditsPerRegion 
-
-
-
 
 Methods in org.apache.hadoop.hbase.regionserver
 with parameters of type MemStoreSizing 
 
@@ -291,11 +278,22 @@
 
 
 ImmutableSegment
+SegmentFactory.createImmutableSegment(MutableSegment segment,
+  MemStoreSizing memstoreSizing) 
+
+
+ImmutableSegment
 SegmentFactory.createImmutableSegmentByFlattening(CSLMImmutableSegment segment,
   CompactingMemStore.IndexType idxType,
   MemStoreSizing memstoreSizing,
   MemStoreCompactionStrategy.Action action) 
 
+
+MutableSegment
+SegmentFactory.createMutableSegment(org.apache.hadoop.conf.Configuration conf,
+CellComparator comparator,
+MemStoreSizing memstoreSizing) 
+
 

[50/51] [partial] hbase-site git commit: Published site at 6a5b4f2a5c188f8eef4f2250b8b7db7dd1e750e4.

2018-08-23 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/checkstyle-aggregate.html
--
diff --git a/checkstyle-aggregate.html b/checkstyle-aggregate.html
index b279072..6098f6a 100644
--- a/checkstyle-aggregate.html
+++ b/checkstyle-aggregate.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase – Checkstyle Results
 
@@ -281,10 +281,10 @@
  Warnings
  Errors
 
-3712
+3714
 0
 0
-15385
+15386
 
 Files
 
@@ -8069,1651 +8069,1656 @@
 0
 4
 
+org/apache/hadoop/hbase/rest/model/TestScannerModel.java
+0
+0
+1
+
 org/apache/hadoop/hbase/rest/model/TestStorageClusterStatusModel.java
 0
 0
 7
-
+
 org/apache/hadoop/hbase/rest/model/VersionModel.java
 0
 0
 3
-
+
 org/apache/hadoop/hbase/rest/provider/JAXBContextResolver.java
 0
 0
 1
-
+
 org/apache/hadoop/hbase/rest/provider/consumer/ProtobufMessageBodyConsumer.java
 0
 0
 1
-
+
 org/apache/hadoop/hbase/rest/provider/producer/PlainTextMessageBodyProducer.java
 0
 0
 1
-
+
 org/apache/hadoop/hbase/rsgroup/IntegrationTestRSGroup.java
 0
 0
 1
-
+
 org/apache/hadoop/hbase/rsgroup/RSGroupAdminServer.java
 0
 0
 2
-
+
 org/apache/hadoop/hbase/rsgroup/RSGroupInfoManagerImpl.java
 0
 0
 2
-
+
 org/apache/hadoop/hbase/security/AbstractHBaseSaslRpcClient.java
 0
 0
 5
-
+
 org/apache/hadoop/hbase/security/AccessDeniedException.java
 0
 0
 1
-
+
 org/apache/hadoop/hbase/security/AuthMethod.java
 0
 0
 2
-
+
 org/apache/hadoop/hbase/security/CryptoAESUnwrapHandler.java
 0
 0
 2
-
+
 org/apache/hadoop/hbase/security/CryptoAESWrapHandler.java
 0
 0
 2
-
+
 org/apache/hadoop/hbase/security/EncryptionUtil.java
 0
 0
 6
-
+
 org/apache/hadoop/hbase/security/HBaseKerberosUtils.java
 0
 0
 2
-
+
 org/apache/hadoop/hbase/security/HBasePolicyProvider.java
 0
 0
 7
-
+
 org/apache/hadoop/hbase/security/HBaseSaslRpcClient.java
 0
 0
 6
-
+
 org/apache/hadoop/hbase/security/HBaseSaslRpcServer.java
 0
 0
 1
-
+
 org/apache/hadoop/hbase/security/HadoopSecurityEnabledUserProviderForTesting.java
 0
 0
 1
-
+
 org/apache/hadoop/hbase/security/NettyHBaseRpcConnectionHeaderHandler.java
 0
 0
 2
-
+
 org/apache/hadoop/hbase/security/NettyHBaseSaslRpcClient.java
 0
 0
 2
-
+
 org/apache/hadoop/hbase/security/NettyHBaseSaslRpcClientHandler.java
 0
 0
 2
-
+
 org/apache/hadoop/hbase/security/SaslChallengeDecoder.java
 0
 0
 2
-
+
 org/apache/hadoop/hbase/security/SaslStatus.java
 0
 0
 2
-
+
 org/apache/hadoop/hbase/security/SaslUnwrapHandler.java
 0
 0
 1
-
+
 org/apache/hadoop/hbase/security/SaslUtil.java
 0
 0
 1
-
+
 org/apache/hadoop/hbase/security/SecurityInfo.java
 0
 0
 1
-
+
 org/apache/hadoop/hbase/security/SecurityUtil.java
 0
 0
 1
-
+
 org/apache/hadoop/hbase/security/Superusers.java
 0
 0
 1
-
+
 org/apache/hadoop/hbase/security/TestSecureIPC.java
 0
 0
 5
-
+
 org/apache/hadoop/hbase/security/TestUser.java
 0
 0
 1
-
+
 org/apache/hadoop/hbase/security/TestUsersOperationsWithSecureHadoop.java
 0
 0
 1
-
+
 org/apache/hadoop/hbase/security/User.java
 0
 0
 6
-
+
 org/apache/hadoop/hbase/security/access/AccessControlClient.java
 0
 0
 48
-
+
 org/apache/hadoop/hbase/security/access/AccessControlFilter.java
 0
 0
 3
-
+
 org/apache/hadoop/hbase/security/access/AccessControlLists.java
 0
 0
 16
-
+
 org/apache/hadoop/hbase/security/access/AccessControlUtil.java
 0
 0
 40
-
+
 org/apache/hadoop/hbase/security/access/AccessController.java
 0
 0
 25
-
+
 org/apache/hadoop/hbase/security/access/AuthResult.java
 0
 0
 4
-
+
 org/apache/hadoop/hbase/security/access/Permission.java
 0
 0
 7
-
+
 org/apache/hadoop/hbase/security/access/SecureTestUtil.java
 0
 0
 8
-
+
 org/apache/hadoop/hbase/security/access/ShadedAccessControlUtil.java
 0
 0
 49
-
+
 org/apache/hadoop/hbase/security/access/TableAuthManager.java
 0
 0
 43
-
+
 org/apache/hadoop/hbase/security/access/TablePermission.java
 0
 0
 10
-
+
 org/apache/hadoop/hbase/security/access/TestAccessControlFilter.java
 0
 0
 1
-
+
 org/apache/hadoop/hbase/security/access/TestAccessController.java
 0
 0
 18
-
+
 org/apache/hadoop/hbase/security/access/TestAccessController2.java
 0
 0
 3
-
+
 org/apache/hadoop/hbase/security/access/TestCellACLWithMultipleVersions.java
 0
 0
 1
-
+
 org/apache/hadoop/hbase/security/access/TestCellACLs.java
 0
 0
 2
-
+
 org/apache/hadoop/hbase/security/access/TestNamespaceCommands.java
 0
 0
 2
-
+
 org/apache/hadoop/hbase/security/access/TestTablePermissions.java
 0
 0
 13
-
+
 org/apache/hadoop/hbase/security/access/TestWithDisabledAuthorization.java
 0
 0
 5
-
+
 org/apache/hadoop/hbase/security/access/UserPermission.java
 0
 0
 3
-
+
 org/apache/hadoop/hbase/security/access/ZKPermissionWatcher.java
 0
 0
 5
-
+
 org/apache/hadoop/hbase/security/token/AuthenticationKey.java
 0
 0
 2
-
+
 org/apache/hadoop/hbase/security/token/AuthenticationTokenIdentifier.java
 0
 0
 2
-
+
 org/apache/hadoop/hbase/security/token/AuthenticationTokenSecretManager.java
 0
 0
 4
-
+
 org/apache/hadoop/hbase/security/token/Authenticati

[46/51] [partial] hbase-site git commit: Published site at 6a5b4f2a5c188f8eef4f2250b8b7db7dd1e750e4.

2018-08-23 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/devapidocs/org/apache/hadoop/hbase/regionserver/CSLMImmutableSegment.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/regionserver/CSLMImmutableSegment.html 
b/devapidocs/org/apache/hadoop/hbase/regionserver/CSLMImmutableSegment.html
index 8f25e59..77384c1 100644
--- a/devapidocs/org/apache/hadoop/hbase/regionserver/CSLMImmutableSegment.html
+++ b/devapidocs/org/apache/hadoop/hbase/regionserver/CSLMImmutableSegment.html
@@ -189,7 +189,8 @@ extends 
 protected 
-CSLMImmutableSegment(Segment segment)
+CSLMImmutableSegment(Segment segment,
+MemStoreSizing memstoreSizing)
 
  Copy C-tor to be used when new CSLMImmutableSegment is being built from a 
Mutable one.
 
@@ -277,13 +278,14 @@ extends 
+
 
 
 
 
 CSLMImmutableSegment
-protected CSLMImmutableSegment(Segment segment)
+protected CSLMImmutableSegment(Segment segment,
+   MemStoreSizing memstoreSizing)
 
  Copy C-tor to be used when new CSLMImmutableSegment is being built from a 
Mutable one.
  This C-tor should be used when active MutableSegment is pushed into the 
compaction
@@ -304,7 +306,7 @@ extends 
 
 indexEntrySize
-protected long indexEntrySize()
+protected long indexEntrySize()
 
 Specified by:
 indexEntrySize in
 class Segment
@@ -317,7 +319,7 @@ extends 
 
 canBeFlattened
-protected boolean canBeFlattened()
+protected boolean canBeFlattened()
 
 Specified by:
 canBeFlattened in
 class ImmutableSegment

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/devapidocs/org/apache/hadoop/hbase/regionserver/ChunkCreator.ChunkType.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/regionserver/ChunkCreator.ChunkType.html 
b/devapidocs/org/apache/hadoop/hbase/regionserver/ChunkCreator.ChunkType.html
index 7352032..3a13db9 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/regionserver/ChunkCreator.ChunkType.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/regionserver/ChunkCreator.ChunkType.html
@@ -249,7 +249,7 @@ the order they are declared.
 
 
 values
-public static ChunkCreator.ChunkType[] values()
+public static ChunkCreator.ChunkType[] values()
 Returns an array containing the constants of this enum 
type, in
 the order they are declared.  This method may be used to iterate
 over the constants as follows:
@@ -269,7 +269,7 @@ for (ChunkCreator.ChunkType c : 
ChunkCreator.ChunkType.values())
 
 
 valueOf
-public static ChunkCreator.ChunkType valueOf(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String name)
+public static ChunkCreator.ChunkType valueOf(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String name)
 Returns the enum constant of this type with the specified 
name.
 The string must match exactly an identifier used to declare an
 enum constant in this type.  (Extraneous whitespace characters are 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/devapidocs/org/apache/hadoop/hbase/regionserver/CompactingMemStore.InMemoryCompactionRunnable.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/regionserver/CompactingMemStore.InMemoryCompactionRunnable.html
 
b/devapidocs/org/apache/hadoop/hbase/regionserver/CompactingMemStore.InMemoryCompactionRunnable.html
index 478c724..0ae4764 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/regionserver/CompactingMemStore.InMemoryCompactionRunnable.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/regionserver/CompactingMemStore.InMemoryCompactionRunnable.html
@@ -117,7 +117,7 @@ var activeTableTab = "activeTableTab";
 
 
 
-private class CompactingMemStore.InMemoryCompactionRunnable
+private class CompactingMemStore.InMemoryCompactionRunnable
 extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true";
 title="class or interface in java.lang">Object
 implements https://docs.oracle.com/javase/8/docs/api/java/lang/Runnable.html?is-external=true";
 title="class or interface in java.lang">Runnable
 The in-memory-flusher thread performs the flush 
asynchronously.
@@ -193,7 +193,7 @@ implements https://docs.oracle.com/javase/8/docs/api/java/lang/Runnable
 
 
 InMemoryCompactionRunnable
-private InMemoryCompactionRunnable()
+private InMemoryCompactionRunnable()
 
 
 
@@ -210,7 +210,7 @@ implements https://docs.oracle.com/javase/8/docs/api/java/lang/Runnable
 
 
 run
-public void run()
+public void run()
 
 Specified by:
 https://docs.oracle.com/javase/8/docs/api/java/lang/Runnable.html?is-external=true#run--";
 title="class or interface in 

[51/51] [partial] hbase-site git commit: Published site at 6a5b4f2a5c188f8eef4f2250b8b7db7dd1e750e4.

2018-08-23 Thread git-site-role
Published site at 6a5b4f2a5c188f8eef4f2250b8b7db7dd1e750e4.


Project: http://git-wip-us.apache.org/repos/asf/hbase-site/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase-site/commit/1ff05a18
Tree: http://git-wip-us.apache.org/repos/asf/hbase-site/tree/1ff05a18
Diff: http://git-wip-us.apache.org/repos/asf/hbase-site/diff/1ff05a18

Branch: refs/heads/asf-site
Commit: 1ff05a186ce60edd330e9bb51fc9401476910a20
Parents: 88e0d1a
Author: jenkins 
Authored: Thu Aug 23 14:47:06 2018 +
Committer: jenkins 
Committed: Thu Aug 23 14:47:06 2018 +

--
 acid-semantics.html | 4 +-
 apache_hbase_reference_guide.pdf| 4 +-
 book.html   | 2 +-
 bulk-loads.html | 4 +-
 checkstyle-aggregate.html   | 35802 +
 checkstyle.rss  |34 +-
 coc.html| 4 +-
 dependencies.html   | 4 +-
 dependency-convergence.html | 4 +-
 dependency-info.html| 4 +-
 dependency-management.html  | 4 +-
 devapidocs/constant-values.html |34 +-
 devapidocs/index-all.html   |56 +-
 .../hadoop/hbase/backup/package-tree.html   | 2 +-
 .../hadoop/hbase/class-use/CellComparator.html  |76 +-
 .../hadoop/hbase/client/package-tree.html   |20 +-
 .../hadoop/hbase/coprocessor/package-tree.html  | 2 +-
 .../hadoop/hbase/filter/package-tree.html   |12 +-
 .../hadoop/hbase/io/hfile/package-tree.html | 8 +-
 .../apache/hadoop/hbase/ipc/package-tree.html   | 4 +-
 .../hadoop/hbase/mapreduce/package-tree.html| 4 +-
 .../hbase/master/MetricsMasterSource.html   |   104 +-
 .../hbase/master/MetricsMasterSourceImpl.html   |22 +-
 .../hbase/master/MetricsMasterWrapper.html  |72 +-
 .../hbase/master/MetricsMasterWrapperImpl.html  |81 +-
 .../hbase/master/balancer/package-tree.html | 2 +-
 .../hadoop/hbase/master/package-tree.html   | 4 +-
 .../hbase/master/procedure/package-tree.html| 4 +-
 .../org/apache/hadoop/hbase/package-tree.html   |16 +-
 .../hadoop/hbase/procedure2/package-tree.html   | 6 +-
 .../hadoop/hbase/quotas/package-tree.html   | 6 +-
 .../hbase/regionserver/AbstractMemStore.html|97 +-
 .../regionserver/CSLMImmutableSegment.html  |12 +-
 .../regionserver/ChunkCreator.ChunkType.html| 4 +-
 ...tingMemStore.InMemoryCompactionRunnable.html | 6 +-
 .../CompactingMemStore.IndexType.html   | 8 +-
 .../hbase/regionserver/CompactingMemStore.html  |   131 +-
 .../hbase/regionserver/CompactionPipeline.html  |36 +-
 .../hbase/regionserver/DefaultMemStore.html |57 +-
 .../HRegion.BatchOperation.Visitor.html | 4 +-
 .../regionserver/HRegion.BatchOperation.html|78 +-
 .../regionserver/HRegion.BulkLoadListener.html  | 8 +-
 .../HRegion.FlushResult.Result.html |10 +-
 .../hbase/regionserver/HRegion.FlushResult.html | 8 +-
 .../HRegion.MutationBatchOperation.html |44 +-
 .../regionserver/HRegion.RegionScannerImpl.html |90 +-
 .../HRegion.ReplayBatchOperation.html   |32 +-
 .../regionserver/HRegion.RowLockContext.html|28 +-
 .../hbase/regionserver/HRegion.RowLockImpl.html |16 +-
 .../hadoop/hbase/regionserver/HRegion.html  |   977 +-
 .../regionserver/HStore.StoreFlusherImpl.html   |34 +-
 .../hadoop/hbase/regionserver/HStore.html   |   294 +-
 .../hbase/regionserver/MutableSegment.html  |20 +-
 .../regionserver/RegionServerAccounting.html|   162 +-
 .../hbase/regionserver/SegmentFactory.html  |57 +-
 .../hbase/regionserver/class-use/CellSet.html   | 5 +-
 .../class-use/ImmutableSegment.html | 8 +-
 .../regionserver/class-use/MemStoreLAB.html |10 +-
 .../regionserver/class-use/MemStoreSize.html|69 +-
 .../regionserver/class-use/MemStoreSizing.html  |76 +-
 .../regionserver/class-use/MutableSegment.html  |13 +-
 .../class-use/RegionServicesForStores.html  |18 +-
 .../hbase/regionserver/class-use/Segment.html   | 3 +-
 ...RegionHandler.PostOpenDeployTasksThread.html |20 +-
 .../regionserver/handler/OpenRegionHandler.html |32 +-
 .../hadoop/hbase/regionserver/package-tree.html |18 +-
 .../hbase/regionserver/wal/package-tree.html| 2 +-
 .../hadoop/hbase/rest/ScannerResource.html  |20 +-
 .../hadoop/hbase/rest/model/package-tree.html   | 2 +-
 .../hbase/security/access/package-tree.html | 2 +-
 .../hadoop/hbase/security/package-tree.html | 2 +-
 .../util/class-use/CancelableProgressable.html  | 2 +-
 .../hbase/util/class-use/PairOfSameT

[30/51] [partial] hbase-site git commit: Published site at 6a5b4f2a5c188f8eef4f2250b8b7db7dd1e750e4.

2018-08-23 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BatchOperation.Visitor.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BatchOperation.Visitor.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BatchOperation.Visitor.html
index db8431b..a8cb7c4 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BatchOperation.Visitor.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BatchOperation.Visitor.html
@@ -885,7766 +885,7797 @@
 877   * @return What the next sequence 
(edit) id should be.
 878   * @throws IOException e
 879   */
-880  private long initialize(final 
CancelableProgressable reporter) throws IOException {
-881
-882//Refuse to open the region if there 
is no column family in the table
-883if 
(htableDescriptor.getColumnFamilyCount() == 0) {
-884  throw new 
DoNotRetryIOException("Table " + 
htableDescriptor.getTableName().getNameAsString()+
-885  " should have at least one 
column family.");
-886}
-887
-888MonitoredTask status = 
TaskMonitor.get().createStatus("Initializing region " + this);
-889long nextSeqId = -1;
-890try {
-891  nextSeqId = 
initializeRegionInternals(reporter, status);
-892  return nextSeqId;
-893} finally {
-894  // nextSeqid will be -1 if the 
initialization fails.
-895  // At least it will be 0 
otherwise.
-896  if (nextSeqId == -1) {
-897status.abort("Exception during 
region " + getRegionInfo().getRegionNameAsString() +
-898  " initialization.");
-899  }
-900}
-901  }
-902
-903  private long 
initializeRegionInternals(final CancelableProgressable reporter,
-904  final MonitoredTask status) throws 
IOException {
-905if (coprocessorHost != null) {
-906  status.setStatus("Running 
coprocessor pre-open hook");
-907  coprocessorHost.preOpen();
-908}
-909
-910// Write HRI to a file in case we 
need to recover hbase:meta
-911// Only the primary replica should 
write .regioninfo
-912if 
(this.getRegionInfo().getReplicaId() == RegionInfo.DEFAULT_REPLICA_ID) {
-913  status.setStatus("Writing region 
info on filesystem");
-914  fs.checkRegionInfoOnFilesystem();
-915}
-916
-917// Initialize all the HStores
-918status.setStatus("Initializing all 
the Stores");
-919long maxSeqId = 
initializeStores(reporter, status);
-920this.mvcc.advanceTo(maxSeqId);
-921if 
(ServerRegionReplicaUtil.shouldReplayRecoveredEdits(this)) {
-922  Collection stores = 
this.stores.values();
-923  try {
-924// update the stores that we are 
replaying
-925LOG.debug("replaying wal for " + 
this.getRegionInfo().getEncodedName());
-926
stores.forEach(HStore::startReplayingFromWAL);
-927// Recover any edits if 
available.
-928maxSeqId = Math.max(maxSeqId,
-929  
replayRecoveredEditsIfAny(this.fs.getRegionDir(), maxSeqIdInStores, reporter, 
status));
-930// Make sure mvcc is up to max.
-931this.mvcc.advanceTo(maxSeqId);
-932  } finally {
-933LOG.debug("stopping wal replay 
for " + this.getRegionInfo().getEncodedName());
-934// update the stores that we are 
done replaying
-935
stores.forEach(HStore::stopReplayingFromWAL);
-936  }
-937}
-938this.lastReplayedOpenRegionSeqId = 
maxSeqId;
-939
-940
this.writestate.setReadOnly(ServerRegionReplicaUtil.isReadOnly(this));
-941this.writestate.flushRequested = 
false;
-942this.writestate.compacting.set(0);
-943
-944if (this.writestate.writesEnabled) 
{
-945  LOG.debug("Cleaning up temporary 
data for " + this.getRegionInfo().getEncodedName());
-946  // Remove temporary data left over 
from old regions
-947  status.setStatus("Cleaning up 
temporary data from old regions");
-948  fs.cleanupTempDir();
-949}
-950
-951if (this.writestate.writesEnabled) 
{
-952  status.setStatus("Cleaning up 
detritus from prior splits");
-953  // Get rid of any splits or merges 
that were lost in-progress.  Clean out
-954  // these directories here on open.  
We may be opening a region that was
-955  // being split but we crashed in 
the middle of it all.
-956  LOG.debug("Cleaning up detritus for 
" + this.getRegionInfo().getEncodedName());
-957  fs.cleanupAnySplitDetritus();
-958  fs.cleanupMergesDir();
-959}
+880  @VisibleForTesting
+881  long initialize(final 
CancelableProgressable reporter) throws IOException {
+882
+883//Refuse to open the region if there 
is no column family in the table
+884if 
(htableDescriptor.getColumnFamilyCount() == 0) {
+885  throw new 
DoNotRetryIOException("Table " + 
htableDescriptor.getTableName().getNameAsString()+
+886  " should have at least one 
column 

[21/51] [partial] hbase-site git commit: Published site at 6a5b4f2a5c188f8eef4f2250b8b7db7dd1e750e4.

2018-08-23 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RegionScannerImpl.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RegionScannerImpl.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RegionScannerImpl.html
index db8431b..a8cb7c4 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RegionScannerImpl.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RegionScannerImpl.html
@@ -885,7766 +885,7797 @@
 877   * @return What the next sequence 
(edit) id should be.
 878   * @throws IOException e
 879   */
-880  private long initialize(final 
CancelableProgressable reporter) throws IOException {
-881
-882//Refuse to open the region if there 
is no column family in the table
-883if 
(htableDescriptor.getColumnFamilyCount() == 0) {
-884  throw new 
DoNotRetryIOException("Table " + 
htableDescriptor.getTableName().getNameAsString()+
-885  " should have at least one 
column family.");
-886}
-887
-888MonitoredTask status = 
TaskMonitor.get().createStatus("Initializing region " + this);
-889long nextSeqId = -1;
-890try {
-891  nextSeqId = 
initializeRegionInternals(reporter, status);
-892  return nextSeqId;
-893} finally {
-894  // nextSeqid will be -1 if the 
initialization fails.
-895  // At least it will be 0 
otherwise.
-896  if (nextSeqId == -1) {
-897status.abort("Exception during 
region " + getRegionInfo().getRegionNameAsString() +
-898  " initialization.");
-899  }
-900}
-901  }
-902
-903  private long 
initializeRegionInternals(final CancelableProgressable reporter,
-904  final MonitoredTask status) throws 
IOException {
-905if (coprocessorHost != null) {
-906  status.setStatus("Running 
coprocessor pre-open hook");
-907  coprocessorHost.preOpen();
-908}
-909
-910// Write HRI to a file in case we 
need to recover hbase:meta
-911// Only the primary replica should 
write .regioninfo
-912if 
(this.getRegionInfo().getReplicaId() == RegionInfo.DEFAULT_REPLICA_ID) {
-913  status.setStatus("Writing region 
info on filesystem");
-914  fs.checkRegionInfoOnFilesystem();
-915}
-916
-917// Initialize all the HStores
-918status.setStatus("Initializing all 
the Stores");
-919long maxSeqId = 
initializeStores(reporter, status);
-920this.mvcc.advanceTo(maxSeqId);
-921if 
(ServerRegionReplicaUtil.shouldReplayRecoveredEdits(this)) {
-922  Collection stores = 
this.stores.values();
-923  try {
-924// update the stores that we are 
replaying
-925LOG.debug("replaying wal for " + 
this.getRegionInfo().getEncodedName());
-926
stores.forEach(HStore::startReplayingFromWAL);
-927// Recover any edits if 
available.
-928maxSeqId = Math.max(maxSeqId,
-929  
replayRecoveredEditsIfAny(this.fs.getRegionDir(), maxSeqIdInStores, reporter, 
status));
-930// Make sure mvcc is up to max.
-931this.mvcc.advanceTo(maxSeqId);
-932  } finally {
-933LOG.debug("stopping wal replay 
for " + this.getRegionInfo().getEncodedName());
-934// update the stores that we are 
done replaying
-935
stores.forEach(HStore::stopReplayingFromWAL);
-936  }
-937}
-938this.lastReplayedOpenRegionSeqId = 
maxSeqId;
-939
-940
this.writestate.setReadOnly(ServerRegionReplicaUtil.isReadOnly(this));
-941this.writestate.flushRequested = 
false;
-942this.writestate.compacting.set(0);
-943
-944if (this.writestate.writesEnabled) 
{
-945  LOG.debug("Cleaning up temporary 
data for " + this.getRegionInfo().getEncodedName());
-946  // Remove temporary data left over 
from old regions
-947  status.setStatus("Cleaning up 
temporary data from old regions");
-948  fs.cleanupTempDir();
-949}
-950
-951if (this.writestate.writesEnabled) 
{
-952  status.setStatus("Cleaning up 
detritus from prior splits");
-953  // Get rid of any splits or merges 
that were lost in-progress.  Clean out
-954  // these directories here on open.  
We may be opening a region that was
-955  // being split but we crashed in 
the middle of it all.
-956  LOG.debug("Cleaning up detritus for 
" + this.getRegionInfo().getEncodedName());
-957  fs.cleanupAnySplitDetritus();
-958  fs.cleanupMergesDir();
-959}
+880  @VisibleForTesting
+881  long initialize(final 
CancelableProgressable reporter) throws IOException {
+882
+883//Refuse to open the region if there 
is no column family in the table
+884if 
(htableDescriptor.getColumnFamilyCount() == 0) {
+885  throw new 
DoNotRetryIOException("Table " + 
htableDescriptor.getTableName().getNameAsString()+
+886  " should have at least one 
column family.");
+887}
+888

[15/51] [partial] hbase-site git commit: Published site at 6a5b4f2a5c188f8eef4f2250b8b7db7dd1e750e4.

2018-08-23 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HStore.StoreFlusherImpl.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HStore.StoreFlusherImpl.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HStore.StoreFlusherImpl.html
index 3559952..bd7445a 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HStore.StoreFlusherImpl.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HStore.StoreFlusherImpl.html
@@ -359,2396 +359,2401 @@
 351switch (inMemoryCompaction) {
 352  case NONE:
 353ms = 
ReflectionUtils.newInstance(DefaultMemStore.class,
-354new Object[]{conf, 
this.comparator});
-355break;
-356  default:
-357Class clz = conf.getClass(MEMSTORE_CLASS_NAME,
-358CompactingMemStore.class, 
CompactingMemStore.class);
-359ms = 
ReflectionUtils.newInstance(clz, new Object[]{conf, this.comparator, this,
-360
this.getHRegion().getRegionServicesForStores(), inMemoryCompaction});
-361}
-362return ms;
-363  }
-364
-365  /**
-366   * Creates the cache config.
-367   * @param family The current column 
family.
-368   */
-369  protected void createCacheConf(final 
ColumnFamilyDescriptor family) {
-370this.cacheConf = new 
CacheConfig(conf, family);
-371  }
-372
-373  /**
-374   * Creates the store engine configured 
for the given Store.
-375   * @param store The store. An 
unfortunate dependency needed due to it
-376   *  being passed to 
coprocessors via the compactor.
-377   * @param conf Store configuration.
-378   * @param kvComparator KVComparator for 
storeFileManager.
-379   * @return StoreEngine to use.
-380   */
-381  protected StoreEngine 
createStoreEngine(HStore store, Configuration conf,
-382  CellComparator kvComparator) throws 
IOException {
-383return StoreEngine.create(store, 
conf, comparator);
-384  }
-385
-386  /**
-387   * @param family
-388   * @return TTL in seconds of the 
specified family
-389   */
-390  public static long 
determineTTLFromFamily(final ColumnFamilyDescriptor family) {
-391// HCD.getTimeToLive returns ttl in 
seconds.  Convert to milliseconds.
-392long ttl = family.getTimeToLive();
-393if (ttl == HConstants.FOREVER) {
-394  // Default is unlimited ttl.
-395  ttl = Long.MAX_VALUE;
-396} else if (ttl == -1) {
-397  ttl = Long.MAX_VALUE;
-398} else {
-399  // Second -> ms adjust for user 
data
-400  ttl *= 1000;
-401}
-402return ttl;
-403  }
-404
-405  @Override
-406  public String getColumnFamilyName() {
-407return 
this.family.getNameAsString();
-408  }
-409
-410  @Override
-411  public TableName getTableName() {
-412return 
this.getRegionInfo().getTable();
-413  }
-414
-415  @Override
-416  public FileSystem getFileSystem() {
-417return this.fs.getFileSystem();
-418  }
-419
-420  public HRegionFileSystem 
getRegionFileSystem() {
-421return this.fs;
-422  }
-423
-424  /* Implementation of 
StoreConfigInformation */
-425  @Override
-426  public long getStoreFileTtl() {
-427// TTL only applies if there's no 
MIN_VERSIONs setting on the column.
-428return 
(this.scanInfo.getMinVersions() == 0) ? this.scanInfo.getTtl() : 
Long.MAX_VALUE;
-429  }
-430
-431  @Override
-432  public long getMemStoreFlushSize() {
-433// TODO: Why is this in here?  The 
flushsize of the region rather than the store?  St.Ack
-434return 
this.region.memstoreFlushSize;
-435  }
-436
-437  @Override
-438  public MemStoreSize getFlushableSize() 
{
-439return 
this.memstore.getFlushableSize();
-440  }
-441
-442  @Override
-443  public MemStoreSize getSnapshotSize() 
{
-444return 
this.memstore.getSnapshotSize();
-445  }
-446
-447  @Override
-448  public long 
getCompactionCheckMultiplier() {
-449return 
this.compactionCheckMultiplier;
-450  }
-451
-452  @Override
-453  public long getBlockingFileCount() {
-454return blockingFileCount;
-455  }
-456  /* End implementation of 
StoreConfigInformation */
-457
-458  /**
-459   * Returns the configured 
bytesPerChecksum value.
-460   * @param conf The configuration
-461   * @return The bytesPerChecksum that is 
set in the configuration
-462   */
-463  public static int 
getBytesPerChecksum(Configuration conf) {
-464return 
conf.getInt(HConstants.BYTES_PER_CHECKSUM,
-465   
HFile.DEFAULT_BYTES_PER_CHECKSUM);
-466  }
-467
-468  /**
-469   * Returns the configured checksum 
algorithm.
-470   * @param conf The configuration
-471   * @return The checksum algorithm that 
is set in the configuration
-472   */
-473  public static ChecksumType 
getChecksumType(Configuration conf) {
-474String checksumName = 
conf.get(HConstants.CHECKSUM_TYPE_NAME);
-475if (checksumName == null) {
-476  

[39/51] [partial] hbase-site git commit: Published site at 6a5b4f2a5c188f8eef4f2250b8b7db7dd1e750e4.

2018-08-23 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/devapidocs/org/apache/hadoop/hbase/regionserver/MutableSegment.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/regionserver/MutableSegment.html 
b/devapidocs/org/apache/hadoop/hbase/regionserver/MutableSegment.html
index c15731a..504d784 100644
--- a/devapidocs/org/apache/hadoop/hbase/regionserver/MutableSegment.html
+++ b/devapidocs/org/apache/hadoop/hbase/regionserver/MutableSegment.html
@@ -179,9 +179,10 @@ extends 
 protected 
-MutableSegment(CellSet cellSet,
+MutableSegment(CellSet cellSet,
   CellComparator comparator,
-  MemStoreLAB memStoreLAB) 
+  MemStoreLAB memStoreLAB,
+  MemStoreSizing memstoreSizing) 
 
 
 
@@ -290,7 +291,7 @@ extends 
+
 
 
 
@@ -298,7 +299,8 @@ extends MutableSegment(CellSet cellSet,
  CellComparator comparator,
- MemStoreLAB memStoreLAB)
+ MemStoreLAB memStoreLAB,
+ MemStoreSizing memstoreSizing)
 
 
 
@@ -315,7 +317,7 @@ extends 
 
 add
-public void add(Cell cell,
+public void add(Cell cell,
 boolean mslabUsed,
 MemStoreSizing memStoreSizing,
 boolean sizeAddedPreOperation)
@@ -333,7 +335,7 @@ extends 
 
 upsert
-public void upsert(Cell cell,
+public void upsert(Cell cell,
long readpoint,
MemStoreSizing memStoreSizing,
boolean sizeAddedPreOperation)
@@ -345,7 +347,7 @@ extends 
 
 setInMemoryFlushed
-public boolean setInMemoryFlushed()
+public boolean setInMemoryFlushed()
 
 
 
@@ -354,7 +356,7 @@ extends 
 
 first
-Cell first()
+Cell first()
 Returns the first cell in the segment
 
 Returns:
@@ -368,7 +370,7 @@ extends 
 
 indexEntrySize
-protected long indexEntrySize()
+protected long indexEntrySize()
 
 Specified by:
 indexEntrySize in
 class Segment

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/devapidocs/org/apache/hadoop/hbase/regionserver/RegionServerAccounting.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/regionserver/RegionServerAccounting.html 
b/devapidocs/org/apache/hadoop/hbase/regionserver/RegionServerAccounting.html
index 0a18af7..0758f66 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/regionserver/RegionServerAccounting.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/regionserver/RegionServerAccounting.html
@@ -18,7 +18,7 @@
 catch(err) {
 }
 //-->
-var methods = 
{"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10};
+var methods = 
{"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10};
 var tabs = {65535:["t0","All Methods"],2:["t2","Instance 
Methods"],8:["t4","Concrete Methods"]};
 var altColor = "altColor";
 var rowColor = "rowColor";
@@ -110,7 +110,7 @@ var activeTableTab = "activeTableTab";
 
 
 @InterfaceAudience.Private
-public class RegionServerAccounting
+public class RegionServerAccounting
 extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true";
 title="class or interface in java.lang">Object
 RegionServerAccounting keeps record of some basic real time 
information about
  the Region Server. Currently, it keeps record the global memstore size and 
global memstore
@@ -169,10 +169,6 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 private https://docs.oracle.com/javase/8/docs/api/java/lang/management/MemoryType.html?is-external=true";
 title="class or interface in java.lang.management">MemoryType
 memType 
 
-
-private https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/ConcurrentMap.html?is-external=true";
 title="class or interface in 
java.util.concurrent">ConcurrentMap
-replayEditsPerRegion 
-
 
 
 
@@ -207,89 +203,69 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 void
-addRegionReplayEditsSize(byte[] regionName,
-MemStoreSize memStoreSize)
-Add memStoreSize to replayEditsPerRegion.
-
-
-
-void
-clearRegionReplayEditsSize(byte[] regionName)
-Clear a region from replayEditsPerRegion.
-
-
-
-void
 decGlobalMemStoreSize(long dataSizeDelta,
  long heapSizeDelta,
  long offHeapSizeDelta) 
 
-
+
 double
 getFlushPressure() 
 
-
+
 long
 getGlobalMemStoreDataSize() 
 
-
+
 long
 getGlobalMemStoreHeapSize() 
 
-
+
 (package private) long
 getGlobalMemStoreLimit() 
 
-
+
 (package private) long
 getGlobalMemStoreLimitLowMark() 
 
-
+
 (package private) float
 getGlobalMemStoreLimitLowMarkPercent() 
 
-
+
 long
 getGlobalMemStoreOffHeapSize() 
 
-
+
 (package private) long
 getGlobalOnHeapMemStoreLimit()

[28/51] [partial] hbase-site git commit: Published site at 6a5b4f2a5c188f8eef4f2250b8b7db7dd1e750e4.

2018-08-23 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BulkLoadListener.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BulkLoadListener.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BulkLoadListener.html
index db8431b..a8cb7c4 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BulkLoadListener.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BulkLoadListener.html
@@ -885,7766 +885,7797 @@
 877   * @return What the next sequence 
(edit) id should be.
 878   * @throws IOException e
 879   */
-880  private long initialize(final 
CancelableProgressable reporter) throws IOException {
-881
-882//Refuse to open the region if there 
is no column family in the table
-883if 
(htableDescriptor.getColumnFamilyCount() == 0) {
-884  throw new 
DoNotRetryIOException("Table " + 
htableDescriptor.getTableName().getNameAsString()+
-885  " should have at least one 
column family.");
-886}
-887
-888MonitoredTask status = 
TaskMonitor.get().createStatus("Initializing region " + this);
-889long nextSeqId = -1;
-890try {
-891  nextSeqId = 
initializeRegionInternals(reporter, status);
-892  return nextSeqId;
-893} finally {
-894  // nextSeqid will be -1 if the 
initialization fails.
-895  // At least it will be 0 
otherwise.
-896  if (nextSeqId == -1) {
-897status.abort("Exception during 
region " + getRegionInfo().getRegionNameAsString() +
-898  " initialization.");
-899  }
-900}
-901  }
-902
-903  private long 
initializeRegionInternals(final CancelableProgressable reporter,
-904  final MonitoredTask status) throws 
IOException {
-905if (coprocessorHost != null) {
-906  status.setStatus("Running 
coprocessor pre-open hook");
-907  coprocessorHost.preOpen();
-908}
-909
-910// Write HRI to a file in case we 
need to recover hbase:meta
-911// Only the primary replica should 
write .regioninfo
-912if 
(this.getRegionInfo().getReplicaId() == RegionInfo.DEFAULT_REPLICA_ID) {
-913  status.setStatus("Writing region 
info on filesystem");
-914  fs.checkRegionInfoOnFilesystem();
-915}
-916
-917// Initialize all the HStores
-918status.setStatus("Initializing all 
the Stores");
-919long maxSeqId = 
initializeStores(reporter, status);
-920this.mvcc.advanceTo(maxSeqId);
-921if 
(ServerRegionReplicaUtil.shouldReplayRecoveredEdits(this)) {
-922  Collection stores = 
this.stores.values();
-923  try {
-924// update the stores that we are 
replaying
-925LOG.debug("replaying wal for " + 
this.getRegionInfo().getEncodedName());
-926
stores.forEach(HStore::startReplayingFromWAL);
-927// Recover any edits if 
available.
-928maxSeqId = Math.max(maxSeqId,
-929  
replayRecoveredEditsIfAny(this.fs.getRegionDir(), maxSeqIdInStores, reporter, 
status));
-930// Make sure mvcc is up to max.
-931this.mvcc.advanceTo(maxSeqId);
-932  } finally {
-933LOG.debug("stopping wal replay 
for " + this.getRegionInfo().getEncodedName());
-934// update the stores that we are 
done replaying
-935
stores.forEach(HStore::stopReplayingFromWAL);
-936  }
-937}
-938this.lastReplayedOpenRegionSeqId = 
maxSeqId;
-939
-940
this.writestate.setReadOnly(ServerRegionReplicaUtil.isReadOnly(this));
-941this.writestate.flushRequested = 
false;
-942this.writestate.compacting.set(0);
-943
-944if (this.writestate.writesEnabled) 
{
-945  LOG.debug("Cleaning up temporary 
data for " + this.getRegionInfo().getEncodedName());
-946  // Remove temporary data left over 
from old regions
-947  status.setStatus("Cleaning up 
temporary data from old regions");
-948  fs.cleanupTempDir();
-949}
-950
-951if (this.writestate.writesEnabled) 
{
-952  status.setStatus("Cleaning up 
detritus from prior splits");
-953  // Get rid of any splits or merges 
that were lost in-progress.  Clean out
-954  // these directories here on open.  
We may be opening a region that was
-955  // being split but we crashed in 
the middle of it all.
-956  LOG.debug("Cleaning up detritus for 
" + this.getRegionInfo().getEncodedName());
-957  fs.cleanupAnySplitDetritus();
-958  fs.cleanupMergesDir();
-959}
+880  @VisibleForTesting
+881  long initialize(final 
CancelableProgressable reporter) throws IOException {
+882
+883//Refuse to open the region if there 
is no column family in the table
+884if 
(htableDescriptor.getColumnFamilyCount() == 0) {
+885  throw new 
DoNotRetryIOException("Table " + 
htableDescriptor.getTableName().getNameAsString()+
+886  " should have at least one 
column family.");
+887}
+888
+889

[20/51] [partial] hbase-site git commit: Published site at 6a5b4f2a5c188f8eef4f2250b8b7db7dd1e750e4.

2018-08-23 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.ReplayBatchOperation.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.ReplayBatchOperation.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.ReplayBatchOperation.html
index db8431b..a8cb7c4 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.ReplayBatchOperation.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.ReplayBatchOperation.html
@@ -885,7766 +885,7797 @@
 877   * @return What the next sequence 
(edit) id should be.
 878   * @throws IOException e
 879   */
-880  private long initialize(final 
CancelableProgressable reporter) throws IOException {
-881
-882//Refuse to open the region if there 
is no column family in the table
-883if 
(htableDescriptor.getColumnFamilyCount() == 0) {
-884  throw new 
DoNotRetryIOException("Table " + 
htableDescriptor.getTableName().getNameAsString()+
-885  " should have at least one 
column family.");
-886}
-887
-888MonitoredTask status = 
TaskMonitor.get().createStatus("Initializing region " + this);
-889long nextSeqId = -1;
-890try {
-891  nextSeqId = 
initializeRegionInternals(reporter, status);
-892  return nextSeqId;
-893} finally {
-894  // nextSeqid will be -1 if the 
initialization fails.
-895  // At least it will be 0 
otherwise.
-896  if (nextSeqId == -1) {
-897status.abort("Exception during 
region " + getRegionInfo().getRegionNameAsString() +
-898  " initialization.");
-899  }
-900}
-901  }
-902
-903  private long 
initializeRegionInternals(final CancelableProgressable reporter,
-904  final MonitoredTask status) throws 
IOException {
-905if (coprocessorHost != null) {
-906  status.setStatus("Running 
coprocessor pre-open hook");
-907  coprocessorHost.preOpen();
-908}
-909
-910// Write HRI to a file in case we 
need to recover hbase:meta
-911// Only the primary replica should 
write .regioninfo
-912if 
(this.getRegionInfo().getReplicaId() == RegionInfo.DEFAULT_REPLICA_ID) {
-913  status.setStatus("Writing region 
info on filesystem");
-914  fs.checkRegionInfoOnFilesystem();
-915}
-916
-917// Initialize all the HStores
-918status.setStatus("Initializing all 
the Stores");
-919long maxSeqId = 
initializeStores(reporter, status);
-920this.mvcc.advanceTo(maxSeqId);
-921if 
(ServerRegionReplicaUtil.shouldReplayRecoveredEdits(this)) {
-922  Collection stores = 
this.stores.values();
-923  try {
-924// update the stores that we are 
replaying
-925LOG.debug("replaying wal for " + 
this.getRegionInfo().getEncodedName());
-926
stores.forEach(HStore::startReplayingFromWAL);
-927// Recover any edits if 
available.
-928maxSeqId = Math.max(maxSeqId,
-929  
replayRecoveredEditsIfAny(this.fs.getRegionDir(), maxSeqIdInStores, reporter, 
status));
-930// Make sure mvcc is up to max.
-931this.mvcc.advanceTo(maxSeqId);
-932  } finally {
-933LOG.debug("stopping wal replay 
for " + this.getRegionInfo().getEncodedName());
-934// update the stores that we are 
done replaying
-935
stores.forEach(HStore::stopReplayingFromWAL);
-936  }
-937}
-938this.lastReplayedOpenRegionSeqId = 
maxSeqId;
-939
-940
this.writestate.setReadOnly(ServerRegionReplicaUtil.isReadOnly(this));
-941this.writestate.flushRequested = 
false;
-942this.writestate.compacting.set(0);
-943
-944if (this.writestate.writesEnabled) 
{
-945  LOG.debug("Cleaning up temporary 
data for " + this.getRegionInfo().getEncodedName());
-946  // Remove temporary data left over 
from old regions
-947  status.setStatus("Cleaning up 
temporary data from old regions");
-948  fs.cleanupTempDir();
-949}
-950
-951if (this.writestate.writesEnabled) 
{
-952  status.setStatus("Cleaning up 
detritus from prior splits");
-953  // Get rid of any splits or merges 
that were lost in-progress.  Clean out
-954  // these directories here on open.  
We may be opening a region that was
-955  // being split but we crashed in 
the middle of it all.
-956  LOG.debug("Cleaning up detritus for 
" + this.getRegionInfo().getEncodedName());
-957  fs.cleanupAnySplitDetritus();
-958  fs.cleanupMergesDir();
-959}
+880  @VisibleForTesting
+881  long initialize(final 
CancelableProgressable reporter) throws IOException {
+882
+883//Refuse to open the region if there 
is no column family in the table
+884if 
(htableDescriptor.getColumnFamilyCount() == 0) {
+885  throw new 
DoNotRetryIOException("Table " + 
htableDescriptor.getTableName().getNameAsString()+
+886  " should have at least one 
column family.");

[45/51] [partial] hbase-site git commit: Published site at 6a5b4f2a5c188f8eef4f2250b8b7db7dd1e750e4.

2018-08-23 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/devapidocs/org/apache/hadoop/hbase/regionserver/DefaultMemStore.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/regionserver/DefaultMemStore.html 
b/devapidocs/org/apache/hadoop/hbase/regionserver/DefaultMemStore.html
index 76cea24..3433f64 100644
--- a/devapidocs/org/apache/hadoop/hbase/regionserver/DefaultMemStore.html
+++ b/devapidocs/org/apache/hadoop/hbase/regionserver/DefaultMemStore.html
@@ -171,7 +171,7 @@ extends AbstractMemStore
-snapshot,
 snapshotId
+regionServices,
 snapshot,
 snapshotId
 
 
 
@@ -197,6 +197,13 @@ extends Constructor.
 
 
+
+DefaultMemStore(org.apache.hadoop.conf.Configuration conf,
+   CellComparator c,
+   RegionServicesForStores regionServices)
+Constructor.
+
+
 
 
 
@@ -374,7 +381,7 @@ extends 
 
 
-
+
 
 DefaultMemStore
 public DefaultMemStore(org.apache.hadoop.conf.Configuration conf,
@@ -386,6 +393,22 @@ extends 
+
+
+
+
+DefaultMemStore
+public DefaultMemStore(org.apache.hadoop.conf.Configuration conf,
+   CellComparator c,
+   RegionServicesForStores regionServices)
+Constructor.
+
+Parameters:
+c - Comparator
+
+
+
 
 
 
@@ -400,7 +423,7 @@ extends 
 
 snapshot
-public MemStoreSnapshot snapshot()
+public MemStoreSnapshot snapshot()
 Creates a snapshot of the current memstore.
  Snapshot must be cleared by call to AbstractMemStore.clearSnapshot(long)
 
@@ -415,7 +438,7 @@ extends 
 
 getFlushableSize
-public MemStoreSize getFlushableSize()
+public MemStoreSize getFlushableSize()
 Description copied from 
interface: MemStore
 Flush will first clear out the data in snapshot if any (It 
will take a second flush
  invocation to clear the current Cell set). If snapshot is empty, current
@@ -432,7 +455,7 @@ extends 
 
 keySize
-protected long keySize()
+protected long keySize()
 
 Specified by:
 keySize in
 class AbstractMemStore
@@ -447,7 +470,7 @@ extends 
 
 heapSize
-protected long heapSize()
+protected long heapSize()
 
 Specified by:
 heapSize in
 class AbstractMemStore
@@ -463,7 +486,7 @@ extends 
 
 getScanners
-public https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
 title="class or interface in java.util">List getScanners(long readPt)
+public https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
 title="class or interface in java.util">List getScanners(long readPt)
   throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true";
 title="class or interface in java.io">IOException
 
 Returns:
@@ -480,7 +503,7 @@ extends 
 
 getSegments
-protected https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
 title="class or interface in java.util">List getSegments()
+protected https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
 title="class or interface in java.util">List getSegments()
  throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true";
 title="class or interface in java.io">IOException
 
 Specified by:
@@ -498,7 +521,7 @@ extends 
 
 getNextRow
-Cell getNextRow(Cell cell)
+Cell getNextRow(Cell cell)
 
 Parameters:
 cell - Find the row that comes after this one.  If null, we 
return the
@@ -514,7 +537,7 @@ extends 
 
 updateLowestUnflushedSequenceIdInWAL
-public void updateLowestUnflushedSequenceIdInWAL(boolean onlyIfMoreRecent)
+public void updateLowestUnflushedSequenceIdInWAL(boolean onlyIfMoreRecent)
 Description copied from 
class: AbstractMemStore
 Updates the wal with the lowest sequence id (oldest entry) 
that is still in memory
 
@@ -532,7 +555,7 @@ extends 
 
 preUpdate
-protected boolean preUpdate(MutableSegment currentActive,
+protected boolean preUpdate(MutableSegment currentActive,
 Cell cell,
 MemStoreSizing memstoreSizing)
 Description copied from 
class: AbstractMemStore
@@ -555,7 +578,7 @@ extends 
 
 postUpdate
-protected void postUpdate(MutableSegment currentActive)
+protected void postUpdate(MutableSegment currentActive)
 Description copied from 
class: AbstractMemStore
 Issue any post update synchronization and tests
 
@@ -572,7 +595,7 @@ extends 
 
 sizeAddedPreOperation
-protected boolean sizeAddedPreOperation()
+protected boolean sizeAddedPreOperation()
 
 Specified by:
 sizeAddedPreOperation in
 class AbstractMemStore
@@ -585,7 +608,7 @@ extends 
 
 size
-public MemStoreSize size()
+public MemStoreSize size()
 
 Returns:
 Total memory occupied by this MemStore. This won't include any size 
occupied by the
@@ -601,7 +624,7 @@ extends 
 
 preFlushSeqIDEstimation
-public long preFlushSeqIDEstimation()
+public long preFlushSeqIDEstimation()
 Description copied from 
interface: MemStore
 This method i

[01/51] [partial] hbase-site git commit: Published site at 6a5b4f2a5c188f8eef4f2250b8b7db7dd1e750e4.

2018-08-23 Thread git-site-role
Repository: hbase-site
Updated Branches:
  refs/heads/asf-site 88e0d1a41 -> 1ff05a186


http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/testdevapidocs/src-html/org/apache/hadoop/hbase/master/TestMasterMetricsWrapper.html
--
diff --git 
a/testdevapidocs/src-html/org/apache/hadoop/hbase/master/TestMasterMetricsWrapper.html
 
b/testdevapidocs/src-html/org/apache/hadoop/hbase/master/TestMasterMetricsWrapper.html
index 532a53d..705e999 100644
--- 
a/testdevapidocs/src-html/org/apache/hadoop/hbase/master/TestMasterMetricsWrapper.html
+++ 
b/testdevapidocs/src-html/org/apache/hadoop/hbase/master/TestMasterMetricsWrapper.html
@@ -28,90 +28,158 @@
 020import static org.junit.Assert.*;
 021
 022import 
java.util.AbstractMap.SimpleImmutableEntry;
-023import 
org.apache.hadoop.hbase.HBaseClassTestRule;
-024import 
org.apache.hadoop.hbase.HBaseTestingUtility;
-025import 
org.apache.hadoop.hbase.quotas.SpaceQuotaSnapshot;
-026import 
org.apache.hadoop.hbase.quotas.SpaceQuotaSnapshot.SpaceQuotaStatus;
-027import 
org.apache.hadoop.hbase.quotas.SpaceViolationPolicy;
-028import 
org.apache.hadoop.hbase.testclassification.MasterTests;
-029import 
org.apache.hadoop.hbase.testclassification.MediumTests;
-030import 
org.apache.hadoop.hbase.util.Threads;
-031import org.junit.AfterClass;
-032import org.junit.BeforeClass;
-033import org.junit.ClassRule;
-034import org.junit.Test;
-035import 
org.junit.experimental.categories.Category;
-036import org.slf4j.Logger;
-037import org.slf4j.LoggerFactory;
-038
-039@Category({MasterTests.class, 
MediumTests.class})
-040public class TestMasterMetricsWrapper {
-041
-042  @ClassRule
-043  public static final HBaseClassTestRule 
CLASS_RULE =
-044  
HBaseClassTestRule.forClass(TestMasterMetricsWrapper.class);
-045
-046  private static final Logger LOG = 
LoggerFactory.getLogger(TestMasterMetricsWrapper.class);
+023import java.util.List;
+024
+025import 
org.apache.hadoop.hbase.HBaseClassTestRule;
+026import 
org.apache.hadoop.hbase.HBaseTestingUtility;
+027import 
org.apache.hadoop.hbase.HColumnDescriptor;
+028import 
org.apache.hadoop.hbase.HTableDescriptor;
+029import 
org.apache.hadoop.hbase.TableName;
+030import 
org.apache.hadoop.hbase.client.RegionInfo;
+031import 
org.apache.hadoop.hbase.master.assignment.RegionStates;
+032import 
org.apache.hadoop.hbase.quotas.SpaceQuotaSnapshot;
+033import 
org.apache.hadoop.hbase.quotas.SpaceQuotaSnapshot.SpaceQuotaStatus;
+034import 
org.apache.hadoop.hbase.quotas.SpaceViolationPolicy;
+035import 
org.apache.hadoop.hbase.testclassification.MasterTests;
+036import 
org.apache.hadoop.hbase.testclassification.MediumTests;
+037import 
org.apache.hadoop.hbase.util.Bytes;
+038import 
org.apache.hadoop.hbase.util.PairOfSameType;
+039import 
org.apache.hadoop.hbase.util.Threads;
+040import org.junit.AfterClass;
+041import org.junit.BeforeClass;
+042import org.junit.ClassRule;
+043import org.junit.Test;
+044import 
org.junit.experimental.categories.Category;
+045import org.slf4j.Logger;
+046import org.slf4j.LoggerFactory;
 047
-048  private static final 
HBaseTestingUtility TEST_UTIL = new HBaseTestingUtility();
-049  private static final int NUM_RS = 4;
+048@Category({MasterTests.class, 
MediumTests.class})
+049public class TestMasterMetricsWrapper {
 050
-051  @BeforeClass
-052  public static void setup() throws 
Exception {
-053TEST_UTIL.startMiniCluster(1, 
NUM_RS);
-054  }
-055
-056  @AfterClass
-057  public static void teardown() throws 
Exception {
-058TEST_UTIL.shutdownMiniCluster();
-059  }
-060
-061  @Test
-062  public void testInfo() {
-063HMaster master = 
TEST_UTIL.getHBaseCluster().getMaster();
-064MetricsMasterWrapperImpl info = new 
MetricsMasterWrapperImpl(master);
-065
assertEquals(master.getSplitPlanCount(), info.getSplitPlanCount(), 0);
-066
assertEquals(master.getMergePlanCount(), info.getMergePlanCount(), 0);
-067assertEquals(master.getAverageLoad(), 
info.getAverageLoad(), 0);
-068assertEquals(master.getClusterId(), 
info.getClusterId());
-069
assertEquals(master.getMasterActiveTime(), info.getActiveTime());
-070
assertEquals(master.getMasterStartTime(), info.getStartTime());
-071
assertEquals(master.getMasterCoprocessors().length, 
info.getCoprocessors().length);
-072
assertEquals(master.getServerManager().getOnlineServersList().size(), 
info.getNumRegionServers());
-073int regionServerCount =
-074  NUM_RS + 
(LoadBalancer.isTablesOnMaster(TEST_UTIL.getConfiguration())? 1: 0);
-075assertEquals(regionServerCount, 
info.getNumRegionServers());
-076
-077String zkServers = 
info.getZookeeperQuorum();
-078
assertEquals(zkServers.split(",").length, 
TEST_UTIL.getZkCluster().getZooKeeperServerNum());
-079
-080final int index = 3;
-081LOG.info("Stopping " + 
TEST_UTIL.getMiniHBaseCluster().getRegionServer(index));
-082
TEST_UTIL.getMiniHBaseCluster().stopRegionServer(index, fa

[16/51] [partial] hbase-site git commit: Published site at 6a5b4f2a5c188f8eef4f2250b8b7db7dd1e750e4.

2018-08-23 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.html 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.html
index db8431b..a8cb7c4 100644
--- a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.html
+++ b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.html
@@ -885,7766 +885,7797 @@
 877   * @return What the next sequence 
(edit) id should be.
 878   * @throws IOException e
 879   */
-880  private long initialize(final 
CancelableProgressable reporter) throws IOException {
-881
-882//Refuse to open the region if there 
is no column family in the table
-883if 
(htableDescriptor.getColumnFamilyCount() == 0) {
-884  throw new 
DoNotRetryIOException("Table " + 
htableDescriptor.getTableName().getNameAsString()+
-885  " should have at least one 
column family.");
-886}
-887
-888MonitoredTask status = 
TaskMonitor.get().createStatus("Initializing region " + this);
-889long nextSeqId = -1;
-890try {
-891  nextSeqId = 
initializeRegionInternals(reporter, status);
-892  return nextSeqId;
-893} finally {
-894  // nextSeqid will be -1 if the 
initialization fails.
-895  // At least it will be 0 
otherwise.
-896  if (nextSeqId == -1) {
-897status.abort("Exception during 
region " + getRegionInfo().getRegionNameAsString() +
-898  " initialization.");
-899  }
-900}
-901  }
-902
-903  private long 
initializeRegionInternals(final CancelableProgressable reporter,
-904  final MonitoredTask status) throws 
IOException {
-905if (coprocessorHost != null) {
-906  status.setStatus("Running 
coprocessor pre-open hook");
-907  coprocessorHost.preOpen();
-908}
-909
-910// Write HRI to a file in case we 
need to recover hbase:meta
-911// Only the primary replica should 
write .regioninfo
-912if 
(this.getRegionInfo().getReplicaId() == RegionInfo.DEFAULT_REPLICA_ID) {
-913  status.setStatus("Writing region 
info on filesystem");
-914  fs.checkRegionInfoOnFilesystem();
-915}
-916
-917// Initialize all the HStores
-918status.setStatus("Initializing all 
the Stores");
-919long maxSeqId = 
initializeStores(reporter, status);
-920this.mvcc.advanceTo(maxSeqId);
-921if 
(ServerRegionReplicaUtil.shouldReplayRecoveredEdits(this)) {
-922  Collection stores = 
this.stores.values();
-923  try {
-924// update the stores that we are 
replaying
-925LOG.debug("replaying wal for " + 
this.getRegionInfo().getEncodedName());
-926
stores.forEach(HStore::startReplayingFromWAL);
-927// Recover any edits if 
available.
-928maxSeqId = Math.max(maxSeqId,
-929  
replayRecoveredEditsIfAny(this.fs.getRegionDir(), maxSeqIdInStores, reporter, 
status));
-930// Make sure mvcc is up to max.
-931this.mvcc.advanceTo(maxSeqId);
-932  } finally {
-933LOG.debug("stopping wal replay 
for " + this.getRegionInfo().getEncodedName());
-934// update the stores that we are 
done replaying
-935
stores.forEach(HStore::stopReplayingFromWAL);
-936  }
-937}
-938this.lastReplayedOpenRegionSeqId = 
maxSeqId;
-939
-940
this.writestate.setReadOnly(ServerRegionReplicaUtil.isReadOnly(this));
-941this.writestate.flushRequested = 
false;
-942this.writestate.compacting.set(0);
-943
-944if (this.writestate.writesEnabled) 
{
-945  LOG.debug("Cleaning up temporary 
data for " + this.getRegionInfo().getEncodedName());
-946  // Remove temporary data left over 
from old regions
-947  status.setStatus("Cleaning up 
temporary data from old regions");
-948  fs.cleanupTempDir();
-949}
-950
-951if (this.writestate.writesEnabled) 
{
-952  status.setStatus("Cleaning up 
detritus from prior splits");
-953  // Get rid of any splits or merges 
that were lost in-progress.  Clean out
-954  // these directories here on open.  
We may be opening a region that was
-955  // being split but we crashed in 
the middle of it all.
-956  LOG.debug("Cleaning up detritus for 
" + this.getRegionInfo().getEncodedName());
-957  fs.cleanupAnySplitDetritus();
-958  fs.cleanupMergesDir();
-959}
+880  @VisibleForTesting
+881  long initialize(final 
CancelableProgressable reporter) throws IOException {
+882
+883//Refuse to open the region if there 
is no column family in the table
+884if 
(htableDescriptor.getColumnFamilyCount() == 0) {
+885  throw new 
DoNotRetryIOException("Table " + 
htableDescriptor.getTableName().getNameAsString()+
+886  " should have at least one 
column family.");
+887}
+888
+889MonitoredTask status = 
TaskMonitor.get().createStatus("Initializing region " + this

[40/51] [partial] hbase-site git commit: Published site at 6a5b4f2a5c188f8eef4f2250b8b7db7dd1e750e4.

2018-08-23 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/devapidocs/org/apache/hadoop/hbase/regionserver/HStore.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/HStore.html 
b/devapidocs/org/apache/hadoop/hbase/regionserver/HStore.html
index b88f63d..21bc534 100644
--- a/devapidocs/org/apache/hadoop/hbase/regionserver/HStore.html
+++ b/devapidocs/org/apache/hadoop/hbase/regionserver/HStore.html
@@ -1749,7 +1749,7 @@ implements 
 
 FIXED_OVERHEAD
-public static final long FIXED_OVERHEAD
+public static final long FIXED_OVERHEAD
 
 
 
@@ -1758,7 +1758,7 @@ implements 
 
 DEEP_OVERHEAD
-public static final long DEEP_OVERHEAD
+public static final long DEEP_OVERHEAD
 
 
 
@@ -1818,7 +1818,7 @@ implements 
 
 createCacheConf
-protected void createCacheConf(ColumnFamilyDescriptor family)
+protected void createCacheConf(ColumnFamilyDescriptor family)
 Creates the cache config.
 
 Parameters:
@@ -1832,7 +1832,7 @@ implements 
 
 createStoreEngine
-protected StoreEngine createStoreEngine(HStore store,
+protected StoreEngine createStoreEngine(HStore store,
  
org.apache.hadoop.conf.Configuration conf,
  CellComparator kvComparator)
   throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true";
 title="class or interface in java.io">IOException
@@ -1856,7 +1856,7 @@ implements 
 
 determineTTLFromFamily
-public static long determineTTLFromFamily(ColumnFamilyDescriptor family)
+public static long determineTTLFromFamily(ColumnFamilyDescriptor family)
 
 Parameters:
 family - 
@@ -1871,7 +1871,7 @@ implements 
 
 getColumnFamilyName
-public https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String getColumnFamilyName()
+public https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String getColumnFamilyName()
 
 Specified by:
 getColumnFamilyName in
 interface Store
@@ -1884,7 +1884,7 @@ implements 
 
 getTableName
-public TableName getTableName()
+public TableName getTableName()
 
 Specified by:
 getTableName in
 interface Store
@@ -1897,7 +1897,7 @@ implements 
 
 getFileSystem
-public org.apache.hadoop.fs.FileSystem getFileSystem()
+public org.apache.hadoop.fs.FileSystem getFileSystem()
 
 Specified by:
 getFileSystem in
 interface Store
@@ -1910,7 +1910,7 @@ implements 
 
 getRegionFileSystem
-public HRegionFileSystem getRegionFileSystem()
+public HRegionFileSystem getRegionFileSystem()
 
 
 
@@ -1919,7 +1919,7 @@ implements 
 
 getStoreFileTtl
-public long getStoreFileTtl()
+public long getStoreFileTtl()
 
 Specified by:
 getStoreFileTtl in
 interface StoreConfigInformation
@@ -1934,7 +1934,7 @@ implements 
 
 getMemStoreFlushSize
-public long getMemStoreFlushSize()
+public long getMemStoreFlushSize()
 
 Specified by:
 getMemStoreFlushSize in
 interface StoreConfigInformation
@@ -1949,7 +1949,7 @@ implements 
 
 getFlushableSize
-public MemStoreSize getFlushableSize()
+public MemStoreSize getFlushableSize()
 
 Specified by:
 getFlushableSize in
 interface Store
@@ -1966,7 +1966,7 @@ implements 
 
 getSnapshotSize
-public MemStoreSize getSnapshotSize()
+public MemStoreSize getSnapshotSize()
 
 Specified by:
 getSnapshotSize in
 interface Store
@@ -1981,7 +1981,7 @@ implements 
 
 getCompactionCheckMultiplier
-public long getCompactionCheckMultiplier()
+public long getCompactionCheckMultiplier()
 
 Specified by:
 getCompactionCheckMultiplier in
 interface StoreConfigInformation
@@ -1998,7 +1998,7 @@ implements 
 
 getBlockingFileCount
-public long getBlockingFileCount()
+public long getBlockingFileCount()
 Description copied from 
interface: StoreConfigInformation
 The number of files required before flushes for this store 
will be blocked.
 
@@ -2013,7 +2013,7 @@ implements 
 
 getBytesPerChecksum
-public static int getBytesPerChecksum(org.apache.hadoop.conf.Configuration conf)
+public static int getBytesPerChecksum(org.apache.hadoop.conf.Configuration conf)
 Returns the configured bytesPerChecksum value.
 
 Parameters:
@@ -2029,7 +2029,7 @@ implements 
 
 getChecksumType
-public static ChecksumType getChecksumType(org.apache.hadoop.conf.Configuration conf)
+public static ChecksumType getChecksumType(org.apache.hadoop.conf.Configuration conf)
 Returns the configured checksum algorithm.
 
 Parameters:
@@ -2045,7 +2045,7 @@ implements 
 
 getCloseCheckInterval
-public static int getCloseCheckInterval()
+public static int getCloseCheckInterval()
 
 Returns:
 how many bytes to write between status checks
@@ -2058,7 +2058,7 @@ implements 
 
 getColumnFamilyDescriptor
-public ColumnFamilyDescriptor getColumnFamilyDescriptor()
+public ColumnFamilyDescriptor getColumnFamilyDescriptor()
 
 Specified by:
 getCol

[31/51] [partial] hbase-site git commit: Published site at 6a5b4f2a5c188f8eef4f2250b8b7db7dd1e750e4.

2018-08-23 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/CompactionPipeline.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/CompactionPipeline.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/CompactionPipeline.html
index 16913e6..e4622c3 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/CompactionPipeline.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/CompactionPipeline.html
@@ -80,272 +80,279 @@
 072  }
 073
 074  public boolean pushHead(MutableSegment 
segment) {
-075ImmutableSegment immutableSegment = 
SegmentFactory.instance().
-076
createImmutableSegment(segment);
-077synchronized (pipeline){
-078  boolean res = 
addFirst(immutableSegment);
-079  readOnlyCopy = new 
LinkedList<>(pipeline);
-080  return res;
-081}
-082  }
-083
-084  public VersionedSegmentsList 
getVersionedList() {
-085synchronized (pipeline){
-086  return new 
VersionedSegmentsList(readOnlyCopy, version);
-087}
-088  }
-089
-090  public VersionedSegmentsList 
getVersionedTail() {
-091synchronized (pipeline){
-092  List 
segmentList = new ArrayList<>();
-093  if(!pipeline.isEmpty()) {
-094segmentList.add(0, 
pipeline.getLast());
-095  }
-096  return new 
VersionedSegmentsList(segmentList, version);
-097}
-098  }
-099
-100  /**
-101   * Swaps the versioned list at the tail 
of the pipeline with a new segment.
-102   * Swapping only if there were no 
changes to the suffix of the list since the version list was
-103   * created.
-104   * @param versionedList suffix of the 
pipeline to be replaced can be tail or all the pipeline
-105   * @param segment new segment to 
replace the suffix. Can be null if the suffix just needs to be
-106   *removed.
-107   * @param closeSuffix whether to close 
the suffix (to release memory), as part of swapping it out
-108   *During index merge op this 
will be false and for compaction it will be true.
-109   * @param updateRegionSize whether to 
update the region size. Update the region size,
-110   * when the 
pipeline is swapped as part of in-memory-flush and
-111   * further 
merge/compaction. Don't update the region size when the
-112   * swap is 
result of the snapshot (flush-to-disk).
-113   * @return true iff swapped tail with 
new segment
-114   */
-115  
@edu.umd.cs.findbugs.annotations.SuppressWarnings(value="VO_VOLATILE_INCREMENT",
-116  justification="Increment is done 
under a synchronize block so safe")
-117  public boolean 
swap(VersionedSegmentsList versionedList, ImmutableSegment segment,
-118  boolean closeSuffix, boolean 
updateRegionSize) {
-119if (versionedList.getVersion() != 
version) {
-120  return false;
-121}
-122List 
suffix;
-123synchronized (pipeline){
-124  if(versionedList.getVersion() != 
version) {
-125return false;
-126  }
-127  suffix = 
versionedList.getStoreSegments();
-128  LOG.debug("Swapping pipeline 
suffix; before={}, new segment={}",
-129  
versionedList.getStoreSegments().size(), segment);
-130  swapSuffix(suffix, segment, 
closeSuffix);
-131  readOnlyCopy = new 
LinkedList<>(pipeline);
-132  version++;
-133}
-134if (updateRegionSize && 
region != null) {
-135  // update the global memstore size 
counter
-136  long suffixDataSize = 
getSegmentsKeySize(suffix);
-137  long newDataSize = 0;
-138  if(segment != null) {
-139newDataSize = 
segment.getDataSize();
-140  }
-141  long dataSizeDelta = suffixDataSize 
- newDataSize;
-142  long suffixHeapSize = 
getSegmentsHeapSize(suffix);
-143  long suffixOffHeapSize = 
getSegmentsOffHeapSize(suffix);
-144  long newHeapSize = 0;
-145  long newOffHeapSize = 0;
-146  if(segment != null) {
-147newHeapSize = 
segment.getHeapSize();
-148newOffHeapSize = 
segment.getOffHeapSize();
-149  }
-150  long offHeapSizeDelta = 
suffixOffHeapSize - newOffHeapSize;
-151  long heapSizeDelta = suffixHeapSize 
- newHeapSize;
-152  
region.addMemStoreSize(-dataSizeDelta, -heapSizeDelta, -offHeapSizeDelta);
-153  LOG.debug("Suffix data size={}, new 
segment data size={}, " + "suffix heap size={},"
-154  + "new segment heap 
size={}" + "suffix off heap size={},"
-155  + "new segment off heap 
size={}", suffixDataSize, newDataSize, suffixHeapSize,
-156  newHeapSize, suffixOffHeapSize, 
newOffHeapSize);
-157}
-158return true;
-159  }
-160
-161  private static long 
getSegmentsHeapSize(List list) {
-162long res = 0;
-163for (Segment segment : list) {
-164  res += segment.getHeapSize();
-165 

[09/51] [partial] hbase-site git commit: Published site at 6a5b4f2a5c188f8eef4f2250b8b7db7dd1e750e4.

2018-08-23 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/testdevapidocs/org/apache/hadoop/hbase/TestIOFencing.BlockCompactionsInPrepRegion.html
--
diff --git 
a/testdevapidocs/org/apache/hadoop/hbase/TestIOFencing.BlockCompactionsInPrepRegion.html
 
b/testdevapidocs/org/apache/hadoop/hbase/TestIOFencing.BlockCompactionsInPrepRegion.html
index 501420c..ee94889 100644
--- 
a/testdevapidocs/org/apache/hadoop/hbase/TestIOFencing.BlockCompactionsInPrepRegion.html
+++ 
b/testdevapidocs/org/apache/hadoop/hbase/TestIOFencing.BlockCompactionsInPrepRegion.html
@@ -233,7 +233,7 @@ extends 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/testdevapidocs/org/apache/hadoop/hbase/TestIOFencing.CompactionBlockerRegion.html
--
diff --git 
a/testdevapidocs/org/apache/hadoop/hbase/TestIOFencing.CompactionBlockerRegion.html
 
b/testdevapidocs/org/apache/hadoop/hbase/TestIOFencing.CompactionBlockerRegion.html
index e2ab629..abb9e50 100644
--- 
a/testdevapidocs/org/apache/hadoop/hbase/TestIOFencing.CompactionBlockerRegion.html
+++ 
b/testdevapidocs/org/apache/hadoop/hbase/TestIOFencing.CompactionBlockerRegion.html
@@ -260,7 +260,7 @@ extends org.apache.hadoop.hbase.regionserver.HRegion
 
 
 Methods inherited from 
class org.apache.hadoop.hbase.regionserver.HRegion
-addRegionToSnapshot, append, append, areWritesEnabled, batchMutate, 
batchMutate, batchMutate, batchReplay, blockUpdates, bulkLoadHFiles, 
bulkLoadHFiles, checkAndMutate, checkAndRowMutate, checkFamilies, 
checkReadOnly, checkReadsEnabled, checkSplit, checkTimestamps, close, close, 
closeRegionOperation, closeRegionOperation, compact, compactStores, 
computeHDFSBlocksDistribution, computeHDFSBlocksDistribution, createHRegion, 
createHRegion, decrementCompactionsQueuedCount, delete, deregisterChildren, 
doRegionCompactionPrep, equals, execService, flush, flushcache, get, get, get, 
getBlockedRequestsCount, getCellComparator, getCheckAndMutateChecksFailed, 
getCheckAndMutateChecksPassed, getCompactionState, getCompactPriority, 
getCoprocessorHost, getCpRequestsCount, getDataInMemoryWithoutWAL, 
getEarliestFlushTimeForAllStores, getEffectiveDurability, getFilesystem, 
getFilteredReadRequestsCount, getHDFSBlocksDistribution, getLoadStatistics, 
getLockedRows, getMaxFlushedSeqId, getMaxStoreSeq
 Id, getMemStoreDataSize, getMemStoreFlushSize, getMemStoreHeapSize, 
getMemStoreOffHeapSize, getMetrics, getMVCC, getNextSequenceId, 
getNumMutationsWithoutWAL, getOldestHfileTs, getOldestSeqIdOfStore, 
getOpenSeqNum, getReadLockCount, getReadPoint, getReadPoint, 
getReadRequestsCount, getRegionDir, getRegionDir, getRegionFileSystem, 
getRegionInfo, getRegionServicesForStores, getReplicationScope, getRowLock, 
getRowLock, getRowLockInternal, getScanner, getScanner, getSmallestReadPoint, 
getSplitPolicy, getStore, getStoreFileList, getStoreFileOpenAndCloseThreadPool, 
getStoreOpenAndCloseThreadPool, getStores, getTableDescriptor, getWAL, 
getWriteRequestsCount, hashCode, hasReferences, heapSize, increment, increment, 
incrementCompactionsQueuedCount, incrementFlushesQueuedCount, initialize, 
instantiateHStore, instantiateRegionScanner, instantiateRegionScanner, 
internalFlushcache, internalFlushCacheAndCommit, internalPrepareFlushCache, 
isAvailable, isClosed, isClosing, isLoadingCfsOnDemandDefau
 lt, isMergeable, isReadOnly, isSplittable, mutateRow, mutateRowsWithLocks, 
onConfigurationChange, openHRegion, openHRegion, openHRegion, openHRegion, 
openHRegion, openHRegion, openHRegion, openHRegion, openHRegion, openHRegion, 
openReadOnlyFileSystemHRegion, prepareDelete, prepareDeleteTimestamps, 
processRowsWithLocks, processRowsWithLocks, processRowsWithLocks, put, 
refreshStoreFiles, refreshStoreFiles, registerChildren, registerService, 
replayRecoveredEditsIfAny, reportCompactionRequestEnd, 
reportCompactionRequestFailure, reportCompactionRequestStart, 
requestCompaction, requestCompaction, requestFlush, restoreEdit, rowIsInRange, 
rowIsInRange, setClosing, setCoprocessorHost, setReadsEnabled, 
setTimeoutForWriteLock, startRegionOperation, startRegionOperation, toString, 
unblockUpdates, waitForFlushes, waitForFlushes, waitForFlushesAndCompactions, 
warmupHRegion, writeRegionOpenMarker
+addRegionToSnapshot, append, append, areWritesEnabled, batchMutate, 
batchMutate, batchMutate, batchReplay, blockUpdates, bulkLoadHFiles, 
bulkLoadHFiles, checkAndMutate, checkAndRowMutate, checkFamilies, 
checkReadOnly, checkReadsEnabled, checkSplit, checkTimestamps, close, close, 
closeRegionOperation, closeRegionOperation, compact, compactStores, 
computeHDFSBlocksDistribution, computeHDFSBlocksDistribution, createHRegion, 
createHRegion, decrementCompactionsQueuedCount, delete, deregisterChildren, 
doRegionCompactionPrep, dropMemStoreContents, equals, execService, flush, 
flushcache, get, get, get, getBlockedRequestsCount, getCellCompar

hbase-site git commit: INFRA-10751 Empty commit

2018-08-23 Thread git-site-role
Repository: hbase-site
Updated Branches:
  refs/heads/asf-site 1ff05a186 -> 455e3292b


INFRA-10751 Empty commit


Project: http://git-wip-us.apache.org/repos/asf/hbase-site/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase-site/commit/455e3292
Tree: http://git-wip-us.apache.org/repos/asf/hbase-site/tree/455e3292
Diff: http://git-wip-us.apache.org/repos/asf/hbase-site/diff/455e3292

Branch: refs/heads/asf-site
Commit: 455e3292b5f1383fb2376f18b6906876fa4df7ac
Parents: 1ff05a1
Author: jenkins 
Authored: Thu Aug 23 14:47:33 2018 +
Committer: jenkins 
Committed: Thu Aug 23 14:47:33 2018 +

--

--




[17/51] [partial] hbase-site git commit: Published site at 6a5b4f2a5c188f8eef4f2250b8b7db7dd1e750e4.

2018-08-23 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.WriteState.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.WriteState.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.WriteState.html
index db8431b..a8cb7c4 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.WriteState.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.WriteState.html
@@ -885,7766 +885,7797 @@
 877   * @return What the next sequence 
(edit) id should be.
 878   * @throws IOException e
 879   */
-880  private long initialize(final 
CancelableProgressable reporter) throws IOException {
-881
-882//Refuse to open the region if there 
is no column family in the table
-883if 
(htableDescriptor.getColumnFamilyCount() == 0) {
-884  throw new 
DoNotRetryIOException("Table " + 
htableDescriptor.getTableName().getNameAsString()+
-885  " should have at least one 
column family.");
-886}
-887
-888MonitoredTask status = 
TaskMonitor.get().createStatus("Initializing region " + this);
-889long nextSeqId = -1;
-890try {
-891  nextSeqId = 
initializeRegionInternals(reporter, status);
-892  return nextSeqId;
-893} finally {
-894  // nextSeqid will be -1 if the 
initialization fails.
-895  // At least it will be 0 
otherwise.
-896  if (nextSeqId == -1) {
-897status.abort("Exception during 
region " + getRegionInfo().getRegionNameAsString() +
-898  " initialization.");
-899  }
-900}
-901  }
-902
-903  private long 
initializeRegionInternals(final CancelableProgressable reporter,
-904  final MonitoredTask status) throws 
IOException {
-905if (coprocessorHost != null) {
-906  status.setStatus("Running 
coprocessor pre-open hook");
-907  coprocessorHost.preOpen();
-908}
-909
-910// Write HRI to a file in case we 
need to recover hbase:meta
-911// Only the primary replica should 
write .regioninfo
-912if 
(this.getRegionInfo().getReplicaId() == RegionInfo.DEFAULT_REPLICA_ID) {
-913  status.setStatus("Writing region 
info on filesystem");
-914  fs.checkRegionInfoOnFilesystem();
-915}
-916
-917// Initialize all the HStores
-918status.setStatus("Initializing all 
the Stores");
-919long maxSeqId = 
initializeStores(reporter, status);
-920this.mvcc.advanceTo(maxSeqId);
-921if 
(ServerRegionReplicaUtil.shouldReplayRecoveredEdits(this)) {
-922  Collection stores = 
this.stores.values();
-923  try {
-924// update the stores that we are 
replaying
-925LOG.debug("replaying wal for " + 
this.getRegionInfo().getEncodedName());
-926
stores.forEach(HStore::startReplayingFromWAL);
-927// Recover any edits if 
available.
-928maxSeqId = Math.max(maxSeqId,
-929  
replayRecoveredEditsIfAny(this.fs.getRegionDir(), maxSeqIdInStores, reporter, 
status));
-930// Make sure mvcc is up to max.
-931this.mvcc.advanceTo(maxSeqId);
-932  } finally {
-933LOG.debug("stopping wal replay 
for " + this.getRegionInfo().getEncodedName());
-934// update the stores that we are 
done replaying
-935
stores.forEach(HStore::stopReplayingFromWAL);
-936  }
-937}
-938this.lastReplayedOpenRegionSeqId = 
maxSeqId;
-939
-940
this.writestate.setReadOnly(ServerRegionReplicaUtil.isReadOnly(this));
-941this.writestate.flushRequested = 
false;
-942this.writestate.compacting.set(0);
-943
-944if (this.writestate.writesEnabled) 
{
-945  LOG.debug("Cleaning up temporary 
data for " + this.getRegionInfo().getEncodedName());
-946  // Remove temporary data left over 
from old regions
-947  status.setStatus("Cleaning up 
temporary data from old regions");
-948  fs.cleanupTempDir();
-949}
-950
-951if (this.writestate.writesEnabled) 
{
-952  status.setStatus("Cleaning up 
detritus from prior splits");
-953  // Get rid of any splits or merges 
that were lost in-progress.  Clean out
-954  // these directories here on open.  
We may be opening a region that was
-955  // being split but we crashed in 
the middle of it all.
-956  LOG.debug("Cleaning up detritus for 
" + this.getRegionInfo().getEncodedName());
-957  fs.cleanupAnySplitDetritus();
-958  fs.cleanupMergesDir();
-959}
+880  @VisibleForTesting
+881  long initialize(final 
CancelableProgressable reporter) throws IOException {
+882
+883//Refuse to open the region if there 
is no column family in the table
+884if 
(htableDescriptor.getColumnFamilyCount() == 0) {
+885  throw new 
DoNotRetryIOException("Table " + 
htableDescriptor.getTableName().getNameAsString()+
+886  " should have at least one 
column family.");
+887}
+888
+889MonitoredTask status = 
Ta

[32/51] [partial] hbase-site git commit: Published site at 6a5b4f2a5c188f8eef4f2250b8b7db7dd1e750e4.

2018-08-23 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/CompactingMemStore.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/CompactingMemStore.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/CompactingMemStore.html
index bf6738e..a62d84e 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/CompactingMemStore.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/CompactingMemStore.html
@@ -74,569 +74,568 @@
 066
 067  private static final Logger LOG = 
LoggerFactory.getLogger(CompactingMemStore.class);
 068  private HStore store;
-069  private RegionServicesForStores 
regionServices;
-070  private CompactionPipeline pipeline;
-071  protected MemStoreCompactor 
compactor;
-072
-073  private long inmemoryFlushSize;   
// the threshold on active size for in-memory flush
-074  private final AtomicBoolean 
inMemoryCompactionInProgress = new AtomicBoolean(false);
-075
-076  // inWalReplay is true while we are 
synchronously replaying the edits from WAL
-077  private boolean inWalReplay = false;
-078
-079  @VisibleForTesting
-080  protected final AtomicBoolean 
allowCompaction = new AtomicBoolean(true);
-081  private boolean compositeSnapshot = 
true;
-082
-083  /**
-084   * Types of indexes (part of immutable 
segments) to be used after flattening,
-085   * compaction, or merge are applied.
-086   */
-087  public enum IndexType {
-088CSLM_MAP,   // ConcurrentSkipLisMap
-089ARRAY_MAP,  // CellArrayMap
-090CHUNK_MAP   // CellChunkMap
-091  }
-092
-093  private IndexType indexType = 
IndexType.ARRAY_MAP;  // default implementation
-094
-095  public static final long DEEP_OVERHEAD 
= ClassSize.align( AbstractMemStore.DEEP_OVERHEAD
-096  + 7 * ClassSize.REFERENCE // 
Store, RegionServicesForStores, CompactionPipeline,
-097  // MemStoreCompactor, 
inMemoryCompactionInProgress,
-098  // allowCompaction, indexType
-099  + Bytes.SIZEOF_LONG   // 
inmemoryFlushSize
-100  + 2 * Bytes.SIZEOF_BOOLEAN// 
compositeSnapshot and inWalReplay
-101  + 2 * ClassSize.ATOMIC_BOOLEAN// 
inMemoryCompactionInProgress and allowCompaction
-102  + CompactionPipeline.DEEP_OVERHEAD 
+ MemStoreCompactor.DEEP_OVERHEAD);
-103
-104  public CompactingMemStore(Configuration 
conf, CellComparator c,
-105  HStore store, 
RegionServicesForStores regionServices,
-106  MemoryCompactionPolicy 
compactionPolicy) throws IOException {
-107super(conf, c);
-108this.store = store;
-109this.regionServices = 
regionServices;
-110this.pipeline = new 
CompactionPipeline(getRegionServices());
-111this.compactor = 
createMemStoreCompactor(compactionPolicy);
-112if 
(conf.getBoolean(MemStoreLAB.USEMSLAB_KEY, MemStoreLAB.USEMSLAB_DEFAULT)) {
-113  // if user requested to work with 
MSLABs (whether on- or off-heap), then the
-114  // immutable segments are going to 
use CellChunkMap as their index
-115  indexType = IndexType.CHUNK_MAP;
-116} else {
-117  indexType = IndexType.ARRAY_MAP;
-118}
-119// initialization of the flush size 
should happen after initialization of the index type
-120// so do not transfer the following 
method
-121initInmemoryFlushSize(conf);
-122LOG.info("Store={}, in-memory flush 
size threshold={}, immutable segments index type={}, " +
-123"compactor={}", 
this.store.getColumnFamilyName(),
-124
StringUtils.byteDesc(this.inmemoryFlushSize), this.indexType,
-125(this.compactor == null? "NULL": 
this.compactor.toString()));
-126  }
-127
-128  @VisibleForTesting
-129  protected MemStoreCompactor 
createMemStoreCompactor(MemoryCompactionPolicy compactionPolicy)
-130  throws IllegalArgumentIOException 
{
-131return new MemStoreCompactor(this, 
compactionPolicy);
-132  }
-133
-134  private void 
initInmemoryFlushSize(Configuration conf) {
-135double factor = 0;
-136long memstoreFlushSize = 
getRegionServices().getMemStoreFlushSize();
-137int numStores = 
getRegionServices().getNumStores();
-138if (numStores <= 1) {
-139  // Family number might also be zero 
in some of our unit test case
-140  numStores = 1;
-141}
-142factor = 
conf.getDouble(IN_MEMORY_FLUSH_THRESHOLD_FACTOR_KEY, 0.0);
-143if(factor != 0.0) {
-144  // multiply by a factor (the same 
factor for all index types)
-145  inmemoryFlushSize = (long) (factor 
* memstoreFlushSize) / numStores;
-146} else {
-147  inmemoryFlushSize = 
IN_MEMORY_FLUSH_MULTIPLIER *
-148  
conf.getLong(MemStoreLAB.CHUNK_SIZE_KEY, MemStoreLAB.CHUNK_SIZE_DEFAULT);
-149  inmemoryFlushSize -= 
ChunkCreator.SIZEOF_CHUNK_HEADER;
-150}
-151  }
-152
-153  /**
-154   * @return Total memory occupied by 
this MemStore. This won't include any size occupied by the
-155   * snapsho

[42/51] [partial] hbase-site git commit: Published site at 6a5b4f2a5c188f8eef4f2250b8b7db7dd1e750e4.

2018-08-23 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.html 
b/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.html
index 1830296..07cb760 100644
--- a/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.html
+++ b/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.html
@@ -18,7 +18,7 @@
 catch(err) {
 }
 //-->
-var methods = 
{"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":10,"i33":10,"i34":10,"i35":10,"i36":10,"i37":10,"i38":10,"i39":10,"i40":10,"i41":9,"i42":9,"i43":9,"i44":9,"i45":10,"i46":10,"i47":10,"i48":10,"i49":10,"i50":10,"i51":10,"i52":10,"i53":10,"i54":10,"i55":10,"i56":10,"i57":10,"i58":10,"i59":10,"i60":10,"i61":9,"i62":10,"i63":10,"i64":10,"i65":10,"i66":10,"i67":10,"i68":10,"i69":10,"i70":10,"i71":10,"i72":10,"i73":10,"i74":10,"i75":10,"i76":10,"i77":10,"i78":10,"i79":10,"i80":10,"i81":10,"i82":10,"i83":10,"i84":10,"i85":10,"i86":10,"i87":10,"i88":10,"i89":10,"i90":10,"i91":10,"i92":10,"i93":10,"i94":9,"i95":10,"i96":10,"i97":10,"i98":10,"i99":10,"i100":10,"i101":10,"i102":10,"i103":10,"i104":10,"i105":10,"i106":10,"i107":9,"i108":10,"i109":10
 
,"i110":10,"i111":10,"i112":10,"i113":10,"i114":41,"i115":41,"i116":10,"i117":10,"i118":10,"i119":10,"i120":10,"i121":10,"i122":10,"i123":10,"i124":10,"i125":10,"i126":10,"i127":10,"i128":10,"i129":10,"i130":10,"i131":10,"i132":10,"i133":10,"i134":10,"i135":10,"i136":10,"i137":10,"i138":10,"i139":10,"i140":9,"i141":10,"i142":10,"i143":10,"i144":10,"i145":10,"i146":10,"i147":10,"i148":10,"i149":42,"i150":10,"i151":10,"i152":10,"i153":10,"i154":10,"i155":10,"i156":10,"i157":10,"i158":10,"i159":10,"i160":10,"i161":10,"i162":10,"i163":10,"i164":10,"i165":10,"i166":10,"i167":10,"i168":10,"i169":10,"i170":10,"i171":9,"i172":10,"i173":10,"i174":10,"i175":10,"i176":10,"i177":10,"i178":10,"i179":10,"i180":9,"i181":10,"i182":10,"i183":9,"i184":9,"i185":9,"i186":9,"i187":9,"i188":9,"i189":9,"i190":9,"i191":9,"i192":9,"i193":10,"i194":10,"i195":10,"i196":10,"i197":10,"i198":10,"i199":10,"i200":10,"i201":10,"i202":9,"i203":10,"i204":10,"i205":10,"i206":10,"i207":10,"i208":10,"i209":10,"i210":10,
 
"i211":10,"i212":10,"i213":10,"i214":10,"i215":10,"i216":10,"i217":10,"i218":10,"i219":10,"i220":10,"i221":10,"i222":10,"i223":10,"i224":10,"i225":10,"i226":10,"i227":10,"i228":10,"i229":10,"i230":10,"i231":10,"i232":10,"i233":9,"i234":9,"i235":10,"i236":10,"i237":10,"i238":10,"i239":10,"i240":10,"i241":10,"i242":10,"i243":10,"i244":10,"i245":10,"i246":10,"i247":9,"i248":10,"i249":10,"i250":10,"i251":10,"i252":10,"i253":10,"i254":10,"i255":9,"i256":10,"i257":10,"i258":10,"i259":10,"i260":10,"i261":9,"i262":10,"i263":10,"i264":10,"i265":10};
+var methods = 
{"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":10,"i33":10,"i34":10,"i35":10,"i36":10,"i37":10,"i38":10,"i39":10,"i40":10,"i41":9,"i42":9,"i43":9,"i44":9,"i45":10,"i46":10,"i47":10,"i48":10,"i49":10,"i50":10,"i51":10,"i52":10,"i53":10,"i54":10,"i55":10,"i56":10,"i57":10,"i58":10,"i59":10,"i60":10,"i61":9,"i62":10,"i63":10,"i64":10,"i65":10,"i66":10,"i67":10,"i68":10,"i69":10,"i70":10,"i71":10,"i72":10,"i73":10,"i74":10,"i75":10,"i76":10,"i77":10,"i78":10,"i79":10,"i80":10,"i81":10,"i82":10,"i83":10,"i84":10,"i85":10,"i86":10,"i87":10,"i88":10,"i89":10,"i90":10,"i91":10,"i92":10,"i93":10,"i94":10,"i95":9,"i96":10,"i97":10,"i98":10,"i99":10,"i100":10,"i101":10,"i102":10,"i103":10,"i104":10,"i105":10,"i106":10,"i107":10,"i108":9,"i109":10
 
,"i110":10,"i111":10,"i112":10,"i113":10,"i114":10,"i115":41,"i116":41,"i117":10,"i118":10,"i119":10,"i120":10,"i121":10,"i122":10,"i123":10,"i124":10,"i125":10,"i126":10,"i127":10,"i128":10,"i129":10,"i130":10,"i131":10,"i132":10,"i133":10,"i134":10,"i135":10,"i136":10,"i137":10,"i138":10,"i139":10,"i140":10,"i141":9,"i142":10,"i143":10,"i144":10,"i145":10,"i146":10,"i147":10,"i148":10,"i149":10,"i150":42,"i151":10,"i152":10,"i153":10,"i154":10,"i155":10,"i156":10,"i157":10,"i158":10,"i159":10,"i160":10,"i161":10,"i162":10,"i163":10,"i164":10,"i165":10,"i166":10,"i167":10,"i168":10,"i169":10,"i170":10,"i171":10,"i172":9,"i173":10,"i174":10,"i175":10,"i176":10,"i177":10,"i178":10,"i179":10,"i180":10,"i181":9,"i182":10,"i183":10,"i184":9,"i185":9,"i186":9,"i187":9,"i188":9,"i189":9,"i190":9,"i191":9,"i192":9,"i193":9,"i194":10,"i195":10,"i196":10,"i197":10,"i198":10,"i199":10,"i200":10,"i

[41/51] [partial] hbase-site git commit: Published site at 6a5b4f2a5c188f8eef4f2250b8b7db7dd1e750e4.

2018-08-23 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/devapidocs/org/apache/hadoop/hbase/regionserver/HStore.StoreFlusherImpl.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/regionserver/HStore.StoreFlusherImpl.html 
b/devapidocs/org/apache/hadoop/hbase/regionserver/HStore.StoreFlusherImpl.html
index 1f561cf..0083ac1 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/regionserver/HStore.StoreFlusherImpl.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/regionserver/HStore.StoreFlusherImpl.html
@@ -117,7 +117,7 @@ var activeTableTab = "activeTableTab";
 
 
 
-private final class HStore.StoreFlusherImpl
+private final class HStore.StoreFlusherImpl
 extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true";
 title="class or interface in java.lang">Object
 implements StoreFlushContext
 
@@ -279,7 +279,7 @@ implements 
 
 tracker
-private final FlushLifeCycleTracker tracker
+private final FlushLifeCycleTracker tracker
 
 
 
@@ -288,7 +288,7 @@ implements 
 
 cacheFlushSeqNum
-private final long cacheFlushSeqNum
+private final long cacheFlushSeqNum
 
 
 
@@ -297,7 +297,7 @@ implements 
 
 snapshot
-private MemStoreSnapshot snapshot
+private MemStoreSnapshot snapshot
 
 
 
@@ -306,7 +306,7 @@ implements 
 
 tempFiles
-private https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
 title="class or interface in 
java.util">List tempFiles
+private https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
 title="class or interface in 
java.util">List tempFiles
 
 
 
@@ -315,7 +315,7 @@ implements 
 
 committedFiles
-private https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
 title="class or interface in 
java.util">List committedFiles
+private https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
 title="class or interface in 
java.util">List committedFiles
 
 
 
@@ -324,7 +324,7 @@ implements 
 
 cacheFlushCount
-private long cacheFlushCount
+private long cacheFlushCount
 
 
 
@@ -333,7 +333,7 @@ implements 
 
 cacheFlushSize
-private long cacheFlushSize
+private long cacheFlushSize
 
 
 
@@ -342,7 +342,7 @@ implements 
 
 outputFileSize
-private long outputFileSize
+private long outputFileSize
 
 
 
@@ -359,7 +359,7 @@ implements 
 
 StoreFlusherImpl
-private StoreFlusherImpl(long cacheFlushSeqNum,
+private StoreFlusherImpl(long cacheFlushSeqNum,
  FlushLifeCycleTracker tracker)
 
 
@@ -377,7 +377,7 @@ implements 
 
 prepare
-public MemStoreSize prepare()
+public MemStoreSize prepare()
 This is not thread safe. The caller should have a lock on 
the region or the store.
  If necessary, the lock can be added with the patch provided in 
HBASE-10087
 
@@ -394,7 +394,7 @@ implements 
 
 flushCache
-public void flushCache(MonitoredTask status)
+public void flushCache(MonitoredTask status)
 throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true";
 title="class or interface in java.io">IOException
 Description copied from 
interface: StoreFlushContext
 Flush the cache (create the new store file)
@@ -415,7 +415,7 @@ implements 
 
 commit
-public boolean commit(MonitoredTask status)
+public boolean commit(MonitoredTask status)
throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true";
 title="class or interface in java.io">IOException
 Description copied from 
interface: StoreFlushContext
 Commit the flush - add the store file to the store and 
clear the
@@ -440,7 +440,7 @@ implements 
 
 getOutputFileSize
-public long getOutputFileSize()
+public long getOutputFileSize()
 
 Specified by:
 getOutputFileSize in
 interface StoreFlushContext
@@ -455,7 +455,7 @@ implements 
 
 getCommittedFiles
-public https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
 title="class or interface in 
java.util">List getCommittedFiles()
+public https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
 title="class or interface in 
java.util">List getCommittedFiles()
 Description copied from 
interface: StoreFlushContext
 Returns the newly committed files from the flush. Called 
only if commit returns true
 
@@ -472,7 +472,7 @@ implements 
 
 replayFlush
-public void replayFlush(https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
 title="class or interface in java.util">ListString> fileNames,
+public void replayFlush(https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
 title="class or interface in java.util">List

[36/51] [partial] hbase-site git commit: Published site at 6a5b4f2a5c188f8eef4f2250b8b7db7dd1e750e4.

2018-08-23 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/devapidocs/src-html/org/apache/hadoop/hbase/master/MetricsMasterWrapper.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/MetricsMasterWrapper.html 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/MetricsMasterWrapper.html
index 1ff1ec5..2894243 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/MetricsMasterWrapper.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/MetricsMasterWrapper.html
@@ -28,128 +28,136 @@
 020
 021import java.util.Map;
 022import java.util.Map.Entry;
-023import 
org.apache.yetus.audience.InterfaceAudience;
-024
-025/**
-026 * This is the interface that will expose 
information to hadoop1/hadoop2 implementations of the
-027 * MetricsMasterSource.
-028 */
-029@InterfaceAudience.Private
-030public interface MetricsMasterWrapper {
-031
-032  /**
-033   * Get ServerName
-034   */
-035  String getServerName();
-036
-037  /**
-038   * Get Average Load
-039   *
-040   * @return Average Load
-041   */
-042  double getAverageLoad();
-043
-044  /**
-045   * Get the Cluster ID
-046   *
-047   * @return Cluster ID
-048   */
-049  String getClusterId();
-050
-051  /**
-052   * Get the ZooKeeper Quorum Info
-053   *
-054   * @return ZooKeeper Quorum Info
-055   */
-056  String getZookeeperQuorum();
-057
-058  /**
-059   * Get the co-processors
-060   *
-061   * @return Co-processors
-062   */
-063  String[] getCoprocessors();
-064
-065  /**
-066   * Get hbase master start time
-067   *
-068   * @return Start time of master in 
milliseconds
-069   */
-070  long getStartTime();
-071
-072  /**
-073   * Get the hbase master active time
-074   *
-075   * @return Time in milliseconds when 
master became active
-076   */
-077  long getActiveTime();
-078
-079  /**
-080   * Whether this master is the active 
master
-081   *
-082   * @return True if this is the active 
master
-083   */
-084  boolean getIsActiveMaster();
-085
-086  /**
-087   * Get the live region servers
-088   *
-089   * @return Live region servers
-090   */
-091  String getRegionServers();
-092
-093  /**
-094   * Get the number of live region 
servers
-095   *
-096   * @return number of Live region 
servers
-097   */
-098
-099  int getNumRegionServers();
-100
-101  /**
-102   * Get the dead region servers
-103   *
-104   * @return Dead region Servers
-105   */
-106  String getDeadRegionServers();
-107
-108  /**
-109   * Get the number of dead region 
servers
-110   *
-111   * @return number of Dead region 
Servers
-112   */
-113  int getNumDeadRegionServers();
-114
-115  /**
-116   * Get the number of master WAL 
files.
-117   */
-118  long getNumWALFiles();
-119
-120  /**
-121   * Get the number of region split plans 
executed.
-122   */
-123  long getSplitPlanCount();
-124
-125  /**
-126   * Get the number of region merge plans 
executed.
-127   */
-128  long getMergePlanCount();
-129
-130  /**
-131   * Gets the space usage and limit for 
each table.
-132   */
-133  
Map> getTableSpaceUtilization();
-134
-135  /**
-136   * Gets the space usage and limit for 
each namespace.
-137   */
-138  
Map> getNamespaceSpaceUtilization();
-139
-140  /**
-141   * Get the time in Millis when the 
master finished initializing/becoming the active master
-142   */
-143  long getMasterInitializationTime();
-144}
+023import 
org.apache.hadoop.hbase.util.PairOfSameType;
+024import 
org.apache.yetus.audience.InterfaceAudience;
+025
+026/**
+027 * This is the interface that will expose 
information to hadoop1/hadoop2 implementations of the
+028 * MetricsMasterSource.
+029 */
+030@InterfaceAudience.Private
+031public interface MetricsMasterWrapper {
+032
+033  /**
+034   * Get ServerName
+035   */
+036  String getServerName();
+037
+038  /**
+039   * Get Average Load
+040   *
+041   * @return Average Load
+042   */
+043  double getAverageLoad();
+044
+045  /**
+046   * Get the Cluster ID
+047   *
+048   * @return Cluster ID
+049   */
+050  String getClusterId();
+051
+052  /**
+053   * Get the ZooKeeper Quorum Info
+054   *
+055   * @return ZooKeeper Quorum Info
+056   */
+057  String getZookeeperQuorum();
+058
+059  /**
+060   * Get the co-processors
+061   *
+062   * @return Co-processors
+063   */
+064  String[] getCoprocessors();
+065
+066  /**
+067   * Get hbase master start time
+068   *
+069   * @return Start time of master in 
milliseconds
+070   */
+071  long getStartTime();
+072
+073  /**
+074   * Get the hbase master active time
+075   *
+076   * @return Time in milliseconds when 
master became active
+077   */
+078  long getActiveTime();
+079
+080  /**
+081   * Whether this master is the active 
master
+082   *
+083   * @return True if this is the active 
master
+084   */
+085  boolean getIsActiveMaster();
+086
+087  /**
+088   * Get the live region servers
+089   *
+090   * @return Live region servers
+091   */
+092

[23/51] [partial] hbase-site git commit: Published site at 6a5b4f2a5c188f8eef4f2250b8b7db7dd1e750e4.

2018-08-23 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.ObservedExceptionsInBatch.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.ObservedExceptionsInBatch.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.ObservedExceptionsInBatch.html
index db8431b..a8cb7c4 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.ObservedExceptionsInBatch.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.ObservedExceptionsInBatch.html
@@ -885,7766 +885,7797 @@
 877   * @return What the next sequence 
(edit) id should be.
 878   * @throws IOException e
 879   */
-880  private long initialize(final 
CancelableProgressable reporter) throws IOException {
-881
-882//Refuse to open the region if there 
is no column family in the table
-883if 
(htableDescriptor.getColumnFamilyCount() == 0) {
-884  throw new 
DoNotRetryIOException("Table " + 
htableDescriptor.getTableName().getNameAsString()+
-885  " should have at least one 
column family.");
-886}
-887
-888MonitoredTask status = 
TaskMonitor.get().createStatus("Initializing region " + this);
-889long nextSeqId = -1;
-890try {
-891  nextSeqId = 
initializeRegionInternals(reporter, status);
-892  return nextSeqId;
-893} finally {
-894  // nextSeqid will be -1 if the 
initialization fails.
-895  // At least it will be 0 
otherwise.
-896  if (nextSeqId == -1) {
-897status.abort("Exception during 
region " + getRegionInfo().getRegionNameAsString() +
-898  " initialization.");
-899  }
-900}
-901  }
-902
-903  private long 
initializeRegionInternals(final CancelableProgressable reporter,
-904  final MonitoredTask status) throws 
IOException {
-905if (coprocessorHost != null) {
-906  status.setStatus("Running 
coprocessor pre-open hook");
-907  coprocessorHost.preOpen();
-908}
-909
-910// Write HRI to a file in case we 
need to recover hbase:meta
-911// Only the primary replica should 
write .regioninfo
-912if 
(this.getRegionInfo().getReplicaId() == RegionInfo.DEFAULT_REPLICA_ID) {
-913  status.setStatus("Writing region 
info on filesystem");
-914  fs.checkRegionInfoOnFilesystem();
-915}
-916
-917// Initialize all the HStores
-918status.setStatus("Initializing all 
the Stores");
-919long maxSeqId = 
initializeStores(reporter, status);
-920this.mvcc.advanceTo(maxSeqId);
-921if 
(ServerRegionReplicaUtil.shouldReplayRecoveredEdits(this)) {
-922  Collection stores = 
this.stores.values();
-923  try {
-924// update the stores that we are 
replaying
-925LOG.debug("replaying wal for " + 
this.getRegionInfo().getEncodedName());
-926
stores.forEach(HStore::startReplayingFromWAL);
-927// Recover any edits if 
available.
-928maxSeqId = Math.max(maxSeqId,
-929  
replayRecoveredEditsIfAny(this.fs.getRegionDir(), maxSeqIdInStores, reporter, 
status));
-930// Make sure mvcc is up to max.
-931this.mvcc.advanceTo(maxSeqId);
-932  } finally {
-933LOG.debug("stopping wal replay 
for " + this.getRegionInfo().getEncodedName());
-934// update the stores that we are 
done replaying
-935
stores.forEach(HStore::stopReplayingFromWAL);
-936  }
-937}
-938this.lastReplayedOpenRegionSeqId = 
maxSeqId;
-939
-940
this.writestate.setReadOnly(ServerRegionReplicaUtil.isReadOnly(this));
-941this.writestate.flushRequested = 
false;
-942this.writestate.compacting.set(0);
-943
-944if (this.writestate.writesEnabled) 
{
-945  LOG.debug("Cleaning up temporary 
data for " + this.getRegionInfo().getEncodedName());
-946  // Remove temporary data left over 
from old regions
-947  status.setStatus("Cleaning up 
temporary data from old regions");
-948  fs.cleanupTempDir();
-949}
-950
-951if (this.writestate.writesEnabled) 
{
-952  status.setStatus("Cleaning up 
detritus from prior splits");
-953  // Get rid of any splits or merges 
that were lost in-progress.  Clean out
-954  // these directories here on open.  
We may be opening a region that was
-955  // being split but we crashed in 
the middle of it all.
-956  LOG.debug("Cleaning up detritus for 
" + this.getRegionInfo().getEncodedName());
-957  fs.cleanupAnySplitDetritus();
-958  fs.cleanupMergesDir();
-959}
+880  @VisibleForTesting
+881  long initialize(final 
CancelableProgressable reporter) throws IOException {
+882
+883//Refuse to open the region if there 
is no column family in the table
+884if 
(htableDescriptor.getColumnFamilyCount() == 0) {
+885  throw new 
DoNotRetryIOException("Table " + 
htableDescriptor.getTableName().getNameAsString()+
+886  " should have at lea

[37/51] [partial] hbase-site git commit: Published site at 6a5b4f2a5c188f8eef4f2250b8b7db7dd1e750e4.

2018-08-23 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html 
b/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html
index 938a4c9..28aa73e 100644
--- a/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html
+++ b/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html
@@ -708,20 +708,20 @@
 
 java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true";
 title="class or interface in java.lang">Enum (implements java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true";
 title="class or interface in java.lang">Comparable, java.io.https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true";
 title="class or interface in java.io">Serializable)
 
-org.apache.hadoop.hbase.regionserver.ChunkCreator.ChunkType
-org.apache.hadoop.hbase.regionserver.MemStoreCompactionStrategy.Action
+org.apache.hadoop.hbase.regionserver.MetricsRegionServerSourceFactoryImpl.FactoryStorage
+org.apache.hadoop.hbase.regionserver.BloomType
+org.apache.hadoop.hbase.regionserver.FlushType
 org.apache.hadoop.hbase.regionserver.TimeRangeTracker.Type
-org.apache.hadoop.hbase.regionserver.ScanType
-org.apache.hadoop.hbase.regionserver.CompactingMemStore.IndexType
+org.apache.hadoop.hbase.regionserver.MemStoreCompactionStrategy.Action
+org.apache.hadoop.hbase.regionserver.DefaultHeapMemoryTuner.StepDirection
+org.apache.hadoop.hbase.regionserver.ScannerContext.LimitScope
+org.apache.hadoop.hbase.regionserver.ChunkCreator.ChunkType
 org.apache.hadoop.hbase.regionserver.Region.Operation
 org.apache.hadoop.hbase.regionserver.ScannerContext.NextState
-org.apache.hadoop.hbase.regionserver.BloomType
-org.apache.hadoop.hbase.regionserver.MetricsRegionServerSourceFactoryImpl.FactoryStorage
-org.apache.hadoop.hbase.regionserver.ScannerContext.LimitScope
 org.apache.hadoop.hbase.regionserver.HRegion.FlushResult.Result
-org.apache.hadoop.hbase.regionserver.FlushType
+org.apache.hadoop.hbase.regionserver.CompactingMemStore.IndexType
 org.apache.hadoop.hbase.regionserver.SplitLogWorker.TaskExecutor.Status
-org.apache.hadoop.hbase.regionserver.DefaultHeapMemoryTuner.StepDirection
+org.apache.hadoop.hbase.regionserver.ScanType
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/devapidocs/org/apache/hadoop/hbase/regionserver/wal/package-tree.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/regionserver/wal/package-tree.html 
b/devapidocs/org/apache/hadoop/hbase/regionserver/wal/package-tree.html
index 19354d1..46651a5 100644
--- a/devapidocs/org/apache/hadoop/hbase/regionserver/wal/package-tree.html
+++ b/devapidocs/org/apache/hadoop/hbase/regionserver/wal/package-tree.html
@@ -247,8 +247,8 @@
 
 java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true";
 title="class or interface in java.lang">Enum (implements java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true";
 title="class or interface in java.lang">Comparable, java.io.https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true";
 title="class or interface in java.io">Serializable)
 
-org.apache.hadoop.hbase.regionserver.wal.CompressionContext.DictionaryIndex
 org.apache.hadoop.hbase.regionserver.wal.ProtobufLogReader.WALHdrResult
+org.apache.hadoop.hbase.regionserver.wal.CompressionContext.DictionaryIndex
 org.apache.hadoop.hbase.regionserver.wal.RingBufferTruck.Type
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/devapidocs/org/apache/hadoop/hbase/rest/ScannerResource.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/rest/ScannerResource.html 
b/devapidocs/org/apache/hadoop/hbase/rest/ScannerResource.html
index 5353aa1..c8fd09e 100644
--- a/devapidocs/org/apache/hadoop/hbase/rest/ScannerResource.html
+++ b/devapidocs/org/apache/hadoop/hbase/rest/ScannerResource.html
@@ -119,7 +119,7 @@ var activeTableTab = "activeTableTab";
 
 
 @InterfaceAudience.Private
-public class ScannerResource
+public class ScannerResource
 extends ResourceBase
 
 
@@ -258,7 +258,7 @@ extends 
 
 LOG
-private static final org.slf4j.Logger LOG
+private static final org.slf4j.Logger LOG
 
 
 
@@ -267,7 +267,7 @@ extends 
 
 scanners
-static final https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true";
 title="class or interface in java.util">MapString,ScannerInstanceResource> 
scanners
+static final https://docs.oracle.com/javase/8/docs/api/java/ut

[29/51] [partial] hbase-site git commit: Published site at 6a5b4f2a5c188f8eef4f2250b8b7db7dd1e750e4.

2018-08-23 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BatchOperation.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BatchOperation.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BatchOperation.html
index db8431b..a8cb7c4 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BatchOperation.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BatchOperation.html
@@ -885,7766 +885,7797 @@
 877   * @return What the next sequence 
(edit) id should be.
 878   * @throws IOException e
 879   */
-880  private long initialize(final 
CancelableProgressable reporter) throws IOException {
-881
-882//Refuse to open the region if there 
is no column family in the table
-883if 
(htableDescriptor.getColumnFamilyCount() == 0) {
-884  throw new 
DoNotRetryIOException("Table " + 
htableDescriptor.getTableName().getNameAsString()+
-885  " should have at least one 
column family.");
-886}
-887
-888MonitoredTask status = 
TaskMonitor.get().createStatus("Initializing region " + this);
-889long nextSeqId = -1;
-890try {
-891  nextSeqId = 
initializeRegionInternals(reporter, status);
-892  return nextSeqId;
-893} finally {
-894  // nextSeqid will be -1 if the 
initialization fails.
-895  // At least it will be 0 
otherwise.
-896  if (nextSeqId == -1) {
-897status.abort("Exception during 
region " + getRegionInfo().getRegionNameAsString() +
-898  " initialization.");
-899  }
-900}
-901  }
-902
-903  private long 
initializeRegionInternals(final CancelableProgressable reporter,
-904  final MonitoredTask status) throws 
IOException {
-905if (coprocessorHost != null) {
-906  status.setStatus("Running 
coprocessor pre-open hook");
-907  coprocessorHost.preOpen();
-908}
-909
-910// Write HRI to a file in case we 
need to recover hbase:meta
-911// Only the primary replica should 
write .regioninfo
-912if 
(this.getRegionInfo().getReplicaId() == RegionInfo.DEFAULT_REPLICA_ID) {
-913  status.setStatus("Writing region 
info on filesystem");
-914  fs.checkRegionInfoOnFilesystem();
-915}
-916
-917// Initialize all the HStores
-918status.setStatus("Initializing all 
the Stores");
-919long maxSeqId = 
initializeStores(reporter, status);
-920this.mvcc.advanceTo(maxSeqId);
-921if 
(ServerRegionReplicaUtil.shouldReplayRecoveredEdits(this)) {
-922  Collection stores = 
this.stores.values();
-923  try {
-924// update the stores that we are 
replaying
-925LOG.debug("replaying wal for " + 
this.getRegionInfo().getEncodedName());
-926
stores.forEach(HStore::startReplayingFromWAL);
-927// Recover any edits if 
available.
-928maxSeqId = Math.max(maxSeqId,
-929  
replayRecoveredEditsIfAny(this.fs.getRegionDir(), maxSeqIdInStores, reporter, 
status));
-930// Make sure mvcc is up to max.
-931this.mvcc.advanceTo(maxSeqId);
-932  } finally {
-933LOG.debug("stopping wal replay 
for " + this.getRegionInfo().getEncodedName());
-934// update the stores that we are 
done replaying
-935
stores.forEach(HStore::stopReplayingFromWAL);
-936  }
-937}
-938this.lastReplayedOpenRegionSeqId = 
maxSeqId;
-939
-940
this.writestate.setReadOnly(ServerRegionReplicaUtil.isReadOnly(this));
-941this.writestate.flushRequested = 
false;
-942this.writestate.compacting.set(0);
-943
-944if (this.writestate.writesEnabled) 
{
-945  LOG.debug("Cleaning up temporary 
data for " + this.getRegionInfo().getEncodedName());
-946  // Remove temporary data left over 
from old regions
-947  status.setStatus("Cleaning up 
temporary data from old regions");
-948  fs.cleanupTempDir();
-949}
-950
-951if (this.writestate.writesEnabled) 
{
-952  status.setStatus("Cleaning up 
detritus from prior splits");
-953  // Get rid of any splits or merges 
that were lost in-progress.  Clean out
-954  // these directories here on open.  
We may be opening a region that was
-955  // being split but we crashed in 
the middle of it all.
-956  LOG.debug("Cleaning up detritus for 
" + this.getRegionInfo().getEncodedName());
-957  fs.cleanupAnySplitDetritus();
-958  fs.cleanupMergesDir();
-959}
+880  @VisibleForTesting
+881  long initialize(final 
CancelableProgressable reporter) throws IOException {
+882
+883//Refuse to open the region if there 
is no column family in the table
+884if 
(htableDescriptor.getColumnFamilyCount() == 0) {
+885  throw new 
DoNotRetryIOException("Table " + 
htableDescriptor.getTableName().getNameAsString()+
+886  " should have at least one 
column family.");
+887}
+888
+889Monito

[33/51] [partial] hbase-site git commit: Published site at 6a5b4f2a5c188f8eef4f2250b8b7db7dd1e750e4.

2018-08-23 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/CompactingMemStore.IndexType.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/CompactingMemStore.IndexType.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/CompactingMemStore.IndexType.html
index bf6738e..a62d84e 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/CompactingMemStore.IndexType.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/CompactingMemStore.IndexType.html
@@ -74,569 +74,568 @@
 066
 067  private static final Logger LOG = 
LoggerFactory.getLogger(CompactingMemStore.class);
 068  private HStore store;
-069  private RegionServicesForStores 
regionServices;
-070  private CompactionPipeline pipeline;
-071  protected MemStoreCompactor 
compactor;
-072
-073  private long inmemoryFlushSize;   
// the threshold on active size for in-memory flush
-074  private final AtomicBoolean 
inMemoryCompactionInProgress = new AtomicBoolean(false);
-075
-076  // inWalReplay is true while we are 
synchronously replaying the edits from WAL
-077  private boolean inWalReplay = false;
-078
-079  @VisibleForTesting
-080  protected final AtomicBoolean 
allowCompaction = new AtomicBoolean(true);
-081  private boolean compositeSnapshot = 
true;
-082
-083  /**
-084   * Types of indexes (part of immutable 
segments) to be used after flattening,
-085   * compaction, or merge are applied.
-086   */
-087  public enum IndexType {
-088CSLM_MAP,   // ConcurrentSkipLisMap
-089ARRAY_MAP,  // CellArrayMap
-090CHUNK_MAP   // CellChunkMap
-091  }
-092
-093  private IndexType indexType = 
IndexType.ARRAY_MAP;  // default implementation
-094
-095  public static final long DEEP_OVERHEAD 
= ClassSize.align( AbstractMemStore.DEEP_OVERHEAD
-096  + 7 * ClassSize.REFERENCE // 
Store, RegionServicesForStores, CompactionPipeline,
-097  // MemStoreCompactor, 
inMemoryCompactionInProgress,
-098  // allowCompaction, indexType
-099  + Bytes.SIZEOF_LONG   // 
inmemoryFlushSize
-100  + 2 * Bytes.SIZEOF_BOOLEAN// 
compositeSnapshot and inWalReplay
-101  + 2 * ClassSize.ATOMIC_BOOLEAN// 
inMemoryCompactionInProgress and allowCompaction
-102  + CompactionPipeline.DEEP_OVERHEAD 
+ MemStoreCompactor.DEEP_OVERHEAD);
-103
-104  public CompactingMemStore(Configuration 
conf, CellComparator c,
-105  HStore store, 
RegionServicesForStores regionServices,
-106  MemoryCompactionPolicy 
compactionPolicy) throws IOException {
-107super(conf, c);
-108this.store = store;
-109this.regionServices = 
regionServices;
-110this.pipeline = new 
CompactionPipeline(getRegionServices());
-111this.compactor = 
createMemStoreCompactor(compactionPolicy);
-112if 
(conf.getBoolean(MemStoreLAB.USEMSLAB_KEY, MemStoreLAB.USEMSLAB_DEFAULT)) {
-113  // if user requested to work with 
MSLABs (whether on- or off-heap), then the
-114  // immutable segments are going to 
use CellChunkMap as their index
-115  indexType = IndexType.CHUNK_MAP;
-116} else {
-117  indexType = IndexType.ARRAY_MAP;
-118}
-119// initialization of the flush size 
should happen after initialization of the index type
-120// so do not transfer the following 
method
-121initInmemoryFlushSize(conf);
-122LOG.info("Store={}, in-memory flush 
size threshold={}, immutable segments index type={}, " +
-123"compactor={}", 
this.store.getColumnFamilyName(),
-124
StringUtils.byteDesc(this.inmemoryFlushSize), this.indexType,
-125(this.compactor == null? "NULL": 
this.compactor.toString()));
-126  }
-127
-128  @VisibleForTesting
-129  protected MemStoreCompactor 
createMemStoreCompactor(MemoryCompactionPolicy compactionPolicy)
-130  throws IllegalArgumentIOException 
{
-131return new MemStoreCompactor(this, 
compactionPolicy);
-132  }
-133
-134  private void 
initInmemoryFlushSize(Configuration conf) {
-135double factor = 0;
-136long memstoreFlushSize = 
getRegionServices().getMemStoreFlushSize();
-137int numStores = 
getRegionServices().getNumStores();
-138if (numStores <= 1) {
-139  // Family number might also be zero 
in some of our unit test case
-140  numStores = 1;
-141}
-142factor = 
conf.getDouble(IN_MEMORY_FLUSH_THRESHOLD_FACTOR_KEY, 0.0);
-143if(factor != 0.0) {
-144  // multiply by a factor (the same 
factor for all index types)
-145  inmemoryFlushSize = (long) (factor 
* memstoreFlushSize) / numStores;
-146} else {
-147  inmemoryFlushSize = 
IN_MEMORY_FLUSH_MULTIPLIER *
-148  
conf.getLong(MemStoreLAB.CHUNK_SIZE_KEY, MemStoreLAB.CHUNK_SIZE_DEFAULT);
-149  inmemoryFlushSize -= 
ChunkCreator.SIZEOF_CHUNK_HEADER;
-150}
-151  }
-152
-153  /**
-154   * @return Total memory occupied by 
this MemStore. This won't include

[25/51] [partial] hbase-site git commit: Published site at 6a5b4f2a5c188f8eef4f2250b8b7db7dd1e750e4.

2018-08-23 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResultImpl.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResultImpl.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResultImpl.html
index db8431b..a8cb7c4 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResultImpl.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResultImpl.html
@@ -885,7766 +885,7797 @@
 877   * @return What the next sequence 
(edit) id should be.
 878   * @throws IOException e
 879   */
-880  private long initialize(final 
CancelableProgressable reporter) throws IOException {
-881
-882//Refuse to open the region if there 
is no column family in the table
-883if 
(htableDescriptor.getColumnFamilyCount() == 0) {
-884  throw new 
DoNotRetryIOException("Table " + 
htableDescriptor.getTableName().getNameAsString()+
-885  " should have at least one 
column family.");
-886}
-887
-888MonitoredTask status = 
TaskMonitor.get().createStatus("Initializing region " + this);
-889long nextSeqId = -1;
-890try {
-891  nextSeqId = 
initializeRegionInternals(reporter, status);
-892  return nextSeqId;
-893} finally {
-894  // nextSeqid will be -1 if the 
initialization fails.
-895  // At least it will be 0 
otherwise.
-896  if (nextSeqId == -1) {
-897status.abort("Exception during 
region " + getRegionInfo().getRegionNameAsString() +
-898  " initialization.");
-899  }
-900}
-901  }
-902
-903  private long 
initializeRegionInternals(final CancelableProgressable reporter,
-904  final MonitoredTask status) throws 
IOException {
-905if (coprocessorHost != null) {
-906  status.setStatus("Running 
coprocessor pre-open hook");
-907  coprocessorHost.preOpen();
-908}
-909
-910// Write HRI to a file in case we 
need to recover hbase:meta
-911// Only the primary replica should 
write .regioninfo
-912if 
(this.getRegionInfo().getReplicaId() == RegionInfo.DEFAULT_REPLICA_ID) {
-913  status.setStatus("Writing region 
info on filesystem");
-914  fs.checkRegionInfoOnFilesystem();
-915}
-916
-917// Initialize all the HStores
-918status.setStatus("Initializing all 
the Stores");
-919long maxSeqId = 
initializeStores(reporter, status);
-920this.mvcc.advanceTo(maxSeqId);
-921if 
(ServerRegionReplicaUtil.shouldReplayRecoveredEdits(this)) {
-922  Collection stores = 
this.stores.values();
-923  try {
-924// update the stores that we are 
replaying
-925LOG.debug("replaying wal for " + 
this.getRegionInfo().getEncodedName());
-926
stores.forEach(HStore::startReplayingFromWAL);
-927// Recover any edits if 
available.
-928maxSeqId = Math.max(maxSeqId,
-929  
replayRecoveredEditsIfAny(this.fs.getRegionDir(), maxSeqIdInStores, reporter, 
status));
-930// Make sure mvcc is up to max.
-931this.mvcc.advanceTo(maxSeqId);
-932  } finally {
-933LOG.debug("stopping wal replay 
for " + this.getRegionInfo().getEncodedName());
-934// update the stores that we are 
done replaying
-935
stores.forEach(HStore::stopReplayingFromWAL);
-936  }
-937}
-938this.lastReplayedOpenRegionSeqId = 
maxSeqId;
-939
-940
this.writestate.setReadOnly(ServerRegionReplicaUtil.isReadOnly(this));
-941this.writestate.flushRequested = 
false;
-942this.writestate.compacting.set(0);
-943
-944if (this.writestate.writesEnabled) 
{
-945  LOG.debug("Cleaning up temporary 
data for " + this.getRegionInfo().getEncodedName());
-946  // Remove temporary data left over 
from old regions
-947  status.setStatus("Cleaning up 
temporary data from old regions");
-948  fs.cleanupTempDir();
-949}
-950
-951if (this.writestate.writesEnabled) 
{
-952  status.setStatus("Cleaning up 
detritus from prior splits");
-953  // Get rid of any splits or merges 
that were lost in-progress.  Clean out
-954  // these directories here on open.  
We may be opening a region that was
-955  // being split but we crashed in 
the middle of it all.
-956  LOG.debug("Cleaning up detritus for 
" + this.getRegionInfo().getEncodedName());
-957  fs.cleanupAnySplitDetritus();
-958  fs.cleanupMergesDir();
-959}
+880  @VisibleForTesting
+881  long initialize(final 
CancelableProgressable reporter) throws IOException {
+882
+883//Refuse to open the region if there 
is no column family in the table
+884if 
(htableDescriptor.getColumnFamilyCount() == 0) {
+885  throw new 
DoNotRetryIOException("Table " + 
htableDescriptor.getTableName().getNameAsString()+
+886  " should have at least one 
column family.");
+887}
+888
+889M

[14/51] [partial] hbase-site git commit: Published site at 6a5b4f2a5c188f8eef4f2250b8b7db7dd1e750e4.

2018-08-23 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HStore.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HStore.html 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HStore.html
index 3559952..bd7445a 100644
--- a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HStore.html
+++ b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HStore.html
@@ -359,2396 +359,2401 @@
 351switch (inMemoryCompaction) {
 352  case NONE:
 353ms = 
ReflectionUtils.newInstance(DefaultMemStore.class,
-354new Object[]{conf, 
this.comparator});
-355break;
-356  default:
-357Class clz = conf.getClass(MEMSTORE_CLASS_NAME,
-358CompactingMemStore.class, 
CompactingMemStore.class);
-359ms = 
ReflectionUtils.newInstance(clz, new Object[]{conf, this.comparator, this,
-360
this.getHRegion().getRegionServicesForStores(), inMemoryCompaction});
-361}
-362return ms;
-363  }
-364
-365  /**
-366   * Creates the cache config.
-367   * @param family The current column 
family.
-368   */
-369  protected void createCacheConf(final 
ColumnFamilyDescriptor family) {
-370this.cacheConf = new 
CacheConfig(conf, family);
-371  }
-372
-373  /**
-374   * Creates the store engine configured 
for the given Store.
-375   * @param store The store. An 
unfortunate dependency needed due to it
-376   *  being passed to 
coprocessors via the compactor.
-377   * @param conf Store configuration.
-378   * @param kvComparator KVComparator for 
storeFileManager.
-379   * @return StoreEngine to use.
-380   */
-381  protected StoreEngine 
createStoreEngine(HStore store, Configuration conf,
-382  CellComparator kvComparator) throws 
IOException {
-383return StoreEngine.create(store, 
conf, comparator);
-384  }
-385
-386  /**
-387   * @param family
-388   * @return TTL in seconds of the 
specified family
-389   */
-390  public static long 
determineTTLFromFamily(final ColumnFamilyDescriptor family) {
-391// HCD.getTimeToLive returns ttl in 
seconds.  Convert to milliseconds.
-392long ttl = family.getTimeToLive();
-393if (ttl == HConstants.FOREVER) {
-394  // Default is unlimited ttl.
-395  ttl = Long.MAX_VALUE;
-396} else if (ttl == -1) {
-397  ttl = Long.MAX_VALUE;
-398} else {
-399  // Second -> ms adjust for user 
data
-400  ttl *= 1000;
-401}
-402return ttl;
-403  }
-404
-405  @Override
-406  public String getColumnFamilyName() {
-407return 
this.family.getNameAsString();
-408  }
-409
-410  @Override
-411  public TableName getTableName() {
-412return 
this.getRegionInfo().getTable();
-413  }
-414
-415  @Override
-416  public FileSystem getFileSystem() {
-417return this.fs.getFileSystem();
-418  }
-419
-420  public HRegionFileSystem 
getRegionFileSystem() {
-421return this.fs;
-422  }
-423
-424  /* Implementation of 
StoreConfigInformation */
-425  @Override
-426  public long getStoreFileTtl() {
-427// TTL only applies if there's no 
MIN_VERSIONs setting on the column.
-428return 
(this.scanInfo.getMinVersions() == 0) ? this.scanInfo.getTtl() : 
Long.MAX_VALUE;
-429  }
-430
-431  @Override
-432  public long getMemStoreFlushSize() {
-433// TODO: Why is this in here?  The 
flushsize of the region rather than the store?  St.Ack
-434return 
this.region.memstoreFlushSize;
-435  }
-436
-437  @Override
-438  public MemStoreSize getFlushableSize() 
{
-439return 
this.memstore.getFlushableSize();
-440  }
-441
-442  @Override
-443  public MemStoreSize getSnapshotSize() 
{
-444return 
this.memstore.getSnapshotSize();
-445  }
-446
-447  @Override
-448  public long 
getCompactionCheckMultiplier() {
-449return 
this.compactionCheckMultiplier;
-450  }
-451
-452  @Override
-453  public long getBlockingFileCount() {
-454return blockingFileCount;
-455  }
-456  /* End implementation of 
StoreConfigInformation */
-457
-458  /**
-459   * Returns the configured 
bytesPerChecksum value.
-460   * @param conf The configuration
-461   * @return The bytesPerChecksum that is 
set in the configuration
-462   */
-463  public static int 
getBytesPerChecksum(Configuration conf) {
-464return 
conf.getInt(HConstants.BYTES_PER_CHECKSUM,
-465   
HFile.DEFAULT_BYTES_PER_CHECKSUM);
-466  }
-467
-468  /**
-469   * Returns the configured checksum 
algorithm.
-470   * @param conf The configuration
-471   * @return The checksum algorithm that 
is set in the configuration
-472   */
-473  public static ChecksumType 
getChecksumType(Configuration conf) {
-474String checksumName = 
conf.get(HConstants.CHECKSUM_TYPE_NAME);
-475if (checksumName == null) {
-476  return 
ChecksumType.getDefaultChecksumType();
-477} else {
-478  return 
Ch

[08/51] [partial] hbase-site git commit: Published site at 6a5b4f2a5c188f8eef4f2250b8b7db7dd1e750e4.

2018-08-23 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/testdevapidocs/org/apache/hadoop/hbase/class-use/HBaseClassTestRule.html
--
diff --git 
a/testdevapidocs/org/apache/hadoop/hbase/class-use/HBaseClassTestRule.html 
b/testdevapidocs/org/apache/hadoop/hbase/class-use/HBaseClassTestRule.html
index bbd876d..1372fe5 100644
--- a/testdevapidocs/org/apache/hadoop/hbase/class-use/HBaseClassTestRule.html
+++ b/testdevapidocs/org/apache/hadoop/hbase/class-use/HBaseClassTestRule.html
@@ -3589,98 +3589,102 @@
 
 
 static HBaseClassTestRule
-TestCreateTableProcedure.CLASS_RULE 
+TestCreateTableProcedureMuitipleRegions.CLASS_RULE 
 
 
 static HBaseClassTestRule
-TestModifyTableProcedure.CLASS_RULE 
+TestCreateTableProcedure.CLASS_RULE 
 
 
 static HBaseClassTestRule
-TestDeleteTableProcedure.CLASS_RULE 
+TestModifyTableProcedure.CLASS_RULE 
 
 
 static HBaseClassTestRule
-TestCreateNamespaceProcedure.CLASS_RULE 
+TestDeleteTableProcedure.CLASS_RULE 
 
 
 static HBaseClassTestRule
-TestServerCrashProcedure.CLASS_RULE 
+TestCreateNamespaceProcedure.CLASS_RULE 
 
 
 static HBaseClassTestRule
-TestModifyNamespaceProcedure.CLASS_RULE 
+TestServerCrashProcedure.CLASS_RULE 
 
 
 static HBaseClassTestRule
-TestFastFailOnProcedureNotRegistered.CLASS_RULE 
+TestModifyNamespaceProcedure.CLASS_RULE 
 
 
 static HBaseClassTestRule
-TestDisableTableProcedure.CLASS_RULE 
+TestFastFailOnProcedureNotRegistered.CLASS_RULE 
 
 
 static HBaseClassTestRule
-TestProcedurePriority.CLASS_RULE 
+TestDisableTableProcedure.CLASS_RULE 
 
 
 static HBaseClassTestRule
-TestMasterProcedureEvents.CLASS_RULE 
+TestProcedurePriority.CLASS_RULE 
 
 
 static HBaseClassTestRule
-TestTruncateTableProcedure.CLASS_RULE 
+TestMasterProcedureEvents.CLASS_RULE 
 
 
 static HBaseClassTestRule
-TestDeleteNamespaceProcedure.CLASS_RULE 
+TestTruncateTableProcedure.CLASS_RULE 
 
 
 static HBaseClassTestRule
-TestMasterObserverPostCalls.CLASS_RULE 
+TestDeleteNamespaceProcedure.CLASS_RULE 
 
 
 static HBaseClassTestRule
-TestWALProcedureStoreOnHDFS.CLASS_RULE 
+TestMasterObserverPostCalls.CLASS_RULE 
 
 
 static HBaseClassTestRule
-TestEnableTableProcedure.CLASS_RULE 
+TestWALProcedureStoreOnHDFS.CLASS_RULE 
 
 
 static HBaseClassTestRule
-TestMasterProcedureScheduler.CLASS_RULE 
+TestEnableTableProcedure.CLASS_RULE 
 
 
 static HBaseClassTestRule
-TestSafemodeBringsDownMaster.CLASS_RULE 
+TestMasterProcedureScheduler.CLASS_RULE 
 
 
 static HBaseClassTestRule
-TestRestoreSnapshotProcedure.CLASS_RULE 
+TestSafemodeBringsDownMaster.CLASS_RULE 
 
 
 static HBaseClassTestRule
-TestMasterFailoverWithProcedures.CLASS_RULE 
+TestRestoreSnapshotProcedure.CLASS_RULE 
 
 
 static HBaseClassTestRule
-TestMasterProcedureWalLease.CLASS_RULE 
+TestMasterFailoverWithProcedures.CLASS_RULE 
 
 
 static HBaseClassTestRule
-TestTableDescriptorModificationFromClient.CLASS_RULE 
+TestMasterProcedureWalLease.CLASS_RULE 
 
 
 static HBaseClassTestRule
-TestDeleteColumnFamilyProcedureFromClient.CLASS_RULE 
+TestTableDescriptorModificationFromClient.CLASS_RULE 
 
 
 static HBaseClassTestRule
-TestCloneSnapshotProcedure.CLASS_RULE 
+TestDeleteColumnFamilyProcedureFromClient.CLASS_RULE 
 
 
 static HBaseClassTestRule
+TestCloneSnapshotProcedure.CLASS_RULE 
+
+
+static HBaseClassTestRule
 TestProcedureAdmin.CLASS_RULE 
 
 
@@ -4515,394 +4519,398 @@
 
 
 static HBaseClassTestRule
-TestHRegionServerBulkLoad.CLASS_RULE 
+TestRecoveredEditsReplayAndAbort.CLASS_RULE 
 
 
 static HBaseClassTestRule
-TestCompactSplitThread.CLASS_RULE 
+TestHRegionServerBulkLoad.CLASS_RULE 
 
 
 static HBaseClassTestRule
-TestOpenSeqNumUnexpectedIncrease.CLASS_RULE 
+TestCompactSplitThread.CLASS_RULE 
 
 
 static HBaseClassTestRule
-TestRegionMergeTransactionOnCluster.CLASS_RULE 
+TestOpenSeqNumUnexpectedIncrease.CLASS_RULE 
 
 
 static HBaseClassTestRule
-TestMobStoreCompaction.CLASS_RULE 
+TestRegionMergeTransactionOnCluster.CLASS_RULE 
 
 
 static HBaseClassTestRule
-TestStripeStoreEngine.CLASS_RULE 
+TestMobStoreCompaction.CLASS_RULE 
 
 
 static HBaseClassTestRule
-TestRowTooBig.CLASS_RULE 
+TestStripeStoreEngine.CLASS_RULE 
 
 
 static HBaseClassTestRule
-TestWalAndCompactingMemStoreFlush.CLASS_RULE 
+TestRowTooBig.CLASS_RULE 
 
 
 static HBaseClassTestRule
-TestWALMonotonicallyIncreasingSeqId.CLASS_RULE 
+TestWalAndCompactingMemStoreFlush.CLASS_RULE 
 
 
 static HBaseClassTestRule
-TestCompaction.CLASS_RULE 
+TestWALMonotonicallyIncreasingSeqId.CLASS_RULE 
 
 
 static HBaseClassTestRule
-TestObservedExceptionsInBatch.CLASS_RULE 
+TestCompaction.CLASS_RULE 
 
 
 static HBaseClassTestRule
-TestDeleteMobTable.CLASS_RULE 
+TestObservedExceptionsInBatch.CLASS_RULE 
 
 
 static HBaseClassTestRule
-TestMultiVersionConcurrencyControlBasic.CLASS_RULE 
+TestDeleteMobTable.CLASS_RULE 
 
 
 static HBaseClassTestRule
-TestBlocksScanned.CLASS_RULE 
+TestMultiVersionConcurrencyControlBasic.CLASS_RULE 
 
 
 static HBaseClassTestRule
-TestCompoundBl

[44/51] [partial] hbase-site git commit: Published site at 6a5b4f2a5c188f8eef4f2250b8b7db7dd1e750e4.

2018-08-23 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.MutationBatchOperation.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.MutationBatchOperation.html
 
b/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.MutationBatchOperation.html
index db892a3..26066fb 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.MutationBatchOperation.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.MutationBatchOperation.html
@@ -118,7 +118,7 @@ var activeTableTab = "activeTableTab";
 
 
 
-static class HRegion.MutationBatchOperation
+static class HRegion.MutationBatchOperation
 extends HRegion.BatchOperation
 Batch of mutation operations. Base class is shared with HRegion.ReplayBatchOperation
 as most
  of the logic is same.
@@ -342,7 +342,7 @@ extends 
 
 nonceGroup
-private long nonceGroup
+private long nonceGroup
 
 
 
@@ -351,7 +351,7 @@ extends 
 
 nonce
-private long nonce
+private long nonce
 
 
 
@@ -368,7 +368,7 @@ extends 
 
 MutationBatchOperation
-public MutationBatchOperation(HRegion region,
+public MutationBatchOperation(HRegion region,
   Mutation[] operations,
   boolean atomic,
   long nonceGroup,
@@ -389,7 +389,7 @@ extends 
 
 getMutation
-public Mutation getMutation(int index)
+public Mutation getMutation(int index)
 
 Specified by:
 getMutation in
 class HRegion.BatchOperation
@@ -402,7 +402,7 @@ extends 
 
 getNonceGroup
-public long getNonceGroup(int index)
+public long getNonceGroup(int index)
 
 Specified by:
 getNonceGroup in
 class HRegion.BatchOperation
@@ -415,7 +415,7 @@ extends 
 
 getNonce
-public long getNonce(int index)
+public long getNonce(int index)
 
 Specified by:
 getNonce in
 class HRegion.BatchOperation
@@ -428,7 +428,7 @@ extends 
 
 getMutationsForCoprocs
-public Mutation[] getMutationsForCoprocs()
+public Mutation[] getMutationsForCoprocs()
 Description copied from 
class: HRegion.BatchOperation
 This method is potentially expensive and useful mostly for 
non-replay CP path.
 
@@ -443,7 +443,7 @@ extends 
 
 isInReplay
-public boolean isInReplay()
+public boolean isInReplay()
 
 Specified by:
 isInReplay in
 class HRegion.BatchOperation
@@ -456,7 +456,7 @@ extends 
 
 getOrigLogSeqNum
-public long getOrigLogSeqNum()
+public long getOrigLogSeqNum()
 
 Specified by:
 getOrigLogSeqNum in
 class HRegion.BatchOperation
@@ -469,7 +469,7 @@ extends 
 
 startRegionOperation
-public void startRegionOperation()
+public void startRegionOperation()
   throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true";
 title="class or interface in java.io">IOException
 
 Specified by:
@@ -485,7 +485,7 @@ extends 
 
 closeRegionOperation
-public void closeRegionOperation()
+public void closeRegionOperation()
   throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true";
 title="class or interface in java.io">IOException
 
 Specified by:
@@ -501,7 +501,7 @@ extends 
 
 checkAndPreparePut
-public void checkAndPreparePut(Put p)
+public void checkAndPreparePut(Put p)
 throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true";
 title="class or interface in java.io">IOException
 Description copied from 
class: HRegion.BatchOperation
 Implement any Put request specific check and prepare logic 
here. Please refer to
@@ -520,7 +520,7 @@ extends 
 
 checkAndPrepare
-public void checkAndPrepare()
+public void checkAndPrepare()
  throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true";
 title="class or interface in java.io">IOException
 Description copied from 
class: HRegion.BatchOperation
 Validates each mutation and prepares a batch for write. If 
necessary (non-replay case), runs
@@ -542,7 +542,7 @@ extends 
 
 prepareMiniBatchOperations
-public void prepareMiniBatchOperations(MiniBatchOperationInProgress miniBatchOp,
+public void prepareMiniBatchOperations(MiniBatchOperationInProgress miniBatchOp,
long timestamp,
https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
 title="class or interface in java.util">List acquiredRowLocks)
 throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true";
 title="class or interface in java.io">IOException
@@ -563,7 +563,7 @@ extends 
 
 buildWALEdits
-public https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
 title="class or interface in java.util">List

[18/51] [partial] hbase-site git commit: Published site at 6a5b4f2a5c188f8eef4f2250b8b7db7dd1e750e4.

2018-08-23 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RowLockImpl.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RowLockImpl.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RowLockImpl.html
index db8431b..a8cb7c4 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RowLockImpl.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RowLockImpl.html
@@ -885,7766 +885,7797 @@
 877   * @return What the next sequence 
(edit) id should be.
 878   * @throws IOException e
 879   */
-880  private long initialize(final 
CancelableProgressable reporter) throws IOException {
-881
-882//Refuse to open the region if there 
is no column family in the table
-883if 
(htableDescriptor.getColumnFamilyCount() == 0) {
-884  throw new 
DoNotRetryIOException("Table " + 
htableDescriptor.getTableName().getNameAsString()+
-885  " should have at least one 
column family.");
-886}
-887
-888MonitoredTask status = 
TaskMonitor.get().createStatus("Initializing region " + this);
-889long nextSeqId = -1;
-890try {
-891  nextSeqId = 
initializeRegionInternals(reporter, status);
-892  return nextSeqId;
-893} finally {
-894  // nextSeqid will be -1 if the 
initialization fails.
-895  // At least it will be 0 
otherwise.
-896  if (nextSeqId == -1) {
-897status.abort("Exception during 
region " + getRegionInfo().getRegionNameAsString() +
-898  " initialization.");
-899  }
-900}
-901  }
-902
-903  private long 
initializeRegionInternals(final CancelableProgressable reporter,
-904  final MonitoredTask status) throws 
IOException {
-905if (coprocessorHost != null) {
-906  status.setStatus("Running 
coprocessor pre-open hook");
-907  coprocessorHost.preOpen();
-908}
-909
-910// Write HRI to a file in case we 
need to recover hbase:meta
-911// Only the primary replica should 
write .regioninfo
-912if 
(this.getRegionInfo().getReplicaId() == RegionInfo.DEFAULT_REPLICA_ID) {
-913  status.setStatus("Writing region 
info on filesystem");
-914  fs.checkRegionInfoOnFilesystem();
-915}
-916
-917// Initialize all the HStores
-918status.setStatus("Initializing all 
the Stores");
-919long maxSeqId = 
initializeStores(reporter, status);
-920this.mvcc.advanceTo(maxSeqId);
-921if 
(ServerRegionReplicaUtil.shouldReplayRecoveredEdits(this)) {
-922  Collection stores = 
this.stores.values();
-923  try {
-924// update the stores that we are 
replaying
-925LOG.debug("replaying wal for " + 
this.getRegionInfo().getEncodedName());
-926
stores.forEach(HStore::startReplayingFromWAL);
-927// Recover any edits if 
available.
-928maxSeqId = Math.max(maxSeqId,
-929  
replayRecoveredEditsIfAny(this.fs.getRegionDir(), maxSeqIdInStores, reporter, 
status));
-930// Make sure mvcc is up to max.
-931this.mvcc.advanceTo(maxSeqId);
-932  } finally {
-933LOG.debug("stopping wal replay 
for " + this.getRegionInfo().getEncodedName());
-934// update the stores that we are 
done replaying
-935
stores.forEach(HStore::stopReplayingFromWAL);
-936  }
-937}
-938this.lastReplayedOpenRegionSeqId = 
maxSeqId;
-939
-940
this.writestate.setReadOnly(ServerRegionReplicaUtil.isReadOnly(this));
-941this.writestate.flushRequested = 
false;
-942this.writestate.compacting.set(0);
-943
-944if (this.writestate.writesEnabled) 
{
-945  LOG.debug("Cleaning up temporary 
data for " + this.getRegionInfo().getEncodedName());
-946  // Remove temporary data left over 
from old regions
-947  status.setStatus("Cleaning up 
temporary data from old regions");
-948  fs.cleanupTempDir();
-949}
-950
-951if (this.writestate.writesEnabled) 
{
-952  status.setStatus("Cleaning up 
detritus from prior splits");
-953  // Get rid of any splits or merges 
that were lost in-progress.  Clean out
-954  // these directories here on open.  
We may be opening a region that was
-955  // being split but we crashed in 
the middle of it all.
-956  LOG.debug("Cleaning up detritus for 
" + this.getRegionInfo().getEncodedName());
-957  fs.cleanupAnySplitDetritus();
-958  fs.cleanupMergesDir();
-959}
+880  @VisibleForTesting
+881  long initialize(final 
CancelableProgressable reporter) throws IOException {
+882
+883//Refuse to open the region if there 
is no column family in the table
+884if 
(htableDescriptor.getColumnFamilyCount() == 0) {
+885  throw new 
DoNotRetryIOException("Table " + 
htableDescriptor.getTableName().getNameAsString()+
+886  " should have at least one 
column family.");
+887}
+888
+889MonitoredTask status 

[22/51] [partial] hbase-site git commit: Published site at 6a5b4f2a5c188f8eef4f2250b8b7db7dd1e750e4.

2018-08-23 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.PrepareFlushResult.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.PrepareFlushResult.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.PrepareFlushResult.html
index db8431b..a8cb7c4 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.PrepareFlushResult.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.PrepareFlushResult.html
@@ -885,7766 +885,7797 @@
 877   * @return What the next sequence 
(edit) id should be.
 878   * @throws IOException e
 879   */
-880  private long initialize(final 
CancelableProgressable reporter) throws IOException {
-881
-882//Refuse to open the region if there 
is no column family in the table
-883if 
(htableDescriptor.getColumnFamilyCount() == 0) {
-884  throw new 
DoNotRetryIOException("Table " + 
htableDescriptor.getTableName().getNameAsString()+
-885  " should have at least one 
column family.");
-886}
-887
-888MonitoredTask status = 
TaskMonitor.get().createStatus("Initializing region " + this);
-889long nextSeqId = -1;
-890try {
-891  nextSeqId = 
initializeRegionInternals(reporter, status);
-892  return nextSeqId;
-893} finally {
-894  // nextSeqid will be -1 if the 
initialization fails.
-895  // At least it will be 0 
otherwise.
-896  if (nextSeqId == -1) {
-897status.abort("Exception during 
region " + getRegionInfo().getRegionNameAsString() +
-898  " initialization.");
-899  }
-900}
-901  }
-902
-903  private long 
initializeRegionInternals(final CancelableProgressable reporter,
-904  final MonitoredTask status) throws 
IOException {
-905if (coprocessorHost != null) {
-906  status.setStatus("Running 
coprocessor pre-open hook");
-907  coprocessorHost.preOpen();
-908}
-909
-910// Write HRI to a file in case we 
need to recover hbase:meta
-911// Only the primary replica should 
write .regioninfo
-912if 
(this.getRegionInfo().getReplicaId() == RegionInfo.DEFAULT_REPLICA_ID) {
-913  status.setStatus("Writing region 
info on filesystem");
-914  fs.checkRegionInfoOnFilesystem();
-915}
-916
-917// Initialize all the HStores
-918status.setStatus("Initializing all 
the Stores");
-919long maxSeqId = 
initializeStores(reporter, status);
-920this.mvcc.advanceTo(maxSeqId);
-921if 
(ServerRegionReplicaUtil.shouldReplayRecoveredEdits(this)) {
-922  Collection stores = 
this.stores.values();
-923  try {
-924// update the stores that we are 
replaying
-925LOG.debug("replaying wal for " + 
this.getRegionInfo().getEncodedName());
-926
stores.forEach(HStore::startReplayingFromWAL);
-927// Recover any edits if 
available.
-928maxSeqId = Math.max(maxSeqId,
-929  
replayRecoveredEditsIfAny(this.fs.getRegionDir(), maxSeqIdInStores, reporter, 
status));
-930// Make sure mvcc is up to max.
-931this.mvcc.advanceTo(maxSeqId);
-932  } finally {
-933LOG.debug("stopping wal replay 
for " + this.getRegionInfo().getEncodedName());
-934// update the stores that we are 
done replaying
-935
stores.forEach(HStore::stopReplayingFromWAL);
-936  }
-937}
-938this.lastReplayedOpenRegionSeqId = 
maxSeqId;
-939
-940
this.writestate.setReadOnly(ServerRegionReplicaUtil.isReadOnly(this));
-941this.writestate.flushRequested = 
false;
-942this.writestate.compacting.set(0);
-943
-944if (this.writestate.writesEnabled) 
{
-945  LOG.debug("Cleaning up temporary 
data for " + this.getRegionInfo().getEncodedName());
-946  // Remove temporary data left over 
from old regions
-947  status.setStatus("Cleaning up 
temporary data from old regions");
-948  fs.cleanupTempDir();
-949}
-950
-951if (this.writestate.writesEnabled) 
{
-952  status.setStatus("Cleaning up 
detritus from prior splits");
-953  // Get rid of any splits or merges 
that were lost in-progress.  Clean out
-954  // these directories here on open.  
We may be opening a region that was
-955  // being split but we crashed in 
the middle of it all.
-956  LOG.debug("Cleaning up detritus for 
" + this.getRegionInfo().getEncodedName());
-957  fs.cleanupAnySplitDetritus();
-958  fs.cleanupMergesDir();
-959}
+880  @VisibleForTesting
+881  long initialize(final 
CancelableProgressable reporter) throws IOException {
+882
+883//Refuse to open the region if there 
is no column family in the table
+884if 
(htableDescriptor.getColumnFamilyCount() == 0) {
+885  throw new 
DoNotRetryIOException("Table " + 
htableDescriptor.getTableName().getNameAsString()+
+886  " should have at least one 
column family.");
+887}

[27/51] [partial] hbase-site git commit: Published site at 6a5b4f2a5c188f8eef4f2250b8b7db7dd1e750e4.

2018-08-23 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.Result.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.Result.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.Result.html
index db8431b..a8cb7c4 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.Result.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.Result.html
@@ -885,7766 +885,7797 @@
 877   * @return What the next sequence 
(edit) id should be.
 878   * @throws IOException e
 879   */
-880  private long initialize(final 
CancelableProgressable reporter) throws IOException {
-881
-882//Refuse to open the region if there 
is no column family in the table
-883if 
(htableDescriptor.getColumnFamilyCount() == 0) {
-884  throw new 
DoNotRetryIOException("Table " + 
htableDescriptor.getTableName().getNameAsString()+
-885  " should have at least one 
column family.");
-886}
-887
-888MonitoredTask status = 
TaskMonitor.get().createStatus("Initializing region " + this);
-889long nextSeqId = -1;
-890try {
-891  nextSeqId = 
initializeRegionInternals(reporter, status);
-892  return nextSeqId;
-893} finally {
-894  // nextSeqid will be -1 if the 
initialization fails.
-895  // At least it will be 0 
otherwise.
-896  if (nextSeqId == -1) {
-897status.abort("Exception during 
region " + getRegionInfo().getRegionNameAsString() +
-898  " initialization.");
-899  }
-900}
-901  }
-902
-903  private long 
initializeRegionInternals(final CancelableProgressable reporter,
-904  final MonitoredTask status) throws 
IOException {
-905if (coprocessorHost != null) {
-906  status.setStatus("Running 
coprocessor pre-open hook");
-907  coprocessorHost.preOpen();
-908}
-909
-910// Write HRI to a file in case we 
need to recover hbase:meta
-911// Only the primary replica should 
write .regioninfo
-912if 
(this.getRegionInfo().getReplicaId() == RegionInfo.DEFAULT_REPLICA_ID) {
-913  status.setStatus("Writing region 
info on filesystem");
-914  fs.checkRegionInfoOnFilesystem();
-915}
-916
-917// Initialize all the HStores
-918status.setStatus("Initializing all 
the Stores");
-919long maxSeqId = 
initializeStores(reporter, status);
-920this.mvcc.advanceTo(maxSeqId);
-921if 
(ServerRegionReplicaUtil.shouldReplayRecoveredEdits(this)) {
-922  Collection stores = 
this.stores.values();
-923  try {
-924// update the stores that we are 
replaying
-925LOG.debug("replaying wal for " + 
this.getRegionInfo().getEncodedName());
-926
stores.forEach(HStore::startReplayingFromWAL);
-927// Recover any edits if 
available.
-928maxSeqId = Math.max(maxSeqId,
-929  
replayRecoveredEditsIfAny(this.fs.getRegionDir(), maxSeqIdInStores, reporter, 
status));
-930// Make sure mvcc is up to max.
-931this.mvcc.advanceTo(maxSeqId);
-932  } finally {
-933LOG.debug("stopping wal replay 
for " + this.getRegionInfo().getEncodedName());
-934// update the stores that we are 
done replaying
-935
stores.forEach(HStore::stopReplayingFromWAL);
-936  }
-937}
-938this.lastReplayedOpenRegionSeqId = 
maxSeqId;
-939
-940
this.writestate.setReadOnly(ServerRegionReplicaUtil.isReadOnly(this));
-941this.writestate.flushRequested = 
false;
-942this.writestate.compacting.set(0);
-943
-944if (this.writestate.writesEnabled) 
{
-945  LOG.debug("Cleaning up temporary 
data for " + this.getRegionInfo().getEncodedName());
-946  // Remove temporary data left over 
from old regions
-947  status.setStatus("Cleaning up 
temporary data from old regions");
-948  fs.cleanupTempDir();
-949}
-950
-951if (this.writestate.writesEnabled) 
{
-952  status.setStatus("Cleaning up 
detritus from prior splits");
-953  // Get rid of any splits or merges 
that were lost in-progress.  Clean out
-954  // these directories here on open.  
We may be opening a region that was
-955  // being split but we crashed in 
the middle of it all.
-956  LOG.debug("Cleaning up detritus for 
" + this.getRegionInfo().getEncodedName());
-957  fs.cleanupAnySplitDetritus();
-958  fs.cleanupMergesDir();
-959}
+880  @VisibleForTesting
+881  long initialize(final 
CancelableProgressable reporter) throws IOException {
+882
+883//Refuse to open the region if there 
is no column family in the table
+884if 
(htableDescriptor.getColumnFamilyCount() == 0) {
+885  throw new 
DoNotRetryIOException("Table " + 
htableDescriptor.getTableName().getNameAsString()+
+886  " should have at least one 
column family.");
+887}

[48/51] [partial] hbase-site git commit: Published site at 6a5b4f2a5c188f8eef4f2250b8b7db7dd1e750e4.

2018-08-23 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/devapidocs/org/apache/hadoop/hbase/mapreduce/package-tree.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/mapreduce/package-tree.html 
b/devapidocs/org/apache/hadoop/hbase/mapreduce/package-tree.html
index 0678f9b..deb19ba 100644
--- a/devapidocs/org/apache/hadoop/hbase/mapreduce/package-tree.html
+++ b/devapidocs/org/apache/hadoop/hbase/mapreduce/package-tree.html
@@ -293,10 +293,10 @@
 
 java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true";
 title="class or interface in java.lang">Enum (implements java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true";
 title="class or interface in java.lang">Comparable, java.io.https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true";
 title="class or interface in java.io">Serializable)
 
-org.apache.hadoop.hbase.mapreduce.SyncTable.SyncMapper.Counter
-org.apache.hadoop.hbase.mapreduce.TableSplit.Version
 org.apache.hadoop.hbase.mapreduce.CellCounter.CellCounterMapper.Counters
+org.apache.hadoop.hbase.mapreduce.SyncTable.SyncMapper.Counter
 org.apache.hadoop.hbase.mapreduce.RowCounter.RowCounterMapper.Counters
+org.apache.hadoop.hbase.mapreduce.TableSplit.Version
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/devapidocs/org/apache/hadoop/hbase/master/MetricsMasterSource.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/master/MetricsMasterSource.html 
b/devapidocs/org/apache/hadoop/hbase/master/MetricsMasterSource.html
index f62c8ef..85c869c 100644
--- a/devapidocs/org/apache/hadoop/hbase/master/MetricsMasterSource.html
+++ b/devapidocs/org/apache/hadoop/hbase/master/MetricsMasterSource.html
@@ -249,6 +249,22 @@ extends 
 
 static https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String
+OFFLINE_REGION_COUNT_DESC 
+
+
+static https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String
+OFFLINE_REGION_COUNT_NAME 
+
+
+static https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String
+ONLINE_REGION_COUNT_DESC 
+
+
+static https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String
+ONLINE_REGION_COUNT_NAME 
+
+
+static https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String
 SERVER_CRASH_METRIC_PREFIX 
 
 
@@ -567,13 +583,39 @@ extends 
 
 
+
+
+
+
+
+ONLINE_REGION_COUNT_NAME
+static final https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String ONLINE_REGION_COUNT_NAME
+
+See Also:
+Constant
 Field Values
+
+
+
+
+
+
+
+
+OFFLINE_REGION_COUNT_NAME
+static final https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String OFFLINE_REGION_COUNT_NAME
+
+See Also:
+Constant
 Field Values
+
+
+
 
 
 
 
 
 CLUSTER_REQUESTS_NAME
-static final https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String CLUSTER_REQUESTS_NAME
+static final https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String CLUSTER_REQUESTS_NAME
 
 See Also:
 Constant
 Field Values
@@ -586,7 +628,7 @@ extends 
 
 MASTER_ACTIVE_TIME_DESC
-static final https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String MASTER_ACTIVE_TIME_DESC
+static final https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String MASTER_ACTIVE_TIME_DESC
 
 See Also:
 Constant
 Field Values
@@ -599,7 +641,7 @@ extends 
 
 MASTER_START_TIME_DESC
-static final https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String MASTER_START_TIME_DESC
+static final https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String MASTER_START_TIME_DESC
 
 See Also:
 Constant
 Field Values
@@ -612,7 +654,7 @@ extends 
 
 MASTER_FINISHED_INITIALIZATION_TIME_DESC
-static final https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String MASTER_FINISHED_INITIALIZATION_TIME_DESC
+static final https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String MASTER

[02/51] [partial] hbase-site git commit: Published site at 6a5b4f2a5c188f8eef4f2250b8b7db7dd1e750e4.

2018-08-23 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/testdevapidocs/org/apache/hadoop/hbase/regionserver/package-summary.html
--
diff --git 
a/testdevapidocs/org/apache/hadoop/hbase/regionserver/package-summary.html 
b/testdevapidocs/org/apache/hadoop/hbase/regionserver/package-summary.html
index fc8d06b..aa6810b 100644
--- a/testdevapidocs/org/apache/hadoop/hbase/regionserver/package-summary.html
+++ b/testdevapidocs/org/apache/hadoop/hbase/regionserver/package-summary.html
@@ -941,519 +941,527 @@
 
 
 
+TestRecoveredEditsReplayAndAbort
+
+HBASE-21031
+ If replay edits fails, we need to make sure memstore is rollbacked
+ And if MSLAB is used, all chunk is released too.
+
+
+
 TestRegionFavoredNodes
 
 Tests the ability to specify favored nodes for a 
region.
 
 
-
+
 TestRegionIncrement
 
 Increments with some concurrency against a region to ensure 
we get the right answer.
 
 
-
+
 TestRegionIncrement.CrossRowCellIncrementer
 
 Increments a random row's Cell count 
times.
 
 
-
+
 TestRegionIncrement.SingleCellIncrementer
 
 Increments a single cell a bunch of times.
 
 
-
+
 TestRegionInfoBuilder
  
 
-
+
 TestRegionMergeTransactionOnCluster
  
 
-
+
 TestRegionMergeTransactionOnCluster.MyMaster
  
 
-
+
 TestRegionMergeTransactionOnCluster.MyMasterRpcServices
  
 
-
+
 TestRegionMove
 
 Test move fails when table disabled
 
 
-
+
 TestRegionOpen
  
 
-
+
 TestRegionReplicaFailover
 
 Tests failover of secondary region replicas.
 
 
-
+
 TestRegionReplicas
 
 Tests for region replicas.
 
 
-
+
 TestRegionReplicasAreDistributed
  
 
-
+
 TestRegionReplicasWithModifyTable
  
 
-
+
 TestRegionReplicasWithRestartScenarios
  
 
-
+
 TestRegionServerAbort
 
 Tests around regionserver shutdown and abort
 
 
-
+
 TestRegionServerAbort.ErrorThrowingHRegion
 
 Throws an exception during store file refresh in order to 
trigger a regionserver abort.
 
 
-
+
 TestRegionServerAbort.StopBlockingRegionObserver
  
 
-
+
 TestRegionServerAccounting
  
 
-
+
 TestRegionServerCrashDisableWAL
 
 Testcase for HBASE-20742
 
 
-
+
 TestRegionServerHostname
 
 Tests for the hostname specification by region server
 
 
-
+
 TestRegionServerMetrics
  
 
-
+
 TestRegionServerNoMaster
 
 Tests on the region server, without the master.
 
 
-
+
 TestRegionServerOnlineConfigChange
 
 Verify that the Online config Changes on the HRegionServer 
side are actually
  happening.
 
 
-
+
 TestRegionServerReadRequestMetrics
  
 
-
+
 TestRegionServerReadRequestMetrics.ScanRegionCoprocessor
  
 
-
+
 TestRegionServerRegionSpaceUseReport
 
 Test class for isolated (non-cluster) tests surrounding the 
report
  of Region space use to the Master by RegionServers.
 
 
-
+
 TestRegionServerReportForDuty
  
 
-
+
 TestRegionServerReportForDuty.MyRegionServer
  
 
-
+
 TestRegionSplitPolicy
  
 
-
+
 TestRemoveRegionMetrics
  
 
-
+
 TestResettingCounters
  
 
-
+
 TestReversibleScanners
 
 Test cases against ReversibleKeyValueScanner
 
 
-
+
 TestRowTooBig
 
 Test case to check HRS throws 
RowTooBigException
  when row size exceeds configured limits.
 
 
-
+
 TestRpcSchedulerFactory
 
 A silly test that does nothing but make sure an 
rpcscheduler factory makes what it says
  it is going to make.
 
 
-
+
 TestRSKilledWhenInitializing
 
 Tests that a regionserver that dies after reporting for 
duty gets removed
  from list of online regions.
 
 
-
+
 TestRSKilledWhenInitializing.RegisterAndDieRegionServer
 
 A RegionServer that reports for duty and then immediately 
dies if it is the first to receive
  the response to a reportForDuty.
 
 
-
+
 TestRSStatusServlet
 
 Tests for the region server status page and its 
template.
 
 
-
+
 TestScanner
 
 Test of a long-lived scanner validating as we go.
 
 
-
+
 TestScannerHeartbeatMessages
 
 Here we test to make sure that scans return the expected 
Results when the server is sending the
  Client heartbeat messages.
 
 
-
+
 TestScannerHeartbeatMessages.HeartbeatHRegion
 
 Custom HRegion class that instantiates 
RegionScanners with configurable sleep times
  between fetches of row Results and/or column family cells.
 
 
-
+
 TestScannerHeartbeatMessages.HeartbeatHRegionServer
 
 Custom HRegionServer instance that instantiates TestScannerHeartbeatMessages.HeartbeatRPCServices
 in place of
  RSRpcServices to allow us to toggle support for heartbeat 
messages
 
 
-
+
 TestScannerHeartbeatMessages.HeartbeatKVHeap
 
 Custom KV Heap that can be configured to sleep/wait in 
between retrievals of column family
  cells.
 
 
-
+
 TestScannerHeartbeatMessages.HeartbeatRegionScanner
 
 Custom RegionScanner that can be configured to sleep 
between retrievals of row Results and/or
  column family cells
 
 
-
+
 TestScannerHeartbeatMessages.HeartbeatReversedKVHeap
 
 Custom reversed KV Heap that can be configured to sleep in 
between retrievals of column family
  cells.
 
 
-
+
 TestScannerHeartbeatMessages.HeartbeatReversedRegionScanner
 
 Custo

[47/51] [partial] hbase-site git commit: Published site at 6a5b4f2a5c188f8eef4f2250b8b7db7dd1e750e4.

2018-08-23 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/devapidocs/org/apache/hadoop/hbase/master/MetricsMasterWrapperImpl.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/master/MetricsMasterWrapperImpl.html 
b/devapidocs/org/apache/hadoop/hbase/master/MetricsMasterWrapperImpl.html
index 018d145..f8393cb 100644
--- a/devapidocs/org/apache/hadoop/hbase/master/MetricsMasterWrapperImpl.html
+++ b/devapidocs/org/apache/hadoop/hbase/master/MetricsMasterWrapperImpl.html
@@ -18,7 +18,7 @@
 catch(err) {
 }
 //-->
-var methods = 
{"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10};
+var methods = 
{"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10};
 var tabs = {65535:["t0","All Methods"],2:["t2","Instance 
Methods"],8:["t4","Concrete Methods"]};
 var altColor = "altColor";
 var rowColor = "rowColor";
@@ -114,7 +114,7 @@ var activeTableTab = "activeTableTab";
 
 
 @InterfaceAudience.Private
-public class MetricsMasterWrapperImpl
+public class MetricsMasterWrapperImpl
 extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true";
 title="class or interface in java.lang">Object
 implements MetricsMasterWrapper
 Impl for exposing HMaster Information through JMX
@@ -249,36 +249,42 @@ implements 
+PairOfSameTypeInteger>
+getRegionCounts()
+Get the online and offline region counts
+
+
+
 https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String
 getRegionServers()
 Get the live region servers
 
 
-
+
 https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String
 getServerName()
 Get ServerName
 
 
-
+
 long
 getSplitPlanCount()
 Get the number of region split plans executed.
 
 
-
+
 long
 getStartTime()
 Get hbase master start time
 
 
-
+
 https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true";
 title="class or interface in java.util">MapString,https://docs.oracle.com/javase/8/docs/api/java/util/Map.Entry.html?is-external=true";
 title="class or interface in java.util">Map.EntryLong,https://docs.oracle.com/javase/8/docs/api/java/lang/Long.html?is-external=true";
 title="class or interface in java.lang">Long>>
 getTableSpaceUtilization()
 Gets the space usage and limit for each table.
 
 
-
+
 https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String
 getZookeeperQuorum()
 Get the ZooKeeper Quorum Info
@@ -312,7 +318,7 @@ implements 
 
 master
-private final HMaster master
+private final HMaster master
 
 
 
@@ -329,7 +335,7 @@ implements 
 
 MetricsMasterWrapperImpl
-public MetricsMasterWrapperImpl(HMaster master)
+public MetricsMasterWrapperImpl(HMaster master)
 
 
 
@@ -346,7 +352,7 @@ implements 
 
 getAverageLoad
-public double getAverageLoad()
+public double getAverageLoad()
 Description copied from 
interface: MetricsMasterWrapper
 Get Average Load
 
@@ -363,7 +369,7 @@ implements 
 
 getSplitPlanCount
-public long getSplitPlanCount()
+public long getSplitPlanCount()
 Description copied from 
interface: MetricsMasterWrapper
 Get the number of region split plans executed.
 
@@ -378,7 +384,7 @@ implements 
 
 getMergePlanCount
-public long getMergePlanCount()
+public long getMergePlanCount()
 Description copied from 
interface: MetricsMasterWrapper
 Get the number of region merge plans executed.
 
@@ -393,7 +399,7 @@ implements 
 
 getMasterInitializationTime
-public long getMasterInitializationTime()
+public long getMasterInitializationTime()
 Description copied from 
interface: MetricsMasterWrapper
 Get the time in Millis when the master finished 
initializing/becoming the active master
 
@@ -408,7 +414,7 @@ implements 
 
 getClusterId
-public https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String getClusterId()
+public https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String getClusterId()
 Description copied from 
interface: MetricsMasterWrapper
 Get the Cluster ID
 
@@ -425,7 +431,7 @@ implements 
 
 getZookeeperQuorum
-public https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title

[34/51] [partial] hbase-site git commit: Published site at 6a5b4f2a5c188f8eef4f2250b8b7db7dd1e750e4.

2018-08-23 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/CompactingMemStore.InMemoryCompactionRunnable.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/CompactingMemStore.InMemoryCompactionRunnable.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/CompactingMemStore.InMemoryCompactionRunnable.html
index bf6738e..a62d84e 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/CompactingMemStore.InMemoryCompactionRunnable.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/CompactingMemStore.InMemoryCompactionRunnable.html
@@ -74,569 +74,568 @@
 066
 067  private static final Logger LOG = 
LoggerFactory.getLogger(CompactingMemStore.class);
 068  private HStore store;
-069  private RegionServicesForStores 
regionServices;
-070  private CompactionPipeline pipeline;
-071  protected MemStoreCompactor 
compactor;
-072
-073  private long inmemoryFlushSize;   
// the threshold on active size for in-memory flush
-074  private final AtomicBoolean 
inMemoryCompactionInProgress = new AtomicBoolean(false);
-075
-076  // inWalReplay is true while we are 
synchronously replaying the edits from WAL
-077  private boolean inWalReplay = false;
-078
-079  @VisibleForTesting
-080  protected final AtomicBoolean 
allowCompaction = new AtomicBoolean(true);
-081  private boolean compositeSnapshot = 
true;
-082
-083  /**
-084   * Types of indexes (part of immutable 
segments) to be used after flattening,
-085   * compaction, or merge are applied.
-086   */
-087  public enum IndexType {
-088CSLM_MAP,   // ConcurrentSkipLisMap
-089ARRAY_MAP,  // CellArrayMap
-090CHUNK_MAP   // CellChunkMap
-091  }
-092
-093  private IndexType indexType = 
IndexType.ARRAY_MAP;  // default implementation
-094
-095  public static final long DEEP_OVERHEAD 
= ClassSize.align( AbstractMemStore.DEEP_OVERHEAD
-096  + 7 * ClassSize.REFERENCE // 
Store, RegionServicesForStores, CompactionPipeline,
-097  // MemStoreCompactor, 
inMemoryCompactionInProgress,
-098  // allowCompaction, indexType
-099  + Bytes.SIZEOF_LONG   // 
inmemoryFlushSize
-100  + 2 * Bytes.SIZEOF_BOOLEAN// 
compositeSnapshot and inWalReplay
-101  + 2 * ClassSize.ATOMIC_BOOLEAN// 
inMemoryCompactionInProgress and allowCompaction
-102  + CompactionPipeline.DEEP_OVERHEAD 
+ MemStoreCompactor.DEEP_OVERHEAD);
-103
-104  public CompactingMemStore(Configuration 
conf, CellComparator c,
-105  HStore store, 
RegionServicesForStores regionServices,
-106  MemoryCompactionPolicy 
compactionPolicy) throws IOException {
-107super(conf, c);
-108this.store = store;
-109this.regionServices = 
regionServices;
-110this.pipeline = new 
CompactionPipeline(getRegionServices());
-111this.compactor = 
createMemStoreCompactor(compactionPolicy);
-112if 
(conf.getBoolean(MemStoreLAB.USEMSLAB_KEY, MemStoreLAB.USEMSLAB_DEFAULT)) {
-113  // if user requested to work with 
MSLABs (whether on- or off-heap), then the
-114  // immutable segments are going to 
use CellChunkMap as their index
-115  indexType = IndexType.CHUNK_MAP;
-116} else {
-117  indexType = IndexType.ARRAY_MAP;
-118}
-119// initialization of the flush size 
should happen after initialization of the index type
-120// so do not transfer the following 
method
-121initInmemoryFlushSize(conf);
-122LOG.info("Store={}, in-memory flush 
size threshold={}, immutable segments index type={}, " +
-123"compactor={}", 
this.store.getColumnFamilyName(),
-124
StringUtils.byteDesc(this.inmemoryFlushSize), this.indexType,
-125(this.compactor == null? "NULL": 
this.compactor.toString()));
-126  }
-127
-128  @VisibleForTesting
-129  protected MemStoreCompactor 
createMemStoreCompactor(MemoryCompactionPolicy compactionPolicy)
-130  throws IllegalArgumentIOException 
{
-131return new MemStoreCompactor(this, 
compactionPolicy);
-132  }
-133
-134  private void 
initInmemoryFlushSize(Configuration conf) {
-135double factor = 0;
-136long memstoreFlushSize = 
getRegionServices().getMemStoreFlushSize();
-137int numStores = 
getRegionServices().getNumStores();
-138if (numStores <= 1) {
-139  // Family number might also be zero 
in some of our unit test case
-140  numStores = 1;
-141}
-142factor = 
conf.getDouble(IN_MEMORY_FLUSH_THRESHOLD_FACTOR_KEY, 0.0);
-143if(factor != 0.0) {
-144  // multiply by a factor (the same 
factor for all index types)
-145  inmemoryFlushSize = (long) (factor 
* memstoreFlushSize) / numStores;
-146} else {
-147  inmemoryFlushSize = 
IN_MEMORY_FLUSH_MULTIPLIER *
-148  
conf.getLong(MemStoreLAB.CHUNK_SIZE_KEY, MemStoreLAB.CHUNK_SIZE_DEFAULT);
-149  inmemoryFlushSize -= 
ChunkCreator.SIZEOF_CHUNK_HEADER;
-150}
-151  }
-152
-

[04/51] [partial] hbase-site git commit: Published site at 6a5b4f2a5c188f8eef4f2250b8b7db7dd1e750e4.

2018-08-23 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/testdevapidocs/org/apache/hadoop/hbase/regionserver/TestHRegion.html
--
diff --git 
a/testdevapidocs/org/apache/hadoop/hbase/regionserver/TestHRegion.html 
b/testdevapidocs/org/apache/hadoop/hbase/regionserver/TestHRegion.html
index edfb4bf..8b112b5 100644
--- a/testdevapidocs/org/apache/hadoop/hbase/regionserver/TestHRegion.html
+++ b/testdevapidocs/org/apache/hadoop/hbase/regionserver/TestHRegion.html
@@ -18,7 +18,7 @@
 catch(err) {
 }
 //-->
-var methods = 
{"i0":9,"i1":10,"i2":10,"i3":10,"i4":9,"i5":9,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":9,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":10,"i33":10,"i34":10,"i35":10,"i36":10,"i37":10,"i38":10,"i39":10,"i40":10,"i41":10,"i42":10,"i43":10,"i44":10,"i45":10,"i46":10,"i47":10,"i48":10,"i49":10,"i50":9,"i51":10,"i52":10,"i53":10,"i54":10,"i55":10,"i56":10,"i57":10,"i58":10,"i59":10,"i60":10,"i61":10,"i62":10,"i63":10,"i64":10,"i65":10,"i66":10,"i67":10,"i68":10,"i69":10,"i70":10,"i71":10,"i72":10,"i73":10,"i74":10,"i75":10,"i76":10,"i77":10,"i78":10,"i79":10,"i80":10,"i81":10,"i82":10,"i83":10,"i84":10,"i85":10,"i86":10,"i87":10,"i88":10,"i89":10,"i90":10,"i91":10,"i92":10,"i93":10,"i94":10,"i95":10,"i96":10,"i97":10,"i98":10,"i99":10,"i100":10,"i101":10,"i102":10,"i103":10,"i104":10,"i105":10,"i106":10,"i107":10,"i108":10,"i109":
 
10,"i110":10,"i111":10,"i112":10,"i113":10,"i114":10,"i115":10,"i116":10,"i117":10,"i118":10,"i119":10,"i120":10,"i121":10,"i122":10,"i123":10,"i124":10,"i125":10,"i126":10,"i127":10,"i128":9,"i129":10};
+var methods = 
{"i0":9,"i1":10,"i2":10,"i3":10,"i4":9,"i5":9,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":9,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":10,"i33":10,"i34":10,"i35":10,"i36":10,"i37":10,"i38":10,"i39":10,"i40":10,"i41":10,"i42":10,"i43":10,"i44":10,"i45":10,"i46":10,"i47":10,"i48":10,"i49":10,"i50":9,"i51":10,"i52":10,"i53":10,"i54":10,"i55":10,"i56":10,"i57":10,"i58":10,"i59":10,"i60":10,"i61":10,"i62":10,"i63":10,"i64":10,"i65":10,"i66":10,"i67":10,"i68":10,"i69":10,"i70":10,"i71":10,"i72":10,"i73":10,"i74":10,"i75":10,"i76":10,"i77":10,"i78":10,"i79":10,"i80":10,"i81":10,"i82":10,"i83":10,"i84":10,"i85":10,"i86":10,"i87":10,"i88":10,"i89":10,"i90":10,"i91":10,"i92":10,"i93":10,"i94":10,"i95":10,"i96":10,"i97":10,"i98":10,"i99":10,"i100":10,"i101":10,"i102":10,"i103":10,"i104":10,"i105":10,"i106":10,"i107":10,"i108":10,"i109":
 
10,"i110":10,"i111":10,"i112":10,"i113":10,"i114":10,"i115":10,"i116":10,"i117":10,"i118":10,"i119":10,"i120":10,"i121":10,"i122":10,"i123":10,"i124":10,"i125":10,"i126":10,"i127":10,"i128":10,"i129":9,"i130":10};
 var tabs = {65535:["t0","All Methods"],1:["t1","Static 
Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]};
 var altColor = "altColor";
 var rowColor = "rowColor";
@@ -644,235 +644,241 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 void
+testFlushAndMemstoreSizeCounting()
+A test case of HBASE-21041
+
+
+
+void
 testFlushCacheWhileScanning()
 Flushes the cache in a thread while scanning.
 
 
-
+
 void
 testFlushedFileWithNoTags() 
 
-
+
 void
 testFlushMarkers() 
 
-
+
 void
 testFlushMarkersWALFail() 
 
-
+
 void
 testFlushResult()
 Test that we get the expected flush results back
 
 
-
+
 void
 testFlushSizeAccounting()
 Test we do not lose data if we fail a flush and then 
close.
 
 
-
+
 void
 testGet_Basic() 
 
-
+
 void
 testGet_Empty() 
 
-
+
 void
 testGet_FamilyChecker() 
 
-
+
 void
 testgetHDFSBlocksDistribution() 
 
-
+
 void
 testGetScanner_WithNoFamilies() 
 
-
+
 void
 testGetScanner_WithNotOkFamilies() 
 
-
+
 void
 testGetScanner_WithOkFamilies() 
 
-
+
 void
 testGetScanner_WithRegionClosed()
 This method tests 
https://issues.apache.org/jira/browse/HBASE-2516.
 
 
-
+
 void
 testGetWhileRegionClose() 
 
-
+
 void
 testGetWithFilter() 
 
-
+
 void
 testHolesInMeta() 
 
-
+
 void
 testIncrementTimestampsAreMonotonic() 
 
-
+
 void
 testIncrWithReadOnlyTable() 
 
-
+
 void
 testIndexesScanWithOneDeletedRow() 
 
-
+
 void
 testLongQualifier()
 Write an HFile block full with Cells whose qualifier that 
are identical between
  0 and Short.MAX_VALUE.
 
 
-
+
 void
 testMemstoreSizeAccountingWithFailedPostBatchMutate() 
 
-
+
 void
 testMemstoreSnapshotSize() 
 
-
+
 void
 testMutateRow_WriteRequestCount() 
 
-
+
 void
 testOpenRegionWrittenToWAL() 
 
-
+
 void
 testParallelAppendWithMemStoreFlush()
 Test case to check append function with memstore 
flushing
 
 
-
+
 void
 testParallelIncrementWithMemStoreFlush()
 Test case to check increment function with memstore 
flushing
 
 
-
+
 void
 test

[43/51] [partial] hbase-site git commit: Published site at 6a5b4f2a5c188f8eef4f2250b8b7db7dd1e750e4.

2018-08-23 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.ReplayBatchOperation.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.ReplayBatchOperation.html
 
b/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.ReplayBatchOperation.html
index b62cd5a..d358bb7 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.ReplayBatchOperation.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.ReplayBatchOperation.html
@@ -118,7 +118,7 @@ var activeTableTab = "activeTableTab";
 
 
 
-static class HRegion.ReplayBatchOperation
+static class HRegion.ReplayBatchOperation
 extends HRegion.BatchOperation
 Batch of mutations for replay. Base class is shared with HRegion.MutationBatchOperation
 as most
  of the logic is same.
@@ -306,7 +306,7 @@ extends 
 
 origLogSeqNum
-private long origLogSeqNum
+private long origLogSeqNum
 
 
 
@@ -323,7 +323,7 @@ extends 
 
 ReplayBatchOperation
-public ReplayBatchOperation(HRegion region,
+public ReplayBatchOperation(HRegion region,
 WALSplitter.MutationReplay[] operations,
 long origLogSeqNum)
 
@@ -342,7 +342,7 @@ extends 
 
 getMutation
-public Mutation getMutation(int index)
+public Mutation getMutation(int index)
 
 Specified by:
 getMutation in
 class HRegion.BatchOperation
@@ -355,7 +355,7 @@ extends 
 
 getNonceGroup
-public long getNonceGroup(int index)
+public long getNonceGroup(int index)
 
 Specified by:
 getNonceGroup in
 class HRegion.BatchOperation
@@ -368,7 +368,7 @@ extends 
 
 getNonce
-public long getNonce(int index)
+public long getNonce(int index)
 
 Specified by:
 getNonce in
 class HRegion.BatchOperation
@@ -381,7 +381,7 @@ extends 
 
 getMutationsForCoprocs
-public Mutation[] getMutationsForCoprocs()
+public Mutation[] getMutationsForCoprocs()
 Description copied from 
class: HRegion.BatchOperation
 This method is potentially expensive and useful mostly for 
non-replay CP path.
 
@@ -396,7 +396,7 @@ extends 
 
 isInReplay
-public boolean isInReplay()
+public boolean isInReplay()
 
 Specified by:
 isInReplay in
 class HRegion.BatchOperation
@@ -409,7 +409,7 @@ extends 
 
 getOrigLogSeqNum
-public long getOrigLogSeqNum()
+public long getOrigLogSeqNum()
 
 Specified by:
 getOrigLogSeqNum in
 class HRegion.BatchOperation
@@ -422,7 +422,7 @@ extends 
 
 startRegionOperation
-public void startRegionOperation()
+public void startRegionOperation()
   throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true";
 title="class or interface in java.io">IOException
 
 Specified by:
@@ -438,7 +438,7 @@ extends 
 
 closeRegionOperation
-public void closeRegionOperation()
+public void closeRegionOperation()
   throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true";
 title="class or interface in java.io">IOException
 
 Specified by:
@@ -454,7 +454,7 @@ extends 
 
 checkAndPreparePut
-protected void checkAndPreparePut(Put p)
+protected void checkAndPreparePut(Put p)
throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true";
 title="class or interface in java.io">IOException
 During replay, there could exist column families which are 
removed between region server
  failure and replay
@@ -472,7 +472,7 @@ extends 
 
 checkAndPrepare
-public void checkAndPrepare()
+public void checkAndPrepare()
  throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true";
 title="class or interface in java.io">IOException
 Description copied from 
class: HRegion.BatchOperation
 Validates each mutation and prepares a batch for write. If 
necessary (non-replay case), runs
@@ -494,7 +494,7 @@ extends 
 
 prepareMiniBatchOperations
-public void prepareMiniBatchOperations(MiniBatchOperationInProgress miniBatchOp,
+public void prepareMiniBatchOperations(MiniBatchOperationInProgress miniBatchOp,
long timestamp,
https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
 title="class or interface in java.util">List acquiredRowLocks)
 throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true";
 title="class or interface in java.io">IOException
@@ -515,7 +515,7 @@ extends 
 
 writeMiniBatchOperationsToMemStore
-public MultiVersionConcurrencyControl.WriteEntry writeMiniBatchOperationsToMemStore(MiniBatchOperationInProgress miniBatchOp,
+public MultiVersionConcurr

[35/51] [partial] hbase-site git commit: Published site at 6a5b4f2a5c188f8eef4f2250b8b7db7dd1e750e4.

2018-08-23 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/AbstractMemStore.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/AbstractMemStore.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/AbstractMemStore.html
index 85262eb..52d6d8e 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/AbstractMemStore.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/AbstractMemStore.html
@@ -62,329 +62,341 @@
 054  // Used to track when to flush
 055  private volatile long 
timeOfOldestEdit;
 056
-057  public final static long FIXED_OVERHEAD 
= (long) ClassSize.OBJECT
-058  + (4 * ClassSize.REFERENCE)
-059  + (2 * Bytes.SIZEOF_LONG); // 
snapshotId, timeOfOldestEdit
-060
-061  public final static long DEEP_OVERHEAD 
= FIXED_OVERHEAD;
+057  protected RegionServicesForStores 
regionServices;
+058
+059  public final static long FIXED_OVERHEAD 
= (long) ClassSize.OBJECT
+060  + (5 * ClassSize.REFERENCE)
+061  + (2 * Bytes.SIZEOF_LONG); // 
snapshotId, timeOfOldestEdit
 062
-063  public static void 
addToScanners(List segments, long readPt,
-064  List 
scanners) {
-065for (Segment item : segments) {
-066  addToScanners(item, readPt, 
scanners);
-067}
-068  }
-069
-070  protected static void 
addToScanners(Segment segment, long readPt,
-071  List 
scanners) {
-072
scanners.add(segment.getScanner(readPt));
-073  }
-074
-075  protected AbstractMemStore(final 
Configuration conf, final CellComparator c) {
-076this.conf = conf;
-077this.comparator = c;
-078resetActive();
-079this.snapshot = 
SegmentFactory.instance().createImmutableSegment(c);
-080this.snapshotId = NO_SNAPSHOT_ID;
-081  }
-082
-083  protected void resetActive() {
-084// Reset heap to not include any 
keys
-085active = 
SegmentFactory.instance().createMutableSegment(conf, comparator);
-086timeOfOldestEdit = Long.MAX_VALUE;
-087  }
-088
-089  /**
-090   * Updates the wal with the lowest 
sequence id (oldest entry) that is still in memory
-091   * @param onlyIfMoreRecent a flag that 
marks whether to update the sequence id no matter what or
-092   *  only if it is 
greater than the previous sequence id
-093   */
-094  public abstract void 
updateLowestUnflushedSequenceIdInWAL(boolean onlyIfMoreRecent);
-095
-096  @Override
-097  public void add(Iterable 
cells, MemStoreSizing memstoreSizing) {
-098for (Cell cell : cells) {
-099  add(cell, memstoreSizing);
-100}
-101  }
-102
-103  @Override
-104  public void add(Cell cell, 
MemStoreSizing memstoreSizing) {
-105doAddOrUpsert(cell, 0, 
memstoreSizing, true);  }
-106
-107  /*
-108   * Inserts the specified Cell into 
MemStore and deletes any existing
-109   * versions of the same 
row/family/qualifier as the specified Cell.
-110   * 

-111 * First, the specified Cell is inserted into the Memstore. -112 *

-113 * If there are any existing Cell in this MemStore with the same row, -114 * family, and qualifier, they are removed. -115 *

-116 * Callers must hold the read lock. -117 * -118 * @param cell the cell to be updated -119 * @param readpoint readpoint below which we can safely remove duplicate KVs -120 * @param memstoreSizing object to accumulate changed size -121 */ -122 private void upsert(Cell cell, long readpoint, MemStoreSizing memstoreSizing) { -123doAddOrUpsert(cell, readpoint, memstoreSizing, false); -124 } -125 -126 private void doAddOrUpsert(Cell cell, long readpoint, MemStoreSizing memstoreSizing, boolean -127 doAdd) { -128MutableSegment currentActive; -129boolean succ = false; -130while (!succ) { -131 currentActive = getActive(); -132 succ = preUpdate(currentActive, cell, memstoreSizing); -133 if (succ) { -134if(doAdd) { -135 doAdd(currentActive, cell, memstoreSizing); -136} else { -137 doUpsert(currentActive, cell, readpoint, memstoreSizing); -138} -139postUpdate(currentActive); -140 } -141} -142 } -143 -144 private void doAdd(MutableSegment currentActive, Cell cell, MemStoreSizing memstoreSizing) { -145Cell toAdd = maybeCloneWithAllocator(currentActive, cell, false); -146boolean mslabUsed = (toAdd != cell); -147// This cell data is backed by the same byte[] where we read request in RPC(See -148// HBASE-15180). By default MSLAB is ON and we might have copied cell to MSLAB area. If -149// not we must do below deep copy. Or else we will keep referring to the bigger chunk of -150// memory and prevent it from getting GCed. -151// Copy to MSLAB would not have happened if -152// 1. MSLAB is turned OFF. See "hbase.hregion.memstore.mslab.enable


[26/51] [partial] hbase-site git commit: Published site at 6a5b4f2a5c188f8eef4f2250b8b7db7dd1e750e4.

2018-08-23 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.html
index db8431b..a8cb7c4 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.html
@@ -885,7766 +885,7797 @@
 877   * @return What the next sequence 
(edit) id should be.
 878   * @throws IOException e
 879   */
-880  private long initialize(final 
CancelableProgressable reporter) throws IOException {
-881
-882//Refuse to open the region if there 
is no column family in the table
-883if 
(htableDescriptor.getColumnFamilyCount() == 0) {
-884  throw new 
DoNotRetryIOException("Table " + 
htableDescriptor.getTableName().getNameAsString()+
-885  " should have at least one 
column family.");
-886}
-887
-888MonitoredTask status = 
TaskMonitor.get().createStatus("Initializing region " + this);
-889long nextSeqId = -1;
-890try {
-891  nextSeqId = 
initializeRegionInternals(reporter, status);
-892  return nextSeqId;
-893} finally {
-894  // nextSeqid will be -1 if the 
initialization fails.
-895  // At least it will be 0 
otherwise.
-896  if (nextSeqId == -1) {
-897status.abort("Exception during 
region " + getRegionInfo().getRegionNameAsString() +
-898  " initialization.");
-899  }
-900}
-901  }
-902
-903  private long 
initializeRegionInternals(final CancelableProgressable reporter,
-904  final MonitoredTask status) throws 
IOException {
-905if (coprocessorHost != null) {
-906  status.setStatus("Running 
coprocessor pre-open hook");
-907  coprocessorHost.preOpen();
-908}
-909
-910// Write HRI to a file in case we 
need to recover hbase:meta
-911// Only the primary replica should 
write .regioninfo
-912if 
(this.getRegionInfo().getReplicaId() == RegionInfo.DEFAULT_REPLICA_ID) {
-913  status.setStatus("Writing region 
info on filesystem");
-914  fs.checkRegionInfoOnFilesystem();
-915}
-916
-917// Initialize all the HStores
-918status.setStatus("Initializing all 
the Stores");
-919long maxSeqId = 
initializeStores(reporter, status);
-920this.mvcc.advanceTo(maxSeqId);
-921if 
(ServerRegionReplicaUtil.shouldReplayRecoveredEdits(this)) {
-922  Collection stores = 
this.stores.values();
-923  try {
-924// update the stores that we are 
replaying
-925LOG.debug("replaying wal for " + 
this.getRegionInfo().getEncodedName());
-926
stores.forEach(HStore::startReplayingFromWAL);
-927// Recover any edits if 
available.
-928maxSeqId = Math.max(maxSeqId,
-929  
replayRecoveredEditsIfAny(this.fs.getRegionDir(), maxSeqIdInStores, reporter, 
status));
-930// Make sure mvcc is up to max.
-931this.mvcc.advanceTo(maxSeqId);
-932  } finally {
-933LOG.debug("stopping wal replay 
for " + this.getRegionInfo().getEncodedName());
-934// update the stores that we are 
done replaying
-935
stores.forEach(HStore::stopReplayingFromWAL);
-936  }
-937}
-938this.lastReplayedOpenRegionSeqId = 
maxSeqId;
-939
-940
this.writestate.setReadOnly(ServerRegionReplicaUtil.isReadOnly(this));
-941this.writestate.flushRequested = 
false;
-942this.writestate.compacting.set(0);
-943
-944if (this.writestate.writesEnabled) 
{
-945  LOG.debug("Cleaning up temporary 
data for " + this.getRegionInfo().getEncodedName());
-946  // Remove temporary data left over 
from old regions
-947  status.setStatus("Cleaning up 
temporary data from old regions");
-948  fs.cleanupTempDir();
-949}
-950
-951if (this.writestate.writesEnabled) 
{
-952  status.setStatus("Cleaning up 
detritus from prior splits");
-953  // Get rid of any splits or merges 
that were lost in-progress.  Clean out
-954  // these directories here on open.  
We may be opening a region that was
-955  // being split but we crashed in 
the middle of it all.
-956  LOG.debug("Cleaning up detritus for 
" + this.getRegionInfo().getEncodedName());
-957  fs.cleanupAnySplitDetritus();
-958  fs.cleanupMergesDir();
-959}
+880  @VisibleForTesting
+881  long initialize(final 
CancelableProgressable reporter) throws IOException {
+882
+883//Refuse to open the region if there 
is no column family in the table
+884if 
(htableDescriptor.getColumnFamilyCount() == 0) {
+885  throw new 
DoNotRetryIOException("Table " + 
htableDescriptor.getTableName().getNameAsString()+
+886  " should have at least one 
column family.");
+887}
+888
+889MonitoredTask status 

[49/51] [partial] hbase-site git commit: Published site at 6a5b4f2a5c188f8eef4f2250b8b7db7dd1e750e4.

2018-08-23 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/checkstyle.rss
--
diff --git a/checkstyle.rss b/checkstyle.rss
index e038ba0..3604af9 100644
--- a/checkstyle.rss
+++ b/checkstyle.rss
@@ -25,8 +25,8 @@ under the License.
 en-us
 ©2007 - 2018 The Apache Software Foundation
 
-  File: 3712,
- Errors: 15385,
+  File: 3714,
+ Errors: 15386,
  Warnings: 0,
  Infos: 0
   
@@ -8656,6 +8656,20 @@ under the License.
   
   
 
+  http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.master.procedure.TestCreateTableProcedureMuitipleRegions.java";>org/apache/hadoop/hbase/master/procedure/TestCreateTableProcedureMuitipleRegions.java
+
+
+  0
+
+
+  0
+
+
+  0
+
+  
+  
+
   http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.replication.regionserver.TestMetricsReplicationSourceFactoryImpl.java";>org/apache/hadoop/hbase/replication/regionserver/TestMetricsReplicationSourceFactoryImpl.java
 
 
@@ -28704,6 +28718,20 @@ under the License.
   
   
 
+  http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.regionserver.TestRecoveredEditsReplayAndAbort.java";>org/apache/hadoop/hbase/regionserver/TestRecoveredEditsReplayAndAbort.java
+
+
+  0
+
+
+  0
+
+
+  0
+
+  
+  
+
   http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.regionserver.TestBulkLoad.java";>org/apache/hadoop/hbase/regionserver/TestBulkLoad.java
 
 
@@ -30547,7 +30575,7 @@ under the License.
   0
 
 
-  0
+  1
 
   
   

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/coc.html
--
diff --git a/coc.html b/coc.html
index 7aee9af..0348c50 100644
--- a/coc.html
+++ b/coc.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase – 
   Code of Conduct Policy
@@ -375,7 +375,7 @@ email to mailto:priv...@hbase.apache.org";>the priv
 https://www.apache.org/";>The Apache Software 
Foundation.
 All rights reserved.  
 
-  Last Published: 
2018-08-22
+  Last Published: 
2018-08-23
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/dependencies.html
--
diff --git a/dependencies.html b/dependencies.html
index ba14184..1d56516 100644
--- a/dependencies.html
+++ b/dependencies.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase – Project Dependencies
 
@@ -440,7 +440,7 @@
 https://www.apache.org/";>The Apache Software 
Foundation.
 All rights reserved.  
 
-  Last Published: 
2018-08-22
+  Last Published: 
2018-08-23
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/dependency-convergence.html
--
diff --git a/dependency-convergence.html b/dependency-convergence.html
index 58a1363..70dce3f 100644
--- a/dependency-convergence.html
+++ b/dependency-convergence.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase – Reactor Dependency Convergence
 
@@ -890,7 +890,7 @@
 https://www.apache.org/";>The Apache Software 
Foundation.
 All rights reserved.  
 
-  Last Published: 
2018-08-22
+  Last Published: 
2018-08-23
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/dependency-info.html
--
diff --git a/dependency-info.html b/dependency-info.html
index 22b1fd0..5161863 100644
--- a/dependency-info.html
+++ b/dependency-info.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase – Dependency Information
 
@@ -313,7 +313,7 @@
 https://www.apache.org/";>The Apache Software 
Foundation.
 All rights reser

[11/51] [partial] hbase-site git commit: Published site at 6a5b4f2a5c188f8eef4f2250b8b7db7dd1e750e4.

2018-08-23 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/handler/OpenRegionHandler.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/handler/OpenRegionHandler.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/handler/OpenRegionHandler.html
index 2709ea3..4a11f27 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/handler/OpenRegionHandler.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/handler/OpenRegionHandler.html
@@ -37,309 +37,299 @@
 029import 
org.apache.hadoop.hbase.executor.EventType;
 030import 
org.apache.hadoop.hbase.regionserver.HRegion;
 031import 
org.apache.hadoop.hbase.regionserver.Region;
-032import 
org.apache.hadoop.hbase.regionserver.RegionServerAccounting;
-033import 
org.apache.hadoop.hbase.regionserver.RegionServerServices;
-034import 
org.apache.hadoop.hbase.regionserver.RegionServerServices.PostOpenDeployContext;
-035import 
org.apache.hadoop.hbase.regionserver.RegionServerServices.RegionStateTransitionContext;
-036import 
org.apache.hadoop.hbase.util.CancelableProgressable;
-037import 
org.apache.yetus.audience.InterfaceAudience;
-038import org.slf4j.Logger;
-039import org.slf4j.LoggerFactory;
-040import 
org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionStateTransition.TransitionCode;
-041/**
-042 * Handles opening of a region on a 
region server.
-043 * 

-044 * This is executed after receiving an OPEN RPC from the master or client. -045 */ -046@InterfaceAudience.Private -047public class OpenRegionHandler extends EventHandler { -048 private static final Logger LOG = LoggerFactory.getLogger(OpenRegionHandler.class); -049 -050 protected final RegionServerServices rsServices; -051 -052 private final RegionInfo regionInfo; -053 private final TableDescriptor htd; -054 private final long masterSystemTime; -055 -056 public OpenRegionHandler(final Server server, -057 final RegionServerServices rsServices, RegionInfo regionInfo, -058 TableDescriptor htd, long masterSystemTime) { -059this(server, rsServices, regionInfo, htd, masterSystemTime, EventType.M_RS_OPEN_REGION); -060 } -061 -062 protected OpenRegionHandler(final Server server, -063 final RegionServerServices rsServices, final RegionInfo regionInfo, -064 final TableDescriptor htd, long masterSystemTime, EventType eventType) { -065super(server, eventType); -066this.rsServices = rsServices; -067this.regionInfo = regionInfo; -068this.htd = htd; -069this.masterSystemTime = masterSystemTime; -070 } -071 -072 public RegionInfo getRegionInfo() { -073return regionInfo; -074 } -075 -076 @Override -077 public void process() throws IOException { -078boolean openSuccessful = false; -079final String regionName = regionInfo.getRegionNameAsString(); -080HRegion region = null; -081 -082try { -083 if (this.server.isStopped() || this.rsServices.isStopping()) { -084return; -085 } -086 final String encodedName = regionInfo.getEncodedName(); -087 -088 // 2 different difficult situations can occur -089 // 1) The opening was cancelled. This is an expected situation -090 // 2) The region is now marked as online while we're suppose to open. This would be a bug. -091 -092 // Check that this region is not already online -093 if (this.rsServices.getRegion(encodedName) != null) { -094LOG.error("Region " + encodedName + -095" was already online when we started processing the opening. " + -096"Marking this new attempt as failed"); -097return; -098 } -099 -100 // Check that we're still supposed to open the region. -101 // If fails, just return. Someone stole the region from under us. -102 if (!isRegionStillOpening()){ -103LOG.error("Region " + encodedName + " opening cancelled"); -104return; -105 } -106 -107 // Open region. After a successful open, failures in subsequent -108 // processing needs to do a close as part of cleanup. -109 region = openRegion(); -110 if (region == null) { -111return; -112 } -113 -114 if (!updateMeta(region, masterSystemTime) || this.server.isStopped() || -115 this.rsServices.isStopping()) { -116return; -117 } -118 -119 if (!isRegionStillOpening()) { -120return; -121 } -122 -123 // Successful region open, and add it to MutableOnlineRegions -124 this.rsServices.addRegion(region); -125 openSuccessful = true; -126 -127 // Done! Successful region open -128 LOG.debug("Opened " + regionName + " on " + this.server.getServerName()); -129} finally { -130 // Do all clean up here -131 if (!openSuccess


hbase git commit: HBASE-19008 Add missing equals or hashCode method(s) to stock Filter implementations

2018-08-23 Thread reidchan
Repository: hbase
Updated Branches:
  refs/heads/master 6a5b4f2a5 -> 72b36e1d9


HBASE-19008 Add missing equals or hashCode method(s) to stock Filter 
implementations

Signed-off-by: Reid Chan 
Signed-off-by: Ted Yu 


Project: http://git-wip-us.apache.org/repos/asf/hbase/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/72b36e1d
Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/72b36e1d
Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/72b36e1d

Branch: refs/heads/master
Commit: 72b36e1d9aa6b10508b6d87a02fbf7d4ead2d2a2
Parents: 6a5b4f2
Author: liubangchen 
Authored: Thu Aug 23 11:19:17 2018 +0800
Committer: Reid Chan 
Committed: Fri Aug 24 00:22:22 2018 +0800

--
 .../hbase/filter/ColumnCountGetFilter.java  | 15 ++
 .../hbase/filter/ColumnPaginationFilter.java| 15 ++
 .../hadoop/hbase/filter/ColumnPrefixFilter.java | 15 ++
 .../hadoop/hbase/filter/ColumnRangeFilter.java  | 16 ++
 .../hadoop/hbase/filter/ColumnValueFilter.java  | 16 ++
 .../hadoop/hbase/filter/CompareFilter.java  | 15 ++
 .../hbase/filter/DependentColumnFilter.java | 16 ++
 .../hadoop/hbase/filter/FamilyFilter.java   | 15 ++
 .../apache/hadoop/hbase/filter/FilterList.java  | 14 ++
 .../hadoop/hbase/filter/FilterListWithAND.java  | 19 +++
 .../hadoop/hbase/filter/FilterListWithOR.java   | 23 +
 .../FirstKeyValueMatchingQualifiersFilter.java  | 15 ++
 .../hadoop/hbase/filter/FuzzyRowFilter.java | 15 ++
 .../hbase/filter/InclusiveStopFilter.java   | 15 ++
 .../hadoop/hbase/filter/KeyOnlyFilter.java  | 16 ++
 .../hbase/filter/MultiRowRangeFilter.java   | 38 ++
 .../filter/MultipleColumnPrefixFilter.java  | 16 ++
 .../apache/hadoop/hbase/filter/PageFilter.java  | 15 ++
 .../hadoop/hbase/filter/PrefixFilter.java   | 15 ++
 .../hadoop/hbase/filter/QualifierFilter.java| 15 ++
 .../hadoop/hbase/filter/RandomRowFilter.java| 15 ++
 .../apache/hadoop/hbase/filter/RowFilter.java   | 15 ++
 .../hbase/filter/SingleColumnValueFilter.java   | 16 ++
 .../apache/hadoop/hbase/filter/SkipFilter.java  | 15 ++
 .../hadoop/hbase/filter/TimestampsFilter.java   | 15 ++
 .../apache/hadoop/hbase/filter/ValueFilter.java | 15 ++
 .../hadoop/hbase/filter/WhileMatchFilter.java   | 15 ++
 .../security/access/AccessControlFilter.java| 23 +
 .../visibility/VisibilityController.java| 17 +++
 .../visibility/VisibilityLabelFilter.java   | 18 +++
 .../hbase/client/ColumnCountOnRowFilter.java| 18 +++
 .../apache/hadoop/hbase/filter/TestFilter.java  |  1 +
 .../hadoop/hbase/filter/TestFilterList.java | 53 
 .../hbase/spark/SparkSQLPushDownFilter.java | 32 
 34 files changed, 607 insertions(+)
--


http://git-wip-us.apache.org/repos/asf/hbase/blob/72b36e1d/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnCountGetFilter.java
--
diff --git 
a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnCountGetFilter.java
 
b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnCountGetFilter.java
index 3aaac36..3cf6675 100644
--- 
a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnCountGetFilter.java
+++ 
b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnCountGetFilter.java
@@ -21,6 +21,7 @@ package org.apache.hadoop.hbase.filter;
 
 import java.io.IOException;
 import java.util.ArrayList;
+import java.util.Objects;
 
 import org.apache.hadoop.hbase.Cell;
 import org.apache.yetus.audience.InterfaceAudience;
@@ -132,4 +133,18 @@ public class ColumnCountGetFilter extends FilterBase {
   public String toString() {
 return this.getClass().getSimpleName() + " " + this.limit;
   }
+
+  @Override
+  public boolean equals(Object obj) {
+if (obj == null || (!(obj instanceof ColumnCountGetFilter))) {
+  return false;
+}
+ColumnCountGetFilter f = (ColumnCountGetFilter) obj;
+return this.areSerializedFieldsEqual(f);
+  }
+
+  @Override
+  public int hashCode() {
+return Objects.hash(this.limit);
+  }
 }

http://git-wip-us.apache.org/repos/asf/hbase/blob/72b36e1d/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnPaginationFilter.java
--
diff --git 
a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnPaginationFilter.java
 
b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnPaginationFilter.java
index c90047d..4f592e9 100644
--- 
a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnPaginationFilter.java
+++ 
b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnPaginationFilter.java
@@ -20,6 +20,7 @@ p

hbase git commit: HBASE-19008 Add missing equals or hashCode method(s) to stock Filter implementations

2018-08-23 Thread reidchan
Repository: hbase
Updated Branches:
  refs/heads/branch-2 cf4d23f8d -> a7a281a64


HBASE-19008 Add missing equals or hashCode method(s) to stock Filter 
implementations

Signed-off-by: Reid Chan 
Signed-off-by: Ted Yu 


Project: http://git-wip-us.apache.org/repos/asf/hbase/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/a7a281a6
Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/a7a281a6
Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/a7a281a6

Branch: refs/heads/branch-2
Commit: a7a281a64404c79767923821e8bd02639cb2f6e6
Parents: cf4d23f
Author: liubangchen 
Authored: Thu Aug 23 11:19:17 2018 +0800
Committer: Reid Chan 
Committed: Fri Aug 24 00:39:35 2018 +0800

--
 .../hbase/filter/ColumnCountGetFilter.java  | 15 ++
 .../hbase/filter/ColumnPaginationFilter.java| 15 ++
 .../hadoop/hbase/filter/ColumnPrefixFilter.java | 15 ++
 .../hadoop/hbase/filter/ColumnRangeFilter.java  | 16 ++
 .../hadoop/hbase/filter/ColumnValueFilter.java  | 16 ++
 .../hadoop/hbase/filter/CompareFilter.java  | 15 ++
 .../hbase/filter/DependentColumnFilter.java | 16 ++
 .../hadoop/hbase/filter/FamilyFilter.java   | 15 ++
 .../apache/hadoop/hbase/filter/FilterList.java  | 14 ++
 .../hadoop/hbase/filter/FilterListWithAND.java  | 19 +++
 .../hadoop/hbase/filter/FilterListWithOR.java   | 23 +
 .../FirstKeyValueMatchingQualifiersFilter.java  | 15 ++
 .../hadoop/hbase/filter/FuzzyRowFilter.java | 15 ++
 .../hbase/filter/InclusiveStopFilter.java   | 15 ++
 .../hadoop/hbase/filter/KeyOnlyFilter.java  | 16 ++
 .../hbase/filter/MultiRowRangeFilter.java   | 38 ++
 .../filter/MultipleColumnPrefixFilter.java  | 16 ++
 .../apache/hadoop/hbase/filter/PageFilter.java  | 15 ++
 .../hadoop/hbase/filter/PrefixFilter.java   | 15 ++
 .../hadoop/hbase/filter/QualifierFilter.java| 15 ++
 .../hadoop/hbase/filter/RandomRowFilter.java| 15 ++
 .../apache/hadoop/hbase/filter/RowFilter.java   | 15 ++
 .../hbase/filter/SingleColumnValueFilter.java   | 16 ++
 .../apache/hadoop/hbase/filter/SkipFilter.java  | 15 ++
 .../hadoop/hbase/filter/TimestampsFilter.java   | 15 ++
 .../apache/hadoop/hbase/filter/ValueFilter.java | 15 ++
 .../hadoop/hbase/filter/WhileMatchFilter.java   | 15 ++
 .../security/access/AccessControlFilter.java| 23 +
 .../visibility/VisibilityController.java| 17 +++
 .../visibility/VisibilityLabelFilter.java   | 18 +++
 .../hbase/client/ColumnCountOnRowFilter.java| 18 +++
 .../apache/hadoop/hbase/filter/TestFilter.java  |  1 +
 .../hadoop/hbase/filter/TestFilterList.java | 53 
 33 files changed, 575 insertions(+)
--


http://git-wip-us.apache.org/repos/asf/hbase/blob/a7a281a6/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnCountGetFilter.java
--
diff --git 
a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnCountGetFilter.java
 
b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnCountGetFilter.java
index 3aaac36..3cf6675 100644
--- 
a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnCountGetFilter.java
+++ 
b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnCountGetFilter.java
@@ -21,6 +21,7 @@ package org.apache.hadoop.hbase.filter;
 
 import java.io.IOException;
 import java.util.ArrayList;
+import java.util.Objects;
 
 import org.apache.hadoop.hbase.Cell;
 import org.apache.yetus.audience.InterfaceAudience;
@@ -132,4 +133,18 @@ public class ColumnCountGetFilter extends FilterBase {
   public String toString() {
 return this.getClass().getSimpleName() + " " + this.limit;
   }
+
+  @Override
+  public boolean equals(Object obj) {
+if (obj == null || (!(obj instanceof ColumnCountGetFilter))) {
+  return false;
+}
+ColumnCountGetFilter f = (ColumnCountGetFilter) obj;
+return this.areSerializedFieldsEqual(f);
+  }
+
+  @Override
+  public int hashCode() {
+return Objects.hash(this.limit);
+  }
 }

http://git-wip-us.apache.org/repos/asf/hbase/blob/a7a281a6/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnPaginationFilter.java
--
diff --git 
a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnPaginationFilter.java
 
b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnPaginationFilter.java
index c90047d..4f592e9 100644
--- 
a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnPaginationFilter.java
+++ 
b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/ColumnPaginationFilter.java
@@ -20,6 +20,7 @@ package org.apache.hadoop.hbase.filter;
 
 import java.io.IOExce

hbase git commit: Update CHANGES.txt for 1.4.7 RC0 (again)

2018-08-23 Thread apurtell
Repository: hbase
Updated Branches:
  refs/heads/branch-1.4 9354d6b45 -> 70c6cba24


Update CHANGES.txt for 1.4.7 RC0 (again)


Project: http://git-wip-us.apache.org/repos/asf/hbase/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/70c6cba2
Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/70c6cba2
Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/70c6cba2

Branch: refs/heads/branch-1.4
Commit: 70c6cba24196f352ae4e156c920fb691a9ff54f2
Parents: 9354d6b
Author: Andrew Purtell 
Authored: Thu Aug 23 11:46:57 2018 -0700
Committer: Andrew Purtell 
Committed: Thu Aug 23 11:46:57 2018 -0700

--
 CHANGES.txt | 3 +++
 1 file changed, 3 insertions(+)
--


http://git-wip-us.apache.org/repos/asf/hbase/blob/70c6cba2/CHANGES.txt
--
diff --git a/CHANGES.txt b/CHANGES.txt
index 46f9f97..04e70d6 100644
--- a/CHANGES.txt
+++ b/CHANGES.txt
@@ -23,6 +23,9 @@ Release Notes - HBase - Version 1.4.7 8/31/2018
 * [HBASE-20930] - MetaScanner.metaScan should use passed variable for meta 
table name rather than TableName.META_TABLE_NAME
 * [HBASE-20935] - HStore.removeCompactedFiles should log in case it is 
unable to delete a file
 
+** Test
+* [HBASE-21076] - refactor TestTableResource to ask for a multi-region 
table instead of relying on a split operation
+
 
 Release Notes - HBase - Version 1.4.6 7/30/2018
 



[hbase] Git Push Summary

2018-08-23 Thread apurtell
Repository: hbase
Updated Tags:  refs/tags/1.4.7RC0 252223dc7 -> 01dda5b5a


hbase git commit: HBASE-21097 Flush pressure assertion may fail in testFlushThroughputTuning

2018-08-23 Thread tedyu
Repository: hbase
Updated Branches:
  refs/heads/master 72b36e1d9 -> 780670ede


HBASE-21097 Flush pressure assertion may fail in testFlushThroughputTuning


Project: http://git-wip-us.apache.org/repos/asf/hbase/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/780670ed
Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/780670ed
Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/780670ed

Branch: refs/heads/master
Commit: 780670ede18ea2c4ff08ef703e2fde35536909e6
Parents: 72b36e1
Author: tedyu 
Authored: Thu Aug 23 11:48:27 2018 -0700
Committer: tedyu 
Committed: Thu Aug 23 11:48:27 2018 -0700

--
 .../regionserver/throttle/TestFlushWithThroughputController.java   | 2 +-
 1 file changed, 1 insertion(+), 1 deletion(-)
--


http://git-wip-us.apache.org/repos/asf/hbase/blob/780670ed/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/throttle/TestFlushWithThroughputController.java
--
diff --git 
a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/throttle/TestFlushWithThroughputController.java
 
b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/throttle/TestFlushWithThroughputController.java
index 1c39646..61f9cd4 100644
--- 
a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/throttle/TestFlushWithThroughputController.java
+++ 
b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/throttle/TestFlushWithThroughputController.java
@@ -66,7 +66,7 @@ public class TestFlushWithThroughputController {
 
   private static final Logger LOG =
   LoggerFactory.getLogger(TestFlushWithThroughputController.class);
-  private static final double EPSILON = 1E-6;
+  private static final double EPSILON = 1.3E-6;
 
   private HBaseTestingUtility hbtu;
   @Rule public TestName testName = new TestName();



hbase git commit: HBASE-21097 Flush pressure assertion may fail in testFlushThroughputTuning

2018-08-23 Thread tedyu
Repository: hbase
Updated Branches:
  refs/heads/branch-2 a7a281a64 -> 87f9b4acc


HBASE-21097 Flush pressure assertion may fail in testFlushThroughputTuning


Project: http://git-wip-us.apache.org/repos/asf/hbase/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/87f9b4ac
Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/87f9b4ac
Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/87f9b4ac

Branch: refs/heads/branch-2
Commit: 87f9b4accda8633aafd60d421657a625a8a16a85
Parents: a7a281a
Author: tedyu 
Authored: Thu Aug 23 11:49:47 2018 -0700
Committer: tedyu 
Committed: Thu Aug 23 11:49:47 2018 -0700

--
 .../regionserver/throttle/TestFlushWithThroughputController.java   | 2 +-
 1 file changed, 1 insertion(+), 1 deletion(-)
--


http://git-wip-us.apache.org/repos/asf/hbase/blob/87f9b4ac/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/throttle/TestFlushWithThroughputController.java
--
diff --git 
a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/throttle/TestFlushWithThroughputController.java
 
b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/throttle/TestFlushWithThroughputController.java
index 1c39646..61f9cd4 100644
--- 
a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/throttle/TestFlushWithThroughputController.java
+++ 
b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/throttle/TestFlushWithThroughputController.java
@@ -66,7 +66,7 @@ public class TestFlushWithThroughputController {
 
   private static final Logger LOG =
   LoggerFactory.getLogger(TestFlushWithThroughputController.class);
-  private static final double EPSILON = 1E-6;
+  private static final double EPSILON = 1.3E-6;
 
   private HBaseTestingUtility hbtu;
   @Rule public TestName testName = new TestName();



hbase git commit: HBASE-21095 The timeout retry logic for several procedures are broken after master restarts

2018-08-23 Thread zhangduo
Repository: hbase
Updated Branches:
  refs/heads/master 780670ede -> aac1a7014


HBASE-21095 The timeout retry logic for several procedures are broken after 
master restarts


Project: http://git-wip-us.apache.org/repos/asf/hbase/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/aac1a701
Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/aac1a701
Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/aac1a701

Branch: refs/heads/master
Commit: aac1a70147ec950c98db5f31d8906ce54547523a
Parents: 780670e
Author: zhangduo 
Authored: Thu Aug 23 22:16:13 2018 +0800
Committer: Duo Zhang 
Committed: Fri Aug 24 10:07:31 2018 +0800

--
 .../master/assignment/AssignmentManager.java| 67 +++-
 .../assignment/TransitRegionStateProcedure.java | 12 ++--
 .../master/procedure/ServerCrashProcedure.java  | 15 +++--
 .../assignment/TestCloseRegionWhileRSCrash.java | 11 +++-
 4 files changed, 75 insertions(+), 30 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/hbase/blob/aac1a701/hbase-server/src/main/java/org/apache/hadoop/hbase/master/assignment/AssignmentManager.java
--
diff --git 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/assignment/AssignmentManager.java
 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/assignment/AssignmentManager.java
index 9b020c8..a91f8e4 100644
--- 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/assignment/AssignmentManager.java
+++ 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/assignment/AssignmentManager.java
@@ -1414,12 +1414,25 @@ public class AssignmentManager implements 
ServerListener {
   //  Region Status update
   //  Should only be called in TransitRegionStateProcedure
   // 

+  private void transitStateAndUpdate(RegionStateNode regionNode, 
RegionState.State newState,
+  RegionState.State... expectedStates) throws IOException {
+RegionState.State state = regionNode.getState();
+regionNode.transitionState(newState, expectedStates);
+boolean succ = false;
+try {
+  regionStateStore.updateRegionLocation(regionNode);
+  succ = true;
+} finally {
+  if (!succ) {
+// revert
+regionNode.setState(state);
+  }
+}
+  }
 
   // should be called within the synchronized block of RegionStateNode
   void regionOpening(RegionStateNode regionNode) throws IOException {
-regionNode.transitionState(State.OPENING, 
RegionStates.STATES_EXPECTED_ON_OPEN);
-regionStateStore.updateRegionLocation(regionNode);
-
+transitStateAndUpdate(regionNode, State.OPENING, 
RegionStates.STATES_EXPECTED_ON_OPEN);
 regionStates.addRegionToServer(regionNode);
 // update the operation count metrics
 metrics.incrementOperationCounter();
@@ -1429,23 +1442,33 @@ public class AssignmentManager implements 
ServerListener {
   // The parameter 'giveUp' means whether we will try to open the region 
again, if it is true, then
   // we will persist the FAILED_OPEN state into hbase:meta.
   void regionFailedOpen(RegionStateNode regionNode, boolean giveUp) throws 
IOException {
-if (regionNode.getRegionLocation() != null) {
-  regionStates.removeRegionFromServer(regionNode.getRegionLocation(), 
regionNode);
-}
+RegionState.State state = regionNode.getState();
+ServerName regionLocation = regionNode.getRegionLocation();
 if (giveUp) {
   regionNode.setState(State.FAILED_OPEN);
   regionNode.setRegionLocation(null);
-  regionStateStore.updateRegionLocation(regionNode);
+  boolean succ = false;
+  try {
+regionStateStore.updateRegionLocation(regionNode);
+succ = true;
+  } finally {
+if (!succ) {
+  // revert
+  regionNode.setState(state);
+  regionNode.setRegionLocation(regionLocation);
+}
+  }
+}
+if (regionLocation != null) {
+  regionStates.removeRegionFromServer(regionLocation, regionNode);
 }
   }
 
   // should be called within the synchronized block of RegionStateNode
   void regionOpened(RegionStateNode regionNode) throws IOException {
-regionNode.transitionState(State.OPEN, 
RegionStates.STATES_EXPECTED_ON_OPEN);
 // TODO: OPENING Updates hbase:meta too... we need to do both here and 
there?
 // That is a lot of hbase:meta writing.
-regionStateStore.updateRegionLocation(regionNode);
-
+transitStateAndUpdate(regionNode, State.OPEN, 
RegionStates.STATES_EXPECTED_ON_OPEN);
 RegionInfo hri = regionNode.getRegionInfo();
 if (isMetaRegion(hri)) {
   // Usually we'd set a table ENABLED at this stage but hbase:meta is 
ALWAYs enabled, it
@@ -1460,7 +1483,7 @@ public class AssignmentManager implements ServerList

hbase git commit: HBASE-20193 Move TestCreateTableProcedure.testMRegions to a separated file

2018-08-23 Thread zhangduo
Repository: hbase
Updated Branches:
  refs/heads/branch-2 87f9b4acc -> b318311df


HBASE-20193 Move TestCreateTableProcedure.testMRegions to a separated file


Project: http://git-wip-us.apache.org/repos/asf/hbase/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/b318311d
Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/b318311d
Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/b318311d

Branch: refs/heads/branch-2
Commit: b318311dfd332160487692bd416fd03da4b35473
Parents: 87f9b4a
Author: zhangduo 
Authored: Wed Aug 22 22:10:58 2018 +0800
Committer: Duo Zhang 
Committed: Fri Aug 24 10:09:23 2018 +0800

--
 .../procedure/TestCreateTableProcedure.java | 25 +---
 ...TestCreateTableProcedureMuitipleRegions.java | 66 
 2 files changed, 69 insertions(+), 22 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/hbase/blob/b318311d/hbase-server/src/test/java/org/apache/hadoop/hbase/master/procedure/TestCreateTableProcedure.java
--
diff --git 
a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/procedure/TestCreateTableProcedure.java
 
b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/procedure/TestCreateTableProcedure.java
index c45cb98..eed1e41 100644
--- 
a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/procedure/TestCreateTableProcedure.java
+++ 
b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/procedure/TestCreateTableProcedure.java
@@ -22,7 +22,6 @@ import static org.junit.Assert.assertTrue;
 import static org.junit.Assert.fail;
 
 import java.io.IOException;
-
 import org.apache.hadoop.conf.Configuration;
 import org.apache.hadoop.fs.FileSystem;
 import org.apache.hadoop.fs.Path;
@@ -49,25 +48,21 @@ import org.junit.Rule;
 import org.junit.Test;
 import org.junit.experimental.categories.Category;
 import org.junit.rules.TestName;
-import org.slf4j.Logger;
-import org.slf4j.LoggerFactory;
 
 import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos;
 
-
 @Category({MasterTests.class, MediumTests.class})
 public class TestCreateTableProcedure extends TestTableDDLProcedureBase {
 
   @ClassRule
   public static final HBaseClassTestRule CLASS_RULE =
-  HBaseClassTestRule.forClass(TestCreateTableProcedure.class);
-
-  private static final Logger LOG = 
LoggerFactory.getLogger(TestCreateTableProcedure.class);
+HBaseClassTestRule.forClass(TestCreateTableProcedure.class);
 
   private static final String F1 = "f1";
   private static final String F2 = "f2";
 
-  @Rule public TestName name = new TestName();
+  @Rule
+  public TestName name = new TestName();
 
   @Test
   public void testSimpleCreate() throws Exception {
@@ -202,20 +197,6 @@ public class TestCreateTableProcedure extends 
TestTableDDLProcedureBase {
 testSimpleCreate(tableName, splitKeys);
   }
 
-  @Test
-  public void testMRegions() throws Exception {
-final byte[][] splitKeys = new byte[500][];
-for (int i = 0; i < splitKeys.length; ++i) {
-  splitKeys[i] = Bytes.toBytes(String.format("%08d", i));
-}
-
-final TableDescriptor htd = MasterProcedureTestingUtility.createHTD(
-  TableName.valueOf("TestMRegions"), F1, F2);
-UTIL.getAdmin().createTableAsync(htd, splitKeys)
-  .get(10, java.util.concurrent.TimeUnit.HOURS);
-LOG.info("TABLE CREATED");
-  }
-
   public static class CreateTableProcedureOnHDFSFailure extends 
CreateTableProcedure {
 private boolean failOnce = false;
 

http://git-wip-us.apache.org/repos/asf/hbase/blob/b318311d/hbase-server/src/test/java/org/apache/hadoop/hbase/master/procedure/TestCreateTableProcedureMuitipleRegions.java
--
diff --git 
a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/procedure/TestCreateTableProcedureMuitipleRegions.java
 
b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/procedure/TestCreateTableProcedureMuitipleRegions.java
new file mode 100644
index 000..2aff487
--- /dev/null
+++ 
b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/procedure/TestCreateTableProcedureMuitipleRegions.java
@@ -0,0 +1,66 @@
+/**
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRA

hbase git commit: HBASE-20193 Move TestCreateTableProcedure.testMRegions to a separated file

2018-08-23 Thread zhangduo
Repository: hbase
Updated Branches:
  refs/heads/branch-2.1 9a3093a88 -> bf21a9dc3


HBASE-20193 Move TestCreateTableProcedure.testMRegions to a separated file


Project: http://git-wip-us.apache.org/repos/asf/hbase/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/bf21a9dc
Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/bf21a9dc
Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/bf21a9dc

Branch: refs/heads/branch-2.1
Commit: bf21a9dc335bd758ec96b0c67a9e359c2091b7b5
Parents: 9a3093a
Author: zhangduo 
Authored: Wed Aug 22 22:10:58 2018 +0800
Committer: Duo Zhang 
Committed: Fri Aug 24 10:09:31 2018 +0800

--
 .../procedure/TestCreateTableProcedure.java | 25 +---
 ...TestCreateTableProcedureMuitipleRegions.java | 66 
 2 files changed, 69 insertions(+), 22 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/hbase/blob/bf21a9dc/hbase-server/src/test/java/org/apache/hadoop/hbase/master/procedure/TestCreateTableProcedure.java
--
diff --git 
a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/procedure/TestCreateTableProcedure.java
 
b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/procedure/TestCreateTableProcedure.java
index c45cb98..eed1e41 100644
--- 
a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/procedure/TestCreateTableProcedure.java
+++ 
b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/procedure/TestCreateTableProcedure.java
@@ -22,7 +22,6 @@ import static org.junit.Assert.assertTrue;
 import static org.junit.Assert.fail;
 
 import java.io.IOException;
-
 import org.apache.hadoop.conf.Configuration;
 import org.apache.hadoop.fs.FileSystem;
 import org.apache.hadoop.fs.Path;
@@ -49,25 +48,21 @@ import org.junit.Rule;
 import org.junit.Test;
 import org.junit.experimental.categories.Category;
 import org.junit.rules.TestName;
-import org.slf4j.Logger;
-import org.slf4j.LoggerFactory;
 
 import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos;
 
-
 @Category({MasterTests.class, MediumTests.class})
 public class TestCreateTableProcedure extends TestTableDDLProcedureBase {
 
   @ClassRule
   public static final HBaseClassTestRule CLASS_RULE =
-  HBaseClassTestRule.forClass(TestCreateTableProcedure.class);
-
-  private static final Logger LOG = 
LoggerFactory.getLogger(TestCreateTableProcedure.class);
+HBaseClassTestRule.forClass(TestCreateTableProcedure.class);
 
   private static final String F1 = "f1";
   private static final String F2 = "f2";
 
-  @Rule public TestName name = new TestName();
+  @Rule
+  public TestName name = new TestName();
 
   @Test
   public void testSimpleCreate() throws Exception {
@@ -202,20 +197,6 @@ public class TestCreateTableProcedure extends 
TestTableDDLProcedureBase {
 testSimpleCreate(tableName, splitKeys);
   }
 
-  @Test
-  public void testMRegions() throws Exception {
-final byte[][] splitKeys = new byte[500][];
-for (int i = 0; i < splitKeys.length; ++i) {
-  splitKeys[i] = Bytes.toBytes(String.format("%08d", i));
-}
-
-final TableDescriptor htd = MasterProcedureTestingUtility.createHTD(
-  TableName.valueOf("TestMRegions"), F1, F2);
-UTIL.getAdmin().createTableAsync(htd, splitKeys)
-  .get(10, java.util.concurrent.TimeUnit.HOURS);
-LOG.info("TABLE CREATED");
-  }
-
   public static class CreateTableProcedureOnHDFSFailure extends 
CreateTableProcedure {
 private boolean failOnce = false;
 

http://git-wip-us.apache.org/repos/asf/hbase/blob/bf21a9dc/hbase-server/src/test/java/org/apache/hadoop/hbase/master/procedure/TestCreateTableProcedureMuitipleRegions.java
--
diff --git 
a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/procedure/TestCreateTableProcedureMuitipleRegions.java
 
b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/procedure/TestCreateTableProcedureMuitipleRegions.java
new file mode 100644
index 000..2aff487
--- /dev/null
+++ 
b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/procedure/TestCreateTableProcedureMuitipleRegions.java
@@ -0,0 +1,66 @@
+/**
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT W

hbase git commit: HBASE-20193 Move TestCreateTableProcedure.testMRegions to a separated file

2018-08-23 Thread zhangduo
Repository: hbase
Updated Branches:
  refs/heads/branch-2.0 89ec91608 -> 131fe0910


HBASE-20193 Move TestCreateTableProcedure.testMRegions to a separated file


Project: http://git-wip-us.apache.org/repos/asf/hbase/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/131fe091
Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/131fe091
Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/131fe091

Branch: refs/heads/branch-2.0
Commit: 131fe09109f7597ad97462afcef017c74065a15d
Parents: 89ec916
Author: zhangduo 
Authored: Wed Aug 22 22:10:58 2018 +0800
Committer: Duo Zhang 
Committed: Fri Aug 24 10:10:04 2018 +0800

--
 .../procedure/TestCreateTableProcedure.java | 25 +---
 ...TestCreateTableProcedureMuitipleRegions.java | 66 
 2 files changed, 69 insertions(+), 22 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/hbase/blob/131fe091/hbase-server/src/test/java/org/apache/hadoop/hbase/master/procedure/TestCreateTableProcedure.java
--
diff --git 
a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/procedure/TestCreateTableProcedure.java
 
b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/procedure/TestCreateTableProcedure.java
index c45cb98..eed1e41 100644
--- 
a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/procedure/TestCreateTableProcedure.java
+++ 
b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/procedure/TestCreateTableProcedure.java
@@ -22,7 +22,6 @@ import static org.junit.Assert.assertTrue;
 import static org.junit.Assert.fail;
 
 import java.io.IOException;
-
 import org.apache.hadoop.conf.Configuration;
 import org.apache.hadoop.fs.FileSystem;
 import org.apache.hadoop.fs.Path;
@@ -49,25 +48,21 @@ import org.junit.Rule;
 import org.junit.Test;
 import org.junit.experimental.categories.Category;
 import org.junit.rules.TestName;
-import org.slf4j.Logger;
-import org.slf4j.LoggerFactory;
 
 import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos;
 
-
 @Category({MasterTests.class, MediumTests.class})
 public class TestCreateTableProcedure extends TestTableDDLProcedureBase {
 
   @ClassRule
   public static final HBaseClassTestRule CLASS_RULE =
-  HBaseClassTestRule.forClass(TestCreateTableProcedure.class);
-
-  private static final Logger LOG = 
LoggerFactory.getLogger(TestCreateTableProcedure.class);
+HBaseClassTestRule.forClass(TestCreateTableProcedure.class);
 
   private static final String F1 = "f1";
   private static final String F2 = "f2";
 
-  @Rule public TestName name = new TestName();
+  @Rule
+  public TestName name = new TestName();
 
   @Test
   public void testSimpleCreate() throws Exception {
@@ -202,20 +197,6 @@ public class TestCreateTableProcedure extends 
TestTableDDLProcedureBase {
 testSimpleCreate(tableName, splitKeys);
   }
 
-  @Test
-  public void testMRegions() throws Exception {
-final byte[][] splitKeys = new byte[500][];
-for (int i = 0; i < splitKeys.length; ++i) {
-  splitKeys[i] = Bytes.toBytes(String.format("%08d", i));
-}
-
-final TableDescriptor htd = MasterProcedureTestingUtility.createHTD(
-  TableName.valueOf("TestMRegions"), F1, F2);
-UTIL.getAdmin().createTableAsync(htd, splitKeys)
-  .get(10, java.util.concurrent.TimeUnit.HOURS);
-LOG.info("TABLE CREATED");
-  }
-
   public static class CreateTableProcedureOnHDFSFailure extends 
CreateTableProcedure {
 private boolean failOnce = false;
 

http://git-wip-us.apache.org/repos/asf/hbase/blob/131fe091/hbase-server/src/test/java/org/apache/hadoop/hbase/master/procedure/TestCreateTableProcedureMuitipleRegions.java
--
diff --git 
a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/procedure/TestCreateTableProcedureMuitipleRegions.java
 
b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/procedure/TestCreateTableProcedureMuitipleRegions.java
new file mode 100644
index 000..2aff487
--- /dev/null
+++ 
b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/procedure/TestCreateTableProcedureMuitipleRegions.java
@@ -0,0 +1,66 @@
+/**
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT W

hbase git commit: HBASE-21101 Remove the waitUntilAllRegionsAssigned call after split in TestTruncateTableProcedure

2018-08-23 Thread zhangduo
Repository: hbase
Updated Branches:
  refs/heads/branch-2 b318311df -> c33af1e85


HBASE-21101 Remove the waitUntilAllRegionsAssigned call after split in 
TestTruncateTableProcedure


Project: http://git-wip-us.apache.org/repos/asf/hbase/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/c33af1e8
Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/c33af1e8
Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/c33af1e8

Branch: refs/heads/branch-2
Commit: c33af1e854844a58a06612f9676e284d0eaf6ffa
Parents: b318311
Author: Duo Zhang 
Authored: Thu Aug 23 18:04:01 2018 +0800
Committer: Duo Zhang 
Committed: Fri Aug 24 10:35:05 2018 +0800

--
 .../hadoop/hbase/master/procedure/TestTruncateTableProcedure.java   | 1 -
 1 file changed, 1 deletion(-)
--


http://git-wip-us.apache.org/repos/asf/hbase/blob/c33af1e8/hbase-server/src/test/java/org/apache/hadoop/hbase/master/procedure/TestTruncateTableProcedure.java
--
diff --git 
a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/procedure/TestTruncateTableProcedure.java
 
b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/procedure/TestTruncateTableProcedure.java
index 6907383..6854c60 100644
--- 
a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/procedure/TestTruncateTableProcedure.java
+++ 
b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/procedure/TestTruncateTableProcedure.java
@@ -352,7 +352,6 @@ public class TestTruncateTableProcedure extends 
TestTableDDLProcedureBase {
   throws IOException, InterruptedException {
 // split a region
 UTIL.getAdmin().split(tableName, new byte[] { '0' });
-UTIL.waitUntilAllRegionsAssigned(tableName);
 
 // wait until split really happens
 UTIL.waitFor(6,



hbase git commit: HBASE-21101 Remove the waitUntilAllRegionsAssigned call after split in TestTruncateTableProcedure

2018-08-23 Thread zhangduo
Repository: hbase
Updated Branches:
  refs/heads/branch-2.1 bf21a9dc3 -> 8a9acd4d2


HBASE-21101 Remove the waitUntilAllRegionsAssigned call after split in 
TestTruncateTableProcedure


Project: http://git-wip-us.apache.org/repos/asf/hbase/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/8a9acd4d
Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/8a9acd4d
Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/8a9acd4d

Branch: refs/heads/branch-2.1
Commit: 8a9acd4d2a3f8a3c3b2aefa072fb328a66f6eab7
Parents: bf21a9d
Author: Duo Zhang 
Authored: Thu Aug 23 18:04:01 2018 +0800
Committer: Duo Zhang 
Committed: Fri Aug 24 10:35:10 2018 +0800

--
 .../hadoop/hbase/master/procedure/TestTruncateTableProcedure.java   | 1 -
 1 file changed, 1 deletion(-)
--


http://git-wip-us.apache.org/repos/asf/hbase/blob/8a9acd4d/hbase-server/src/test/java/org/apache/hadoop/hbase/master/procedure/TestTruncateTableProcedure.java
--
diff --git 
a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/procedure/TestTruncateTableProcedure.java
 
b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/procedure/TestTruncateTableProcedure.java
index 6907383..6854c60 100644
--- 
a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/procedure/TestTruncateTableProcedure.java
+++ 
b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/procedure/TestTruncateTableProcedure.java
@@ -352,7 +352,6 @@ public class TestTruncateTableProcedure extends 
TestTableDDLProcedureBase {
   throws IOException, InterruptedException {
 // split a region
 UTIL.getAdmin().split(tableName, new byte[] { '0' });
-UTIL.waitUntilAllRegionsAssigned(tableName);
 
 // wait until split really happens
 UTIL.waitFor(6,



hbase git commit: HBASE-21101 Remove the waitUntilAllRegionsAssigned call after split in TestTruncateTableProcedure

2018-08-23 Thread zhangduo
Repository: hbase
Updated Branches:
  refs/heads/branch-2.0 131fe0910 -> ed235c97e


HBASE-21101 Remove the waitUntilAllRegionsAssigned call after split in 
TestTruncateTableProcedure


Project: http://git-wip-us.apache.org/repos/asf/hbase/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/ed235c97
Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/ed235c97
Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/ed235c97

Branch: refs/heads/branch-2.0
Commit: ed235c97ebed368247c473bc82d94834a780fb1b
Parents: 131fe09
Author: Duo Zhang 
Authored: Thu Aug 23 18:04:01 2018 +0800
Committer: Duo Zhang 
Committed: Fri Aug 24 10:35:14 2018 +0800

--
 .../hadoop/hbase/master/procedure/TestTruncateTableProcedure.java   | 1 -
 1 file changed, 1 deletion(-)
--


http://git-wip-us.apache.org/repos/asf/hbase/blob/ed235c97/hbase-server/src/test/java/org/apache/hadoop/hbase/master/procedure/TestTruncateTableProcedure.java
--
diff --git 
a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/procedure/TestTruncateTableProcedure.java
 
b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/procedure/TestTruncateTableProcedure.java
index 21557fb..7fa2a9e 100644
--- 
a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/procedure/TestTruncateTableProcedure.java
+++ 
b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/procedure/TestTruncateTableProcedure.java
@@ -352,7 +352,6 @@ public class TestTruncateTableProcedure extends 
TestTableDDLProcedureBase {
   throws IOException, InterruptedException {
 // split a region
 UTIL.getAdmin().split(tableName, new byte[] { '0' });
-UTIL.waitUntilAllRegionsAssigned(tableName);
 
 // wait until split really happens
 UTIL.waitFor(6,



hbase git commit: HBASE-21017 Add debug log for finding out race where we update region state to OPEN accidentally

2018-08-23 Thread zhangduo
Repository: hbase
Updated Branches:
  refs/heads/master aac1a7014 -> a452487a9


HBASE-21017 Add debug log for finding out race where we update region state to 
OPEN accidentally


Project: http://git-wip-us.apache.org/repos/asf/hbase/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/a452487a
Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/a452487a
Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/a452487a

Branch: refs/heads/master
Commit: a452487a9b82bfd33bc10683c3f8b8ae74d58883
Parents: aac1a70
Author: Duo Zhang 
Authored: Fri Aug 24 11:45:30 2018 +0800
Committer: Duo Zhang 
Committed: Fri Aug 24 11:48:04 2018 +0800

--
 .../apache/hadoop/hbase/master/assignment/RegionStateStore.java| 2 +-
 1 file changed, 1 insertion(+), 1 deletion(-)
--


http://git-wip-us.apache.org/repos/asf/hbase/blob/a452487a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/assignment/RegionStateStore.java
--
diff --git 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/assignment/RegionStateStore.java
 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/assignment/RegionStateStore.java
index 48ec4fb..0e1ce71 100644
--- 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/assignment/RegionStateStore.java
+++ 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/assignment/RegionStateStore.java
@@ -196,7 +196,7 @@ public class RegionStateStore {
 .setType(Cell.Type.Put)
 .setValue(Bytes.toBytes(state.name()))
 .build());
-LOG.info(info.toString());
+LOG.info(info.toString(), new Exception());
 updateRegionLocation(regionInfo, state, put);
   }