http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c0fcd7f3/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.html ---------------------------------------------------------------------- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.html b/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.html index af42219..a9770f8 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.html @@ -61,32 +61,32 @@ <span class="sourceLineNo">053</span>import org.apache.hadoop.conf.Configuration;<a name="line.53"></a> <span class="sourceLineNo">054</span>import org.apache.hadoop.fs.FileSystem;<a name="line.54"></a> <span class="sourceLineNo">055</span>import org.apache.hadoop.fs.Path;<a name="line.55"></a> -<span class="sourceLineNo">056</span>import org.apache.hadoop.hbase.client.TableDescriptor;<a name="line.56"></a> -<span class="sourceLineNo">057</span>import org.apache.hadoop.hbase.ClusterStatus;<a name="line.57"></a> -<span class="sourceLineNo">058</span>import org.apache.hadoop.hbase.CoordinatedStateException;<a name="line.58"></a> -<span class="sourceLineNo">059</span>import org.apache.hadoop.hbase.CoordinatedStateManager;<a name="line.59"></a> -<span class="sourceLineNo">060</span>import org.apache.hadoop.hbase.DoNotRetryIOException;<a name="line.60"></a> -<span class="sourceLineNo">061</span>import org.apache.hadoop.hbase.HBaseIOException;<a name="line.61"></a> -<span class="sourceLineNo">062</span>import org.apache.hadoop.hbase.HBaseInterfaceAudience;<a name="line.62"></a> -<span class="sourceLineNo">063</span>import org.apache.hadoop.hbase.HColumnDescriptor;<a name="line.63"></a> -<span class="sourceLineNo">064</span>import org.apache.hadoop.hbase.HConstants;<a name="line.64"></a> -<span class="sourceLineNo">065</span>import org.apache.hadoop.hbase.HRegionInfo;<a name="line.65"></a> -<span class="sourceLineNo">066</span>import org.apache.hadoop.hbase.HTableDescriptor;<a name="line.66"></a> -<span class="sourceLineNo">067</span>import org.apache.hadoop.hbase.MasterNotRunningException;<a name="line.67"></a> -<span class="sourceLineNo">068</span>import org.apache.hadoop.hbase.MetaTableAccessor;<a name="line.68"></a> -<span class="sourceLineNo">069</span>import org.apache.hadoop.hbase.NamespaceDescriptor;<a name="line.69"></a> -<span class="sourceLineNo">070</span>import org.apache.hadoop.hbase.PleaseHoldException;<a name="line.70"></a> -<span class="sourceLineNo">071</span>import org.apache.hadoop.hbase.ProcedureInfo;<a name="line.71"></a> -<span class="sourceLineNo">072</span>import org.apache.hadoop.hbase.ServerLoad;<a name="line.72"></a> -<span class="sourceLineNo">073</span>import org.apache.hadoop.hbase.ServerName;<a name="line.73"></a> -<span class="sourceLineNo">074</span>import org.apache.hadoop.hbase.TableDescriptors;<a name="line.74"></a> -<span class="sourceLineNo">075</span>import org.apache.hadoop.hbase.TableName;<a name="line.75"></a> -<span class="sourceLineNo">076</span>import org.apache.hadoop.hbase.TableNotDisabledException;<a name="line.76"></a> -<span class="sourceLineNo">077</span>import org.apache.hadoop.hbase.TableNotFoundException;<a name="line.77"></a> -<span class="sourceLineNo">078</span>import org.apache.hadoop.hbase.UnknownRegionException;<a name="line.78"></a> -<span class="sourceLineNo">079</span>import org.apache.hadoop.hbase.classification.InterfaceAudience;<a name="line.79"></a> -<span class="sourceLineNo">080</span>import org.apache.hadoop.hbase.client.MasterSwitchType;<a name="line.80"></a> -<span class="sourceLineNo">081</span>import org.apache.hadoop.hbase.client.Result;<a name="line.81"></a> +<span class="sourceLineNo">056</span>import org.apache.hadoop.hbase.ClusterStatus;<a name="line.56"></a> +<span class="sourceLineNo">057</span>import org.apache.hadoop.hbase.CoordinatedStateException;<a name="line.57"></a> +<span class="sourceLineNo">058</span>import org.apache.hadoop.hbase.CoordinatedStateManager;<a name="line.58"></a> +<span class="sourceLineNo">059</span>import org.apache.hadoop.hbase.DoNotRetryIOException;<a name="line.59"></a> +<span class="sourceLineNo">060</span>import org.apache.hadoop.hbase.HBaseIOException;<a name="line.60"></a> +<span class="sourceLineNo">061</span>import org.apache.hadoop.hbase.HBaseInterfaceAudience;<a name="line.61"></a> +<span class="sourceLineNo">062</span>import org.apache.hadoop.hbase.HColumnDescriptor;<a name="line.62"></a> +<span class="sourceLineNo">063</span>import org.apache.hadoop.hbase.HConstants;<a name="line.63"></a> +<span class="sourceLineNo">064</span>import org.apache.hadoop.hbase.HRegionInfo;<a name="line.64"></a> +<span class="sourceLineNo">065</span>import org.apache.hadoop.hbase.HTableDescriptor;<a name="line.65"></a> +<span class="sourceLineNo">066</span>import org.apache.hadoop.hbase.MasterNotRunningException;<a name="line.66"></a> +<span class="sourceLineNo">067</span>import org.apache.hadoop.hbase.MetaTableAccessor;<a name="line.67"></a> +<span class="sourceLineNo">068</span>import org.apache.hadoop.hbase.NamespaceDescriptor;<a name="line.68"></a> +<span class="sourceLineNo">069</span>import org.apache.hadoop.hbase.PleaseHoldException;<a name="line.69"></a> +<span class="sourceLineNo">070</span>import org.apache.hadoop.hbase.ProcedureInfo;<a name="line.70"></a> +<span class="sourceLineNo">071</span>import org.apache.hadoop.hbase.ServerLoad;<a name="line.71"></a> +<span class="sourceLineNo">072</span>import org.apache.hadoop.hbase.ServerName;<a name="line.72"></a> +<span class="sourceLineNo">073</span>import org.apache.hadoop.hbase.TableDescriptors;<a name="line.73"></a> +<span class="sourceLineNo">074</span>import org.apache.hadoop.hbase.TableName;<a name="line.74"></a> +<span class="sourceLineNo">075</span>import org.apache.hadoop.hbase.TableNotDisabledException;<a name="line.75"></a> +<span class="sourceLineNo">076</span>import org.apache.hadoop.hbase.TableNotFoundException;<a name="line.76"></a> +<span class="sourceLineNo">077</span>import org.apache.hadoop.hbase.UnknownRegionException;<a name="line.77"></a> +<span class="sourceLineNo">078</span>import org.apache.hadoop.hbase.classification.InterfaceAudience;<a name="line.78"></a> +<span class="sourceLineNo">079</span>import org.apache.hadoop.hbase.client.MasterSwitchType;<a name="line.79"></a> +<span class="sourceLineNo">080</span>import org.apache.hadoop.hbase.client.Result;<a name="line.80"></a> +<span class="sourceLineNo">081</span>import org.apache.hadoop.hbase.client.TableDescriptor;<a name="line.81"></a> <span class="sourceLineNo">082</span>import org.apache.hadoop.hbase.client.TableState;<a name="line.82"></a> <span class="sourceLineNo">083</span>import org.apache.hadoop.hbase.coprocessor.BypassCoprocessorException;<a name="line.83"></a> <span class="sourceLineNo">084</span>import org.apache.hadoop.hbase.coprocessor.CoprocessorHost;<a name="line.84"></a> @@ -172,45 +172,45 @@ <span class="sourceLineNo">164</span>import org.apache.hadoop.hbase.replication.regionserver.Replication;<a name="line.164"></a> <span class="sourceLineNo">165</span>import org.apache.hadoop.hbase.security.AccessDeniedException;<a name="line.165"></a> <span class="sourceLineNo">166</span>import org.apache.hadoop.hbase.security.UserProvider;<a name="line.166"></a> -<span class="sourceLineNo">167</span>import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;<a name="line.167"></a> -<span class="sourceLineNo">168</span>import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse.CompactionState;<a name="line.168"></a> -<span class="sourceLineNo">169</span>import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionServerInfo;<a name="line.169"></a> -<span class="sourceLineNo">170</span>import org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.Quotas;<a name="line.170"></a> -<span class="sourceLineNo">171</span>import org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.SpaceViolationPolicy;<a name="line.171"></a> -<span class="sourceLineNo">172</span>import org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription;<a name="line.172"></a> -<span class="sourceLineNo">173</span>import org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos;<a name="line.173"></a> -<span class="sourceLineNo">174</span>import org.apache.hadoop.hbase.util.Addressing;<a name="line.174"></a> -<span class="sourceLineNo">175</span>import org.apache.hadoop.hbase.util.Bytes;<a name="line.175"></a> -<span class="sourceLineNo">176</span>import org.apache.hadoop.hbase.util.CompressionTest;<a name="line.176"></a> -<span class="sourceLineNo">177</span>import org.apache.hadoop.hbase.util.EncryptionTest;<a name="line.177"></a> -<span class="sourceLineNo">178</span>import org.apache.hadoop.hbase.util.FSUtils;<a name="line.178"></a> -<span class="sourceLineNo">179</span>import org.apache.hadoop.hbase.util.HFileArchiveUtil;<a name="line.179"></a> -<span class="sourceLineNo">180</span>import org.apache.hadoop.hbase.util.HasThread;<a name="line.180"></a> -<span class="sourceLineNo">181</span>import org.apache.hadoop.hbase.util.IdLock;<a name="line.181"></a> -<span class="sourceLineNo">182</span>import org.apache.hadoop.hbase.util.ModifyRegionUtils;<a name="line.182"></a> -<span class="sourceLineNo">183</span>import org.apache.hadoop.hbase.util.Pair;<a name="line.183"></a> -<span class="sourceLineNo">184</span>import org.apache.hadoop.hbase.util.Threads;<a name="line.184"></a> -<span class="sourceLineNo">185</span>import org.apache.hadoop.hbase.util.VersionInfo;<a name="line.185"></a> -<span class="sourceLineNo">186</span>import org.apache.hadoop.hbase.util.ZKDataMigrator;<a name="line.186"></a> -<span class="sourceLineNo">187</span>import org.apache.hadoop.hbase.zookeeper.DrainingServerTracker;<a name="line.187"></a> -<span class="sourceLineNo">188</span>import org.apache.hadoop.hbase.zookeeper.LoadBalancerTracker;<a name="line.188"></a> -<span class="sourceLineNo">189</span>import org.apache.hadoop.hbase.zookeeper.MasterAddressTracker;<a name="line.189"></a> -<span class="sourceLineNo">190</span>import org.apache.hadoop.hbase.zookeeper.MasterMaintenanceModeTracker;<a name="line.190"></a> -<span class="sourceLineNo">191</span>import org.apache.hadoop.hbase.zookeeper.RegionNormalizerTracker;<a name="line.191"></a> -<span class="sourceLineNo">192</span>import org.apache.hadoop.hbase.zookeeper.RegionServerTracker;<a name="line.192"></a> -<span class="sourceLineNo">193</span>import org.apache.hadoop.hbase.zookeeper.SplitOrMergeTracker;<a name="line.193"></a> -<span class="sourceLineNo">194</span>import org.apache.hadoop.hbase.zookeeper.ZKClusterId;<a name="line.194"></a> -<span class="sourceLineNo">195</span>import org.apache.hadoop.hbase.zookeeper.ZKUtil;<a name="line.195"></a> -<span class="sourceLineNo">196</span>import org.apache.hadoop.hbase.zookeeper.ZooKeeperWatcher;<a name="line.196"></a> -<span class="sourceLineNo">197</span>import org.apache.zookeeper.KeeperException;<a name="line.197"></a> -<span class="sourceLineNo">198</span>import org.eclipse.jetty.server.Server;<a name="line.198"></a> -<span class="sourceLineNo">199</span>import org.eclipse.jetty.server.ServerConnector;<a name="line.199"></a> -<span class="sourceLineNo">200</span>import org.eclipse.jetty.servlet.ServletHolder;<a name="line.200"></a> -<span class="sourceLineNo">201</span>import org.eclipse.jetty.webapp.WebAppContext;<a name="line.201"></a> -<span class="sourceLineNo">202</span><a name="line.202"></a> -<span class="sourceLineNo">203</span>import org.apache.hadoop.hbase.shaded.com.google.common.annotations.VisibleForTesting;<a name="line.203"></a> -<span class="sourceLineNo">204</span>import org.apache.hadoop.hbase.shaded.com.google.common.collect.Lists;<a name="line.204"></a> -<span class="sourceLineNo">205</span>import org.apache.hadoop.hbase.shaded.com.google.common.collect.Maps;<a name="line.205"></a> +<span class="sourceLineNo">167</span>import org.apache.hadoop.hbase.shaded.com.google.common.annotations.VisibleForTesting;<a name="line.167"></a> +<span class="sourceLineNo">168</span>import org.apache.hadoop.hbase.shaded.com.google.common.collect.Lists;<a name="line.168"></a> +<span class="sourceLineNo">169</span>import org.apache.hadoop.hbase.shaded.com.google.common.collect.Maps;<a name="line.169"></a> +<span class="sourceLineNo">170</span>import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;<a name="line.170"></a> +<span class="sourceLineNo">171</span>import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse.CompactionState;<a name="line.171"></a> +<span class="sourceLineNo">172</span>import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionServerInfo;<a name="line.172"></a> +<span class="sourceLineNo">173</span>import org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.Quotas;<a name="line.173"></a> +<span class="sourceLineNo">174</span>import org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.SpaceViolationPolicy;<a name="line.174"></a> +<span class="sourceLineNo">175</span>import org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription;<a name="line.175"></a> +<span class="sourceLineNo">176</span>import org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos;<a name="line.176"></a> +<span class="sourceLineNo">177</span>import org.apache.hadoop.hbase.util.Addressing;<a name="line.177"></a> +<span class="sourceLineNo">178</span>import org.apache.hadoop.hbase.util.Bytes;<a name="line.178"></a> +<span class="sourceLineNo">179</span>import org.apache.hadoop.hbase.util.CompressionTest;<a name="line.179"></a> +<span class="sourceLineNo">180</span>import org.apache.hadoop.hbase.util.EncryptionTest;<a name="line.180"></a> +<span class="sourceLineNo">181</span>import org.apache.hadoop.hbase.util.FSUtils;<a name="line.181"></a> +<span class="sourceLineNo">182</span>import org.apache.hadoop.hbase.util.HFileArchiveUtil;<a name="line.182"></a> +<span class="sourceLineNo">183</span>import org.apache.hadoop.hbase.util.HasThread;<a name="line.183"></a> +<span class="sourceLineNo">184</span>import org.apache.hadoop.hbase.util.IdLock;<a name="line.184"></a> +<span class="sourceLineNo">185</span>import org.apache.hadoop.hbase.util.ModifyRegionUtils;<a name="line.185"></a> +<span class="sourceLineNo">186</span>import org.apache.hadoop.hbase.util.Pair;<a name="line.186"></a> +<span class="sourceLineNo">187</span>import org.apache.hadoop.hbase.util.Threads;<a name="line.187"></a> +<span class="sourceLineNo">188</span>import org.apache.hadoop.hbase.util.VersionInfo;<a name="line.188"></a> +<span class="sourceLineNo">189</span>import org.apache.hadoop.hbase.util.ZKDataMigrator;<a name="line.189"></a> +<span class="sourceLineNo">190</span>import org.apache.hadoop.hbase.zookeeper.DrainingServerTracker;<a name="line.190"></a> +<span class="sourceLineNo">191</span>import org.apache.hadoop.hbase.zookeeper.LoadBalancerTracker;<a name="line.191"></a> +<span class="sourceLineNo">192</span>import org.apache.hadoop.hbase.zookeeper.MasterAddressTracker;<a name="line.192"></a> +<span class="sourceLineNo">193</span>import org.apache.hadoop.hbase.zookeeper.MasterMaintenanceModeTracker;<a name="line.193"></a> +<span class="sourceLineNo">194</span>import org.apache.hadoop.hbase.zookeeper.RegionNormalizerTracker;<a name="line.194"></a> +<span class="sourceLineNo">195</span>import org.apache.hadoop.hbase.zookeeper.RegionServerTracker;<a name="line.195"></a> +<span class="sourceLineNo">196</span>import org.apache.hadoop.hbase.zookeeper.SplitOrMergeTracker;<a name="line.196"></a> +<span class="sourceLineNo">197</span>import org.apache.hadoop.hbase.zookeeper.ZKClusterId;<a name="line.197"></a> +<span class="sourceLineNo">198</span>import org.apache.hadoop.hbase.zookeeper.ZKUtil;<a name="line.198"></a> +<span class="sourceLineNo">199</span>import org.apache.hadoop.hbase.zookeeper.ZooKeeperWatcher;<a name="line.199"></a> +<span class="sourceLineNo">200</span>import org.apache.zookeeper.KeeperException;<a name="line.200"></a> +<span class="sourceLineNo">201</span>import org.eclipse.jetty.server.Server;<a name="line.201"></a> +<span class="sourceLineNo">202</span>import org.eclipse.jetty.server.ServerConnector;<a name="line.202"></a> +<span class="sourceLineNo">203</span>import org.eclipse.jetty.servlet.ServletHolder;<a name="line.203"></a> +<span class="sourceLineNo">204</span>import org.eclipse.jetty.webapp.WebAppContext;<a name="line.204"></a> +<span class="sourceLineNo">205</span><a name="line.205"></a> <span class="sourceLineNo">206</span>import com.google.protobuf.Descriptors;<a name="line.206"></a> <span class="sourceLineNo">207</span>import com.google.protobuf.Service;<a name="line.207"></a> <span class="sourceLineNo">208</span><a name="line.208"></a> @@ -1945,14 +1945,14 @@ <span class="sourceLineNo">1937</span> }<a name="line.1937"></a> <span class="sourceLineNo">1938</span><a name="line.1938"></a> <span class="sourceLineNo">1939</span> // 3. blocking file count<a name="line.1939"></a> -<span class="sourceLineNo">1940</span> String sbfc = htd.getConfigurationValue(HStore.BLOCKING_STOREFILES_KEY);<a name="line.1940"></a> -<span class="sourceLineNo">1941</span> if (sbfc != null) {<a name="line.1941"></a> -<span class="sourceLineNo">1942</span> blockingFileCount = Integer.parseInt(sbfc);<a name="line.1942"></a> +<span class="sourceLineNo">1940</span> sv = hcd.getConfigurationValue(HStore.BLOCKING_STOREFILES_KEY);<a name="line.1940"></a> +<span class="sourceLineNo">1941</span> if (sv != null) {<a name="line.1941"></a> +<span class="sourceLineNo">1942</span> blockingFileCount = Integer.parseInt(sv);<a name="line.1942"></a> <span class="sourceLineNo">1943</span> }<a name="line.1943"></a> <span class="sourceLineNo">1944</span> if (blockingFileCount < 1000) {<a name="line.1944"></a> <span class="sourceLineNo">1945</span> message =<a name="line.1945"></a> -<span class="sourceLineNo">1946</span> "blocking file count '" + HStore.BLOCKING_STOREFILES_KEY + "' " + blockingFileCount<a name="line.1946"></a> -<span class="sourceLineNo">1947</span> + " is below recommended minimum of 1000";<a name="line.1947"></a> +<span class="sourceLineNo">1946</span> "Blocking file count '" + HStore.BLOCKING_STOREFILES_KEY + "' " + blockingFileCount<a name="line.1946"></a> +<span class="sourceLineNo">1947</span> + " is below recommended minimum of 1000 for column family "+ hcd.getNameAsString();<a name="line.1947"></a> <span class="sourceLineNo">1948</span> throw new IOException(message);<a name="line.1948"></a> <span class="sourceLineNo">1949</span> }<a name="line.1949"></a> <span class="sourceLineNo">1950</span> }<a name="line.1950"></a>
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c0fcd7f3/devapidocs/src-html/org/apache/hadoop/hbase/master/ServerManager.html ---------------------------------------------------------------------- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/master/ServerManager.html b/devapidocs/src-html/org/apache/hadoop/hbase/master/ServerManager.html index 7ee3fb6..9112346 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/master/ServerManager.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/master/ServerManager.html @@ -1036,15 +1036,22 @@ <span class="sourceLineNo">1028</span> flushedSequenceIdByRegion.remove(encodedName);<a name="line.1028"></a> <span class="sourceLineNo">1029</span> }<a name="line.1029"></a> <span class="sourceLineNo">1030</span><a name="line.1030"></a> -<span class="sourceLineNo">1031</span> /**<a name="line.1031"></a> -<span class="sourceLineNo">1032</span> * Called by delete table and similar to notify the ServerManager that a region was removed.<a name="line.1032"></a> -<span class="sourceLineNo">1033</span> */<a name="line.1033"></a> -<span class="sourceLineNo">1034</span> public void removeRegions(final List<HRegionInfo> regions) {<a name="line.1034"></a> -<span class="sourceLineNo">1035</span> for (HRegionInfo hri: regions) {<a name="line.1035"></a> -<span class="sourceLineNo">1036</span> removeRegion(hri);<a name="line.1036"></a> -<span class="sourceLineNo">1037</span> }<a name="line.1037"></a> -<span class="sourceLineNo">1038</span> }<a name="line.1038"></a> -<span class="sourceLineNo">1039</span>}<a name="line.1039"></a> +<span class="sourceLineNo">1031</span> @VisibleForTesting<a name="line.1031"></a> +<span class="sourceLineNo">1032</span> public boolean isRegionInServerManagerStates(final HRegionInfo hri) {<a name="line.1032"></a> +<span class="sourceLineNo">1033</span> final byte[] encodedName = hri.getEncodedNameAsBytes();<a name="line.1033"></a> +<span class="sourceLineNo">1034</span> return (storeFlushedSequenceIdsByRegion.containsKey(encodedName)<a name="line.1034"></a> +<span class="sourceLineNo">1035</span> || flushedSequenceIdByRegion.containsKey(encodedName));<a name="line.1035"></a> +<span class="sourceLineNo">1036</span> }<a name="line.1036"></a> +<span class="sourceLineNo">1037</span><a name="line.1037"></a> +<span class="sourceLineNo">1038</span> /**<a name="line.1038"></a> +<span class="sourceLineNo">1039</span> * Called by delete table and similar to notify the ServerManager that a region was removed.<a name="line.1039"></a> +<span class="sourceLineNo">1040</span> */<a name="line.1040"></a> +<span class="sourceLineNo">1041</span> public void removeRegions(final List<HRegionInfo> regions) {<a name="line.1041"></a> +<span class="sourceLineNo">1042</span> for (HRegionInfo hri: regions) {<a name="line.1042"></a> +<span class="sourceLineNo">1043</span> removeRegion(hri);<a name="line.1043"></a> +<span class="sourceLineNo">1044</span> }<a name="line.1044"></a> +<span class="sourceLineNo">1045</span> }<a name="line.1045"></a> +<span class="sourceLineNo">1046</span>}<a name="line.1046"></a> http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c0fcd7f3/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/MergeTableRegionsProcedure.html ---------------------------------------------------------------------- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/MergeTableRegionsProcedure.html b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/MergeTableRegionsProcedure.html index 6422e20..35a66e0 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/MergeTableRegionsProcedure.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/MergeTableRegionsProcedure.html @@ -521,7 +521,7 @@ <span class="sourceLineNo">513</span> throws IOException {<a name="line.513"></a> <span class="sourceLineNo">514</span> GetRegionInfoResponse response =<a name="line.514"></a> <span class="sourceLineNo">515</span> Util.getRegionInfoResponse(env, rs.getServerName(), rs.getRegion());<a name="line.515"></a> -<span class="sourceLineNo">516</span> return response.hasSplittable() && response.getSplittable();<a name="line.516"></a> +<span class="sourceLineNo">516</span> return response.hasMergeable() && response.getMergeable();<a name="line.516"></a> <span class="sourceLineNo">517</span> }<a name="line.517"></a> <span class="sourceLineNo">518</span><a name="line.518"></a> <span class="sourceLineNo">519</span> /**<a name="line.519"></a>