hbase git commit: HBASE-17343 Changing the default type of CompactingMemStore to be BASIC

2017-05-09 Thread anastasia
Repository: hbase
Updated Branches:
  refs/heads/master 3bcc475ee -> c5cc81d8e


HBASE-17343 Changing the default type of CompactingMemStore to be BASIC


Project: http://git-wip-us.apache.org/repos/asf/hbase/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/c5cc81d8
Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/c5cc81d8
Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/c5cc81d8

Branch: refs/heads/master
Commit: c5cc81d8e31ba76833adf25b6c357205745c23ad
Parents: 3bcc475
Author: anastas 
Authored: Wed May 10 09:02:26 2017 +0300
Committer: anastas 
Committed: Wed May 10 09:02:26 2017 +0300

--
 .../org/apache/hadoop/hbase/regionserver/CompactingMemStore.java   | 2 +-
 1 file changed, 1 insertion(+), 1 deletion(-)
--


http://git-wip-us.apache.org/repos/asf/hbase/blob/c5cc81d8/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/CompactingMemStore.java
--
diff --git 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/CompactingMemStore.java
 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/CompactingMemStore.java
index b244997..aaf60e3 100644
--- 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/CompactingMemStore.java
+++ 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/CompactingMemStore.java
@@ -56,7 +56,7 @@ public class CompactingMemStore extends AbstractMemStore {
   public static final String COMPACTING_MEMSTORE_TYPE_KEY =
   "hbase.hregion.compacting.memstore.type";
   public static final String COMPACTING_MEMSTORE_TYPE_DEFAULT =
-  String.valueOf(MemoryCompactionPolicy.NONE);
+  String.valueOf(MemoryCompactionPolicy.BASIC);
   // Default fraction of in-memory-flush size w.r.t. flush-to-disk size
   public static final String IN_MEMORY_FLUSH_THRESHOLD_FACTOR_KEY =
   "hbase.memstore.inmemoryflush.threshold.factor";



hbase git commit: HBASE-8758 Error in RegionCoprocessorHost class preScanner method documentation

2017-05-09 Thread chia7712
Repository: hbase
Updated Branches:
  refs/heads/branch-1.3 6545b45b1 -> 286394ba6


HBASE-8758 Error in RegionCoprocessorHost class preScanner method documentation

Signed-off-by: Chia-Ping Tsai 


Project: http://git-wip-us.apache.org/repos/asf/hbase/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/286394ba
Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/286394ba
Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/286394ba

Branch: refs/heads/branch-1.3
Commit: 286394ba636f3be0f27ced00d58227b48ff290e8
Parents: 6545b45
Author: Roman Nikitchenko 
Authored: Wed May 10 11:30:16 2017 +0800
Committer: Chia-Ping Tsai 
Committed: Wed May 10 11:37:55 2017 +0800

--
 .../apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java| 2 +-
 1 file changed, 1 insertion(+), 1 deletion(-)
--


http://git-wip-us.apache.org/repos/asf/hbase/blob/286394ba/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java
--
diff --git 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java
 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java
index 814370c..1ed866a 100644
--- 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java
+++ 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java
@@ -1288,7 +1288,7 @@ public class RegionCoprocessorHost
   /**
* @param scan the Scan specification
* @return scanner id to return to client if default operation should be
-   * bypassed, false otherwise
+   * bypassed, null otherwise
* @exception IOException Exception
*/
   public RegionScanner preScannerOpen(final Scan scan) throws IOException {



hbase git commit: HBASE-8758 Error in RegionCoprocessorHost class preScanner method documentation

2017-05-09 Thread chia7712
Repository: hbase
Updated Branches:
  refs/heads/branch-1.2 afb67a8d9 -> b6d1b19a3


HBASE-8758 Error in RegionCoprocessorHost class preScanner method documentation

Signed-off-by: Chia-Ping Tsai 


Project: http://git-wip-us.apache.org/repos/asf/hbase/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/b6d1b19a
Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/b6d1b19a
Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/b6d1b19a

Branch: refs/heads/branch-1.2
Commit: b6d1b19a3e5a645c7906fb00870775598f9d1514
Parents: afb67a8
Author: Roman Nikitchenko 
Authored: Wed May 10 11:30:16 2017 +0800
Committer: Chia-Ping Tsai 
Committed: Wed May 10 11:36:53 2017 +0800

--
 .../apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java| 2 +-
 1 file changed, 1 insertion(+), 1 deletion(-)
--


http://git-wip-us.apache.org/repos/asf/hbase/blob/b6d1b19a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java
--
diff --git 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java
 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java
index 814370c..1ed866a 100644
--- 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java
+++ 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java
@@ -1288,7 +1288,7 @@ public class RegionCoprocessorHost
   /**
* @param scan the Scan specification
* @return scanner id to return to client if default operation should be
-   * bypassed, false otherwise
+   * bypassed, null otherwise
* @exception IOException Exception
*/
   public RegionScanner preScannerOpen(final Scan scan) throws IOException {



hbase git commit: HBASE-8758 Error in RegionCoprocessorHost class preScanner method documentation

2017-05-09 Thread chia7712
Repository: hbase
Updated Branches:
  refs/heads/master 815b0f853 -> 3bcc475ee


HBASE-8758 Error in RegionCoprocessorHost class preScanner method documentation

Signed-off-by: Chia-Ping Tsai 


Project: http://git-wip-us.apache.org/repos/asf/hbase/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/3bcc475e
Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/3bcc475e
Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/3bcc475e

Branch: refs/heads/master
Commit: 3bcc475ee245a083f39b346c18dfdd2892aae745
Parents: 815b0f8
Author: Roman Nikitchenko 
Authored: Wed May 10 11:30:16 2017 +0800
Committer: Chia-Ping Tsai 
Committed: Wed May 10 11:35:45 2017 +0800

--
 .../apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java| 2 +-
 1 file changed, 1 insertion(+), 1 deletion(-)
--


http://git-wip-us.apache.org/repos/asf/hbase/blob/3bcc475e/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java
--
diff --git 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java
 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java
index 64823b6..0abc988 100644
--- 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java
+++ 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java
@@ -1254,7 +1254,7 @@ public class RegionCoprocessorHost
   /**
* @param scan the Scan specification
* @return scanner id to return to client if default operation should be
-   * bypassed, false otherwise
+   * bypassed, null otherwise
* @exception IOException Exception
*/
   public RegionScanner preScannerOpen(final Scan scan) throws IOException {



hbase git commit: HBASE-8758 Error in RegionCoprocessorHost class preScanner method documentation

2017-05-09 Thread chia7712
Repository: hbase
Updated Branches:
  refs/heads/branch-1 8d51c8371 -> ea89047ab


HBASE-8758 Error in RegionCoprocessorHost class preScanner method documentation

Signed-off-by: Chia-Ping Tsai 


Project: http://git-wip-us.apache.org/repos/asf/hbase/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/ea89047a
Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/ea89047a
Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/ea89047a

Branch: refs/heads/branch-1
Commit: ea89047abf20b1dbf55dd1fa758a5545984157be
Parents: 8d51c83
Author: Roman Nikitchenko 
Authored: Wed May 10 11:30:16 2017 +0800
Committer: Chia-Ping Tsai 
Committed: Wed May 10 11:34:18 2017 +0800

--
 .../apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java| 2 +-
 1 file changed, 1 insertion(+), 1 deletion(-)
--


http://git-wip-us.apache.org/repos/asf/hbase/blob/ea89047a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java
--
diff --git 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java
 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java
index 8adc3a4..74ecbc8 100644
--- 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java
+++ 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java
@@ -1290,7 +1290,7 @@ public class RegionCoprocessorHost
   /**
* @param scan the Scan specification
* @return scanner id to return to client if default operation should be
-   * bypassed, false otherwise
+   * bypassed, null otherwise
* @exception IOException Exception
*/
   public RegionScanner preScannerOpen(final Scan scan) throws IOException {



[1/2] hbase git commit: HBASE-17928 Shell tool to clear compaction queues (Guangxu Cheng)

2017-05-09 Thread tedyu
Repository: hbase
Updated Branches:
  refs/heads/master b67f6fecc -> 815b0f853


http://git-wip-us.apache.org/repos/asf/hbase/blob/815b0f85/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/RegionServerObserver.java
--
diff --git 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/RegionServerObserver.java
 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/RegionServerObserver.java
index 9fbf4db..140bdbe 100644
--- 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/RegionServerObserver.java
+++ 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/RegionServerObserver.java
@@ -171,4 +171,20 @@ public interface RegionServerObserver extends Coprocessor {
   default void postReplicateLogEntries(
   final ObserverContext ctx,
   List entries, CellScanner cells) throws IOException {}
+
+  /**
+   * This will be called before clearing compaction queues
+   * @param ctx the environment to interact with the framework and region 
server.
+   */
+  default void preClearCompactionQueues(
+  final ObserverContext ctx)
+  throws IOException {}
+
+  /**
+   * This will be called after clearing compaction queues
+   * @param ctx the environment to interact with the framework and region 
server.
+   */
+  default void postClearCompactionQueues(
+  final ObserverContext ctx)
+  throws IOException {}
 }

http://git-wip-us.apache.org/repos/asf/hbase/blob/815b0f85/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/CompactSplitThread.java
--
diff --git 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/CompactSplitThread.java
 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/CompactSplitThread.java
index 5356ee1..a74c4cb 100644
--- 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/CompactSplitThread.java
+++ 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/CompactSplitThread.java
@@ -672,4 +672,12 @@ public class CompactSplitThread implements 
CompactionRequestor, PropagatingConfi
   void shutdownLongCompactions(){
 this.longCompactions.shutdown();
   }
+
+  public void clearLongCompactionsQueue() {
+longCompactions.getQueue().clear();
+  }
+
+  public void clearShortCompactionsQueue() {
+shortCompactions.getQueue().clear();
+  }
 }

http://git-wip-us.apache.org/repos/asf/hbase/blob/815b0f85/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RSRpcServices.java
--
diff --git 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RSRpcServices.java
 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RSRpcServices.java
index 5deb6ba..34401f2 100644
--- 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RSRpcServices.java
+++ 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RSRpcServices.java
@@ -31,6 +31,7 @@ import java.util.*;
 import java.util.Map.Entry;
 import java.util.concurrent.ConcurrentHashMap;
 import java.util.concurrent.ConcurrentMap;
+import java.util.concurrent.atomic.AtomicBoolean;
 import java.util.concurrent.atomic.AtomicLong;
 import java.util.concurrent.atomic.LongAdder;
 
@@ -110,6 +111,8 @@ import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
 import org.apache.hadoop.hbase.shaded.protobuf.RequestConverter;
 import org.apache.hadoop.hbase.shaded.protobuf.ResponseConverter;
 import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.AdminService;
+import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ClearCompactionQueuesRequest;
+import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ClearCompactionQueuesResponse;
 import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CloseRegionForSplitOrMergeRequest;
 import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CloseRegionForSplitOrMergeResponse;
 import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CloseRegionRequest;
@@ -268,6 +271,8 @@ public class RSRpcServices implements HBaseRPCErrorHandler,
*/
   private final long minimumScanTimeLimitDelta;
 
+  final AtomicBoolean clearCompactionQueues = new AtomicBoolean(false);
+
   /**
* An Rpc callback for closing a RegionScanner.
*/
@@ -1611,6 +1616,44 @@ public class RSRpcServices implements 
HBaseRPCErrorHandler,
 return builder.build();
   }
 
+  @Override
+  @QosPriority(priority=HConstants.ADMIN_QOS)
+  public ClearCompactionQueuesResponse clearCompactionQueues(RpcController 
controller,
+ClearCompactionQueuesRequest request) throws ServiceException {
+LOG.debug("Client=" + RpcServer.getRequestUserName() + "/" + 
RpcServer.getRemoteAddress()
++ " clear compactions queue");
+ClearCompactionQueuesResponse.Builder respB

[2/2] hbase git commit: HBASE-17928 Shell tool to clear compaction queues (Guangxu Cheng)

2017-05-09 Thread tedyu
HBASE-17928 Shell tool to clear compaction queues (Guangxu Cheng)


Project: http://git-wip-us.apache.org/repos/asf/hbase/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/815b0f85
Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/815b0f85
Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/815b0f85

Branch: refs/heads/master
Commit: 815b0f853b8005535219d8e3381d5a44c06887f7
Parents: b67f6fe
Author: tedyu 
Authored: Tue May 9 18:32:38 2017 -0700
Committer: tedyu 
Committed: Tue May 9 18:32:38 2017 -0700

--
 .../org/apache/hadoop/hbase/client/Admin.java   |   11 +
 .../apache/hadoop/hbase/client/HBaseAdmin.java  |   22 +
 .../hbase/shaded/protobuf/RequestConverter.java |   10 +
 .../shaded/protobuf/generated/AdminProtos.java  | 1356 +++---
 .../src/main/protobuf/Admin.proto   |   10 +
 .../hbase/coprocessor/RegionServerObserver.java |   16 +
 .../hbase/regionserver/CompactSplitThread.java  |8 +
 .../hbase/regionserver/RSRpcServices.java   |   43 +
 .../RegionServerCoprocessorHost.java|   20 +
 .../hbase/security/access/AccessController.java |6 +
 .../hadoop/hbase/master/MockRegionServer.java   |8 +
 hbase-shell/src/main/ruby/hbase/admin.rb|   26 +
 hbase-shell/src/main/ruby/shell.rb  |1 +
 .../shell/commands/clear_compaction_queues.rb   |   41 +
 14 files changed, 1418 insertions(+), 160 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/hbase/blob/815b0f85/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Admin.java
--
diff --git 
a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Admin.java 
b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Admin.java
index 414c5ac..2efc9cb 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Admin.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Admin.java
@@ -24,6 +24,7 @@ import java.util.ArrayList;
 import java.util.Collection;
 import java.util.List;
 import java.util.Map;
+import java.util.Set;
 import java.util.concurrent.Future;
 import java.util.regex.Pattern;
 
@@ -2026,4 +2027,14 @@ public interface Admin extends Abortable, Closeable {
* @throws IOException if a remote or network exception occurs
*/
   void disableTableReplication(final TableName tableName) throws IOException;
+
+  /**
+   * Clear compacting queues on a regionserver.
+   * @param sn the region server name
+   * @param queues the set of queue name
+   * @throws IOException if a remote or network exception occurs
+   * @throws InterruptedException
+   */
+  void clearCompactionQueues(final ServerName sn, final Set queues)
+throws IOException, InterruptedException;
 }

http://git-wip-us.apache.org/repos/asf/hbase/blob/815b0f85/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HBaseAdmin.java
--
diff --git 
a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HBaseAdmin.java 
b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HBaseAdmin.java
index ac5c239..9192b10 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HBaseAdmin.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HBaseAdmin.java
@@ -92,6 +92,7 @@ import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
 import org.apache.hadoop.hbase.shaded.protobuf.RequestConverter;
 import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos;
 import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.AdminService;
+import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ClearCompactionQueuesRequest;
 import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CloseRegionRequest;
 import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CloseRegionResponse;
 import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CompactRegionRequest;
@@ -4227,4 +4228,25 @@ public class HBaseAdmin implements Admin {
 
 return otherConf;
   }
+
+  @Override
+  public void clearCompactionQueues(final ServerName sn, final Set 
queues)
+throws IOException, InterruptedException {
+if (queues == null || queues.size() == 0) {
+  throw new IllegalArgumentException("queues cannot be null or empty");
+}
+final AdminService.BlockingInterface admin = this.connection.getAdmin(sn);
+Callable callable = new Callable() {
+  @Override
+  public Void call() throws Exception {
+// TODO: There is no timeout on this controller. Set one!
+HBaseRpcController controller = rpcControllerFactory.newController();
+ClearCompactionQueuesRequest request =
+RequestConverter.buildClearCompactionQueuesRequest(queues);
+admin.clearCompact

hbase git commit: HBASE-15199 (addendum) - When JRUBY_HOME is specified, update CLASSPATH and HBASE_OPTS when jruby is needed

2017-05-09 Thread busbey
Repository: hbase
Updated Branches:
  refs/heads/master c38bf1244 -> b67f6fecc


HBASE-15199 (addendum) - When JRUBY_HOME is specified, update CLASSPATH and 
HBASE_OPTS when jruby is needed

Signed-off-by: Sean Busbey 
Signed-off-by: tedyu 
Signed-off-by: Jerry He 


Project: http://git-wip-us.apache.org/repos/asf/hbase/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/b67f6fec
Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/b67f6fec
Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/b67f6fec

Branch: refs/heads/master
Commit: b67f6fecc173ff1272284f3e47f95d493fab331d
Parents: c38bf12
Author: Xiang Li 
Authored: Sat May 6 11:37:13 2017 +0800
Committer: Sean Busbey 
Committed: Tue May 9 16:12:45 2017 -0500

--
 bin/hbase | 52 +++-
 bin/hbase.cmd | 44 
 2 files changed, 51 insertions(+), 45 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/hbase/blob/b67f6fec/bin/hbase
--
diff --git a/bin/hbase b/bin/hbase
index 579c413..fa70f6e 100755
--- a/bin/hbase
+++ b/bin/hbase
@@ -294,32 +294,34 @@ if [ "$AUTH_AS_SERVER" == "true" ] || [ "$COMMAND" = 
"hbck" ]; then
 fi
 
 # for jruby
-# (1) when JRUBY_HOME is specified explicitly, eg. export 
JRUBY_HOME=/usr/local/share/jruby
-# for all commands, CLASSPATH and HBASE_OPTS are updated according to 
JRUBY_HOME specified
-# (2) when JRUBY_HOME is not specified explicitly
-# A. for jruby_cmds defined below, add JRUBY_PACKAGED_WITH_HBASE into 
CLASSPATH
-# B. for other commands, do nothing
-
-if [ "$JRUBY_HOME" != "" ]; then  # JRUBY_HOME is specified explicitly, eg. 
export JRUBY_HOME=/usr/local/share/jruby
-  # add jruby.jar into CLASSPATH
-  CLASSPATH="$JRUBY_HOME/lib/jruby.jar:$CLASSPATH"
-
-  # add jruby to HBASE_OPTS
-  HBASE_OPTS="$HBASE_OPTS -Djruby.home=$JRUBY_HOME -Djruby.lib=$JRUBY_HOME/lib"
-
-else  # JRUBY_HOME is not specified explicitly
-  # check if the commmand needs jruby
-  declare -a jruby_cmds=("shell" "org.jruby.Main")
-  for cmd in "${jruby_cmds[@]}"; do
-if [[ $cmd == "$COMMAND" ]]; then
-  jruby_needed=true
-  break
-fi
-  done
+# (1) for the commands which need jruby (see jruby_cmds defined below)
+# A. when JRUBY_HOME is specified explicitly, eg. export 
JRUBY_HOME=/usr/local/share/jruby
+#CLASSPATH and HBASE_OPTS are updated according to JRUBY_HOME specified
+# B. when JRUBY_HOME is not specified explicitly
+#add jruby packaged with HBase to CLASSPATH
+# (2) for other commands, do nothing
+
+# check if the commmand needs jruby
+declare -a jruby_cmds=("shell" "org.jruby.Main")
+for cmd in "${jruby_cmds[@]}"; do
+  if [[ $cmd == "$COMMAND" ]]; then
+jruby_needed=true
+break
+  fi
+done
+
+# the command needs jruby
+if [[ $jruby_needed ]]; then
+  if [ "$JRUBY_HOME" != "" ]; then  # JRUBY_HOME is specified explicitly, eg. 
export JRUBY_HOME=/usr/local/share/jruby
+# add jruby.jar into CLASSPATH
+CLASSPATH="$JRUBY_HOME/lib/jruby.jar:$CLASSPATH"
+
+# add jruby to HBASE_OPTS
+HBASE_OPTS="$HBASE_OPTS -Djruby.home=$JRUBY_HOME 
-Djruby.lib=$JRUBY_HOME/lib"
 
-  # add JRUBY_PACKAGED_WITH_HBASE to CLASSPATH when jruby is needed
-  JRUBY_PACKAGED_WITH_HBASE="$HBASE_HOME/lib/ruby/*.jar"
-  if [[ $jruby_needed ]]; then  # the command needs jruby
+  else  # JRUBY_HOME is not specified explicitly
+# add jruby packaged with HBase to CLASSPATH
+JRUBY_PACKAGED_WITH_HBASE="$HBASE_HOME/lib/ruby/*.jar"
 for jruby_jar in $JRUBY_PACKAGED_WITH_HBASE; do
   CLASSPATH=$jruby_jar:$CLASSPATH;
 done

http://git-wip-us.apache.org/repos/asf/hbase/blob/b67f6fec/bin/hbase.cmd
--
diff --git a/bin/hbase.cmd b/bin/hbase.cmd
index 44cb9d3..67a3bc1 100644
--- a/bin/hbase.cmd
+++ b/bin/hbase.cmd
@@ -263,30 +263,34 @@ if defined service_entry (
 )
 
 @rem for jruby
-@rem (1) when JRUBY_HOME is defined
-@rem for all commands, CLASSPATH and HBASE_OPTS are updated according to 
JRUBY_HOME specified
-@rem (2) when JRUBY_HOME is not defined
-@rem A. for jruby-commands defined below, add JRUBY_PACKAGED_WITH_HBASE 
into CLASSPATH
-@rem B. for other commands, do nothing
-
-@rem JRUBY_HOME is defined
-if defined JRUBY_HOME (
-  set CLASSPATH=%CLASSPATH%;%JRUBY_HOME%\lib\jruby.jar
-  set HBASE_OPTS=%HBASE_OPTS% -Djruby.home="%JRUBY_HOME%" 
-Djruby.lib="%JRUBY_HOME%\lib"
+@rem (1) for the commands which need jruby (see jruby-commands defined below)
+@rem A. when JRUBY_HOME is defined
+@remCLASSPATH and HBASE_OPTS are updated according to JRUBY_HOME 
defined
+@rem B. when JRUBY_HOME is not defined
+@remadd jruby packaged with HBase to CLASSPATH
+@rem (2) for other commands,

hbase git commit: HBASE-18017 Reduce frequency of setStoragePolicy failure warnings

2017-05-09 Thread apurtell
Repository: hbase
Updated Branches:
  refs/heads/master 951b23a44 -> c38bf1244


HBASE-18017 Reduce frequency of setStoragePolicy failure warnings


Project: http://git-wip-us.apache.org/repos/asf/hbase/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/c38bf124
Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/c38bf124
Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/c38bf124

Branch: refs/heads/master
Commit: c38bf12444aca77c7cb12637147c07dc711acbe9
Parents: 951b23a
Author: Andrew Purtell 
Authored: Mon May 8 15:05:59 2017 -0700
Committer: Andrew Purtell 
Committed: Tue May 9 12:08:05 2017 -0700

--
 .../org/apache/hadoop/hbase/util/FSUtils.java   | 67 +---
 1 file changed, 44 insertions(+), 23 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/hbase/blob/c38bf124/hbase-server/src/main/java/org/apache/hadoop/hbase/util/FSUtils.java
--
diff --git 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/util/FSUtils.java 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/util/FSUtils.java
index c78ba06..284b786 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/util/FSUtils.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/util/FSUtils.java
@@ -152,6 +152,9 @@ public abstract class FSUtils {
 setStoragePolicy(fs, path, storagePolicy);
   }
 
+  private static final Map warningMap =
+  new ConcurrentHashMap();
+
   /**
* Sets storage policy for given path.
* If the passed path is a directory, we'll set the storage policy for all 
files
@@ -187,17 +190,20 @@ public abstract class FSUtils {
 try {
   distributed = isDistributedFileSystem(fs);
 } catch (IOException ioe) {
-  // This should NEVER happen.
-  LOG.warn("Failed setStoragePolicy=" + trimmedStoragePolicy + " on path=" 
+
-  path + "; failed isDFS test", ioe);
+  if (!warningMap.containsKey(fs)) {
+warningMap.put(fs, true);
+LOG.warn("FileSystem isn't an instance of DistributedFileSystem; 
presuming it doesn't "
++ "support setStoragePolicy. Unable to set storagePolicy=" + 
trimmedStoragePolicy
++ " on path=" + path);
+  } else if (LOG.isDebugEnabled()) {
+LOG.debug("FileSystem isn't an instance of DistributedFileSystem; 
presuming it doesn't "
++ "support setStoragePolicy. Unable to set storagePolicy=" + 
trimmedStoragePolicy
++ " on path=" + path);
+  }
   return;
 }
 if (distributed) {
   invokeSetStoragePolicy(fs, path, trimmedStoragePolicy);
-} else {
-  LOG.info("FileSystem isn't an instance of DistributedFileSystem; 
presuming it doesn't " +
-  "support setStoragePolicy. Unable to set storagePolicy=" + 
trimmedStoragePolicy +
-  " on path=" + path);
 }
   }
 
@@ -209,41 +215,56 @@ public abstract class FSUtils {
 Method m = null;
 try {
   m = fs.getClass().getDeclaredMethod("setStoragePolicy",
-  new Class[] { Path.class, String.class });
+new Class[] { Path.class, String.class });
   m.setAccessible(true);
 } catch (NoSuchMethodException e) {
-  LOG.info("FileSystem doesn't support setStoragePolicy; HDFS-6584 not 
available "
-  + "(hadoop-2.6.0+): " + e.getMessage());
+  final String msg = "FileSystem doesn't support setStoragePolicy; 
HDFS-6584 not available";
+  if (!warningMap.containsKey(fs)) {
+warningMap.put(fs, true);
+LOG.warn(msg, e);
+  } else if (LOG.isDebugEnabled()) {
+LOG.debug(msg, e);
+  }
+  m = null;
 } catch (SecurityException e) {
-  LOG.info("Don't have access to setStoragePolicy on FileSystems; 
HDFS-6584 not available "
-  + "(hadoop-2.6.0+): ", e);
+  final String msg = "No access to setStoragePolicy on FileSystem; 
HDFS-6584 not available";
+  if (!warningMap.containsKey(fs)) {
+warningMap.put(fs, true);
+LOG.warn(msg, e);
+  } else if (LOG.isDebugEnabled()) {
+LOG.debug(msg, e);
+  }
   m = null; // could happen on setAccessible()
 }
 if (m != null) {
   try {
 m.invoke(fs, path, storagePolicy);
-LOG.info("Set storagePolicy=" + storagePolicy + " for path=" + path);
+if (LOG.isDebugEnabled()) {
+  LOG.debug("Set storagePolicy=" + storagePolicy + " for path=" + 
path);
+}
   } catch (Exception e) {
+// This swallows FNFE, should we be throwing it? seems more likely to 
indicate dev
+// misuse than a runtime problem with HDFS.
+if (!warningMap.containsKey(fs)) {
+  warningMap.put(fs, true);
+  LOG.warn("Unable to set storagePolicy=" + storagePolicy + " for 
path=" + path, e);
+} else if (LOG.isDebugEnabled()) {
+  LOG.

hbase git commit: HBASE-11013: Clone Snapshots on Secure Cluster Should provide option to apply Retained User Permissions

2017-05-09 Thread tedyu
Repository: hbase
Updated Branches:
  refs/heads/master ac1024af2 -> 951b23a44


HBASE-11013: Clone Snapshots on Secure Cluster Should provide option to apply 
Retained User Permissions

Signed-off-by: tedyu 


Project: http://git-wip-us.apache.org/repos/asf/hbase/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/951b23a4
Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/951b23a4
Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/951b23a4

Branch: refs/heads/master
Commit: 951b23a44cd90ae4afed9b255de0e678fbfba946
Parents: ac1024a
Author: huzheng 
Authored: Mon May 8 21:01:47 2017 +0800
Committer: tedyu 
Committed: Tue May 9 09:32:48 2017 -0700

--
 .../org/apache/hadoop/hbase/client/Admin.java   |  13 +
 .../apache/hadoop/hbase/client/HBaseAdmin.java  |  25 +-
 .../hbase/security/access/TablePermission.java  |   4 +
 .../shaded/protobuf/generated/MasterProtos.java | 597 +++
 .../src/main/protobuf/Master.proto  |   1 +
 .../org/apache/hadoop/hbase/master/HMaster.java |   5 +-
 .../hadoop/hbase/master/MasterRpcServices.java  |   4 +-
 .../procedure/CloneSnapshotProcedure.java   |  21 +-
 .../hbase/master/snapshot/SnapshotManager.java  |  17 +-
 .../master/snapshot/TakeSnapshotHandler.java|   3 +
 .../security/access/AccessControlLists.java |   2 +-
 .../snapshot/SnapshotDescriptionUtils.java  |  97 ++-
 .../hbase/client/TestSnapshotWithAcl.java   | 203 +++
 .../hbase/security/access/SecureTestUtil.java   |   2 +-
 hbase-shell/src/main/ruby/hbase/admin.rb|   4 +-
 hbase-shell/src/main/ruby/hbase_constants.rb|   1 +
 .../main/ruby/shell/commands/clone_snapshot.rb  |  11 +-
 .../hbase/client/TestReplicationShell.java  |   1 -
 18 files changed, 727 insertions(+), 284 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/hbase/blob/951b23a4/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Admin.java
--
diff --git 
a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Admin.java 
b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Admin.java
index 8a9dc61..414c5ac 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Admin.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Admin.java
@@ -1534,6 +1534,19 @@ public interface Admin extends Abortable, Closeable {
 
   /**
* Create a new table by cloning the snapshot content.
+   * @param snapshotName name of the snapshot to be cloned
+   * @param tableName name of the table where the snapshot will be restored
+   * @param restoreAcl true to clone acl into newly created table
+   * @throws IOException if a remote or network exception occurs
+   * @throws TableExistsException if table to be created already exists
+   * @throws RestoreSnapshotException if snapshot failed to be cloned
+   * @throws IllegalArgumentException if the specified table has not a valid 
name
+   */
+  void cloneSnapshot(final String snapshotName, final TableName tableName, 
final boolean restoreAcl)
+  throws IOException, TableExistsException, RestoreSnapshotException;
+
+  /**
+   * Create a new table by cloning the snapshot content.
*
* @param snapshotName name of the snapshot to be cloned
* @param tableName name of the table where the snapshot will be restored

http://git-wip-us.apache.org/repos/asf/hbase/blob/951b23a4/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HBaseAdmin.java
--
diff --git 
a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HBaseAdmin.java 
b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HBaseAdmin.java
index 0c62688..ac5c239 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HBaseAdmin.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HBaseAdmin.java
@@ -2572,7 +2572,7 @@ public class HBaseAdmin implements Admin {
 try {
   // Restore snapshot
   get(
-internalRestoreSnapshotAsync(snapshotName, tableName),
+internalRestoreSnapshotAsync(snapshotName, tableName, false),
 syncWaitTimeout,
 TimeUnit.MILLISECONDS);
 } catch (IOException e) {
@@ -2581,7 +2581,7 @@ public class HBaseAdmin implements Admin {
   if (takeFailSafeSnapshot) {
 try {
   get(
-internalRestoreSnapshotAsync(failSafeSnapshotSnapshotName, 
tableName),
+internalRestoreSnapshotAsync(failSafeSnapshotSnapshotName, 
tableName, false),
 syncWaitTimeout,
 TimeUnit.MILLISECONDS);
   String msg = "Restore snapshot=" + snapshotName +
@@ -2624,7 +2624,7 @@ public class HBaseAdmin implements Admin {
   throw new TableNotDisabledException(tableName);
 }
 
-return internalRest

[50/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-05-09 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/checkstyle-aggregate.html
--
diff --git a/checkstyle-aggregate.html b/checkstyle-aggregate.html
index 153e8e6..27b828c 100644
--- a/checkstyle-aggregate.html
+++ b/checkstyle-aggregate.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase – Checkstyle Results
 
@@ -286,10 +286,10 @@
  Warnings
  Errors
 
-2156
+2159
 0
 0
-14354
+14355
 
 Files
 
@@ -2457,7 +2457,7 @@
 org/apache/hadoop/hbase/ipc/NettyRpcServer.java
 0
 0
-4
+3
 
 org/apache/hadoop/hbase/ipc/PriorityFunction.java
 0
@@ -2517,7 +2517,7 @@
 org/apache/hadoop/hbase/ipc/RpcServer.java
 0
 0
-57
+53
 
 org/apache/hadoop/hbase/ipc/RpcServerFactory.java
 0
@@ -2529,4511 +2529,4516 @@
 0
 4
 
+org/apache/hadoop/hbase/ipc/ServerCall.java
+0
+0
+3
+
 org/apache/hadoop/hbase/ipc/SimpleRpcScheduler.java
 0
 0
 8
-
+
 org/apache/hadoop/hbase/ipc/SimpleRpcServer.java
 0
 0
-25
-
+26
+
 org/apache/hadoop/hbase/mapred/Driver.java
 0
 0
 4
-
+
 org/apache/hadoop/hbase/mapred/GroupingTableMap.java
 0
 0
 11
-
+
 org/apache/hadoop/hbase/mapred/HRegionPartitioner.java
 0
 0
 4
-
+
 org/apache/hadoop/hbase/mapred/IdentityTableMap.java
 0
 0
 8
-
+
 org/apache/hadoop/hbase/mapred/IdentityTableReduce.java
 0
 0
 7
-
+
 org/apache/hadoop/hbase/mapred/MultiTableSnapshotInputFormat.java
 0
 0
 5
-
+
 org/apache/hadoop/hbase/mapred/RowCounter.java
 0
 0
 7
-
+
 org/apache/hadoop/hbase/mapred/TableInputFormat.java
 0
 0
 1
-
+
 org/apache/hadoop/hbase/mapred/TableInputFormatBase.java
 0
 0
 6
-
+
 org/apache/hadoop/hbase/mapred/TableMap.java
 0
 0
 1
-
+
 org/apache/hadoop/hbase/mapred/TableMapReduceUtil.java
 0
 0
 11
-
+
 org/apache/hadoop/hbase/mapred/TableOutputFormat.java
 0
 0
 2
-
+
 org/apache/hadoop/hbase/mapred/TableRecordReader.java
 0
 0
 5
-
+
 org/apache/hadoop/hbase/mapred/TableRecordReaderImpl.java
 0
 0
 7
-
+
 org/apache/hadoop/hbase/mapred/TableReduce.java
 0
 0
 1
-
+
 org/apache/hadoop/hbase/mapred/TableSnapshotInputFormat.java
 0
 0
 4
-
+
 org/apache/hadoop/hbase/mapred/TableSplit.java
 0
 0
 17
-
+
 org/apache/hadoop/hbase/mapreduce/CellCounter.java
 0
 0
 10
-
+
 org/apache/hadoop/hbase/mapreduce/CellCreator.java
 0
 0
 6
-
+
 org/apache/hadoop/hbase/mapreduce/CopyTable.java
 0
 0
 12
-
+
 org/apache/hadoop/hbase/mapreduce/DefaultVisibilityExpressionResolver.java
 0
 0
 2
-
+
 org/apache/hadoop/hbase/mapreduce/Driver.java
 0
 0
 3
-
+
 org/apache/hadoop/hbase/mapreduce/Export.java
 0
 0
 7
-
+
 org/apache/hadoop/hbase/mapreduce/GroupingTableMapper.java
 0
 0
 3
-
+
 org/apache/hadoop/hbase/mapreduce/HFileInputFormat.java
 0
 0
 1
-
+
 org/apache/hadoop/hbase/mapreduce/HFileOutputFormat2.java
 0
 0
 5
-
+
 org/apache/hadoop/hbase/mapreduce/HRegionPartitioner.java
 0
 0
 4
-
+
 org/apache/hadoop/hbase/mapreduce/HashTable.java
 0
 0
 2
-
+
 org/apache/hadoop/hbase/mapreduce/IdentityTableMapper.java
 0
 0
 1
-
+
 org/apache/hadoop/hbase/mapreduce/IdentityTableReducer.java
 0
 0
 1
-
+
 org/apache/hadoop/hbase/mapreduce/Import.java
 0
 0
 9
-
+
 org/apache/hadoop/hbase/mapreduce/ImportTsv.java
 0
 0
 19
-
+
 org/apache/hadoop/hbase/mapreduce/IndexBuilder.java
 0
 0
 1
-
+
 org/apache/hadoop/hbase/mapreduce/JarFinder.java
 0
 0
 1
-
+
 org/apache/hadoop/hbase/mapreduce/JobUtil.java
 0
 0
 4
-
+
 org/apache/hadoop/hbase/mapreduce/KeyValueSortReducer.java
 0
 0
 4
-
+
 org/apache/hadoop/hbase/mapreduce/LoadIncrementalHFiles.java
 0
 0
 6
-
+
 org/apache/hadoop/hbase/mapreduce/MultiHFileOutputFormat.java
 0
 0
 7
-
+
 org/apache/hadoop/hbase/mapreduce/MultiTableInputFormat.java
 0
 0
 1
-
+
 org/apache/hadoop/hbase/mapreduce/MultiTableInputFormatBase.java
 0
 0
 5
-
+
 org/apache/hadoop/hbase/mapreduce/MultiTableOutputFormat.java
 0
 0
 3
-
+
 org/apache/hadoop/hbase/mapreduce/MultiTableSnapshotInputFormat.java
 0
 0
 1
-
+
 org/apache/hadoop/hbase/mapreduce/MultiTableSnapshotInputFormatImpl.java
 0
 0
 12
-
+
 org/apache/hadoop/hbase/mapreduce/MultithreadedTableMapper.java
 0
 0
 7
-
+
 org/apache/hadoop/hbase/mapreduce/PutCombiner.java
 0
 0
 3
-
+
 org/apache/hadoop/hbase/mapreduce/PutSortReducer.java
 0
 0
 2
-
+
 org/apache/hadoop/hbase/mapreduce/ResultSerialization.java
 0
 0
 1
-
+
 org/apache/hadoop/hbase/mapreduce/RowCounter.java
 0
 0
 6
-
+
 org/apache/hadoop/hbase/mapreduce/SampleUploader.java
 0
 0
 1
-
+
 org/apache/hadoop/hbase/mapreduce/SimpleTotalOrderPartitioner.java
 0
 0
 6
-
+
 org/apache/hadoop/hbase/mapreduce/SyncTable.java
 0
 0
 6
-
+
 org/apache/hadoop/hbase/mapreduce/TableInputFormat.java
 0
 0
 4
-
+
 org/apache/hadoop/hbase/mapreduce/TableInputFormatBase.java
 0
 0
 15
-
+
 org/apache/hadoop/hbase/mapreduce/TableMapReduceUtil.java
 0
 0
 54
-
+
 org/apache/hadoop/hbase/mapreduce/TableMapper.java
 0
 0
 1
-
+
 org/apache/hadoop/hbase/mapreduce/TableOutputCommitter.java
 0
 0
 1
-
+
 org/apache/hadoop/hbase/mapreduce/TableOutputFormat.java
 0
 0
 4
-
+
 org/apache/hadoop/h

[38/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-05-09 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/org/apache/hadoop/hbase/ipc/class-use/BufferChain.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/ipc/class-use/BufferChain.html 
b/devapidocs/org/apache/hadoop/hbase/ipc/class-use/BufferChain.html
index 627009f..f851c7e 100644
--- a/devapidocs/org/apache/hadoop/hbase/ipc/class-use/BufferChain.html
+++ b/devapidocs/org/apache/hadoop/hbase/ipc/class-use/BufferChain.html
@@ -106,7 +106,7 @@
 
 
 protected BufferChain
-RpcServer.Call.response
+ServerCall.response
 Chain of buffers to send as response.
 
 
@@ -121,7 +121,7 @@
 
 
 protected BufferChain
-RpcServer.Call.wrapWithSasl(BufferChain bc) 
+ServerCall.wrapWithSasl(BufferChain bc) 
 
 
 
@@ -141,7 +141,7 @@
 
 
 protected BufferChain
-RpcServer.Call.wrapWithSasl(BufferChain bc) 
+ServerCall.wrapWithSasl(BufferChain bc) 
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/org/apache/hadoop/hbase/ipc/class-use/CellBlockBuilder.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/ipc/class-use/CellBlockBuilder.html 
b/devapidocs/org/apache/hadoop/hbase/ipc/class-use/CellBlockBuilder.html
index d132558..1fb183e 100644
--- a/devapidocs/org/apache/hadoop/hbase/ipc/class-use/CellBlockBuilder.html
+++ b/devapidocs/org/apache/hadoop/hbase/ipc/class-use/CellBlockBuilder.html
@@ -114,6 +114,10 @@
 
 
 protected CellBlockBuilder
+ServerCall.cellBlockBuilder 
+
+
+protected CellBlockBuilder
 RpcServer.cellBlockBuilder 
 
 
@@ -130,6 +134,58 @@
  Codec codec,
  
org.apache.hadoop.io.compress.CompressionCodec compressor) 
 
+
+NettyServerCall(int id,
+   
org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService service,
+   
org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor md,
+   
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader header,
+   
org.apache.hadoop.hbase.shaded.com.google.protobuf.Message param,
+   CellScanner cellScanner,
+   RpcServer.Connection connection,
+   long size,
+   org.apache.htrace.TraceInfo tinfo,
+   http://docs.oracle.com/javase/8/docs/api/java/net/InetAddress.html?is-external=true";
 title="class or interface in java.net">InetAddress remoteAddress,
+   long receiveTime,
+   int timeout,
+   ByteBufferPool reservoir,
+   CellBlockBuilder cellBlockBuilder,
+   RpcServer.CallCleanup reqCleanup) 
+
+
+ServerCall(int id,
+  
org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService service,
+  
org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor md,
+  
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader header,
+  
org.apache.hadoop.hbase.shaded.com.google.protobuf.Message param,
+  CellScanner cellScanner,
+  RpcServer.Connection connection,
+  long size,
+  org.apache.htrace.TraceInfo tinfo,
+  http://docs.oracle.com/javase/8/docs/api/java/net/InetAddress.html?is-external=true";
 title="class or interface in java.net">InetAddress remoteAddress,
+  long receiveTime,
+  int timeout,
+  ByteBufferPool reservoir,
+  CellBlockBuilder cellBlockBuilder,
+  RpcServer.CallCleanup reqCleanup) 
+
+
+SimpleServerCall(int id,
+
org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService service,
+
org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor md,
+
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader header,
+
org.apache.hadoop.hbase.shaded.com.google.protobuf.Message param,
+CellScanner cellScanner,
+RpcServer.Connection connection,
+long size,
+org.apache.htrace.TraceInfo tinfo,
+http://docs.oracle.com/javase/8/docs/api/java/net/InetAddress.html?is-external=true";
 title="class or interface in java.net">InetAddress remoteAddress,
+long receiveTime,
+int timeout,
+ByteBufferPool reservoir,
+CellBlockBuilder cellBlockBuilder,
+RpcServer.CallCleanup reqCleanup,
+SimpleRpcServer.Responder responder) 
+
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/org/apache/hadoop/hbase/ipc/class-use/NettyRpcServer.Call.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/ipc/class-use/NettyRpcServer.Call.html 
b/devapidocs/org/apache/hadoop/hbase/ipc/class-use/NettyRpcServer.Call.ht

[30/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-05-09 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.Call.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.Call.html 
b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.Call.html
deleted file mode 100644
index 27e0dee..000
--- a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.Call.html
+++ /dev/null
@@ -1,612 +0,0 @@
-http://www.w3.org/TR/html4/loose.dtd";>
-
-
-Source code
-
-
-
-
-001/**
-002 * Licensed to the Apache Software 
Foundation (ASF) under one
-003 * or more contributor license 
agreements.  See the NOTICE file
-004 * distributed with this work for 
additional information
-005 * regarding copyright ownership.  The 
ASF licenses this file
-006 * to you under the Apache License, 
Version 2.0 (the
-007 * "License"); you may not use this file 
except in compliance
-008 * with the License.  You may obtain a 
copy of the License at
-009 *
-010 * 
http://www.apache.org/licenses/LICENSE-2.0
-011 *
-012 * Unless required by applicable law or 
agreed to in writing, software
-013 * distributed under the License is 
distributed on an "AS IS" BASIS,
-014 * WITHOUT WARRANTIES OR CONDITIONS OF 
ANY KIND, either express or implied.
-015 * See the License for the specific 
language governing permissions and
-016 * limitations under the License.
-017 */
-018
-019package org.apache.hadoop.hbase.ipc;
-020
-021import 
io.netty.bootstrap.ServerBootstrap;
-022import io.netty.buffer.ByteBuf;
-023import 
io.netty.buffer.PooledByteBufAllocator;
-024import io.netty.buffer.Unpooled;
-025import io.netty.channel.Channel;
-026import io.netty.channel.ChannelFuture;
-027import 
io.netty.channel.ChannelFutureListener;
-028import 
io.netty.channel.ChannelHandlerContext;
-029import 
io.netty.channel.ChannelInboundHandlerAdapter;
-030import 
io.netty.channel.ChannelInitializer;
-031import io.netty.channel.ChannelOption;
-032import 
io.netty.channel.ChannelOutboundHandlerAdapter;
-033import 
io.netty.channel.ChannelPipeline;
-034import io.netty.channel.ChannelPromise;
-035import io.netty.channel.EventLoopGroup;
-036import 
io.netty.channel.epoll.EpollEventLoopGroup;
-037import 
io.netty.channel.epoll.EpollServerSocketChannel;
-038import 
io.netty.channel.group.ChannelGroup;
-039import 
io.netty.channel.group.DefaultChannelGroup;
-040import 
io.netty.channel.nio.NioEventLoopGroup;
-041import 
io.netty.channel.socket.SocketChannel;
-042import 
io.netty.channel.socket.nio.NioServerSocketChannel;
-043import 
io.netty.handler.codec.ByteToMessageDecoder;
-044import 
io.netty.handler.codec.LengthFieldBasedFrameDecoder;
-045import 
io.netty.util.concurrent.GlobalEventExecutor;
-046
-047import java.io.IOException;
-048import java.io.InterruptedIOException;
-049import java.net.InetAddress;
-050import java.net.InetSocketAddress;
-051import java.nio.ByteBuffer;
-052import java.util.Arrays;
-053import java.util.List;
-054import 
java.util.concurrent.CountDownLatch;
-055
-056import org.apache.commons.logging.Log;
-057import 
org.apache.commons.logging.LogFactory;
-058import 
org.apache.hadoop.conf.Configuration;
-059import 
org.apache.hadoop.hbase.CellScanner;
-060import 
org.apache.hadoop.hbase.HConstants;
-061import org.apache.hadoop.hbase.Server;
-062import 
org.apache.hadoop.hbase.classification.InterfaceStability;
-063import 
org.apache.hadoop.hbase.monitoring.MonitoredRPCHandler;
-064import 
org.apache.hadoop.hbase.nio.ByteBuff;
-065import 
org.apache.hadoop.hbase.nio.SingleByteBuff;
-066import 
org.apache.hadoop.hbase.security.AccessDeniedException;
-067import 
org.apache.hadoop.hbase.security.AuthMethod;
-068import 
org.apache.hadoop.hbase.security.HBasePolicyProvider;
-069import 
org.apache.hadoop.hbase.security.SaslStatus;
-070import 
org.apache.hadoop.hbase.security.SaslUtil;
-071import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService;
-072import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor;
-073import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.Message;
-074import 
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader;
-075import 
org.apache.hadoop.hbase.util.Bytes;
-076import 
org.apache.hadoop.hbase.util.JVM;
-077import 
org.apache.hadoop.hbase.util.Pair;
-078import 
org.apache.hadoop.io.IntWritable;
-079import 
org.apache.hadoop.security.authorize.ServiceAuthorizationManager;
-080import org.apache.htrace.TraceInfo;
-081
-082/**
-083 * An RPC server with Netty4 
implementation.
-084 *
-085 */
-086public class NettyRpcServer extends 
RpcServer {
-087
-088  public static final Log LOG = 
LogFactory.getLog(NettyRpcServer.class);
-089
-090  protected final InetSocketAddress 
bindAddress;
-091
-092  private final CountDownLatch closed = 
new CountDownLatch(1);
-093  private final Channel serverChannel;
-094  private final ChannelGroup allChanne

hbase-site git commit: INFRA-10751 Empty commit

2017-05-09 Thread git-site-role
Repository: hbase-site
Updated Branches:
  refs/heads/asf-site 7ef4c5a9a -> 8678c699d


INFRA-10751 Empty commit


Project: http://git-wip-us.apache.org/repos/asf/hbase-site/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase-site/commit/8678c699
Tree: http://git-wip-us.apache.org/repos/asf/hbase-site/tree/8678c699
Diff: http://git-wip-us.apache.org/repos/asf/hbase-site/diff/8678c699

Branch: refs/heads/asf-site
Commit: 8678c699dd903f277455fb29daf542d7d01e676e
Parents: 7ef4c5a
Author: jenkins 
Authored: Tue May 9 15:00:39 2017 +
Committer: jenkins 
Committed: Tue May 9 15:00:39 2017 +

--

--




[21/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-05-09 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.ByteBuffByteInput.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.ByteBuffByteInput.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.ByteBuffByteInput.html
index 5cc356a..9e1c66c 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.ByteBuffByteInput.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.ByteBuffByteInput.html
@@ -70,2037 +70,1559 @@
 062import 
org.apache.hadoop.hbase.client.VersionInfoUtil;
 063import 
org.apache.hadoop.hbase.codec.Codec;
 064import 
org.apache.hadoop.hbase.conf.ConfigurationObserver;
-065import 
org.apache.hadoop.hbase.exceptions.RegionMovedException;
-066import 
org.apache.hadoop.hbase.exceptions.RequestTooBigException;
-067import 
org.apache.hadoop.hbase.io.ByteBufferListOutputStream;
-068import 
org.apache.hadoop.hbase.io.ByteBufferOutputStream;
-069import 
org.apache.hadoop.hbase.io.ByteBufferPool;
-070import 
org.apache.hadoop.hbase.io.crypto.aes.CryptoAES;
-071import 
org.apache.hadoop.hbase.monitoring.MonitoredRPCHandler;
-072import 
org.apache.hadoop.hbase.monitoring.TaskMonitor;
-073import 
org.apache.hadoop.hbase.nio.ByteBuff;
-074import 
org.apache.hadoop.hbase.nio.MultiByteBuff;
-075import 
org.apache.hadoop.hbase.nio.SingleByteBuff;
-076import 
org.apache.hadoop.hbase.regionserver.RSRpcServices;
-077import 
org.apache.hadoop.hbase.security.AccessDeniedException;
-078import 
org.apache.hadoop.hbase.security.AuthMethod;
-079import 
org.apache.hadoop.hbase.security.HBaseSaslRpcServer;
-080import 
org.apache.hadoop.hbase.security.HBaseSaslRpcServer.SaslDigestCallbackHandler;
-081import 
org.apache.hadoop.hbase.security.HBaseSaslRpcServer.SaslGssCallbackHandler;
-082import 
org.apache.hadoop.hbase.security.SaslStatus;
-083import 
org.apache.hadoop.hbase.security.SaslUtil;
-084import 
org.apache.hadoop.hbase.security.User;
-085import 
org.apache.hadoop.hbase.security.UserProvider;
-086import 
org.apache.hadoop.hbase.security.token.AuthenticationTokenSecretManager;
-087import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService;
-088import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteInput;
-089import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString;
-090import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream;
-091import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream;
-092import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor;
-093import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.Message;
-094import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException;
-095import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.TextFormat;
-096import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.UnsafeByteOperations;
-097import 
org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
-098import 
org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos;
-099import 
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.VersionInfo;
-100import 
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos;
-101import 
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.CellBlockMeta;
-102import 
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ConnectionHeader;
-103import 
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ExceptionResponse;
-104import 
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader;
-105import 
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ResponseHeader;
-106import 
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation;
-107import 
org.apache.hadoop.hbase.util.ByteBufferUtils;
-108import 
org.apache.hadoop.hbase.util.Bytes;
-109import 
org.apache.hadoop.hbase.util.Pair;
-110import 
org.apache.hadoop.io.BytesWritable;
-111import org.apache.hadoop.io.Writable;
-112import 
org.apache.hadoop.io.WritableUtils;
-113import 
org.apache.hadoop.io.compress.CompressionCodec;
-114import 
org.apache.hadoop.security.UserGroupInformation;
-115import 
org.apache.hadoop.security.UserGroupInformation.AuthenticationMethod;
-116import 
org.apache.hadoop.security.authorize.AuthorizationException;
-117import 
org.apache.hadoop.security.authorize.PolicyProvider;
-118import 
org.apache.hadoop.security.authorize.ProxyUsers;
-119import 
org.apache.hadoop.security.authorize.ServiceAuthorizationManager;
-120import 
org.apache.hadoop.security.token.SecretManager;
-121import 
org.apache.hadoop.security.token.SecretManager.InvalidToken;
-122import 
org.apache.hadoop.security.token.TokenIdentifier;
-123import 
org.apache.hadoop.util.StringUtils;
-124import org.apache.htrace.TraceInfo;
-125import 
org.codehaus.jackson.map.ObjectMapper;
-126
-127import 
com.google.common.annotations.V

[32/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-05-09 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html 
b/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html
index 49a36a3..ff4d17b 100644
--- a/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html
+++ b/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html
@@ -683,20 +683,20 @@
 
 java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true";
 title="class or interface in java.lang">Enum (implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true";
 title="class or interface in java.lang">Comparable, java.io.http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true";
 title="class or interface in java.io">Serializable)
 
-org.apache.hadoop.hbase.regionserver.ScanType
-org.apache.hadoop.hbase.regionserver.FlushType
-org.apache.hadoop.hbase.regionserver.StoreScanner.StoreScannerCompactionRace
-org.apache.hadoop.hbase.regionserver.DefaultHeapMemoryTuner.StepDirection
-org.apache.hadoop.hbase.regionserver.MemStoreCompactor.Action
-org.apache.hadoop.hbase.regionserver.ScannerContext.NextState
-org.apache.hadoop.hbase.regionserver.SplitLogWorker.TaskExecutor.Status
+org.apache.hadoop.hbase.regionserver.Region.Operation
+org.apache.hadoop.hbase.regionserver.ImmutableSegment.Type
 org.apache.hadoop.hbase.regionserver.RegionOpeningState
 org.apache.hadoop.hbase.regionserver.BloomType
+org.apache.hadoop.hbase.regionserver.ScannerContext.NextState
+org.apache.hadoop.hbase.regionserver.DefaultHeapMemoryTuner.StepDirection
+org.apache.hadoop.hbase.regionserver.SplitLogWorker.TaskExecutor.Status
+org.apache.hadoop.hbase.regionserver.ScannerContext.LimitScope
+org.apache.hadoop.hbase.regionserver.FlushType
 org.apache.hadoop.hbase.regionserver.MetricsRegionServerSourceFactoryImpl.FactoryStorage
-org.apache.hadoop.hbase.regionserver.Region.Operation
+org.apache.hadoop.hbase.regionserver.ScanType
+org.apache.hadoop.hbase.regionserver.StoreScanner.StoreScannerCompactionRace
 org.apache.hadoop.hbase.regionserver.Region.FlushResult.Result
-org.apache.hadoop.hbase.regionserver.ImmutableSegment.Type
-org.apache.hadoop.hbase.regionserver.ScannerContext.LimitScope
+org.apache.hadoop.hbase.regionserver.MemStoreCompactor.Action
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/org/apache/hadoop/hbase/regionserver/querymatcher/package-tree.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/regionserver/querymatcher/package-tree.html
 
b/devapidocs/org/apache/hadoop/hbase/regionserver/querymatcher/package-tree.html
index 80fd563..d4c8f41 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/regionserver/querymatcher/package-tree.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/regionserver/querymatcher/package-tree.html
@@ -125,10 +125,10 @@
 
 java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true";
 title="class or interface in java.lang">Enum (implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true";
 title="class or interface in java.lang">Comparable, java.io.http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true";
 title="class or interface in java.io">Serializable)
 
-org.apache.hadoop.hbase.regionserver.querymatcher.StripeCompactionScanQueryMatcher.DropDeletesInOutput
-org.apache.hadoop.hbase.regionserver.querymatcher.ScanQueryMatcher.MatchCode
 org.apache.hadoop.hbase.regionserver.querymatcher.DeleteTracker.DeleteCompare
 org.apache.hadoop.hbase.regionserver.querymatcher.DeleteTracker.DeleteResult
+org.apache.hadoop.hbase.regionserver.querymatcher.ScanQueryMatcher.MatchCode
+org.apache.hadoop.hbase.regionserver.querymatcher.StripeCompactionScanQueryMatcher.DropDeletesInOutput
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/org/apache/hadoop/hbase/rest/model/package-tree.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/rest/model/package-tree.html 
b/devapidocs/org/apache/hadoop/hbase/rest/model/package-tree.html
index dbfbf42..d9efa05 100644
--- a/devapidocs/org/apache/hadoop/hbase/rest/model/package-tree.html
+++ b/devapidocs/org/apache/hadoop/hbase/rest/model/package-tree.html
@@ -110,8 +110,8 @@
 
 java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true";
 title="class or interface in java.lang">Enum (implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true";
 title="class or interface in java.lang">Comparable, java.io.http://docs.oracle.com/javase/8/docs/

[11/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-05-09 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Listener.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Listener.html 
b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Listener.html
index 2987e7b..7f61b54 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Listener.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Listener.html
@@ -135,1305 +135,1251 @@
 127  private Listener listener = null;
 128  protected Responder responder = null;
 129
-130  /**
-131   * Datastructure that holds all 
necessary to a method invocation and then afterward, carries
-132   * the result.
-133   */
-134  @InterfaceStability.Evolving
-135  public class Call extends 
RpcServer.Call {
-136
-137protected Responder responder;
+130  /** Listens on the socket. Creates jobs 
for the handler threads*/
+131  private class Listener extends Thread 
{
+132
+133private ServerSocketChannel 
acceptChannel = null; //the accept channel
+134private Selector selector = null; 
//the selector that we use for the server
+135private Reader[] readers = null;
+136private int currentReader = 0;
+137private final int 
readerPendingConnectionQueueLength;
 138
-139
@edu.umd.cs.findbugs.annotations.SuppressWarnings(value="NP_NULL_ON_SOME_PATH",
-140justification="Can't figure why 
this complaint is happening... see below")
-141Call(int id, final BlockingService 
service, final MethodDescriptor md,
-142RequestHeader header, Message 
param, CellScanner cellScanner,
-143RpcServer.Connection connection, 
long size, TraceInfo tinfo,
-144final InetAddress remoteAddress, 
int timeout, CallCleanup reqCleanup,
-145Responder responder) {
-146  super(id, service, md, header, 
param, cellScanner, connection, size,
-147  tinfo, remoteAddress, timeout, 
reqCleanup);
-148  this.responder = responder;
-149}
+139private ExecutorService readPool;
+140
+141public Listener(final String name) 
throws IOException {
+142  super(name);
+143  // The backlog of requests that we 
will have the serversocket carry.
+144  int backlogLength = 
conf.getInt("hbase.ipc.server.listen.queue.size", 128);
+145  readerPendingConnectionQueueLength 
=
+146  
conf.getInt("hbase.ipc.server.read.connection-queue.size", 100);
+147  // Create a new server socket and 
set to non blocking mode
+148  acceptChannel = 
ServerSocketChannel.open();
+149  
acceptChannel.configureBlocking(false);
 150
-151/**
-152 * Call is done. Execution happened 
and we returned results to client. It is now safe to
-153 * cleanup.
-154 */
-155
@edu.umd.cs.findbugs.annotations.SuppressWarnings(value="IS2_INCONSISTENT_SYNC",
-156justification="Presume the lock 
on processing request held by caller is protection enough")
-157@Override
-158void done() {
-159  super.done();
-160  this.getConnection().decRpcCount(); 
// Say that we're done with this call.
-161}
-162
-163@Override
-164public long disconnectSince() {
-165  if 
(!getConnection().isConnectionOpen()) {
-166return System.currentTimeMillis() 
- timestamp;
-167  } else {
-168return -1L;
-169  }
-170}
-171
-172@Override
-173public synchronized void 
sendResponseIfReady() throws IOException {
-174  // set param null to reduce memory 
pressure
-175  this.param = null;
-176  this.responder.doRespond(this);
+151  // Bind the server socket to the 
binding addrees (can be different from the default interface)
+152  bind(acceptChannel.socket(), 
bindAddress, backlogLength);
+153  port = 
acceptChannel.socket().getLocalPort(); //Could be an ephemeral port
+154  address = 
(InetSocketAddress)acceptChannel.socket().getLocalSocketAddress();
+155  // create a selector;
+156  selector = Selector.open();
+157
+158  readers = new 
Reader[readThreads];
+159  // Why this executor thing? Why not 
like hadoop just start up all the threads? I suppose it
+160  // has an advantage in that it is 
easy to shutdown the pool.
+161  readPool = 
Executors.newFixedThreadPool(readThreads,
+162new 
ThreadFactoryBuilder().setNameFormat(
+163  
"RpcServer.reader=%d,bindAddress=" + bindAddress.getHostName() +
+164  ",port=" + 
port).setDaemon(true)
+165
.setUncaughtExceptionHandler(Threads.LOGGING_EXCEPTION_HANDLER).build());
+166  for (int i = 0; i < readThreads; 
++i) {
+167Reader reader = new Reader();
+168readers[i] = reader;
+169readPool.execute(reader);
+170  }
+171  LOG.info(getName() + ": started " + 
readThreads + " reader(s) listening on port=" + port);
+172
+173  // Register accepts on the serv

[08/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-05-09 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleServerCall.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleServerCall.html 
b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleServerCall.html
new file mode 100644
index 000..31df5ee
--- /dev/null
+++ b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleServerCall.html
@@ -0,0 +1,151 @@
+http://www.w3.org/TR/html4/loose.dtd";>
+
+
+Source code
+
+
+
+
+001/**
+002 * Licensed to the Apache Software 
Foundation (ASF) under one
+003 * or more contributor license 
agreements.  See the NOTICE file
+004 * distributed with this work for 
additional information
+005 * regarding copyright ownership.  The 
ASF licenses this file
+006 * to you under the Apache License, 
Version 2.0 (the
+007 * "License"); you may not use this file 
except in compliance
+008 * with the License.  You may obtain a 
copy of the License at
+009 *
+010 * 
http://www.apache.org/licenses/LICENSE-2.0
+011 *
+012 * Unless required by applicable law or 
agreed to in writing, software
+013 * distributed under the License is 
distributed on an "AS IS" BASIS,
+014 * WITHOUT WARRANTIES OR CONDITIONS OF 
ANY KIND, either express or implied.
+015 * See the License for the specific 
language governing permissions and
+016 * limitations under the License.
+017 */
+018package org.apache.hadoop.hbase.ipc;
+019
+020import java.io.IOException;
+021import java.net.InetAddress;
+022
+023import 
org.apache.hadoop.hbase.CellScanner;
+024import 
org.apache.hadoop.hbase.classification.InterfaceAudience;
+025import 
org.apache.hadoop.hbase.io.ByteBufferPool;
+026import 
org.apache.hadoop.hbase.ipc.RpcServer.CallCleanup;
+027import 
org.apache.hadoop.hbase.ipc.SimpleRpcServer.Connection;
+028import 
org.apache.hadoop.hbase.ipc.SimpleRpcServer.Responder;
+029import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService;
+030import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor;
+031import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.Message;
+032import 
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader;
+033import org.apache.htrace.TraceInfo;
+034
+035/**
+036 * Datastructure that holds all necessary 
to a method invocation and then afterward, carries the
+037 * result.
+038 */
+039@InterfaceAudience.Private
+040class SimpleServerCall extends ServerCall 
{
+041
+042  long lastSentTime;
+043
+044  final Responder responder;
+045
+046  
@edu.umd.cs.findbugs.annotations.SuppressWarnings(value = 
"NP_NULL_ON_SOME_PATH",
+047  justification = "Can't figure why 
this complaint is happening... see below")
+048  SimpleServerCall(int id, final 
BlockingService service, final MethodDescriptor md,
+049  RequestHeader header, Message 
param, CellScanner cellScanner, RpcServer.Connection connection,
+050  long size, TraceInfo tinfo, final 
InetAddress remoteAddress, long receiveTime, int timeout,
+051  ByteBufferPool reservoir, 
CellBlockBuilder cellBlockBuilder, CallCleanup reqCleanup,
+052  Responder responder) {
+053super(id, service, md, header, param, 
cellScanner, connection, size, tinfo, remoteAddress,
+054receiveTime, timeout, reservoir, 
cellBlockBuilder, reqCleanup);
+055this.responder = responder;
+056  }
+057
+058  /**
+059   * Call is done. Execution happened and 
we returned results to client. It is now safe to cleanup.
+060   */
+061  
@edu.umd.cs.findbugs.annotations.SuppressWarnings(value = 
"IS2_INCONSISTENT_SYNC",
+062  justification = "Presume the lock 
on processing request held by caller is protection enough")
+063  @Override
+064  void done() {
+065super.done();
+066this.getConnection().decRpcCount(); 
// Say that we're done with this call.
+067  }
+068
+069  @Override
+070  public synchronized void 
sendResponseIfReady() throws IOException {
+071// set param null to reduce memory 
pressure
+072this.param = null;
+073this.responder.doRespond(this);
+074  }
+075
+076  Connection getConnection() {
+077return (Connection) 
this.connection;
+078  }
+079}
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+



[47/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-05-09 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/org/apache/hadoop/hbase/backup/package-tree.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/backup/package-tree.html 
b/devapidocs/org/apache/hadoop/hbase/backup/package-tree.html
index 1f603c7..72dbc96 100644
--- a/devapidocs/org/apache/hadoop/hbase/backup/package-tree.html
+++ b/devapidocs/org/apache/hadoop/hbase/backup/package-tree.html
@@ -165,10 +165,10 @@
 
 java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true";
 title="class or interface in java.lang">Enum (implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true";
 title="class or interface in java.lang">Comparable, java.io.http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true";
 title="class or interface in java.io">Serializable)
 
-org.apache.hadoop.hbase.backup.BackupRestoreConstants.BackupCommand
+org.apache.hadoop.hbase.backup.BackupInfo.BackupPhase
 org.apache.hadoop.hbase.backup.BackupInfo.BackupState
 org.apache.hadoop.hbase.backup.BackupType
-org.apache.hadoop.hbase.backup.BackupInfo.BackupPhase
+org.apache.hadoop.hbase.backup.BackupRestoreConstants.BackupCommand
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/org/apache/hadoop/hbase/class-use/CellScanner.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/class-use/CellScanner.html 
b/devapidocs/org/apache/hadoop/hbase/class-use/CellScanner.html
index 20110fb..3fa62ce 100644
--- a/devapidocs/org/apache/hadoop/hbase/class-use/CellScanner.html
+++ b/devapidocs/org/apache/hadoop/hbase/class-use/CellScanner.html
@@ -441,7 +441,7 @@
 
 
 protected CellScanner
-RpcServer.Call.cellScanner 
+ServerCall.cellScanner 
 
 
 
@@ -474,7 +474,7 @@
 
 
 CellScanner
-RpcServer.Call.getCellScanner() 
+ServerCall.getCellScanner() 
 
 
 CellScanner
@@ -680,7 +680,7 @@
 int timeout) 
 
 
-abstract RpcServer.Call
+abstract ServerCall
 RpcServer.Connection.createCall(int id,
   
org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService service,
   
org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor md,
@@ -695,7 +695,7 @@
   RpcServer.CallCleanup reqCleanup) 
 
 
-RpcServer.Call
+ServerCall
 SimpleRpcServer.Connection.createCall(int id,
   
org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService service,
   
org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor md,
@@ -710,7 +710,7 @@
   RpcServer.CallCleanup reqCleanup) 
 
 
-RpcServer.Call
+ServerCall
 NettyRpcServer.NettyConnection.createCall(int id,
   
org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService service,
   
org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor md,
@@ -774,7 +774,7 @@
 
 
 void
-RpcServer.Call.setResponse(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message m,
+ServerCall.setResponse(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message m,
CellScanner cells,
http://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true";
 title="class or interface in java.lang">Throwable t,
http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in 
java.lang">String errorMsg) 
@@ -797,49 +797,6 @@
 
 
 
-Call(int id,
-
org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService service,
-
org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor md,
-
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader header,
-org.apache.hadoop.hbase.shaded.com.google.protobuf.Message param,
-CellScanner cellScanner,
-RpcServer.Connection connection,
-long size,
-org.apache.htrace.TraceInfo tinfo,
-http://docs.oracle.com/javase/8/docs/api/java/net/InetAddress.html?is-external=true";
 title="class or interface in java.net">InetAddress remoteAddress,
-int timeout,
-RpcServer.CallCleanup reqCleanup) 
-
-
-Call(int id,
-
org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService service,
-
org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor md,
-
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader header,
-org.apache.hadoop.hbase.shaded.com.google.protobuf.Message param,
-CellScanner cellScanner,
-RpcServer.Connection connection,
-long size,
-org.apache.htrace.TraceInfo tinfo,
-http://docs.oracle.com/javase/8/docs/api/java/net/InetAddress.html?is-external=true";
 title="class or interface in java.net">InetAddress remoteAddress,
-int timeout,
-RpcServer.CallCleanup reqCleanup) 
-
-
-Call(int id,
-
org.a

[48/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-05-09 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/index-all.html
--
diff --git a/devapidocs/index-all.html b/devapidocs/index-all.html
index 3a10992..6d90f3b 100644
--- a/devapidocs/index-all.html
+++ b/devapidocs/index-all.html
@@ -8235,16 +8235,12 @@
  
 call(BlockingService,
 Descriptors.MethodDescriptor, Message, CellScanner, long, MonitoredRPCHandler, 
long, int) - Method in class org.apache.hadoop.hbase.ipc.NettyRpcServer
  
-Call(int,
 BlockingService, Descriptors.MethodDescriptor, RPCProtos.RequestHeader, 
Message, CellScanner, RpcServer.Connection, long, TraceInfo, InetAddress, int, 
RpcServer.CallCleanup) - Constructor for class 
org.apache.hadoop.hbase.ipc.NettyRpcServer.Call
- 
 call
 - Variable in class org.apache.hadoop.hbase.ipc.NettyRpcServer.CallWriteListener
  
 call(RpcCall,
 MonitoredRPCHandler) - Method in class 
org.apache.hadoop.hbase.ipc.RpcServer
 
 This is a server side method, which is invoked over 
RPC.
 
-Call(int,
 BlockingService, Descriptors.MethodDescriptor, RPCProtos.RequestHeader, 
Message, CellScanner, RpcServer.Connection, long, TraceInfo, InetAddress, int, 
RpcServer.CallCleanup) - Constructor for class 
org.apache.hadoop.hbase.ipc.RpcServer.Call
- 
 call(BlockingService,
 Descriptors.MethodDescriptor, Message, CellScanner, long, 
MonitoredRPCHandler) - Method in interface 
org.apache.hadoop.hbase.ipc.RpcServerInterface
 
 Deprecated.
@@ -8263,8 +8259,6 @@
  
 call(BlockingService,
 Descriptors.MethodDescriptor, Message, CellScanner, long, MonitoredRPCHandler, 
long, int) - Method in class org.apache.hadoop.hbase.ipc.SimpleRpcServer
  
-Call(int,
 BlockingService, Descriptors.MethodDescriptor, RPCProtos.RequestHeader, 
Message, CellScanner, RpcServer.Connection, long, TraceInfo, InetAddress, int, 
RpcServer.CallCleanup, SimpleRpcServer.Responder) - Constructor for 
class org.apache.hadoop.hbase.ipc.SimpleRpcServer.Call
- 
 call()
 - Method in class org.apache.hadoop.hbase.master.AssignCallable
  
 call(MasterObserver,
 ObserverContext) - Method in 
class org.apache.hadoop.hbase.master.MasterCoprocessorHost.CoprocessorOperation
@@ -8580,7 +8574,7 @@
 
 callWithRetries(RetryingCallable,
 int) - Method in class org.apache.hadoop.hbase.client.RpcRetryingCallerImpl
  
-CallWriteListener(NettyRpcServer.Call)
 - Constructor for class org.apache.hadoop.hbase.ipc.NettyRpcServer.CallWriteListener
+CallWriteListener(NettyServerCall)
 - Constructor for class org.apache.hadoop.hbase.ipc.NettyRpcServer.CallWriteListener
  
 calSleepTimeMs()
 - Method in class org.apache.hadoop.hbase.io.hadoopbackport.ThrottledInputStream
  
@@ -9005,6 +8999,8 @@
  
 cellBlockBuilder
 - Variable in class org.apache.hadoop.hbase.ipc.RpcServer
  
+cellBlockBuilder
 - Variable in class org.apache.hadoop.hbase.ipc.ServerCall
+ 
 CellBlockBuilder.ByteBufferOutputStreamSupplier 
- Class in org.apache.hadoop.hbase.ipc
  
 CellBlockBuilder.ByteBufOutputStreamSupplier - 
Class in org.apache.hadoop.hbase.ipc
@@ -9017,7 +9013,7 @@
 
 How much we think the decompressor will expand the original 
compressed content.
 
-cellBlockStream
 - Variable in class org.apache.hadoop.hbase.ipc.RpcServer.Call
+cellBlockStream
 - Variable in class org.apache.hadoop.hbase.ipc.ServerCall
  
 CellCodec 
- Class in org.apache.hadoop.hbase.codec
 
@@ -9220,7 +9216,7 @@
 
 cellScanner()
 - Method in class org.apache.hadoop.hbase.ipc.HBaseRpcControllerImpl
  
-cellScanner
 - Variable in class org.apache.hadoop.hbase.ipc.RpcServer.Call
+cellScanner
 - Variable in class org.apache.hadoop.hbase.ipc.ServerCall
  
 CellScanner(Iterator)
 - Constructor for class org.apache.hadoop.hbase.mapreduce.SyncTable.SyncMapper.CellScanner
  
@@ -10942,7 +10938,7 @@
 
 Do the necessary cleanup after the call if needed.
 
-cleanup()
 - Method in class org.apache.hadoop.hbase.ipc.RpcServer.Call
+cleanup()
 - Method in class org.apache.hadoop.hbase.ipc.ServerCall
  
 cleanup(Mapper.Context) 
- Method in class org.apache.hadoop.hbase.mapreduce.HashTable.HashMapper
  
@@ -14498,6 +14494,8 @@
 
 compareTo(WALKey)
 - Method in class org.apache.hadoop.hbase.wal.WALKey
  
+compareTo(WALSplitter.MutationReplay)
 - Method in class org.apache.hadoop.hbase.wal.WALSplitter.MutationReplay
+ 
 compareToCurrentToken(Cell)
 - Method in class org.apache.hadoop.hbase.codec.prefixtree.decode.PrefixTreeArraySearcher
 
 Compare only the bytes within the window of the current 
token
@@ -15763,10 +15761,10 @@
  
 connection
 - Variable in class org.apache.hadoop.hbase.ipc.NettyRpcServer.MessageDecoder
  
-connection
 - Variable in class org.apache.hadoop.hbase.ipc.RpcServer.Call
- 
 Connection()
 - Constructor for class org.apache.hadoop.hbase.ipc.RpcServer.Connection
  
+connection
 - Variable in class org.apache.hadoop.hbase.ipc.ServerCall
+ 
 Connection(SocketCh

[45/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-05-09 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/org/apache/hadoop/hbase/ipc/NettyServerCall.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/ipc/NettyServerCall.html 
b/devapidocs/org/apache/hadoop/hbase/ipc/NettyServerCall.html
new file mode 100644
index 000..43d126f
--- /dev/null
+++ b/devapidocs/org/apache/hadoop/hbase/ipc/NettyServerCall.html
@@ -0,0 +1,374 @@
+http://www.w3.org/TR/html4/loose.dtd";>
+
+
+
+
+
+NettyServerCall (Apache HBase 2.0.0-SNAPSHOT API)
+
+
+
+
+
+var methods = {"i0":10,"i1":10,"i2":10};
+var tabs = {65535:["t0","All Methods"],2:["t2","Instance 
Methods"],8:["t4","Concrete Methods"]};
+var altColor = "altColor";
+var rowColor = "rowColor";
+var tableTab = "tableTab";
+var activeTableTab = "activeTableTab";
+
+
+JavaScript is disabled on your browser.
+
+
+
+
+
+Skip navigation links
+
+
+
+
+Overview
+Package
+Class
+Use
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+Prev Class
+Next Class
+
+
+Frames
+No Frames
+
+
+All Classes
+
+
+
+
+
+
+
+Summary: 
+Nested | 
+Field | 
+Constr | 
+Method
+
+
+Detail: 
+Field | 
+Constr | 
+Method
+
+
+
+
+
+
+
+
+org.apache.hadoop.hbase.ipc
+Class NettyServerCall
+
+
+
+http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true";
 title="class or interface in java.lang">java.lang.Object
+
+
+org.apache.hadoop.hbase.ipc.ServerCall
+
+
+org.apache.hadoop.hbase.ipc.NettyServerCall
+
+
+
+
+
+
+
+
+
+All Implemented Interfaces:
+RpcCall, RpcCallContext
+
+
+
+@InterfaceAudience.Private
+class NettyServerCall
+extends ServerCall
+Datastructure that holds all necessary to a method 
invocation and then afterward, carries the
+ result.
+
+
+
+
+
+
+
+
+
+
+
+Field Summary
+
+
+
+
+Fields inherited from class org.apache.hadoop.hbase.ipc.ServerCall
+cellBlockBuilder,
 cellBlockStream,
 cellScanner,
 connection,
 deadline,
 header,
 id, isError,
 md, param,
 receiveTime, remoteAddress,
 reqCleanup,
 reservoir,
 response,
 rpcCallback,
 service,
 size,
 startTime,
 timeout,
 tinfo,
 user
+
+
+
+
+
+
+
+
+Constructor Summary
+
+Constructors 
+
+Constructor and Description
+
+
+NettyServerCall(int id,
+   
org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService service,
+   
org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor md,
+   
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader header,
+   
org.apache.hadoop.hbase.shaded.com.google.protobuf.Message param,
+   CellScanner cellScanner,
+   RpcServer.Connection connection,
+   long size,
+   org.apache.htrace.TraceInfo tinfo,
+   http://docs.oracle.com/javase/8/docs/api/java/net/InetAddress.html?is-external=true";
 title="class or interface in java.net">InetAddress remoteAddress,
+   long receiveTime,
+   int timeout,
+   ByteBufferPool reservoir,
+   CellBlockBuilder cellBlockBuilder,
+   RpcServer.CallCleanup reqCleanup) 
+
+
+
+
+
+
+
+
+
+Method Summary
+
+All Methods Instance Methods Concrete Methods 
+
+Modifier and Type
+Method and Description
+
+
+(package private) NettyRpcServer.NettyConnection
+getConnection() 
+
+
+void
+sendResponseIfReady()
+If we have a response, and delay is not set, then respond 
immediately.
+
+
+
+void
+sendResponseIfReady(io.netty.channel.ChannelFutureListener listener) 
+
+
+
+
+
+
+Methods inherited from class org.apache.hadoop.hbase.ipc.ServerCall
+cleanup,
 createHeaderAndMessageBytes,
 disconnectSince,
 done,
 getCellScanner,
 getClientVersionInfo,
 getDeadline,
 getHeader,
 getMethod, 
getParam,
 getPriority,
 getReceiveTime,
 getRemoteAddress,
 getRemotePort,
 getRequestUser,
 getRequestUserName,
 getResponseBlockSize,
 getResponseCellSize, getResponseExceptionSize,
 getService,
 getSize,
 getStartTime,
 getTimeout,
 getTraceInfo,
 incrementResponseBlockSize,
 incrementResponseCellSize,
 incrementResponseExceptionSize, 
isClientCellBlockSupported,
 isRetryImmediatelySupported,
 setCallBack,
 setConnectionHeaderResponse,
 setExceptionResponse,
 setResponse,
 setSaslTokenResponse,
 setStartTime,
 toShortString,
 toString,
 wrapWithSasl
+
+
+
+
+
+Methods inherited from class java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true";
 title="class or interface in java.lang">Object
+http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#cl

[49/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-05-09 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/checkstyle.rss
--
diff --git a/checkstyle.rss b/checkstyle.rss
index b77bb56..cda80bb 100644
--- a/checkstyle.rss
+++ b/checkstyle.rss
@@ -25,8 +25,8 @@ under the License.
 en-us
 ©2007 - 2017 The Apache Software Foundation
 
-  File: 2156,
- Errors: 14354,
+  File: 2159,
+ Errors: 14355,
  Warnings: 0,
  Infos: 0
   
@@ -11787,7 +11787,7 @@ under the License.
   0
 
 
-  25
+  26
 
   
   
@@ -15824,6 +15824,20 @@ under the License.
   
   
 
+  http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.ipc.SimpleServerCall.java";>org/apache/hadoop/hbase/ipc/SimpleServerCall.java
+
+
+  0
+
+
+  0
+
+
+  0
+
+  
+  
+
   http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.shaded.com.google.protobuf.Any.java";>org/apache/hadoop/hbase/shaded/com/google/protobuf/Any.java
 
 
@@ -16967,7 +16981,7 @@ under the License.
   0
 
 
-  57
+  53
 
   
   
@@ -20271,7 +20285,7 @@ under the License.
   0
 
 
-  4
+  3
 
   
   
@@ -23183,7 +23197,7 @@ under the License.
   0
 
 
-  117
+  119
 
   
   
@@ -23454,6 +23468,20 @@ under the License.
   
   
 
+  http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.ipc.ServerCall.java";>org/apache/hadoop/hbase/ipc/ServerCall.java
+
+
+  0
+
+
+  0
+
+
+  3
+
+  
+  
+
   http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.regionserver.LeaseListener.java";>org/apache/hadoop/hbase/regionserver/LeaseListener.java
 
 
@@ -30160,6 +30188,20 @@ under the License.
   
   
 
+  http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.ipc.NettyServerCall.java";>org/apache/hadoop/hbase/ipc/NettyServerCall.java
+
+
+  0
+
+
+  0
+
+
+  0
+
+  
+  
+
   http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.ZooKeeperConnectionException.java";>org/apache/hadoop/hbase/ZooKeeperConnectionException.java
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/coc.html
--
diff --git a/coc.html b/coc.html
index 1cdd55d..c9d7435 100644
--- a/coc.html
+++ b/coc.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase – 
   Code of Conduct Policy
@@ -380,7 +380,7 @@ email to mailto:priv...@hbase.apache.org";>the priv
 https://www.apache.org/";>The Apache Software 
Foundation.
 All rights reserved.  
 
-  Last Published: 
2017-05-06
+  Last Published: 
2017-05-09
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/cygwin.html
--
diff --git a/cygwin.html b/cygwin.html
index ad6e1eb..83a8ea1 100644
--- a/cygwin.html
+++ b/cygwin.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase – Installing Apache HBase (TM) on Windows using 
Cygwin
 
@@ -679,7 +679,7 @@ Now your HBase server is running, start 
coding and build that next
 https://www.apache.org/";>The Apache Software 
Foundation.
 All rights reserved.  
 
-  Last Published: 
2017-05-06
+  Last Published: 
2017-05-09
 
 
 

http://git-wip-us.apache.org/repos/asf/hb

[43/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-05-09 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/org/apache/hadoop/hbase/ipc/RpcServer.CallCleanup.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/ipc/RpcServer.CallCleanup.html 
b/devapidocs/org/apache/hadoop/hbase/ipc/RpcServer.CallCleanup.html
index 2d1c621..09e77fb 100644
--- a/devapidocs/org/apache/hadoop/hbase/ipc/RpcServer.CallCleanup.html
+++ b/devapidocs/org/apache/hadoop/hbase/ipc/RpcServer.CallCleanup.html
@@ -49,7 +49,7 @@ var activeTableTab = "activeTableTab";
 
 
 
-Prev Class
+Prev Class
 Next Class
 
 
@@ -110,7 +110,7 @@ var activeTableTab = "activeTableTab";
 
 
 http://docs.oracle.com/javase/8/docs/api/java/lang/FunctionalInterface.html?is-external=true";
 title="class or interface in java.lang">@FunctionalInterface
-protected static interface RpcServer.CallCleanup
+protected static interface RpcServer.CallCleanup
 
 
 
@@ -154,7 +154,7 @@ protected static interface 
 
 run
-void run()
+void run()
 
 
 
@@ -185,7 +185,7 @@ protected static interface 
 
-Prev Class
+Prev Class
 Next Class
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/org/apache/hadoop/hbase/ipc/RpcServer.Connection.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/ipc/RpcServer.Connection.html 
b/devapidocs/org/apache/hadoop/hbase/ipc/RpcServer.Connection.html
index 2319d90..9239da0 100644
--- a/devapidocs/org/apache/hadoop/hbase/ipc/RpcServer.Connection.html
+++ b/devapidocs/org/apache/hadoop/hbase/ipc/RpcServer.Connection.html
@@ -121,7 +121,7 @@ var activeTableTab = "activeTableTab";
 
 
 
-public abstract class RpcServer.Connection
+public abstract class RpcServer.Connection
 extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true";
 title="class or interface in java.lang">Object
 implements http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.html?is-external=true";
 title="class or interface in java.io">Closeable
 Reads calls from a connection and queues them for 
handling.
@@ -156,7 +156,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.h
 authenticatedWithFallback 
 
 
-protected RpcServer.Call
+protected ServerCall
 authFailedCall 
 
 
@@ -224,7 +224,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.h
 SASL_CALLID 
 
 
-protected RpcServer.Call
+protected ServerCall
 saslCall 
 
 
@@ -240,7 +240,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.h
 service 
 
 
-protected RpcServer.Call
+protected ServerCall
 setConnectionHeaderResponseCall 
 
 
@@ -312,7 +312,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.h
 authorizeConnection() 
 
 
-abstract RpcServer.Call
+abstract ServerCall
 createCall(int id,
   
org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService service,
   
org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor md,
@@ -458,7 +458,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.h
 
 
 connectionPreambleRead
-protected boolean connectionPreambleRead
+protected boolean connectionPreambleRead
 
 
 
@@ -467,7 +467,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.h
 
 
 connectionHeaderRead
-protected boolean connectionHeaderRead
+protected boolean connectionHeaderRead
 
 
 
@@ -476,7 +476,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.h
 
 
 callCleanup
-protected RpcServer.CallCleanup callCleanup
+protected RpcServer.CallCleanup callCleanup
 
 
 
@@ -485,7 +485,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.h
 
 
 hostAddress
-protected http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String hostAddress
+protected http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String hostAddress
 
 
 
@@ -494,7 +494,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.h
 
 
 remotePort
-protected int remotePort
+protected int remotePort
 
 
 
@@ -503,7 +503,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.h
 
 
 addr
-protected http://docs.oracle.com/javase/8/docs/api/java/net/InetAddress.html?is-external=true";
 title="class or interface in java.net">InetAddress addr
+protected http://docs.oracle.com/javase/8/docs/api/java/net/InetAddress.html?is-external=true";
 title="class or interface in java.net">InetAddress addr
 
 
 
@@ -512,7 +512,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.h
 
 
 connectionHeader
-protected org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ConnectionHeader
 connectionHeader
+protected org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ConnectionHea

[16/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-05-09 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/src-html/org/apache/hadoop/hbase/ipc/ServerCall.html
--
diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/ServerCall.html 
b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/ServerCall.html
new file mode 100644
index 000..6ad1e1b
--- /dev/null
+++ b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/ServerCall.html
@@ -0,0 +1,599 @@
+http://www.w3.org/TR/html4/loose.dtd";>
+
+
+Source code
+
+
+
+
+001/**
+002 * Licensed to the Apache Software 
Foundation (ASF) under one
+003 * or more contributor license 
agreements.  See the NOTICE file
+004 * distributed with this work for 
additional information
+005 * regarding copyright ownership.  The 
ASF licenses this file
+006 * to you under the Apache License, 
Version 2.0 (the
+007 * "License"); you may not use this file 
except in compliance
+008 * with the License.  You may obtain a 
copy of the License at
+009 *
+010 * 
http://www.apache.org/licenses/LICENSE-2.0
+011 *
+012 * Unless required by applicable law or 
agreed to in writing, software
+013 * distributed under the License is 
distributed on an "AS IS" BASIS,
+014 * WITHOUT WARRANTIES OR CONDITIONS OF 
ANY KIND, either express or implied.
+015 * See the License for the specific 
language governing permissions and
+016 * limitations under the License.
+017 */
+018package org.apache.hadoop.hbase.ipc;
+019
+020import java.io.IOException;
+021import java.net.InetAddress;
+022import java.nio.ByteBuffer;
+023import java.util.ArrayList;
+024import java.util.List;
+025
+026import 
org.apache.hadoop.hbase.CellScanner;
+027import 
org.apache.hadoop.hbase.DoNotRetryIOException;
+028import 
org.apache.hadoop.hbase.classification.InterfaceAudience;
+029import 
org.apache.hadoop.hbase.exceptions.RegionMovedException;
+030import 
org.apache.hadoop.hbase.io.ByteBufferListOutputStream;
+031import 
org.apache.hadoop.hbase.io.ByteBufferPool;
+032import 
org.apache.hadoop.hbase.ipc.RpcServer.CallCleanup;
+033import 
org.apache.hadoop.hbase.ipc.RpcServer.Connection;
+034import 
org.apache.hadoop.hbase.security.User;
+035import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService;
+036import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream;
+037import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor;
+038import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.Message;
+039import 
org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
+040import 
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.VersionInfo;
+041import 
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.CellBlockMeta;
+042import 
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ExceptionResponse;
+043import 
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader;
+044import 
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ResponseHeader;
+045import 
org.apache.hadoop.hbase.util.ByteBufferUtils;
+046import 
org.apache.hadoop.hbase.util.Bytes;
+047import 
org.apache.hadoop.util.StringUtils;
+048import org.apache.htrace.TraceInfo;
+049
+050/**
+051 * Datastructure that holds all necessary 
to a method invocation and then afterward, carries
+052 * the result.
+053 */
+054@InterfaceAudience.Private
+055abstract class ServerCall implements 
RpcCall {
+056
+057  protected final int id; 
// the client's call id
+058  protected final BlockingService 
service;
+059  protected final MethodDescriptor md;
+060  protected final RequestHeader header;
+061  protected Message param;
  // the parameter passed
+062  // Optional cell data passed outside of 
protobufs.
+063  protected final CellScanner 
cellScanner;
+064  protected final Connection connection;  
// connection to client
+065  protected final long receiveTime;  
// the time received when response is null
+066 // the 
time served when response is not null
+067  protected final int timeout;
+068  protected long startTime;
+069  protected final long deadline;// the 
deadline to handle this call, if exceed we can drop it.
+070
+071  protected final ByteBufferPool 
reservoir;
+072
+073  protected final CellBlockBuilder 
cellBlockBuilder;
+074
+075  /**
+076   * Chain of buffers to send as 
response.
+077   */
+078  protected BufferChain response;
+079
+080  protected final long size;  
// size of current call
+081  protected boolean isError;
+082  protected final TraceInfo tinfo;
+083  protected ByteBufferListOutputStream 
cellBlockStream = null;
+084  protected CallCleanup reqCleanup = 
null;
+085
+086  protected User user;
+087  protected final InetAddress 
remoteAddress;
+088  protected RpcCallback rpcCallback;
+089
+090  private long responseCellSize = 0;
+091  private long responseBlockSize = 0;
+092  // cumul

[20/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-05-09 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.Call.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.Call.html 
b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.Call.html
deleted file mode 100644
index 5cc356a..000
--- a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.Call.html
+++ /dev/null
@@ -1,2167 +0,0 @@
-http://www.w3.org/TR/html4/loose.dtd";>
-
-
-Source code
-
-
-
-
-001/**
-002 * Licensed to the Apache Software 
Foundation (ASF) under one
-003 * or more contributor license 
agreements.  See the NOTICE file
-004 * distributed with this work for 
additional information
-005 * regarding copyright ownership.  The 
ASF licenses this file
-006 * to you under the Apache License, 
Version 2.0 (the
-007 * "License"); you may not use this file 
except in compliance
-008 * with the License.  You may obtain a 
copy of the License at
-009 *
-010 * 
http://www.apache.org/licenses/LICENSE-2.0
-011 *
-012 * Unless required by applicable law or 
agreed to in writing, software
-013 * distributed under the License is 
distributed on an "AS IS" BASIS,
-014 * WITHOUT WARRANTIES OR CONDITIONS OF 
ANY KIND, either express or implied.
-015 * See the License for the specific 
language governing permissions and
-016 * limitations under the License.
-017 */
-018
-019package org.apache.hadoop.hbase.ipc;
-020
-021import static 
org.apache.hadoop.fs.CommonConfigurationKeysPublic.HADOOP_SECURITY_AUTHORIZATION;
-022
-023import java.io.ByteArrayInputStream;
-024import java.io.ByteArrayOutputStream;
-025import java.io.Closeable;
-026import java.io.DataOutputStream;
-027import java.io.IOException;
-028import java.net.InetAddress;
-029import java.net.InetSocketAddress;
-030import java.nio.ByteBuffer;
-031import java.nio.channels.Channels;
-032import 
java.nio.channels.GatheringByteChannel;
-033import 
java.nio.channels.ReadableByteChannel;
-034import 
java.nio.channels.WritableByteChannel;
-035import 
java.security.GeneralSecurityException;
-036import 
java.security.PrivilegedExceptionAction;
-037import java.util.ArrayList;
-038import java.util.HashMap;
-039import java.util.List;
-040import java.util.Map;
-041import java.util.Properties;
-042import 
java.util.concurrent.atomic.LongAdder;
-043
-044import javax.security.sasl.Sasl;
-045import 
javax.security.sasl.SaslException;
-046import javax.security.sasl.SaslServer;
-047
-048import 
org.apache.commons.crypto.cipher.CryptoCipherFactory;
-049import 
org.apache.commons.crypto.random.CryptoRandom;
-050import 
org.apache.commons.crypto.random.CryptoRandomFactory;
-051import org.apache.commons.logging.Log;
-052import 
org.apache.commons.logging.LogFactory;
-053import 
org.apache.hadoop.conf.Configuration;
-054import 
org.apache.hadoop.hbase.CallQueueTooBigException;
-055import 
org.apache.hadoop.hbase.CellScanner;
-056import 
org.apache.hadoop.hbase.DoNotRetryIOException;
-057import 
org.apache.hadoop.hbase.HBaseInterfaceAudience;
-058import 
org.apache.hadoop.hbase.HConstants;
-059import org.apache.hadoop.hbase.Server;
-060import 
org.apache.hadoop.hbase.classification.InterfaceAudience;
-061import 
org.apache.hadoop.hbase.classification.InterfaceStability;
-062import 
org.apache.hadoop.hbase.client.VersionInfoUtil;
-063import 
org.apache.hadoop.hbase.codec.Codec;
-064import 
org.apache.hadoop.hbase.conf.ConfigurationObserver;
-065import 
org.apache.hadoop.hbase.exceptions.RegionMovedException;
-066import 
org.apache.hadoop.hbase.exceptions.RequestTooBigException;
-067import 
org.apache.hadoop.hbase.io.ByteBufferListOutputStream;
-068import 
org.apache.hadoop.hbase.io.ByteBufferOutputStream;
-069import 
org.apache.hadoop.hbase.io.ByteBufferPool;
-070import 
org.apache.hadoop.hbase.io.crypto.aes.CryptoAES;
-071import 
org.apache.hadoop.hbase.monitoring.MonitoredRPCHandler;
-072import 
org.apache.hadoop.hbase.monitoring.TaskMonitor;
-073import 
org.apache.hadoop.hbase.nio.ByteBuff;
-074import 
org.apache.hadoop.hbase.nio.MultiByteBuff;
-075import 
org.apache.hadoop.hbase.nio.SingleByteBuff;
-076import 
org.apache.hadoop.hbase.regionserver.RSRpcServices;
-077import 
org.apache.hadoop.hbase.security.AccessDeniedException;
-078import 
org.apache.hadoop.hbase.security.AuthMethod;
-079import 
org.apache.hadoop.hbase.security.HBaseSaslRpcServer;
-080import 
org.apache.hadoop.hbase.security.HBaseSaslRpcServer.SaslDigestCallbackHandler;
-081import 
org.apache.hadoop.hbase.security.HBaseSaslRpcServer.SaslGssCallbackHandler;
-082import 
org.apache.hadoop.hbase.security.SaslStatus;
-083import 
org.apache.hadoop.hbase.security.SaslUtil;
-084import 
org.apache.hadoop.hbase.security.User;
-085import 
org.apache.hadoop.hbase.security.UserProvider;
-086import 
org.apache.hadoop.hbase.security.token.AuthenticationTokenSecretManager;
-087import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingSe

[42/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-05-09 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/org/apache/hadoop/hbase/ipc/RpcServer.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/ipc/RpcServer.html 
b/devapidocs/org/apache/hadoop/hbase/ipc/RpcServer.html
index 737b9b3..307b468 100644
--- a/devapidocs/org/apache/hadoop/hbase/ipc/RpcServer.html
+++ b/devapidocs/org/apache/hadoop/hbase/ipc/RpcServer.html
@@ -119,7 +119,7 @@ var activeTableTab = "activeTableTab";
 
 @InterfaceAudience.LimitedPrivate(value={"Coprocesssor","Phoenix"})
  @InterfaceStability.Evolving
-public abstract class RpcServer
+public abstract class RpcServer
 extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true";
 title="class or interface in java.lang">Object
 implements RpcServerInterface, ConfigurationObserver
 An RPC server that hosts protobuf described Services.
@@ -153,17 +153,10 @@ implements RpcServer.ByteBuffByteInput 
 
 
-class 
-RpcServer.Call
-Datastructure that holds all necessary to a method 
invocation and then afterward, carries
- the result.
-
-
-
 protected static interface 
 RpcServer.CallCleanup 
 
-
+
 class 
 RpcServer.Connection
 Reads calls from a connection and queues them for 
handling.
@@ -639,8 +632,8 @@ implements 
 protected void
-setupResponse(http://docs.oracle.com/javase/8/docs/api/java/io/ByteArrayOutputStream.html?is-external=true";
 title="class or interface in java.io">ByteArrayOutputStream response,
- RpcServer.Call call,
+setupResponse(http://docs.oracle.com/javase/8/docs/api/java/io/ByteArrayOutputStream.html?is-external=true";
 title="class or interface in java.io">ByteArrayOutputStream response,
+ ServerCall call,
  http://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true";
 title="class or interface in java.lang">Throwable t,
  http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String error)
 Setup response for the RPC Call.
@@ -681,7 +674,7 @@ implements 
 
 LOG
-public static final org.apache.commons.logging.Log LOG
+public static final org.apache.commons.logging.Log LOG
 
 
 
@@ -690,7 +683,7 @@ implements 
 
 CALL_QUEUE_TOO_BIG_EXCEPTION
-protected static final CallQueueTooBigException CALL_QUEUE_TOO_BIG_EXCEPTION
+protected static final CallQueueTooBigException CALL_QUEUE_TOO_BIG_EXCEPTION
 
 
 
@@ -699,7 +692,7 @@ implements 
 
 authorize
-private final boolean authorize
+private final boolean authorize
 
 
 
@@ -708,7 +701,7 @@ implements 
 
 isSecurityEnabled
-protected boolean isSecurityEnabled
+protected boolean isSecurityEnabled
 
 
 
@@ -717,7 +710,7 @@ implements 
 
 CURRENT_VERSION
-public static final byte CURRENT_VERSION
+public static final byte CURRENT_VERSION
 
 See Also:
 Constant
 Field Values
@@ -730,7 +723,7 @@ implements 
 
 FALLBACK_TO_INSECURE_CLIENT_AUTH
-public static final http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String FALLBACK_TO_INSECURE_CLIENT_AUTH
+public static final http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String FALLBACK_TO_INSECURE_CLIENT_AUTH
 Whether we allow a fallback to SIMPLE auth for insecure 
clients when security is enabled.
 
 See Also:
@@ -744,7 +737,7 @@ implements 
 
 DEFAULT_MAX_CALLQUEUE_LENGTH_PER_HANDLER
-protected static final int DEFAULT_MAX_CALLQUEUE_LENGTH_PER_HANDLER
+protected static final int DEFAULT_MAX_CALLQUEUE_LENGTH_PER_HANDLER
 How many calls/handler are allowed in the queue.
 
 See Also:
@@ -758,7 +751,7 @@ implements 
 
 cellBlockBuilder
-protected final CellBlockBuilder cellBlockBuilder
+protected final CellBlockBuilder cellBlockBuilder
 
 
 
@@ -767,7 +760,7 @@ implements 
 
 AUTH_FAILED_FOR
-protected static final http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String AUTH_FAILED_FOR
+protected static final http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String AUTH_FAILED_FOR
 
 See Also:
 Constant
 Field Values
@@ -780,7 +773,7 @@ implements 
 
 AUTH_SUCCESSFUL_FOR
-protected static final http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String AUTH_SUCCESSFUL_FOR
+protected static final http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String AUTH_SUCCESSFUL_FOR
 
 See Also:
 Constant
 Field Values
@@ -793,7 +786,7 @@ implements 
 
 AUDITLOG
-protected static final org.apache.commons.logging.Log AUDITLOG
+protected static final org.apache.commons.logging.Log AUDITLOG
 
 
 
@@ -802,7 +795,7 @@ implements 
 
 secretManager
-protected org.a

[17/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-05-09 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.html
--
diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.html 
b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.html
index 5cc356a..9e1c66c 100644
--- a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.html
+++ b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.html
@@ -70,2037 +70,1559 @@
 062import 
org.apache.hadoop.hbase.client.VersionInfoUtil;
 063import 
org.apache.hadoop.hbase.codec.Codec;
 064import 
org.apache.hadoop.hbase.conf.ConfigurationObserver;
-065import 
org.apache.hadoop.hbase.exceptions.RegionMovedException;
-066import 
org.apache.hadoop.hbase.exceptions.RequestTooBigException;
-067import 
org.apache.hadoop.hbase.io.ByteBufferListOutputStream;
-068import 
org.apache.hadoop.hbase.io.ByteBufferOutputStream;
-069import 
org.apache.hadoop.hbase.io.ByteBufferPool;
-070import 
org.apache.hadoop.hbase.io.crypto.aes.CryptoAES;
-071import 
org.apache.hadoop.hbase.monitoring.MonitoredRPCHandler;
-072import 
org.apache.hadoop.hbase.monitoring.TaskMonitor;
-073import 
org.apache.hadoop.hbase.nio.ByteBuff;
-074import 
org.apache.hadoop.hbase.nio.MultiByteBuff;
-075import 
org.apache.hadoop.hbase.nio.SingleByteBuff;
-076import 
org.apache.hadoop.hbase.regionserver.RSRpcServices;
-077import 
org.apache.hadoop.hbase.security.AccessDeniedException;
-078import 
org.apache.hadoop.hbase.security.AuthMethod;
-079import 
org.apache.hadoop.hbase.security.HBaseSaslRpcServer;
-080import 
org.apache.hadoop.hbase.security.HBaseSaslRpcServer.SaslDigestCallbackHandler;
-081import 
org.apache.hadoop.hbase.security.HBaseSaslRpcServer.SaslGssCallbackHandler;
-082import 
org.apache.hadoop.hbase.security.SaslStatus;
-083import 
org.apache.hadoop.hbase.security.SaslUtil;
-084import 
org.apache.hadoop.hbase.security.User;
-085import 
org.apache.hadoop.hbase.security.UserProvider;
-086import 
org.apache.hadoop.hbase.security.token.AuthenticationTokenSecretManager;
-087import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService;
-088import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteInput;
-089import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString;
-090import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream;
-091import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream;
-092import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor;
-093import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.Message;
-094import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException;
-095import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.TextFormat;
-096import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.UnsafeByteOperations;
-097import 
org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
-098import 
org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos;
-099import 
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.VersionInfo;
-100import 
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos;
-101import 
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.CellBlockMeta;
-102import 
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ConnectionHeader;
-103import 
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ExceptionResponse;
-104import 
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader;
-105import 
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ResponseHeader;
-106import 
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation;
-107import 
org.apache.hadoop.hbase.util.ByteBufferUtils;
-108import 
org.apache.hadoop.hbase.util.Bytes;
-109import 
org.apache.hadoop.hbase.util.Pair;
-110import 
org.apache.hadoop.io.BytesWritable;
-111import org.apache.hadoop.io.Writable;
-112import 
org.apache.hadoop.io.WritableUtils;
-113import 
org.apache.hadoop.io.compress.CompressionCodec;
-114import 
org.apache.hadoop.security.UserGroupInformation;
-115import 
org.apache.hadoop.security.UserGroupInformation.AuthenticationMethod;
-116import 
org.apache.hadoop.security.authorize.AuthorizationException;
-117import 
org.apache.hadoop.security.authorize.PolicyProvider;
-118import 
org.apache.hadoop.security.authorize.ProxyUsers;
-119import 
org.apache.hadoop.security.authorize.ServiceAuthorizationManager;
-120import 
org.apache.hadoop.security.token.SecretManager;
-121import 
org.apache.hadoop.security.token.SecretManager.InvalidToken;
-122import 
org.apache.hadoop.security.token.TokenIdentifier;
-123import 
org.apache.hadoop.util.StringUtils;
-124import org.apache.htrace.TraceInfo;
-125import 
org.codehaus.jackson.map.ObjectMapper;
-126
-127import 
com.google.common.annotations.VisibleForTesting;
-128
-129/**
-130 * An RPC server that hosts protobuf 
described Services.
-

[35/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-05-09 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/org/apache/hadoop/hbase/master/HMaster.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/master/HMaster.html 
b/devapidocs/org/apache/hadoop/hbase/master/HMaster.html
index bb60410..8967608 100644
--- a/devapidocs/org/apache/hadoop/hbase/master/HMaster.html
+++ b/devapidocs/org/apache/hadoop/hbase/master/HMaster.html
@@ -128,7 +128,7 @@ var activeTableTab = "activeTableTab";
 
 
 @InterfaceAudience.LimitedPrivate(value="Tools")
-public class HMaster
+public class HMaster
 extends HRegionServer
 implements MasterServices
 HMaster is the "master server" for HBase. An HBase cluster 
has one active
@@ -1382,7 +1382,7 @@ implements 
 
 LOG
-private static final org.apache.commons.logging.Log LOG
+private static final org.apache.commons.logging.Log LOG
 
 
 
@@ -1391,7 +1391,7 @@ implements 
 
 MASTER
-public static final http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String MASTER
+public static final http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String MASTER
 
 See Also:
 Constant
 Field Values
@@ -1404,7 +1404,7 @@ implements 
 
 activeMasterManager
-private final ActiveMasterManager activeMasterManager
+private final ActiveMasterManager activeMasterManager
 
 
 
@@ -1413,7 +1413,7 @@ implements 
 
 regionServerTracker
-RegionServerTracker regionServerTracker
+RegionServerTracker regionServerTracker
 
 
 
@@ -1422,7 +1422,7 @@ implements 
 
 drainingServerTracker
-private DrainingServerTracker 
drainingServerTracker
+private DrainingServerTracker 
drainingServerTracker
 
 
 
@@ -1431,7 +1431,7 @@ implements 
 
 loadBalancerTracker
-LoadBalancerTracker loadBalancerTracker
+LoadBalancerTracker loadBalancerTracker
 
 
 
@@ -1440,7 +1440,7 @@ implements 
 
 splitOrMergeTracker
-private SplitOrMergeTracker splitOrMergeTracker
+private SplitOrMergeTracker splitOrMergeTracker
 
 
 
@@ -1449,7 +1449,7 @@ implements 
 
 regionNormalizerTracker
-private RegionNormalizerTracker 
regionNormalizerTracker
+private RegionNormalizerTracker 
regionNormalizerTracker
 
 
 
@@ -1458,7 +1458,7 @@ implements 
 
 maintenanceModeTracker
-private MasterMaintenanceModeTracker maintenanceModeTracker
+private MasterMaintenanceModeTracker maintenanceModeTracker
 
 
 
@@ -1467,7 +1467,7 @@ implements 
 
 clusterSchemaService
-private ClusterSchemaService clusterSchemaService
+private ClusterSchemaService clusterSchemaService
 
 
 
@@ -1476,7 +1476,7 @@ implements 
 
 metricsMaster
-final MetricsMaster metricsMaster
+final MetricsMaster metricsMaster
 
 
 
@@ -1485,7 +1485,7 @@ implements 
 
 fileSystemManager
-private MasterFileSystem fileSystemManager
+private MasterFileSystem fileSystemManager
 
 
 
@@ -1494,7 +1494,7 @@ implements 
 
 walManager
-private MasterWalManager walManager
+private MasterWalManager walManager
 
 
 
@@ -1503,7 +1503,7 @@ implements 
 
 serverManager
-private volatile ServerManager serverManager
+private volatile ServerManager serverManager
 
 
 
@@ -1512,7 +1512,7 @@ implements 
 
 assignmentManager
-private AssignmentManager assignmentManager
+private AssignmentManager assignmentManager
 
 
 
@@ -1521,7 +1521,7 @@ implements 
 
 replicationManager
-private ReplicationManager replicationManager
+private ReplicationManager replicationManager
 
 
 
@@ -1530,7 +1530,7 @@ implements 
 
 rsFatals
-MemoryBoundedLogMessageBuffer rsFatals
+MemoryBoundedLogMessageBuffer rsFatals
 
 
 
@@ -1539,7 +1539,7 @@ implements 
 
 activeMaster
-private volatile boolean activeMaster
+private volatile boolean activeMaster
 
 
 
@@ -1548,7 +1548,7 @@ implements 
 
 initialized
-private final ProcedureEvent initialized
+private final ProcedureEvent initialized
 
 
 
@@ -1557,7 +1557,7 @@ implements 
 
 serviceStarted
-volatile boolean serviceStarted
+volatile boolean serviceStarted
 
 
 
@@ -1566,7 +1566,7 @@ implements 
 
 serverCrashProcessingEnabled
-private final ProcedureEvent serverCrashProcessingEnabled
+private final ProcedureEvent serverCrashProcessingEnabled
 
 
 
@@ -1575,7 +1575,7 @@ implements 
 
 maxBlancingTime
-private final int maxBlancingTime
+private final int maxBlancingTime
 
 
 
@@ -1584,7 +1584,7 @@ implements 
 
 maxRitPercent
-private final double maxRitPercent
+private final double maxRitPercent
 
 
 
@@ -1593,7 +1593,7 @@ implements 
 
 lockManager
-private final LockManager lockManager
+private final LockManager lockManager
 
 
 
@@ -1602,7 +1602,7 @@ implements 
 
 balancer
-private LoadBalancer balancer
+private LoadBalancer balancer
 
 
 
@@ -1611,7 +1611,7 @@ implements 
 
 normalizer
-private RegionNormalizer normalizer
+private RegionNormalizer normalizer
 
 
 
@@ -1620,7 +1620,7 @@ implements 
 
 balancerChore
-private BalancerChore balancerChore
+private BalancerChore balancerChore
 
 
 
@@ -1629,7 +1

[36/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-05-09 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/org/apache/hadoop/hbase/master/HMaster.InitializationMonitor.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/master/HMaster.InitializationMonitor.html 
b/devapidocs/org/apache/hadoop/hbase/master/HMaster.InitializationMonitor.html
index 0791ec1..6476d55 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/master/HMaster.InitializationMonitor.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/master/HMaster.InitializationMonitor.html
@@ -122,7 +122,7 @@ var activeTableTab = "activeTableTab";
 
 
 
-private static class HMaster.InitializationMonitor
+private static class HMaster.InitializationMonitor
 extends HasThread
 Protection against zombie master. Started once Master 
accepts active responsibility and
  starts taking over responsibilities. Allows a finite time window before 
giving up ownership.
@@ -250,7 +250,7 @@ extends 
 
 TIMEOUT_KEY
-public static final http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String TIMEOUT_KEY
+public static final http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String TIMEOUT_KEY
 The amount of time in milliseconds to sleep before checking 
initialization status.
 
 See Also:
@@ -264,7 +264,7 @@ extends 
 
 TIMEOUT_DEFAULT
-public static final long TIMEOUT_DEFAULT
+public static final long TIMEOUT_DEFAULT
 
 
 
@@ -273,7 +273,7 @@ extends 
 
 HALT_KEY
-public static final http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String HALT_KEY
+public static final http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String HALT_KEY
 When timeout expired and initialization has not complete, 
call http://docs.oracle.com/javase/8/docs/api/java/lang/System.html?is-external=true#exit-int-";
 title="class or interface in java.lang">System.exit(int) when
  true, do nothing otherwise.
 
@@ -288,7 +288,7 @@ extends 
 
 HALT_DEFAULT
-public static final boolean HALT_DEFAULT
+public static final boolean HALT_DEFAULT
 
 See Also:
 Constant
 Field Values
@@ -301,7 +301,7 @@ extends 
 
 master
-private final HMaster master
+private final HMaster master
 
 
 
@@ -310,7 +310,7 @@ extends 
 
 timeout
-private final long timeout
+private final long timeout
 
 
 
@@ -319,7 +319,7 @@ extends 
 
 haltOnTimeout
-private final boolean haltOnTimeout
+private final boolean haltOnTimeout
 
 
 
@@ -336,7 +336,7 @@ extends 
 
 InitializationMonitor
-InitializationMonitor(HMaster master)
+InitializationMonitor(HMaster master)
 Creates a Thread that monitors the HMaster.isInitialized()
 state.
 
 
@@ -354,7 +354,7 @@ extends 
 
 run
-public void run()
+public void run()
 
 Specified by:
 http://docs.oracle.com/javase/8/docs/api/java/lang/Runnable.html?is-external=true#run--";
 title="class or interface in java.lang">run in 
interface http://docs.oracle.com/javase/8/docs/api/java/lang/Runnable.html?is-external=true";
 title="class or interface in java.lang">Runnable

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/org/apache/hadoop/hbase/master/HMaster.PeriodicDoMetrics.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/master/HMaster.PeriodicDoMetrics.html 
b/devapidocs/org/apache/hadoop/hbase/master/HMaster.PeriodicDoMetrics.html
index 6537ea5..c1296d6 100644
--- a/devapidocs/org/apache/hadoop/hbase/master/HMaster.PeriodicDoMetrics.html
+++ b/devapidocs/org/apache/hadoop/hbase/master/HMaster.PeriodicDoMetrics.html
@@ -122,7 +122,7 @@ var activeTableTab = "activeTableTab";
 
 
 
-private static class HMaster.PeriodicDoMetrics
+private static class HMaster.PeriodicDoMetrics
 extends ScheduledChore
 
 
@@ -220,7 +220,7 @@ extends 
 
 server
-private final HMaster server
+private final HMaster server
 
 
 
@@ -237,7 +237,7 @@ extends 
 
 PeriodicDoMetrics
-public PeriodicDoMetrics(int doMetricsInterval,
+public PeriodicDoMetrics(int doMetricsInterval,
  HMaster server)
 
 
@@ -255,7 +255,7 @@ extends 
 
 chore
-protected void chore()
+protected void chore()
 Description copied from 
class: ScheduledChore
 The task to execute on each scheduled execution of the 
Chore
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/org/apache/hadoop/hbase/master/HMaster.RedirectServlet.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/master/HMaster.RedirectServlet.html 
b/devapidocs/org/apache/hadoop/hbase/master/HMaster.RedirectServlet.html
index a619b44..6ec0266 100644
--- a/devapidocs/org/apache/hadoop/hbase/master/HMaster.RedirectServlet.html
+++ b/devapidocs/org/apache

[37/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-05-09 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/org/apache/hadoop/hbase/ipc/class-use/RpcServer.Connection.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/ipc/class-use/RpcServer.Connection.html 
b/devapidocs/org/apache/hadoop/hbase/ipc/class-use/RpcServer.Connection.html
index 3856a94..b7d3911 100644
--- a/devapidocs/org/apache/hadoop/hbase/ipc/class-use/RpcServer.Connection.html
+++ b/devapidocs/org/apache/hadoop/hbase/ipc/class-use/RpcServer.Connection.html
@@ -129,7 +129,7 @@
 
 
 protected RpcServer.Connection
-RpcServer.Call.connection 
+ServerCall.connection 
 
 
 
@@ -141,7 +141,7 @@
 
 
 
-abstract RpcServer.Call
+abstract ServerCall
 RpcServer.Connection.createCall(int id,
   
org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService service,
   
org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor md,
@@ -156,7 +156,7 @@
   RpcServer.CallCleanup reqCleanup) 
 
 
-RpcServer.Call
+ServerCall
 SimpleRpcServer.Connection.createCall(int id,
   
org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService service,
   
org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor md,
@@ -171,7 +171,7 @@
   RpcServer.CallCleanup reqCleanup) 
 
 
-RpcServer.Call
+ServerCall
 NettyRpcServer.NettyConnection.createCall(int id,
   
org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService service,
   
org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor md,
@@ -194,47 +194,56 @@
 
 
 
-Call(int id,
-
org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService service,
-
org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor md,
-
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader header,
-org.apache.hadoop.hbase.shaded.com.google.protobuf.Message param,
-CellScanner cellScanner,
-RpcServer.Connection connection,
-long size,
-org.apache.htrace.TraceInfo tinfo,
-http://docs.oracle.com/javase/8/docs/api/java/net/InetAddress.html?is-external=true";
 title="class or interface in java.net">InetAddress remoteAddress,
-int timeout,
-RpcServer.CallCleanup reqCleanup) 
+NettyServerCall(int id,
+   
org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService service,
+   
org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor md,
+   
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader header,
+   
org.apache.hadoop.hbase.shaded.com.google.protobuf.Message param,
+   CellScanner cellScanner,
+   RpcServer.Connection connection,
+   long size,
+   org.apache.htrace.TraceInfo tinfo,
+   http://docs.oracle.com/javase/8/docs/api/java/net/InetAddress.html?is-external=true";
 title="class or interface in java.net">InetAddress remoteAddress,
+   long receiveTime,
+   int timeout,
+   ByteBufferPool reservoir,
+   CellBlockBuilder cellBlockBuilder,
+   RpcServer.CallCleanup reqCleanup) 
 
 
-Call(int id,
-
org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService service,
-
org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor md,
-
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader header,
-org.apache.hadoop.hbase.shaded.com.google.protobuf.Message param,
-CellScanner cellScanner,
-RpcServer.Connection connection,
-long size,
-org.apache.htrace.TraceInfo tinfo,
-http://docs.oracle.com/javase/8/docs/api/java/net/InetAddress.html?is-external=true";
 title="class or interface in java.net">InetAddress remoteAddress,
-int timeout,
-RpcServer.CallCleanup reqCleanup) 
+ServerCall(int id,
+  
org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService service,
+  
org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor md,
+  
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader header,
+  
org.apache.hadoop.hbase.shaded.com.google.protobuf.Message param,
+  CellScanner cellScanner,
+  RpcServer.Connection connection,
+  long size,
+  org.apache.htrace.TraceInfo tinfo,
+  http://docs.oracle.com/javase/8/docs/api/java/net/InetAddress.html?is-external=true";
 title="class or interface in java.net">InetAddress remoteAddress,
+  long receiveTime,
+  int timeout,
+  ByteBufferPool reservoir,
+  CellBlockBuilder cellBlockBuilder,
+  RpcServer.CallCleanup reqCleanup) 
 
 
-Call(int id,
-
org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService service,
-
org.apache.hadoop.hbase.shaded.com

[03/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-05-09 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/RSRpcServices.RegionScannerCloseCallBack.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/RSRpcServices.RegionScannerCloseCallBack.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/RSRpcServices.RegionScannerCloseCallBack.html
index 17d1bcb..6dfd1d4 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/RSRpcServices.RegionScannerCloseCallBack.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/RSRpcServices.RegionScannerCloseCallBack.html
@@ -35,863 +35,863 @@
 027import java.net.InetSocketAddress;
 028import java.net.UnknownHostException;
 029import java.nio.ByteBuffer;
-030import java.util.ArrayList;
-031import java.util.Collections;
-032import java.util.HashMap;
-033import java.util.Iterator;
-034import java.util.List;
-035import java.util.Map;
-036import java.util.Map.Entry;
-037import java.util.NavigableMap;
-038import java.util.Set;
-039import java.util.TreeSet;
-040import 
java.util.concurrent.ConcurrentHashMap;
-041import 
java.util.concurrent.ConcurrentMap;
-042import 
java.util.concurrent.atomic.AtomicLong;
-043import 
java.util.concurrent.atomic.LongAdder;
-044
-045import 
org.apache.commons.lang.mutable.MutableObject;
-046import org.apache.commons.logging.Log;
-047import 
org.apache.commons.logging.LogFactory;
-048import 
org.apache.hadoop.conf.Configuration;
-049import org.apache.hadoop.fs.Path;
-050import 
org.apache.hadoop.hbase.ByteBufferCell;
-051import org.apache.hadoop.hbase.Cell;
-052import 
org.apache.hadoop.hbase.CellScannable;
-053import 
org.apache.hadoop.hbase.CellScanner;
-054import 
org.apache.hadoop.hbase.CellUtil;
-055import 
org.apache.hadoop.hbase.DoNotRetryIOException;
-056import 
org.apache.hadoop.hbase.DroppedSnapshotException;
-057import 
org.apache.hadoop.hbase.HBaseIOException;
-058import 
org.apache.hadoop.hbase.HConstants;
-059import 
org.apache.hadoop.hbase.HRegionInfo;
-060import 
org.apache.hadoop.hbase.HTableDescriptor;
-061import 
org.apache.hadoop.hbase.MultiActionResultTooLarge;
-062import 
org.apache.hadoop.hbase.NotServingRegionException;
-063import 
org.apache.hadoop.hbase.ServerName;
-064import 
org.apache.hadoop.hbase.TableName;
-065import 
org.apache.hadoop.hbase.UnknownScannerException;
-066import 
org.apache.hadoop.hbase.classification.InterfaceAudience;
-067import 
org.apache.hadoop.hbase.client.Append;
-068import 
org.apache.hadoop.hbase.client.ConnectionUtils;
-069import 
org.apache.hadoop.hbase.client.Delete;
-070import 
org.apache.hadoop.hbase.client.Durability;
-071import 
org.apache.hadoop.hbase.client.Get;
-072import 
org.apache.hadoop.hbase.client.Increment;
-073import 
org.apache.hadoop.hbase.client.Mutation;
-074import 
org.apache.hadoop.hbase.client.Put;
-075import 
org.apache.hadoop.hbase.client.RegionReplicaUtil;
-076import 
org.apache.hadoop.hbase.client.Result;
-077import 
org.apache.hadoop.hbase.client.RowMutations;
-078import 
org.apache.hadoop.hbase.client.Scan;
-079import 
org.apache.hadoop.hbase.client.VersionInfoUtil;
-080import 
org.apache.hadoop.hbase.conf.ConfigurationObserver;
-081import 
org.apache.hadoop.hbase.exceptions.FailedSanityCheckException;
-082import 
org.apache.hadoop.hbase.exceptions.OutOfOrderScannerNextException;
-083import 
org.apache.hadoop.hbase.exceptions.ScannerResetException;
-084import 
org.apache.hadoop.hbase.filter.ByteArrayComparable;
-085import 
org.apache.hadoop.hbase.filter.CompareFilter.CompareOp;
-086import 
org.apache.hadoop.hbase.ipc.HBaseRPCErrorHandler;
-087import 
org.apache.hadoop.hbase.ipc.HBaseRpcController;
-088import 
org.apache.hadoop.hbase.ipc.PriorityFunction;
-089import 
org.apache.hadoop.hbase.ipc.QosPriority;
-090import 
org.apache.hadoop.hbase.ipc.RpcCallContext;
-091import 
org.apache.hadoop.hbase.ipc.RpcCallback;
-092import 
org.apache.hadoop.hbase.ipc.RpcServer;
-093import 
org.apache.hadoop.hbase.ipc.RpcServer.BlockingServiceAndInterface;
-094import 
org.apache.hadoop.hbase.ipc.RpcServerFactory;
-095import 
org.apache.hadoop.hbase.ipc.RpcServerInterface;
-096import 
org.apache.hadoop.hbase.ipc.ServerNotRunningYetException;
-097import 
org.apache.hadoop.hbase.ipc.ServerRpcController;
-098import 
org.apache.hadoop.hbase.master.MasterRpcServices;
-099import 
org.apache.hadoop.hbase.quotas.OperationQuota;
-100import 
org.apache.hadoop.hbase.quotas.RegionServerQuotaManager;
-101import 
org.apache.hadoop.hbase.regionserver.HRegion.RegionScannerImpl;
-102import 
org.apache.hadoop.hbase.regionserver.Leases.Lease;
-103import 
org.apache.hadoop.hbase.regionserver.Leases.LeaseStillHeldException;
-104import 
org.apache.hadoop.hbase.regionserver.Region.Operation;
-105import 
org.apache.hadoop.hbase.regionserver.ScannerContext.LimitScope;
-106import 
org.apache.hadoop.hbase.regionserver.handler.OpenMetaHandler;
-107import 
org.

[05/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-05-09 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.RedirectServlet.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.RedirectServlet.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.RedirectServlet.html
index 66fe094..17bc96c 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.RedirectServlet.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.RedirectServlet.html
@@ -57,3290 +57,3303 @@
 049import org.apache.commons.logging.Log;
 050import 
org.apache.commons.logging.LogFactory;
 051import 
org.apache.hadoop.conf.Configuration;
-052import org.apache.hadoop.fs.Path;
-053import 
org.apache.hadoop.hbase.ClusterStatus;
-054import 
org.apache.hadoop.hbase.CoordinatedStateException;
-055import 
org.apache.hadoop.hbase.CoordinatedStateManager;
-056import 
org.apache.hadoop.hbase.DoNotRetryIOException;
-057import 
org.apache.hadoop.hbase.HBaseIOException;
-058import 
org.apache.hadoop.hbase.HBaseInterfaceAudience;
-059import 
org.apache.hadoop.hbase.HColumnDescriptor;
-060import 
org.apache.hadoop.hbase.HConstants;
-061import 
org.apache.hadoop.hbase.HRegionInfo;
-062import 
org.apache.hadoop.hbase.HTableDescriptor;
-063import 
org.apache.hadoop.hbase.MasterNotRunningException;
-064import 
org.apache.hadoop.hbase.MetaTableAccessor;
-065import 
org.apache.hadoop.hbase.NamespaceDescriptor;
-066import 
org.apache.hadoop.hbase.PleaseHoldException;
-067import 
org.apache.hadoop.hbase.ProcedureInfo;
-068import 
org.apache.hadoop.hbase.ScheduledChore;
-069import 
org.apache.hadoop.hbase.ServerLoad;
-070import 
org.apache.hadoop.hbase.ServerName;
-071import 
org.apache.hadoop.hbase.TableDescriptors;
-072import 
org.apache.hadoop.hbase.TableName;
-073import 
org.apache.hadoop.hbase.TableNotDisabledException;
-074import 
org.apache.hadoop.hbase.TableNotFoundException;
-075import 
org.apache.hadoop.hbase.UnknownRegionException;
-076import 
org.apache.hadoop.hbase.classification.InterfaceAudience;
-077import 
org.apache.hadoop.hbase.client.MasterSwitchType;
-078import 
org.apache.hadoop.hbase.client.Result;
-079import 
org.apache.hadoop.hbase.client.TableState;
-080import 
org.apache.hadoop.hbase.coprocessor.BypassCoprocessorException;
-081import 
org.apache.hadoop.hbase.coprocessor.CoprocessorHost;
-082import 
org.apache.hadoop.hbase.exceptions.DeserializationException;
-083import 
org.apache.hadoop.hbase.exceptions.MergeRegionException;
-084import 
org.apache.hadoop.hbase.executor.ExecutorType;
-085import 
org.apache.hadoop.hbase.favored.FavoredNodesManager;
-086import 
org.apache.hadoop.hbase.favored.FavoredNodesPromoter;
-087import 
org.apache.hadoop.hbase.http.InfoServer;
-088import 
org.apache.hadoop.hbase.ipc.CoprocessorRpcUtils;
-089import 
org.apache.hadoop.hbase.ipc.RpcServer;
-090import 
org.apache.hadoop.hbase.ipc.ServerNotRunningYetException;
-091import 
org.apache.hadoop.hbase.master.MasterRpcServices.BalanceSwitchMode;
-092import 
org.apache.hadoop.hbase.master.balancer.BalancerChore;
-093import 
org.apache.hadoop.hbase.master.balancer.BaseLoadBalancer;
-094import 
org.apache.hadoop.hbase.master.balancer.ClusterStatusChore;
-095import 
org.apache.hadoop.hbase.master.balancer.LoadBalancerFactory;
-096import 
org.apache.hadoop.hbase.master.cleaner.HFileCleaner;
-097import 
org.apache.hadoop.hbase.master.cleaner.LogCleaner;
-098import 
org.apache.hadoop.hbase.master.cleaner.ReplicationMetaCleaner;
-099import 
org.apache.hadoop.hbase.master.cleaner.ReplicationZKNodeCleaner;
-100import 
org.apache.hadoop.hbase.master.cleaner.ReplicationZKNodeCleanerChore;
-101import 
org.apache.hadoop.hbase.master.locking.LockManager;
-102import 
org.apache.hadoop.hbase.master.normalizer.NormalizationPlan;
-103import 
org.apache.hadoop.hbase.master.normalizer.NormalizationPlan.PlanType;
-104import 
org.apache.hadoop.hbase.master.normalizer.RegionNormalizer;
-105import 
org.apache.hadoop.hbase.master.normalizer.RegionNormalizerChore;
-106import 
org.apache.hadoop.hbase.master.normalizer.RegionNormalizerFactory;
-107import 
org.apache.hadoop.hbase.master.procedure.AddColumnFamilyProcedure;
-108import 
org.apache.hadoop.hbase.master.procedure.CreateTableProcedure;
-109import 
org.apache.hadoop.hbase.master.procedure.DeleteColumnFamilyProcedure;
-110import 
org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure;
-111import 
org.apache.hadoop.hbase.master.procedure.DisableTableProcedure;
-112import 
org.apache.hadoop.hbase.master.procedure.EnableTableProcedure;
-113import 
org.apache.hadoop.hbase.master.procedure.MasterProcedureConstants;
-114import 
org.apache.hadoop.hbase.master.procedure.MasterProcedureEnv;
-115import 
org.apache.hadoop.hbase.master.procedure.MasterProcedureScheduler;
-116import 
org.apache.hadoop.hbase.master.procedure.MasterProcedureUtil;
-117import 
org.apache.hadoop.hbase.master.p

[51/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-05-09 Thread git-site-role
Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.


Project: http://git-wip-us.apache.org/repos/asf/hbase-site/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase-site/commit/7ef4c5a9
Tree: http://git-wip-us.apache.org/repos/asf/hbase-site/tree/7ef4c5a9
Diff: http://git-wip-us.apache.org/repos/asf/hbase-site/diff/7ef4c5a9

Branch: refs/heads/asf-site
Commit: 7ef4c5a9a2c4f8fd9b2456b7dfd0b36ab6e45ad4
Parents: baff481
Author: jenkins 
Authored: Tue May 9 15:00:02 2017 +
Committer: jenkins 
Committed: Tue May 9 15:00:02 2017 +

--
 acid-semantics.html | 4 +-
 apache_hbase_reference_guide.pdf| 4 +-
 apache_hbase_reference_guide.pdfmarks   | 4 +-
 bulk-loads.html | 4 +-
 checkstyle-aggregate.html   | 34006 +
 checkstyle.rss  |54 +-
 coc.html| 4 +-
 cygwin.html | 4 +-
 dependencies.html   | 4 +-
 dependency-convergence.html | 4 +-
 dependency-info.html| 4 +-
 dependency-management.html  | 4 +-
 devapidocs/allclasses-frame.html| 3 +
 devapidocs/allclasses-noframe.html  | 3 +
 devapidocs/constant-values.html | 6 +-
 devapidocs/index-all.html   |   238 +-
 .../hadoop/hbase/backup/package-tree.html   | 4 +-
 .../hadoop/hbase/class-use/CellScanner.html |   109 +-
 .../class-use/InterfaceAudience.Private.html|42 +-
 .../class-use/InterfaceStability.Evolving.html  |27 +-
 .../hbase/classification/package-tree.html  | 4 +-
 .../hadoop/hbase/client/package-tree.html   |18 +-
 .../hadoop/hbase/filter/package-tree.html   | 8 +-
 .../class-use/ByteBufferListOutputStream.html   | 2 +-
 .../hbase/io/class-use/ByteBufferPool.html  |64 +
 .../hadoop/hbase/io/hfile/package-tree.html | 6 +-
 .../org/apache/hadoop/hbase/ipc/CallRunner.html | 4 +-
 .../hadoop/hbase/ipc/NettyRpcServer.Call.html   |   395 -
 .../ipc/NettyRpcServer.CallWriteListener.html   |18 +-
 .../NettyRpcServer.ConnectionHeaderHandler.html | 8 +-
 .../hbase/ipc/NettyRpcServer.Initializer.html   | 8 +-
 .../ipc/NettyRpcServer.MessageDecoder.html  |16 +-
 .../ipc/NettyRpcServer.MessageEncoder.html  | 6 +-
 .../ipc/NettyRpcServer.NettyConnection.html |36 +-
 .../apache/hadoop/hbase/ipc/NettyRpcServer.html |55 +-
 .../hadoop/hbase/ipc/NettyServerCall.html   |   374 +
 .../hadoop/hbase/ipc/PriorityFunction.html  | 4 +-
 .../hadoop/hbase/ipc/RWQueueRpcExecutor.html| 4 +-
 .../org/apache/hadoop/hbase/ipc/RpcCall.html|50 +-
 .../apache/hadoop/hbase/ipc/RpcCallContext.html | 2 +-
 .../RpcServer.BlockingServiceAndInterface.html  |12 +-
 .../hbase/ipc/RpcServer.ByteBuffByteInput.html  |24 +-
 .../apache/hadoop/hbase/ipc/RpcServer.Call.html |  1453 -
 .../hadoop/hbase/ipc/RpcServer.CallCleanup.html | 8 +-
 .../hadoop/hbase/ipc/RpcServer.Connection.html  |   144 +-
 .../org/apache/hadoop/hbase/ipc/RpcServer.html  |   189 +-
 .../org/apache/hadoop/hbase/ipc/ServerCall.html |  1477 +
 .../hbase/ipc/ServerNotRunningYetException.html | 4 +-
 .../hadoop/hbase/ipc/SimpleRpcServer.Call.html  |   429 -
 .../hbase/ipc/SimpleRpcServer.Connection.html   |82 +-
 .../ipc/SimpleRpcServer.ConnectionManager.html  |40 +-
 .../ipc/SimpleRpcServer.Listener.Reader.html|14 +-
 .../hbase/ipc/SimpleRpcServer.Listener.html |30 +-
 .../hbase/ipc/SimpleRpcServer.Responder.html|38 +-
 .../hadoop/hbase/ipc/SimpleRpcServer.html   |45 +-
 .../hadoop/hbase/ipc/SimpleServerCall.html  |   420 +
 .../hbase/ipc/StoppedRpcClientException.html| 4 +-
 .../hadoop/hbase/ipc/class-use/BufferChain.html | 6 +-
 .../hbase/ipc/class-use/CellBlockBuilder.html   |56 +
 .../ipc/class-use/NettyRpcServer.Call.html  |   178 -
 .../NettyRpcServer.NettyConnection.html | 2 +-
 .../hbase/ipc/class-use/NettyServerCall.html|   178 +
 .../hadoop/hbase/ipc/class-use/RpcCall.html |16 +-
 .../hbase/ipc/class-use/RpcCallContext.html |16 +-
 .../hadoop/hbase/ipc/class-use/RpcCallback.html | 4 +-
 .../hbase/ipc/class-use/RpcServer.Call.html |   278 -
 .../ipc/class-use/RpcServer.CallCleanup.html|91 +-
 .../ipc/class-use/RpcServer.Connection.html |91 +-
 .../hadoop/hbase/ipc/class-use/ServerCall.html  |   278 +
 .../ipc/class-use/SimpleRpcServer.Call.html |   186 -
 .../class-use/SimpleRpcServer.Connection.html   | 2 +-
 .../class-use/SimpleRpcServer.Responder.html|35 +-
 .../hbase/ipc/class-use/SimpleServerCall.html   |   1

[46/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-05-09 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/org/apache/hadoop/hbase/ipc/NettyRpcServer.Call.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/ipc/NettyRpcServer.Call.html 
b/devapidocs/org/apache/hadoop/hbase/ipc/NettyRpcServer.Call.html
deleted file mode 100644
index 8e2f829..000
--- a/devapidocs/org/apache/hadoop/hbase/ipc/NettyRpcServer.Call.html
+++ /dev/null
@@ -1,395 +0,0 @@
-http://www.w3.org/TR/html4/loose.dtd";>
-
-
-
-
-
-NettyRpcServer.Call (Apache HBase 2.0.0-SNAPSHOT API)
-
-
-
-
-
-var methods = {"i0":10,"i1":10,"i2":10,"i3":10};
-var tabs = {65535:["t0","All Methods"],2:["t2","Instance 
Methods"],8:["t4","Concrete Methods"]};
-var altColor = "altColor";
-var rowColor = "rowColor";
-var tableTab = "tableTab";
-var activeTableTab = "activeTableTab";
-
-
-JavaScript is disabled on your browser.
-
-
-
-
-
-Skip navigation links
-
-
-
-
-Overview
-Package
-Class
-Use
-Tree
-Deprecated
-Index
-Help
-
-
-
-
-Prev Class
-Next Class
-
-
-Frames
-No Frames
-
-
-All Classes
-
-
-
-
-
-
-
-Summary: 
-Nested | 
-Field | 
-Constr | 
-Method
-
-
-Detail: 
-Field | 
-Constr | 
-Method
-
-
-
-
-
-
-
-
-org.apache.hadoop.hbase.ipc
-Class 
NettyRpcServer.Call
-
-
-
-http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true";
 title="class or interface in java.lang">java.lang.Object
-
-
-org.apache.hadoop.hbase.ipc.RpcServer.Call
-
-
-org.apache.hadoop.hbase.ipc.NettyRpcServer.Call
-
-
-
-
-
-
-
-
-
-All Implemented Interfaces:
-RpcCall, RpcCallContext
-
-
-Enclosing class:
-NettyRpcServer
-
-
-
-@InterfaceStability.Evolving
-public class NettyRpcServer.Call
-extends RpcServer.Call
-Datastructure that holds all necessary to a method 
invocation and then afterward, carries the
- result.
-
-
-
-
-
-
-
-
-
-
-
-Field Summary
-
-
-
-
-Fields inherited from class org.apache.hadoop.hbase.ipc.RpcServer.Call
-cellBlockStream,
 cellScanner,
 connection,
 deadline,
 header,
 id,
 isError,
 md,
 param,
 remoteAddress,
 reqCleanup, response,
 rpcCallback,
 service,
 size,
 startTime,
 timeout,
 timestamp,
 tinfo,
 user
-
-
-
-
-
-
-
-
-Constructor Summary
-
-Constructors 
-
-Constructor and Description
-
-
-Call(int id,
-
org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService service,
-
org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor md,
-
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader header,
-org.apache.hadoop.hbase.shaded.com.google.protobuf.Message param,
-CellScanner cellScanner,
-RpcServer.Connection connection,
-long size,
-org.apache.htrace.TraceInfo tinfo,
-http://docs.oracle.com/javase/8/docs/api/java/net/InetAddress.html?is-external=true";
 title="class or interface in java.net">InetAddress remoteAddress,
-int timeout,
-RpcServer.CallCleanup reqCleanup) 
-
-
-
-
-
-
-
-
-
-Method Summary
-
-All Methods Instance Methods Concrete Methods 
-
-Modifier and Type
-Method and Description
-
-
-long
-disconnectSince()
-Check if the caller who made this IPC call has 
disconnected.
-
-
-
-(package private) NettyRpcServer.NettyConnection
-getConnection() 
-
-
-void
-sendResponseIfReady()
-If we have a response, and delay is not set, then respond 
immediately.
-
-
-
-void
-sendResponseIfReady(io.netty.channel.ChannelFutureListener listener) 
-
-
-
-
-
-
-Methods inherited from class org.apache.hadoop.hbase.ipc.RpcServer.Call
-cleanup,
 createHeaderAndMessageBytes,
 done,
 getCellScanner,
 getClientVersionInfo,
 getDeadline,
 getHeader,
 getMethod, href="../../../../../org/apache/hadoop/hbase/ipc/RpcServer.Call.html#getParam--">getParam,
 > href="../../../../../org/apache/hadoop/hbase/ipc/RpcServer.Call.html#getPriority--">getPriority,
 > href="../../../../../org/apache/hadoop/hbase/ipc/RpcServer.Call.html#getReceiveTime--">getReceiveTime,
 > href="../../../../../org/apache/hadoop/hbase/ipc/RpcServer.Call.html#getRemoteAddress--">getRemoteAddress,
 > href="../../../../../org/apache/hadoop/hbase/ipc/RpcServer.Call.html#getRemotePort--">getRemotePort,
 > href="../../../../../org/apache/hadoop/hbase/ipc/RpcServer.Call.html#getRequestUser--">getRequestUser,
 > href="../../../../../org/apache/hadoop/hbase/ipc/RpcServer.Call.html#getRequestUserName--">getRequestUserName,
 > href="../../../../../org/apache/hadoop/hbase/ipc/RpcServer.Call.html#getResponseBlockSize--">getResponseBlockSize,
 > href="../../../../../org/apache/hadoop/hbase/ipc/RpcServer.Call.html#getRes

[15/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-05-09 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Call.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Call.html 
b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Call.html
deleted file mode 100644
index 2987e7b..000
--- a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Call.html
+++ /dev/null
@@ -1,1500 +0,0 @@
-http://www.w3.org/TR/html4/loose.dtd";>
-
-
-Source code
-
-
-
-
-001/**
-002 * Licensed to the Apache Software 
Foundation (ASF) under one
-003 * or more contributor license 
agreements.  See the NOTICE file
-004 * distributed with this work for 
additional information
-005 * regarding copyright ownership.  The 
ASF licenses this file
-006 * to you under the Apache License, 
Version 2.0 (the
-007 * "License"); you may not use this file 
except in compliance
-008 * with the License.  You may obtain a 
copy of the License at
-009 *
-010 * 
http://www.apache.org/licenses/LICENSE-2.0
-011 *
-012 * Unless required by applicable law or 
agreed to in writing, software
-013 * distributed under the License is 
distributed on an "AS IS" BASIS,
-014 * WITHOUT WARRANTIES OR CONDITIONS OF 
ANY KIND, either express or implied.
-015 * See the License for the specific 
language governing permissions and
-016 * limitations under the License.
-017 */
-018
-019package org.apache.hadoop.hbase.ipc;
-020
-021import java.io.IOException;
-022import java.io.InputStream;
-023import java.net.BindException;
-024import java.net.InetAddress;
-025import java.net.InetSocketAddress;
-026import java.net.ServerSocket;
-027import java.net.Socket;
-028import java.net.SocketException;
-029import java.net.UnknownHostException;
-030import java.nio.ByteBuffer;
-031import 
java.nio.channels.CancelledKeyException;
-032import 
java.nio.channels.ClosedChannelException;
-033import 
java.nio.channels.GatheringByteChannel;
-034import 
java.nio.channels.ReadableByteChannel;
-035import java.nio.channels.SelectionKey;
-036import java.nio.channels.Selector;
-037import 
java.nio.channels.ServerSocketChannel;
-038import java.nio.channels.SocketChannel;
-039import java.util.ArrayList;
-040import java.util.Arrays;
-041import java.util.Collections;
-042import java.util.Iterator;
-043import java.util.List;
-044import java.util.Set;
-045import java.util.Timer;
-046import java.util.TimerTask;
-047import 
java.util.concurrent.ConcurrentHashMap;
-048import 
java.util.concurrent.ConcurrentLinkedDeque;
-049import 
java.util.concurrent.ExecutorService;
-050import java.util.concurrent.Executors;
-051import 
java.util.concurrent.LinkedBlockingQueue;
-052import 
java.util.concurrent.atomic.AtomicInteger;
-053import 
java.util.concurrent.atomic.LongAdder;
-054import java.util.concurrent.locks.Lock;
-055import 
java.util.concurrent.locks.ReentrantLock;
-056
-057import 
org.apache.hadoop.conf.Configuration;
-058import 
org.apache.hadoop.hbase.CellScanner;
-059import 
org.apache.hadoop.hbase.DoNotRetryIOException;
-060import 
org.apache.hadoop.hbase.HBaseIOException;
-061import 
org.apache.hadoop.hbase.HConstants;
-062import org.apache.hadoop.hbase.Server;
-063import 
org.apache.hadoop.hbase.classification.InterfaceAudience;
-064import 
org.apache.hadoop.hbase.classification.InterfaceStability;
-065import 
org.apache.hadoop.hbase.client.VersionInfoUtil;
-066import 
org.apache.hadoop.hbase.exceptions.RequestTooBigException;
-067import 
org.apache.hadoop.hbase.monitoring.MonitoredRPCHandler;
-068import 
org.apache.hadoop.hbase.nio.ByteBuff;
-069import 
org.apache.hadoop.hbase.nio.SingleByteBuff;
-070import 
org.apache.hadoop.hbase.security.AccessDeniedException;
-071import 
org.apache.hadoop.hbase.security.AuthMethod;
-072import 
org.apache.hadoop.hbase.security.HBasePolicyProvider;
-073import 
org.apache.hadoop.hbase.security.SaslStatus;
-074import 
org.apache.hadoop.hbase.security.SaslUtil;
-075import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService;
-076import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream;
-077import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor;
-078import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.Message;
-079import 
org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
-080import 
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader;
-081import 
org.apache.hadoop.hbase.util.Bytes;
-082import 
org.apache.hadoop.hbase.util.Pair;
-083import 
org.apache.hadoop.hbase.util.Threads;
-084import org.apache.hadoop.io.IOUtils;
-085import 
org.apache.hadoop.io.IntWritable;
-086import 
org.apache.hadoop.security.authorize.ServiceAuthorizationManager;
-087import 
org.apache.hadoop.util.StringUtils;
-088import org.apache.htrace.TraceInfo;
-089
-090import 
com.google.common.util.concurrent.ThreadFactoryBuilder;
-091
-092/**
-093 * The 

[24/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-05-09 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.NettyConnection.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.NettyConnection.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.NettyConnection.html
index 27e0dee..109b5f3 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.NettyConnection.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.NettyConnection.html
@@ -67,157 +67,157 @@
 059import 
org.apache.hadoop.hbase.CellScanner;
 060import 
org.apache.hadoop.hbase.HConstants;
 061import org.apache.hadoop.hbase.Server;
-062import 
org.apache.hadoop.hbase.classification.InterfaceStability;
-063import 
org.apache.hadoop.hbase.monitoring.MonitoredRPCHandler;
-064import 
org.apache.hadoop.hbase.nio.ByteBuff;
-065import 
org.apache.hadoop.hbase.nio.SingleByteBuff;
-066import 
org.apache.hadoop.hbase.security.AccessDeniedException;
-067import 
org.apache.hadoop.hbase.security.AuthMethod;
-068import 
org.apache.hadoop.hbase.security.HBasePolicyProvider;
-069import 
org.apache.hadoop.hbase.security.SaslStatus;
-070import 
org.apache.hadoop.hbase.security.SaslUtil;
-071import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService;
-072import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor;
-073import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.Message;
-074import 
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader;
-075import 
org.apache.hadoop.hbase.util.Bytes;
-076import 
org.apache.hadoop.hbase.util.JVM;
-077import 
org.apache.hadoop.hbase.util.Pair;
-078import 
org.apache.hadoop.io.IntWritable;
-079import 
org.apache.hadoop.security.authorize.ServiceAuthorizationManager;
-080import org.apache.htrace.TraceInfo;
-081
-082/**
-083 * An RPC server with Netty4 
implementation.
-084 *
-085 */
-086public class NettyRpcServer extends 
RpcServer {
-087
-088  public static final Log LOG = 
LogFactory.getLog(NettyRpcServer.class);
-089
-090  protected final InetSocketAddress 
bindAddress;
-091
-092  private final CountDownLatch closed = 
new CountDownLatch(1);
-093  private final Channel serverChannel;
-094  private final ChannelGroup allChannels 
= new DefaultChannelGroup(GlobalEventExecutor.INSTANCE);;
-095
-096  public NettyRpcServer(final Server 
server, final String name,
-097  final 
List services,
-098  final InetSocketAddress 
bindAddress, Configuration conf,
-099  RpcScheduler scheduler) throws 
IOException {
-100super(server, name, services, 
bindAddress, conf, scheduler);
-101this.bindAddress = bindAddress;
-102boolean useEpoll = useEpoll(conf);
-103int workerCount = 
conf.getInt("hbase.netty.rpc.server.worker.count",
-104
Runtime.getRuntime().availableProcessors() / 4);
-105EventLoopGroup bossGroup = null;
-106EventLoopGroup workerGroup = null;
-107if (useEpoll) {
-108  bossGroup = new 
EpollEventLoopGroup(1);
-109  workerGroup = new 
EpollEventLoopGroup(workerCount);
-110} else {
-111  bossGroup = new 
NioEventLoopGroup(1);
-112  workerGroup = new 
NioEventLoopGroup(workerCount);
-113}
-114ServerBootstrap bootstrap = new 
ServerBootstrap();
-115bootstrap.group(bossGroup, 
workerGroup);
-116if (useEpoll) {
-117  
bootstrap.channel(EpollServerSocketChannel.class);
-118} else {
-119  
bootstrap.channel(NioServerSocketChannel.class);
-120}
-121
bootstrap.childOption(ChannelOption.TCP_NODELAY, tcpNoDelay);
-122
bootstrap.childOption(ChannelOption.SO_KEEPALIVE, tcpKeepAlive);
-123
bootstrap.childOption(ChannelOption.ALLOCATOR,
-124
PooledByteBufAllocator.DEFAULT);
-125bootstrap.childHandler(new 
Initializer(maxRequestSize));
-126
-127try {
-128  serverChannel = 
bootstrap.bind(this.bindAddress).sync().channel();
-129  LOG.info("NettyRpcServer bind to 
address=" + serverChannel.localAddress()
-130  + ", 
hbase.netty.rpc.server.worker.count=" + workerCount
-131  + ", useEpoll=" + useEpoll);
-132  allChannels.add(serverChannel);
-133} catch (InterruptedException e) {
-134  throw new 
InterruptedIOException(e.getMessage());
-135}
-136initReconfigurable(conf);
-137this.scheduler.init(new 
RpcSchedulerContext(this));
-138  }
-139
-140  private static boolean 
useEpoll(Configuration conf) {
-141// Config to enable native 
transport.
-142boolean epollEnabled = 
conf.getBoolean("hbase.rpc.server.nativetransport",
-143true);
-144// Use the faster native epoll 
transport mechanism on linux if enabled
-145return epollEnabled && 
JVM.isLinux() && JVM.isAmd64();
-146  }
-147
-148  @Override
-149  public synchronized void start() {
-150if (started) {
-151  return;
-152

[10/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-05-09 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Responder.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Responder.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Responder.html
index 2987e7b..7f61b54 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Responder.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Responder.html
@@ -135,1305 +135,1251 @@
 127  private Listener listener = null;
 128  protected Responder responder = null;
 129
-130  /**
-131   * Datastructure that holds all 
necessary to a method invocation and then afterward, carries
-132   * the result.
-133   */
-134  @InterfaceStability.Evolving
-135  public class Call extends 
RpcServer.Call {
-136
-137protected Responder responder;
+130  /** Listens on the socket. Creates jobs 
for the handler threads*/
+131  private class Listener extends Thread 
{
+132
+133private ServerSocketChannel 
acceptChannel = null; //the accept channel
+134private Selector selector = null; 
//the selector that we use for the server
+135private Reader[] readers = null;
+136private int currentReader = 0;
+137private final int 
readerPendingConnectionQueueLength;
 138
-139
@edu.umd.cs.findbugs.annotations.SuppressWarnings(value="NP_NULL_ON_SOME_PATH",
-140justification="Can't figure why 
this complaint is happening... see below")
-141Call(int id, final BlockingService 
service, final MethodDescriptor md,
-142RequestHeader header, Message 
param, CellScanner cellScanner,
-143RpcServer.Connection connection, 
long size, TraceInfo tinfo,
-144final InetAddress remoteAddress, 
int timeout, CallCleanup reqCleanup,
-145Responder responder) {
-146  super(id, service, md, header, 
param, cellScanner, connection, size,
-147  tinfo, remoteAddress, timeout, 
reqCleanup);
-148  this.responder = responder;
-149}
+139private ExecutorService readPool;
+140
+141public Listener(final String name) 
throws IOException {
+142  super(name);
+143  // The backlog of requests that we 
will have the serversocket carry.
+144  int backlogLength = 
conf.getInt("hbase.ipc.server.listen.queue.size", 128);
+145  readerPendingConnectionQueueLength 
=
+146  
conf.getInt("hbase.ipc.server.read.connection-queue.size", 100);
+147  // Create a new server socket and 
set to non blocking mode
+148  acceptChannel = 
ServerSocketChannel.open();
+149  
acceptChannel.configureBlocking(false);
 150
-151/**
-152 * Call is done. Execution happened 
and we returned results to client. It is now safe to
-153 * cleanup.
-154 */
-155
@edu.umd.cs.findbugs.annotations.SuppressWarnings(value="IS2_INCONSISTENT_SYNC",
-156justification="Presume the lock 
on processing request held by caller is protection enough")
-157@Override
-158void done() {
-159  super.done();
-160  this.getConnection().decRpcCount(); 
// Say that we're done with this call.
-161}
-162
-163@Override
-164public long disconnectSince() {
-165  if 
(!getConnection().isConnectionOpen()) {
-166return System.currentTimeMillis() 
- timestamp;
-167  } else {
-168return -1L;
-169  }
-170}
-171
-172@Override
-173public synchronized void 
sendResponseIfReady() throws IOException {
-174  // set param null to reduce memory 
pressure
-175  this.param = null;
-176  this.responder.doRespond(this);
+151  // Bind the server socket to the 
binding addrees (can be different from the default interface)
+152  bind(acceptChannel.socket(), 
bindAddress, backlogLength);
+153  port = 
acceptChannel.socket().getLocalPort(); //Could be an ephemeral port
+154  address = 
(InetSocketAddress)acceptChannel.socket().getLocalSocketAddress();
+155  // create a selector;
+156  selector = Selector.open();
+157
+158  readers = new 
Reader[readThreads];
+159  // Why this executor thing? Why not 
like hadoop just start up all the threads? I suppose it
+160  // has an advantage in that it is 
easy to shutdown the pool.
+161  readPool = 
Executors.newFixedThreadPool(readThreads,
+162new 
ThreadFactoryBuilder().setNameFormat(
+163  
"RpcServer.reader=%d,bindAddress=" + bindAddress.getHostName() +
+164  ",port=" + 
port).setDaemon(true)
+165
.setUncaughtExceptionHandler(Threads.LOGGING_EXCEPTION_HANDLER).build());
+166  for (int i = 0; i < readThreads; 
++i) {
+167Reader reader = new Reader();
+168readers[i] = reader;
+169readPool.execute(reader);
+170  }
+171  LOG.info(getName() + ": started " + 
readThreads + " reader(s) listening on port=" + port);
+172
+173  // Register accepts on th

[12/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-05-09 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Listener.Reader.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Listener.Reader.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Listener.Reader.html
index 2987e7b..7f61b54 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Listener.Reader.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Listener.Reader.html
@@ -135,1305 +135,1251 @@
 127  private Listener listener = null;
 128  protected Responder responder = null;
 129
-130  /**
-131   * Datastructure that holds all 
necessary to a method invocation and then afterward, carries
-132   * the result.
-133   */
-134  @InterfaceStability.Evolving
-135  public class Call extends 
RpcServer.Call {
-136
-137protected Responder responder;
+130  /** Listens on the socket. Creates jobs 
for the handler threads*/
+131  private class Listener extends Thread 
{
+132
+133private ServerSocketChannel 
acceptChannel = null; //the accept channel
+134private Selector selector = null; 
//the selector that we use for the server
+135private Reader[] readers = null;
+136private int currentReader = 0;
+137private final int 
readerPendingConnectionQueueLength;
 138
-139
@edu.umd.cs.findbugs.annotations.SuppressWarnings(value="NP_NULL_ON_SOME_PATH",
-140justification="Can't figure why 
this complaint is happening... see below")
-141Call(int id, final BlockingService 
service, final MethodDescriptor md,
-142RequestHeader header, Message 
param, CellScanner cellScanner,
-143RpcServer.Connection connection, 
long size, TraceInfo tinfo,
-144final InetAddress remoteAddress, 
int timeout, CallCleanup reqCleanup,
-145Responder responder) {
-146  super(id, service, md, header, 
param, cellScanner, connection, size,
-147  tinfo, remoteAddress, timeout, 
reqCleanup);
-148  this.responder = responder;
-149}
+139private ExecutorService readPool;
+140
+141public Listener(final String name) 
throws IOException {
+142  super(name);
+143  // The backlog of requests that we 
will have the serversocket carry.
+144  int backlogLength = 
conf.getInt("hbase.ipc.server.listen.queue.size", 128);
+145  readerPendingConnectionQueueLength 
=
+146  
conf.getInt("hbase.ipc.server.read.connection-queue.size", 100);
+147  // Create a new server socket and 
set to non blocking mode
+148  acceptChannel = 
ServerSocketChannel.open();
+149  
acceptChannel.configureBlocking(false);
 150
-151/**
-152 * Call is done. Execution happened 
and we returned results to client. It is now safe to
-153 * cleanup.
-154 */
-155
@edu.umd.cs.findbugs.annotations.SuppressWarnings(value="IS2_INCONSISTENT_SYNC",
-156justification="Presume the lock 
on processing request held by caller is protection enough")
-157@Override
-158void done() {
-159  super.done();
-160  this.getConnection().decRpcCount(); 
// Say that we're done with this call.
-161}
-162
-163@Override
-164public long disconnectSince() {
-165  if 
(!getConnection().isConnectionOpen()) {
-166return System.currentTimeMillis() 
- timestamp;
-167  } else {
-168return -1L;
-169  }
-170}
-171
-172@Override
-173public synchronized void 
sendResponseIfReady() throws IOException {
-174  // set param null to reduce memory 
pressure
-175  this.param = null;
-176  this.responder.doRespond(this);
+151  // Bind the server socket to the 
binding addrees (can be different from the default interface)
+152  bind(acceptChannel.socket(), 
bindAddress, backlogLength);
+153  port = 
acceptChannel.socket().getLocalPort(); //Could be an ephemeral port
+154  address = 
(InetSocketAddress)acceptChannel.socket().getLocalSocketAddress();
+155  // create a selector;
+156  selector = Selector.open();
+157
+158  readers = new 
Reader[readThreads];
+159  // Why this executor thing? Why not 
like hadoop just start up all the threads? I suppose it
+160  // has an advantage in that it is 
easy to shutdown the pool.
+161  readPool = 
Executors.newFixedThreadPool(readThreads,
+162new 
ThreadFactoryBuilder().setNameFormat(
+163  
"RpcServer.reader=%d,bindAddress=" + bindAddress.getHostName() +
+164  ",port=" + 
port).setDaemon(true)
+165
.setUncaughtExceptionHandler(Threads.LOGGING_EXCEPTION_HANDLER).build());
+166  for (int i = 0; i < readThreads; 
++i) {
+167Reader reader = new Reader();
+168readers[i] = reader;
+169readPool.execute(reader);
+170  }
+171  LOG.info(getName() + ": started " + 
readThreads + " reader(s) listening on port=" + port);
+172
+173 

[31/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-05-09 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/org/apache/hadoop/hbase/tmpl/regionserver/RSStatusTmpl.ImplData.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/tmpl/regionserver/RSStatusTmpl.ImplData.html
 
b/devapidocs/org/apache/hadoop/hbase/tmpl/regionserver/RSStatusTmpl.ImplData.html
index 19716c1..6034ffe 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/tmpl/regionserver/RSStatusTmpl.ImplData.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/tmpl/regionserver/RSStatusTmpl.ImplData.html
@@ -293,76 +293,76 @@ extends org.jamon.AbstractTemplateProxy.ImplData
 private HRegionServer m_regionServer
 
 
-
+
 
 
 
 
-m_format
-private http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String m_format
+m_filter
+private http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String m_filter
 
 
-
+
 
 
 
 
-m_format__IsNotDefault
-private boolean m_format__IsNotDefault
+m_filter__IsNotDefault
+private boolean m_filter__IsNotDefault
 
 
-
+
 
 
 
 
-m_bcv
-private http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String m_bcv
+m_bcn
+private http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String m_bcn
 
 
-
+
 
 
 
 
-m_bcv__IsNotDefault
-private boolean m_bcv__IsNotDefault
+m_bcn__IsNotDefault
+private boolean m_bcn__IsNotDefault
 
 
-
+
 
 
 
 
-m_bcn
-private http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String m_bcn
+m_format
+private http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String m_format
 
 
-
+
 
 
 
 
-m_bcn__IsNotDefault
-private boolean m_bcn__IsNotDefault
+m_format__IsNotDefault
+private boolean m_format__IsNotDefault
 
 
-
+
 
 
 
 
-m_filter
-private http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String m_filter
+m_bcv
+private http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String m_bcv
 
 
-
+
 
 
 
 
-m_filter__IsNotDefault
-private boolean m_filter__IsNotDefault
+m_bcv__IsNotDefault
+private boolean m_bcv__IsNotDefault
 
 
 
@@ -408,112 +408,112 @@ extends org.jamon.AbstractTemplateProxy.ImplData
 public HRegionServer getRegionServer()
 
 
-
+
 
 
 
 
-setFormat
-public void setFormat(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String format)
+setFilter
+public void setFilter(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String filter)
 
 
-
+
 
 
 
 
-getFormat
-public http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String getFormat()
+getFilter
+public http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String getFilter()
 
 
-
+
 
 
 
 
-getFormat__IsNotDefault
-public boolean getFormat__IsNotDefault()
+getFilter__IsNotDefault
+public boolean getFilter__IsNotDefault()
 
 
-
+
 
 
 
 
-setBcv
-public void setBcv(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String bcv)
+setBcn
+public void setBcn(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String bcn)
 
 
-
+
 
 
 
 
-getBcv
-public http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String getBcv()
+getBcn
+public http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String getBcn()
 
 
-
+
 
 
 
 
-getBcv__IsNotDefault
-public boolean getBcv__IsNotDefault()
+getBcn__IsNotDefault
+public boolean getBcn__IsNotDefault()
 
 
-
+
 
 
 
 
-setBcn
-public void setBcn(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String bcn)
+setFormat
+public void setFormat(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String format)
 
 
-
+
 
 
 
 
-getBcn
-public http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String getBcn()
+getFormat
+public http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">

[23/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-05-09 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.html 
b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.html
index 27e0dee..109b5f3 100644
--- a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.html
+++ b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.html
@@ -67,157 +67,157 @@
 059import 
org.apache.hadoop.hbase.CellScanner;
 060import 
org.apache.hadoop.hbase.HConstants;
 061import org.apache.hadoop.hbase.Server;
-062import 
org.apache.hadoop.hbase.classification.InterfaceStability;
-063import 
org.apache.hadoop.hbase.monitoring.MonitoredRPCHandler;
-064import 
org.apache.hadoop.hbase.nio.ByteBuff;
-065import 
org.apache.hadoop.hbase.nio.SingleByteBuff;
-066import 
org.apache.hadoop.hbase.security.AccessDeniedException;
-067import 
org.apache.hadoop.hbase.security.AuthMethod;
-068import 
org.apache.hadoop.hbase.security.HBasePolicyProvider;
-069import 
org.apache.hadoop.hbase.security.SaslStatus;
-070import 
org.apache.hadoop.hbase.security.SaslUtil;
-071import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService;
-072import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor;
-073import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.Message;
-074import 
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader;
-075import 
org.apache.hadoop.hbase.util.Bytes;
-076import 
org.apache.hadoop.hbase.util.JVM;
-077import 
org.apache.hadoop.hbase.util.Pair;
-078import 
org.apache.hadoop.io.IntWritable;
-079import 
org.apache.hadoop.security.authorize.ServiceAuthorizationManager;
-080import org.apache.htrace.TraceInfo;
-081
-082/**
-083 * An RPC server with Netty4 
implementation.
-084 *
-085 */
-086public class NettyRpcServer extends 
RpcServer {
-087
-088  public static final Log LOG = 
LogFactory.getLog(NettyRpcServer.class);
-089
-090  protected final InetSocketAddress 
bindAddress;
-091
-092  private final CountDownLatch closed = 
new CountDownLatch(1);
-093  private final Channel serverChannel;
-094  private final ChannelGroup allChannels 
= new DefaultChannelGroup(GlobalEventExecutor.INSTANCE);;
-095
-096  public NettyRpcServer(final Server 
server, final String name,
-097  final 
List services,
-098  final InetSocketAddress 
bindAddress, Configuration conf,
-099  RpcScheduler scheduler) throws 
IOException {
-100super(server, name, services, 
bindAddress, conf, scheduler);
-101this.bindAddress = bindAddress;
-102boolean useEpoll = useEpoll(conf);
-103int workerCount = 
conf.getInt("hbase.netty.rpc.server.worker.count",
-104
Runtime.getRuntime().availableProcessors() / 4);
-105EventLoopGroup bossGroup = null;
-106EventLoopGroup workerGroup = null;
-107if (useEpoll) {
-108  bossGroup = new 
EpollEventLoopGroup(1);
-109  workerGroup = new 
EpollEventLoopGroup(workerCount);
-110} else {
-111  bossGroup = new 
NioEventLoopGroup(1);
-112  workerGroup = new 
NioEventLoopGroup(workerCount);
-113}
-114ServerBootstrap bootstrap = new 
ServerBootstrap();
-115bootstrap.group(bossGroup, 
workerGroup);
-116if (useEpoll) {
-117  
bootstrap.channel(EpollServerSocketChannel.class);
-118} else {
-119  
bootstrap.channel(NioServerSocketChannel.class);
-120}
-121
bootstrap.childOption(ChannelOption.TCP_NODELAY, tcpNoDelay);
-122
bootstrap.childOption(ChannelOption.SO_KEEPALIVE, tcpKeepAlive);
-123
bootstrap.childOption(ChannelOption.ALLOCATOR,
-124
PooledByteBufAllocator.DEFAULT);
-125bootstrap.childHandler(new 
Initializer(maxRequestSize));
-126
-127try {
-128  serverChannel = 
bootstrap.bind(this.bindAddress).sync().channel();
-129  LOG.info("NettyRpcServer bind to 
address=" + serverChannel.localAddress()
-130  + ", 
hbase.netty.rpc.server.worker.count=" + workerCount
-131  + ", useEpoll=" + useEpoll);
-132  allChannels.add(serverChannel);
-133} catch (InterruptedException e) {
-134  throw new 
InterruptedIOException(e.getMessage());
-135}
-136initReconfigurable(conf);
-137this.scheduler.init(new 
RpcSchedulerContext(this));
-138  }
-139
-140  private static boolean 
useEpoll(Configuration conf) {
-141// Config to enable native 
transport.
-142boolean epollEnabled = 
conf.getBoolean("hbase.rpc.server.nativetransport",
-143true);
-144// Use the faster native epoll 
transport mechanism on linux if enabled
-145return epollEnabled && 
JVM.isLinux() && JVM.isAmd64();
-146  }
-147
-148  @Override
-149  public synchronized void start() {
-150if (started) {
-151  return;
-152}
-153authTokenSecretMgr = 
createSecretManager();
-154if (authTokenSecretM

[26/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-05-09 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.MessageDecoder.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.MessageDecoder.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.MessageDecoder.html
index 27e0dee..109b5f3 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.MessageDecoder.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.MessageDecoder.html
@@ -67,157 +67,157 @@
 059import 
org.apache.hadoop.hbase.CellScanner;
 060import 
org.apache.hadoop.hbase.HConstants;
 061import org.apache.hadoop.hbase.Server;
-062import 
org.apache.hadoop.hbase.classification.InterfaceStability;
-063import 
org.apache.hadoop.hbase.monitoring.MonitoredRPCHandler;
-064import 
org.apache.hadoop.hbase.nio.ByteBuff;
-065import 
org.apache.hadoop.hbase.nio.SingleByteBuff;
-066import 
org.apache.hadoop.hbase.security.AccessDeniedException;
-067import 
org.apache.hadoop.hbase.security.AuthMethod;
-068import 
org.apache.hadoop.hbase.security.HBasePolicyProvider;
-069import 
org.apache.hadoop.hbase.security.SaslStatus;
-070import 
org.apache.hadoop.hbase.security.SaslUtil;
-071import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService;
-072import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor;
-073import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.Message;
-074import 
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader;
-075import 
org.apache.hadoop.hbase.util.Bytes;
-076import 
org.apache.hadoop.hbase.util.JVM;
-077import 
org.apache.hadoop.hbase.util.Pair;
-078import 
org.apache.hadoop.io.IntWritable;
-079import 
org.apache.hadoop.security.authorize.ServiceAuthorizationManager;
-080import org.apache.htrace.TraceInfo;
-081
-082/**
-083 * An RPC server with Netty4 
implementation.
-084 *
-085 */
-086public class NettyRpcServer extends 
RpcServer {
-087
-088  public static final Log LOG = 
LogFactory.getLog(NettyRpcServer.class);
-089
-090  protected final InetSocketAddress 
bindAddress;
-091
-092  private final CountDownLatch closed = 
new CountDownLatch(1);
-093  private final Channel serverChannel;
-094  private final ChannelGroup allChannels 
= new DefaultChannelGroup(GlobalEventExecutor.INSTANCE);;
-095
-096  public NettyRpcServer(final Server 
server, final String name,
-097  final 
List services,
-098  final InetSocketAddress 
bindAddress, Configuration conf,
-099  RpcScheduler scheduler) throws 
IOException {
-100super(server, name, services, 
bindAddress, conf, scheduler);
-101this.bindAddress = bindAddress;
-102boolean useEpoll = useEpoll(conf);
-103int workerCount = 
conf.getInt("hbase.netty.rpc.server.worker.count",
-104
Runtime.getRuntime().availableProcessors() / 4);
-105EventLoopGroup bossGroup = null;
-106EventLoopGroup workerGroup = null;
-107if (useEpoll) {
-108  bossGroup = new 
EpollEventLoopGroup(1);
-109  workerGroup = new 
EpollEventLoopGroup(workerCount);
-110} else {
-111  bossGroup = new 
NioEventLoopGroup(1);
-112  workerGroup = new 
NioEventLoopGroup(workerCount);
-113}
-114ServerBootstrap bootstrap = new 
ServerBootstrap();
-115bootstrap.group(bossGroup, 
workerGroup);
-116if (useEpoll) {
-117  
bootstrap.channel(EpollServerSocketChannel.class);
-118} else {
-119  
bootstrap.channel(NioServerSocketChannel.class);
-120}
-121
bootstrap.childOption(ChannelOption.TCP_NODELAY, tcpNoDelay);
-122
bootstrap.childOption(ChannelOption.SO_KEEPALIVE, tcpKeepAlive);
-123
bootstrap.childOption(ChannelOption.ALLOCATOR,
-124
PooledByteBufAllocator.DEFAULT);
-125bootstrap.childHandler(new 
Initializer(maxRequestSize));
-126
-127try {
-128  serverChannel = 
bootstrap.bind(this.bindAddress).sync().channel();
-129  LOG.info("NettyRpcServer bind to 
address=" + serverChannel.localAddress()
-130  + ", 
hbase.netty.rpc.server.worker.count=" + workerCount
-131  + ", useEpoll=" + useEpoll);
-132  allChannels.add(serverChannel);
-133} catch (InterruptedException e) {
-134  throw new 
InterruptedIOException(e.getMessage());
-135}
-136initReconfigurable(conf);
-137this.scheduler.init(new 
RpcSchedulerContext(this));
-138  }
-139
-140  private static boolean 
useEpoll(Configuration conf) {
-141// Config to enable native 
transport.
-142boolean epollEnabled = 
conf.getBoolean("hbase.rpc.server.nativetransport",
-143true);
-144// Use the faster native epoll 
transport mechanism on linux if enabled
-145return epollEnabled && 
JVM.isLinux() && JVM.isAmd64();
-146  }
-147
-148  @Override
-149  public synchronized void start() {
-150if (started) {
-151  return;
-152}
-15

[33/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-05-09 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/org/apache/hadoop/hbase/regionserver/RSRpcServices.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/RSRpcServices.html 
b/devapidocs/org/apache/hadoop/hbase/regionserver/RSRpcServices.html
index d98892a..c1b5dd1 100644
--- a/devapidocs/org/apache/hadoop/hbase/regionserver/RSRpcServices.html
+++ b/devapidocs/org/apache/hadoop/hbase/regionserver/RSRpcServices.html
@@ -118,7 +118,7 @@ var activeTableTab = "activeTableTab";
 
 
 @InterfaceAudience.Private
-public class RSRpcServices
+public class RSRpcServices
 extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true";
 title="class or interface in java.lang">Object
 implements HBaseRPCErrorHandler, 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.AdminService.BlockingInterface,
 
org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.ClientService.BlockingInterface,
 PriorityFunction, ConfigurationObserver
 Implements the regionserver RPC services.
@@ -826,7 +826,7 @@ implements 
 
 LOG
-protected static final org.apache.commons.logging.Log LOG
+protected static final org.apache.commons.logging.Log LOG
 
 
 
@@ -835,7 +835,7 @@ implements 
 
 REGION_SERVER_RPC_SCHEDULER_FACTORY_CLASS
-public static final http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String REGION_SERVER_RPC_SCHEDULER_FACTORY_CLASS
+public static final http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String REGION_SERVER_RPC_SCHEDULER_FACTORY_CLASS
 RPC scheduler to use for the region server.
 
 See Also:
@@ -849,7 +849,7 @@ implements 
 
 REGION_SERVER_RPC_MINIMUM_SCAN_TIME_LIMIT_DELTA
-private static final http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String REGION_SERVER_RPC_MINIMUM_SCAN_TIME_LIMIT_DELTA
+private static final http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String REGION_SERVER_RPC_MINIMUM_SCAN_TIME_LIMIT_DELTA
 Minimum allowable time limit delta (in milliseconds) that 
can be enforced during scans. This
  configuration exists to prevent the scenario where a time limit is specified 
to be so
  restrictive that the time limit is reached immediately (before any cells are 
scanned).
@@ -865,7 +865,7 @@ implements 
 
 DEFAULT_REGION_SERVER_RPC_MINIMUM_SCAN_TIME_LIMIT_DELTA
-private static final long DEFAULT_REGION_SERVER_RPC_MINIMUM_SCAN_TIME_LIMIT_DELTA
+private static final long DEFAULT_REGION_SERVER_RPC_MINIMUM_SCAN_TIME_LIMIT_DELTA
 Default value of REGION_SERVER_RPC_MINIMUM_SCAN_TIME_LIMIT_DELTA
 
 See Also:
@@ -879,7 +879,7 @@ implements 
 
 requestCount
-final http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/atomic/LongAdder.html?is-external=true";
 title="class or interface in java.util.concurrent.atomic">LongAdder requestCount
+final http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/atomic/LongAdder.html?is-external=true";
 title="class or interface in java.util.concurrent.atomic">LongAdder requestCount
 
 
 
@@ -888,7 +888,7 @@ implements 
 
 rpcGetRequestCount
-final http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/atomic/LongAdder.html?is-external=true";
 title="class or interface in java.util.concurrent.atomic">LongAdder rpcGetRequestCount
+final http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/atomic/LongAdder.html?is-external=true";
 title="class or interface in java.util.concurrent.atomic">LongAdder rpcGetRequestCount
 
 
 
@@ -897,7 +897,7 @@ implements 
 
 rpcScanRequestCount
-final http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/atomic/LongAdder.html?is-external=true";
 title="class or interface in java.util.concurrent.atomic">LongAdder rpcScanRequestCount
+final http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/atomic/LongAdder.html?is-external=true";
 title="class or interface in java.util.concurrent.atomic">LongAdder rpcScanRequestCount
 
 
 
@@ -906,7 +906,7 @@ implements 
 
 rpcMultiRequestCount
-final http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/atomic/LongAdder.html?is-external=true";
 title="class or interface in java.util.concurrent.atomic">LongAdder rpcMultiRequestCount
+final http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/atomic/LongAdder.html?is-external=true";
 title="class or interface in java.util.concurrent.atomic">LongAdder rpcMultiRequestCount
 
 
 
@@ -915,7 +915,7 @@ implements 
 
 rpcMutateRequestCount
-final http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/atomic/LongAdder.html?is-external=true";
 title="class or interface in java.util.concurrent.atomic">LongAdder rpcMutateRequestCount
+final http://docs.oracl

[14/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-05-09 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Connection.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Connection.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Connection.html
index 2987e7b..7f61b54 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Connection.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Connection.html
@@ -135,1305 +135,1251 @@
 127  private Listener listener = null;
 128  protected Responder responder = null;
 129
-130  /**
-131   * Datastructure that holds all 
necessary to a method invocation and then afterward, carries
-132   * the result.
-133   */
-134  @InterfaceStability.Evolving
-135  public class Call extends 
RpcServer.Call {
-136
-137protected Responder responder;
+130  /** Listens on the socket. Creates jobs 
for the handler threads*/
+131  private class Listener extends Thread 
{
+132
+133private ServerSocketChannel 
acceptChannel = null; //the accept channel
+134private Selector selector = null; 
//the selector that we use for the server
+135private Reader[] readers = null;
+136private int currentReader = 0;
+137private final int 
readerPendingConnectionQueueLength;
 138
-139
@edu.umd.cs.findbugs.annotations.SuppressWarnings(value="NP_NULL_ON_SOME_PATH",
-140justification="Can't figure why 
this complaint is happening... see below")
-141Call(int id, final BlockingService 
service, final MethodDescriptor md,
-142RequestHeader header, Message 
param, CellScanner cellScanner,
-143RpcServer.Connection connection, 
long size, TraceInfo tinfo,
-144final InetAddress remoteAddress, 
int timeout, CallCleanup reqCleanup,
-145Responder responder) {
-146  super(id, service, md, header, 
param, cellScanner, connection, size,
-147  tinfo, remoteAddress, timeout, 
reqCleanup);
-148  this.responder = responder;
-149}
+139private ExecutorService readPool;
+140
+141public Listener(final String name) 
throws IOException {
+142  super(name);
+143  // The backlog of requests that we 
will have the serversocket carry.
+144  int backlogLength = 
conf.getInt("hbase.ipc.server.listen.queue.size", 128);
+145  readerPendingConnectionQueueLength 
=
+146  
conf.getInt("hbase.ipc.server.read.connection-queue.size", 100);
+147  // Create a new server socket and 
set to non blocking mode
+148  acceptChannel = 
ServerSocketChannel.open();
+149  
acceptChannel.configureBlocking(false);
 150
-151/**
-152 * Call is done. Execution happened 
and we returned results to client. It is now safe to
-153 * cleanup.
-154 */
-155
@edu.umd.cs.findbugs.annotations.SuppressWarnings(value="IS2_INCONSISTENT_SYNC",
-156justification="Presume the lock 
on processing request held by caller is protection enough")
-157@Override
-158void done() {
-159  super.done();
-160  this.getConnection().decRpcCount(); 
// Say that we're done with this call.
-161}
-162
-163@Override
-164public long disconnectSince() {
-165  if 
(!getConnection().isConnectionOpen()) {
-166return System.currentTimeMillis() 
- timestamp;
-167  } else {
-168return -1L;
-169  }
-170}
-171
-172@Override
-173public synchronized void 
sendResponseIfReady() throws IOException {
-174  // set param null to reduce memory 
pressure
-175  this.param = null;
-176  this.responder.doRespond(this);
+151  // Bind the server socket to the 
binding addrees (can be different from the default interface)
+152  bind(acceptChannel.socket(), 
bindAddress, backlogLength);
+153  port = 
acceptChannel.socket().getLocalPort(); //Could be an ephemeral port
+154  address = 
(InetSocketAddress)acceptChannel.socket().getLocalSocketAddress();
+155  // create a selector;
+156  selector = Selector.open();
+157
+158  readers = new 
Reader[readThreads];
+159  // Why this executor thing? Why not 
like hadoop just start up all the threads? I suppose it
+160  // has an advantage in that it is 
easy to shutdown the pool.
+161  readPool = 
Executors.newFixedThreadPool(readThreads,
+162new 
ThreadFactoryBuilder().setNameFormat(
+163  
"RpcServer.reader=%d,bindAddress=" + bindAddress.getHostName() +
+164  ",port=" + 
port).setDaemon(true)
+165
.setUncaughtExceptionHandler(Threads.LOGGING_EXCEPTION_HANDLER).build());
+166  for (int i = 0; i < readThreads; 
++i) {
+167Reader reader = new Reader();
+168readers[i] = reader;
+169readPool.execute(reader);
+170  }
+171  LOG.info(getName() + ": started " + 
readThreads + " reader(s) listening on port=" + port);
+172
+173  // Register accepts 

[39/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-05-09 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Listener.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Listener.html 
b/devapidocs/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Listener.html
index 85f0adc..5c85b2d 100644
--- a/devapidocs/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Listener.html
+++ b/devapidocs/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Listener.html
@@ -122,7 +122,7 @@ var activeTableTab = "activeTableTab";
 
 
 
-private class SimpleRpcServer.Listener
+private class SimpleRpcServer.Listener
 extends http://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html?is-external=true";
 title="class or interface in java.lang">Thread
 Listens on the socket. Creates jobs for the handler 
threads
 
@@ -296,7 +296,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html?
 
 
 acceptChannel
-private http://docs.oracle.com/javase/8/docs/api/java/nio/channels/ServerSocketChannel.html?is-external=true";
 title="class or interface in java.nio.channels">ServerSocketChannel acceptChannel
+private http://docs.oracle.com/javase/8/docs/api/java/nio/channels/ServerSocketChannel.html?is-external=true";
 title="class or interface in java.nio.channels">ServerSocketChannel acceptChannel
 
 
 
@@ -305,7 +305,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html?
 
 
 selector
-private http://docs.oracle.com/javase/8/docs/api/java/nio/channels/Selector.html?is-external=true";
 title="class or interface in java.nio.channels">Selector selector
+private http://docs.oracle.com/javase/8/docs/api/java/nio/channels/Selector.html?is-external=true";
 title="class or interface in java.nio.channels">Selector selector
 
 
 
@@ -314,7 +314,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html?
 
 
 readers
-private SimpleRpcServer.Listener.Reader[] readers
+private SimpleRpcServer.Listener.Reader[] readers
 
 
 
@@ -323,7 +323,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html?
 
 
 currentReader
-private int currentReader
+private int currentReader
 
 
 
@@ -332,7 +332,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html?
 
 
 readerPendingConnectionQueueLength
-private final int readerPendingConnectionQueueLength
+private final int readerPendingConnectionQueueLength
 
 
 
@@ -341,7 +341,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html?
 
 
 readPool
-private http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/ExecutorService.html?is-external=true";
 title="class or interface in java.util.concurrent">ExecutorService readPool
+private http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/ExecutorService.html?is-external=true";
 title="class or interface in java.util.concurrent">ExecutorService readPool
 
 
 
@@ -358,7 +358,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html?
 
 
 Listener
-public Listener(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String name)
+public Listener(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String name)
  throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true";
 title="class or interface in java.io">IOException
 
 Throws:
@@ -380,7 +380,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html?
 
 
 run
-public void run()
+public void run()
 
 Specified by:
 http://docs.oracle.com/javase/8/docs/api/java/lang/Runnable.html?is-external=true#run--";
 title="class or interface in java.lang">run in 
interface http://docs.oracle.com/javase/8/docs/api/java/lang/Runnable.html?is-external=true";
 title="class or interface in java.lang">Runnable
@@ -395,7 +395,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html?
 
 
 closeCurrentConnection
-private void closeCurrentConnection(http://docs.oracle.com/javase/8/docs/api/java/nio/channels/SelectionKey.html?is-external=true";
 title="class or interface in java.nio.channels">SelectionKey key,
+private void closeCurrentConnection(http://docs.oracle.com/javase/8/docs/api/java/nio/channels/SelectionKey.html?is-external=true";
 title="class or interface in java.nio.channels">SelectionKey key,
 http://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true";
 title="class or interface in java.lang">Throwable e)
 
 
@@ -405,7 +405,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html?
 
 
 getAddress
-http://docs.oracle.com/javase/8/docs/api/java/net/InetSocketAddress.html?is-external=true";
 title="class or interface in java.net">InetSocketAddress getAddress()
+http://docs.oracle.com/javase/8/docs/api/java/net/

[02/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-05-09 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/RSRpcServices.RegionScannerHolder.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/RSRpcServices.RegionScannerHolder.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/RSRpcServices.RegionScannerHolder.html
index 17d1bcb..6dfd1d4 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/RSRpcServices.RegionScannerHolder.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/RSRpcServices.RegionScannerHolder.html
@@ -35,863 +35,863 @@
 027import java.net.InetSocketAddress;
 028import java.net.UnknownHostException;
 029import java.nio.ByteBuffer;
-030import java.util.ArrayList;
-031import java.util.Collections;
-032import java.util.HashMap;
-033import java.util.Iterator;
-034import java.util.List;
-035import java.util.Map;
-036import java.util.Map.Entry;
-037import java.util.NavigableMap;
-038import java.util.Set;
-039import java.util.TreeSet;
-040import 
java.util.concurrent.ConcurrentHashMap;
-041import 
java.util.concurrent.ConcurrentMap;
-042import 
java.util.concurrent.atomic.AtomicLong;
-043import 
java.util.concurrent.atomic.LongAdder;
-044
-045import 
org.apache.commons.lang.mutable.MutableObject;
-046import org.apache.commons.logging.Log;
-047import 
org.apache.commons.logging.LogFactory;
-048import 
org.apache.hadoop.conf.Configuration;
-049import org.apache.hadoop.fs.Path;
-050import 
org.apache.hadoop.hbase.ByteBufferCell;
-051import org.apache.hadoop.hbase.Cell;
-052import 
org.apache.hadoop.hbase.CellScannable;
-053import 
org.apache.hadoop.hbase.CellScanner;
-054import 
org.apache.hadoop.hbase.CellUtil;
-055import 
org.apache.hadoop.hbase.DoNotRetryIOException;
-056import 
org.apache.hadoop.hbase.DroppedSnapshotException;
-057import 
org.apache.hadoop.hbase.HBaseIOException;
-058import 
org.apache.hadoop.hbase.HConstants;
-059import 
org.apache.hadoop.hbase.HRegionInfo;
-060import 
org.apache.hadoop.hbase.HTableDescriptor;
-061import 
org.apache.hadoop.hbase.MultiActionResultTooLarge;
-062import 
org.apache.hadoop.hbase.NotServingRegionException;
-063import 
org.apache.hadoop.hbase.ServerName;
-064import 
org.apache.hadoop.hbase.TableName;
-065import 
org.apache.hadoop.hbase.UnknownScannerException;
-066import 
org.apache.hadoop.hbase.classification.InterfaceAudience;
-067import 
org.apache.hadoop.hbase.client.Append;
-068import 
org.apache.hadoop.hbase.client.ConnectionUtils;
-069import 
org.apache.hadoop.hbase.client.Delete;
-070import 
org.apache.hadoop.hbase.client.Durability;
-071import 
org.apache.hadoop.hbase.client.Get;
-072import 
org.apache.hadoop.hbase.client.Increment;
-073import 
org.apache.hadoop.hbase.client.Mutation;
-074import 
org.apache.hadoop.hbase.client.Put;
-075import 
org.apache.hadoop.hbase.client.RegionReplicaUtil;
-076import 
org.apache.hadoop.hbase.client.Result;
-077import 
org.apache.hadoop.hbase.client.RowMutations;
-078import 
org.apache.hadoop.hbase.client.Scan;
-079import 
org.apache.hadoop.hbase.client.VersionInfoUtil;
-080import 
org.apache.hadoop.hbase.conf.ConfigurationObserver;
-081import 
org.apache.hadoop.hbase.exceptions.FailedSanityCheckException;
-082import 
org.apache.hadoop.hbase.exceptions.OutOfOrderScannerNextException;
-083import 
org.apache.hadoop.hbase.exceptions.ScannerResetException;
-084import 
org.apache.hadoop.hbase.filter.ByteArrayComparable;
-085import 
org.apache.hadoop.hbase.filter.CompareFilter.CompareOp;
-086import 
org.apache.hadoop.hbase.ipc.HBaseRPCErrorHandler;
-087import 
org.apache.hadoop.hbase.ipc.HBaseRpcController;
-088import 
org.apache.hadoop.hbase.ipc.PriorityFunction;
-089import 
org.apache.hadoop.hbase.ipc.QosPriority;
-090import 
org.apache.hadoop.hbase.ipc.RpcCallContext;
-091import 
org.apache.hadoop.hbase.ipc.RpcCallback;
-092import 
org.apache.hadoop.hbase.ipc.RpcServer;
-093import 
org.apache.hadoop.hbase.ipc.RpcServer.BlockingServiceAndInterface;
-094import 
org.apache.hadoop.hbase.ipc.RpcServerFactory;
-095import 
org.apache.hadoop.hbase.ipc.RpcServerInterface;
-096import 
org.apache.hadoop.hbase.ipc.ServerNotRunningYetException;
-097import 
org.apache.hadoop.hbase.ipc.ServerRpcController;
-098import 
org.apache.hadoop.hbase.master.MasterRpcServices;
-099import 
org.apache.hadoop.hbase.quotas.OperationQuota;
-100import 
org.apache.hadoop.hbase.quotas.RegionServerQuotaManager;
-101import 
org.apache.hadoop.hbase.regionserver.HRegion.RegionScannerImpl;
-102import 
org.apache.hadoop.hbase.regionserver.Leases.Lease;
-103import 
org.apache.hadoop.hbase.regionserver.Leases.LeaseStillHeldException;
-104import 
org.apache.hadoop.hbase.regionserver.Region.Operation;
-105import 
org.apache.hadoop.hbase.regionserver.ScannerContext.LimitScope;
-106import 
org.apache.hadoop.hbase.regionserver.handler.OpenMetaHandler;
-107import 
org.apache.hadoop.hbase.regionserver.ha

[44/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-05-09 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/org/apache/hadoop/hbase/ipc/RpcServer.Call.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/ipc/RpcServer.Call.html 
b/devapidocs/org/apache/hadoop/hbase/ipc/RpcServer.Call.html
deleted file mode 100644
index eff339f..000
--- a/devapidocs/org/apache/hadoop/hbase/ipc/RpcServer.Call.html
+++ /dev/null
@@ -1,1453 +0,0 @@
-http://www.w3.org/TR/html4/loose.dtd";>
-
-
-
-
-
-RpcServer.Call (Apache HBase 2.0.0-SNAPSHOT API)
-
-
-
-
-
-var methods = 
{"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":10,"i33":10,"i34":10,"i35":10,"i36":10,"i37":10,"i38":10,"i39":10};
-var tabs = {65535:["t0","All Methods"],2:["t2","Instance 
Methods"],8:["t4","Concrete Methods"]};
-var altColor = "altColor";
-var rowColor = "rowColor";
-var tableTab = "tableTab";
-var activeTableTab = "activeTableTab";
-
-
-JavaScript is disabled on your browser.
-
-
-
-
-
-Skip navigation links
-
-
-
-
-Overview
-Package
-Class
-Use
-Tree
-Deprecated
-Index
-Help
-
-
-
-
-Prev Class
-Next Class
-
-
-Frames
-No Frames
-
-
-All Classes
-
-
-
-
-
-
-
-Summary: 
-Nested | 
-Field | 
-Constr | 
-Method
-
-
-Detail: 
-Field | 
-Constr | 
-Method
-
-
-
-
-
-
-
-
-org.apache.hadoop.hbase.ipc
-Class RpcServer.Call
-
-
-
-http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true";
 title="class or interface in java.lang">java.lang.Object
-
-
-org.apache.hadoop.hbase.ipc.RpcServer.Call
-
-
-
-
-
-
-
-All Implemented Interfaces:
-RpcCall, RpcCallContext
-
-
-Direct Known Subclasses:
-NettyRpcServer.Call, SimpleRpcServer.Call
-
-
-Enclosing class:
-RpcServer
-
-
-
-@InterfaceStability.Evolving
- @InterfaceAudience.Private
-public abstract class RpcServer.Call
-extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true";
 title="class or interface in java.lang">Object
-implements RpcCall
-Datastructure that holds all necessary to a method 
invocation and then afterward, carries
- the result.
-
-
-
-
-
-
-
-
-
-
-
-Field Summary
-
-Fields 
-
-Modifier and Type
-Field and Description
-
-
-protected ByteBufferListOutputStream
-cellBlockStream 
-
-
-protected CellScanner
-cellScanner 
-
-
-protected RpcServer.Connection
-connection 
-
-
-protected long
-deadline 
-
-
-private long
-exceptionSize 
-
-
-protected 
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader
-header 
-
-
-protected int
-id 
-
-
-protected boolean
-isError 
-
-
-protected 
org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor
-md 
-
-
-protected 
org.apache.hadoop.hbase.shaded.com.google.protobuf.Message
-param 
-
-
-protected http://docs.oracle.com/javase/8/docs/api/java/net/InetAddress.html?is-external=true";
 title="class or interface in java.net">InetAddress
-remoteAddress 
-
-
-protected RpcServer.CallCleanup
-reqCleanup 
-
-
-protected BufferChain
-response
-Chain of buffers to send as response.
-
-
-
-private long
-responseBlockSize 
-
-
-private long
-responseCellSize 
-
-
-private boolean
-retryImmediatelySupported 
-
-
-protected RpcCallback
-rpcCallback 
-
-
-protected 
org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService
-service 
-
-
-protected long
-size 
-
-
-protected long
-startTime 
-
-
-protected int
-timeout 
-
-
-protected long
-timestamp 
-
-
-protected org.apache.htrace.TraceInfo
-tinfo 
-
-
-protected User
-user 
-
-
-
-
-
-
-
-
-
-Constructor Summary
-
-Constructors 
-
-Constructor and Description
-
-
-Call(int id,
-
org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService service,
-
org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor md,
-
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader header,
-org.apache.hadoop.hbase.shaded.com.google.protobuf.Message param,
-CellScanner cellScanner,
-RpcServer.Connection connection,
-long size,
-org.apache.htrace.TraceInfo tinfo,
-http://docs.oracle.com/javase/8/docs/api/java/net/InetAddress.html?is-external=true";
 title="class or interface in java.net">InetAddress remoteAddress,
-int timeout,
-RpcServer.CallCleanup reqCleanup) 
-
-
-
-
-
-
-
-
-
-Method Summary
-
-All Methods Instance Methods Concrete Methods 

[34/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-05-09 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/org/apache/hadoop/hbase/master/package-tree.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/master/package-tree.html 
b/devapidocs/org/apache/hadoop/hbase/master/package-tree.html
index a96a542..c18daae 100644
--- a/devapidocs/org/apache/hadoop/hbase/master/package-tree.html
+++ b/devapidocs/org/apache/hadoop/hbase/master/package-tree.html
@@ -322,11 +322,11 @@
 
 java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true";
 title="class or interface in java.lang">Enum (implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true";
 title="class or interface in java.lang">Comparable, java.io.http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true";
 title="class or interface in java.io">Serializable)
 
-org.apache.hadoop.hbase.master.RegionState.State
-org.apache.hadoop.hbase.master.MetricsMasterSourceFactoryImpl.FactoryStorage
-org.apache.hadoop.hbase.master.MasterRpcServices.BalanceSwitchMode
-org.apache.hadoop.hbase.master.SplitLogManager.TerminationStatus
 org.apache.hadoop.hbase.master.SplitLogManager.ResubmitDirective
+org.apache.hadoop.hbase.master.SplitLogManager.TerminationStatus
+org.apache.hadoop.hbase.master.MasterRpcServices.BalanceSwitchMode
+org.apache.hadoop.hbase.master.MetricsMasterSourceFactoryImpl.FactoryStorage
+org.apache.hadoop.hbase.master.RegionState.State
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/org/apache/hadoop/hbase/master/procedure/package-tree.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/master/procedure/package-tree.html 
b/devapidocs/org/apache/hadoop/hbase/master/procedure/package-tree.html
index 4460f73..ff790b4 100644
--- a/devapidocs/org/apache/hadoop/hbase/master/procedure/package-tree.html
+++ b/devapidocs/org/apache/hadoop/hbase/master/procedure/package-tree.html
@@ -174,8 +174,8 @@
 
 java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true";
 title="class or interface in java.lang">Enum (implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true";
 title="class or interface in java.lang">Comparable, java.io.http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true";
 title="class or interface in java.io">Serializable)
 
-org.apache.hadoop.hbase.master.procedure.TableProcedureInterface.TableOperationType
 org.apache.hadoop.hbase.master.procedure.ServerProcedureInterface.ServerOperationType
+org.apache.hadoop.hbase.master.procedure.TableProcedureInterface.TableOperationType
 org.apache.hadoop.hbase.master.procedure.DisableTableProcedure.MarkRegionOfflineOpResult
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/org/apache/hadoop/hbase/package-tree.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/package-tree.html 
b/devapidocs/org/apache/hadoop/hbase/package-tree.html
index 2cc1064..d7279ca 100644
--- a/devapidocs/org/apache/hadoop/hbase/package-tree.html
+++ b/devapidocs/org/apache/hadoop/hbase/package-tree.html
@@ -422,16 +422,16 @@
 
 java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true";
 title="class or interface in java.lang">Enum (implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true";
 title="class or interface in java.lang">Comparable, java.io.http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true";
 title="class or interface in java.io">Serializable)
 
-org.apache.hadoop.hbase.KeyValue.Type
+org.apache.hadoop.hbase.HConstants.Modify
+org.apache.hadoop.hbase.CompatibilitySingletonFactory.SingletonStorage
 org.apache.hadoop.hbase.KeepDeletedCells
+org.apache.hadoop.hbase.HConstants.OperationStatusCode
+org.apache.hadoop.hbase.Coprocessor.State
 org.apache.hadoop.hbase.MetaTableAccessor.QueryType
 org.apache.hadoop.hbase.MemoryCompactionPolicy
-org.apache.hadoop.hbase.HConstants.OperationStatusCode
 org.apache.hadoop.hbase.HealthChecker.HealthCheckerExitStatus
-org.apache.hadoop.hbase.HConstants.Modify
 org.apache.hadoop.hbase.ProcedureState
-org.apache.hadoop.hbase.CompatibilitySingletonFactory.SingletonStorage
-org.apache.hadoop.hbase.Coprocessor.State
+org.apache.hadoop.hbase.KeyValue.Type
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/org/apache/hadoop/hbase/procedure2/package-tree.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/procedure2/package-tree.html 
b/devapidocs/org/apache/hadoop/hbase/procedure2/package-tree.html
index cbe5806..2992239 1

[40/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-05-09 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/org/apache/hadoop/hbase/ipc/ServerNotRunningYetException.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/ipc/ServerNotRunningYetException.html 
b/devapidocs/org/apache/hadoop/hbase/ipc/ServerNotRunningYetException.html
index b113a5a..63453fa 100644
--- a/devapidocs/org/apache/hadoop/hbase/ipc/ServerNotRunningYetException.html
+++ b/devapidocs/org/apache/hadoop/hbase/ipc/ServerNotRunningYetException.html
@@ -43,7 +43,7 @@
 
 
 
-Prev Class
+Prev Class
 Next Class
 
 
@@ -223,7 +223,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/io/IOException.ht
 
 
 
-Prev Class
+Prev Class
 Next Class
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Call.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Call.html 
b/devapidocs/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Call.html
deleted file mode 100644
index 59c6329..000
--- a/devapidocs/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Call.html
+++ /dev/null
@@ -1,429 +0,0 @@
-http://www.w3.org/TR/html4/loose.dtd";>
-
-
-
-
-
-SimpleRpcServer.Call (Apache HBase 2.0.0-SNAPSHOT API)
-
-
-
-
-
-var methods = {"i0":10,"i1":10,"i2":10,"i3":10};
-var tabs = {65535:["t0","All Methods"],2:["t2","Instance 
Methods"],8:["t4","Concrete Methods"]};
-var altColor = "altColor";
-var rowColor = "rowColor";
-var tableTab = "tableTab";
-var activeTableTab = "activeTableTab";
-
-
-JavaScript is disabled on your browser.
-
-
-
-
-
-Skip navigation links
-
-
-
-
-Overview
-Package
-Class
-Use
-Tree
-Deprecated
-Index
-Help
-
-
-
-
-Prev Class
-Next Class
-
-
-Frames
-No Frames
-
-
-All Classes
-
-
-
-
-
-
-
-Summary: 
-Nested | 
-Field | 
-Constr | 
-Method
-
-
-Detail: 
-Field | 
-Constr | 
-Method
-
-
-
-
-
-
-
-
-org.apache.hadoop.hbase.ipc
-Class 
SimpleRpcServer.Call
-
-
-
-http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true";
 title="class or interface in java.lang">java.lang.Object
-
-
-org.apache.hadoop.hbase.ipc.RpcServer.Call
-
-
-org.apache.hadoop.hbase.ipc.SimpleRpcServer.Call
-
-
-
-
-
-
-
-
-
-All Implemented Interfaces:
-RpcCall, RpcCallContext
-
-
-Enclosing class:
-SimpleRpcServer
-
-
-
-@InterfaceStability.Evolving
-public class SimpleRpcServer.Call
-extends RpcServer.Call
-Datastructure that holds all necessary to a method 
invocation and then afterward, carries
- the result.
-
-
-
-
-
-
-
-
-
-
-
-Field Summary
-
-Fields 
-
-Modifier and Type
-Field and Description
-
-
-protected SimpleRpcServer.Responder
-responder 
-
-
-
-
-
-
-Fields inherited from class org.apache.hadoop.hbase.ipc.RpcServer.Call
-cellBlockStream,
 cellScanner,
 connection,
 deadline,
 header,
 id,
 isError,
 md,
 param,
 remoteAddress,
 reqCleanup, response,
 rpcCallback,
 service,
 size,
 startTime,
 timeout,
 timestamp,
 tinfo,
 user
-
-
-
-
-
-
-
-
-Constructor Summary
-
-Constructors 
-
-Constructor and Description
-
-
-Call(int id,
-
org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService service,
-
org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor md,
-
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader header,
-org.apache.hadoop.hbase.shaded.com.google.protobuf.Message param,
-CellScanner cellScanner,
-RpcServer.Connection connection,
-long size,
-org.apache.htrace.TraceInfo tinfo,
-http://docs.oracle.com/javase/8/docs/api/java/net/InetAddress.html?is-external=true";
 title="class or interface in java.net">InetAddress remoteAddress,
-int timeout,
-RpcServer.CallCleanup reqCleanup,
-SimpleRpcServer.Responder responder) 
-
-
-
-
-
-
-
-
-
-Method Summary
-
-All Methods Instance Methods Concrete Methods 
-
-Modifier and Type
-Method and Description
-
-
-long
-disconnectSince()
-Check if the caller who made this IPC call has 
disconnected.
-
-
-
-(package private) void
-done()
-Call is done.
-
-
-
-(package private) SimpleRpcServer.Connection
-getConnection() 
-
-
-void
-sendResponseIfReady()
-Send the response of this RPC call.
-
-
-
-
-
-
-
-Methods inherited from class org.apache.hadoop.hbase.ipc.RpcServer.Call
-cleanup,
 createHeaderAndMessageBytes,
 getCellScanner,
 getClientVersionInfo,
 getDeadline,
 getHeader,
 getMethod,
 get
 Param, getPriority,
 getReceiveTime,
 getRemoteAddress,
 getRemotePort

[18/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-05-09 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.Connection.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.Connection.html 
b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.Connection.html
index 5cc356a..9e1c66c 100644
--- a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.Connection.html
+++ b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.Connection.html
@@ -70,2037 +70,1559 @@
 062import 
org.apache.hadoop.hbase.client.VersionInfoUtil;
 063import 
org.apache.hadoop.hbase.codec.Codec;
 064import 
org.apache.hadoop.hbase.conf.ConfigurationObserver;
-065import 
org.apache.hadoop.hbase.exceptions.RegionMovedException;
-066import 
org.apache.hadoop.hbase.exceptions.RequestTooBigException;
-067import 
org.apache.hadoop.hbase.io.ByteBufferListOutputStream;
-068import 
org.apache.hadoop.hbase.io.ByteBufferOutputStream;
-069import 
org.apache.hadoop.hbase.io.ByteBufferPool;
-070import 
org.apache.hadoop.hbase.io.crypto.aes.CryptoAES;
-071import 
org.apache.hadoop.hbase.monitoring.MonitoredRPCHandler;
-072import 
org.apache.hadoop.hbase.monitoring.TaskMonitor;
-073import 
org.apache.hadoop.hbase.nio.ByteBuff;
-074import 
org.apache.hadoop.hbase.nio.MultiByteBuff;
-075import 
org.apache.hadoop.hbase.nio.SingleByteBuff;
-076import 
org.apache.hadoop.hbase.regionserver.RSRpcServices;
-077import 
org.apache.hadoop.hbase.security.AccessDeniedException;
-078import 
org.apache.hadoop.hbase.security.AuthMethod;
-079import 
org.apache.hadoop.hbase.security.HBaseSaslRpcServer;
-080import 
org.apache.hadoop.hbase.security.HBaseSaslRpcServer.SaslDigestCallbackHandler;
-081import 
org.apache.hadoop.hbase.security.HBaseSaslRpcServer.SaslGssCallbackHandler;
-082import 
org.apache.hadoop.hbase.security.SaslStatus;
-083import 
org.apache.hadoop.hbase.security.SaslUtil;
-084import 
org.apache.hadoop.hbase.security.User;
-085import 
org.apache.hadoop.hbase.security.UserProvider;
-086import 
org.apache.hadoop.hbase.security.token.AuthenticationTokenSecretManager;
-087import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService;
-088import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteInput;
-089import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString;
-090import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream;
-091import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream;
-092import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor;
-093import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.Message;
-094import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException;
-095import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.TextFormat;
-096import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.UnsafeByteOperations;
-097import 
org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
-098import 
org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos;
-099import 
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.VersionInfo;
-100import 
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos;
-101import 
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.CellBlockMeta;
-102import 
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ConnectionHeader;
-103import 
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ExceptionResponse;
-104import 
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader;
-105import 
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ResponseHeader;
-106import 
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation;
-107import 
org.apache.hadoop.hbase.util.ByteBufferUtils;
-108import 
org.apache.hadoop.hbase.util.Bytes;
-109import 
org.apache.hadoop.hbase.util.Pair;
-110import 
org.apache.hadoop.io.BytesWritable;
-111import org.apache.hadoop.io.Writable;
-112import 
org.apache.hadoop.io.WritableUtils;
-113import 
org.apache.hadoop.io.compress.CompressionCodec;
-114import 
org.apache.hadoop.security.UserGroupInformation;
-115import 
org.apache.hadoop.security.UserGroupInformation.AuthenticationMethod;
-116import 
org.apache.hadoop.security.authorize.AuthorizationException;
-117import 
org.apache.hadoop.security.authorize.PolicyProvider;
-118import 
org.apache.hadoop.security.authorize.ProxyUsers;
-119import 
org.apache.hadoop.security.authorize.ServiceAuthorizationManager;
-120import 
org.apache.hadoop.security.token.SecretManager;
-121import 
org.apache.hadoop.security.token.SecretManager.InvalidToken;
-122import 
org.apache.hadoop.security.token.TokenIdentifier;
-123import 
org.apache.hadoop.util.StringUtils;
-124import org.apache.htrace.TraceInfo;
-125import 
org.codehaus.jackson.map.ObjectMapper;
-126
-127import 
com.google.common.annotations.VisibleForTesting;
-128
-129/**
-130 * 

[25/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-05-09 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.MessageEncoder.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.MessageEncoder.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.MessageEncoder.html
index 27e0dee..109b5f3 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.MessageEncoder.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.MessageEncoder.html
@@ -67,157 +67,157 @@
 059import 
org.apache.hadoop.hbase.CellScanner;
 060import 
org.apache.hadoop.hbase.HConstants;
 061import org.apache.hadoop.hbase.Server;
-062import 
org.apache.hadoop.hbase.classification.InterfaceStability;
-063import 
org.apache.hadoop.hbase.monitoring.MonitoredRPCHandler;
-064import 
org.apache.hadoop.hbase.nio.ByteBuff;
-065import 
org.apache.hadoop.hbase.nio.SingleByteBuff;
-066import 
org.apache.hadoop.hbase.security.AccessDeniedException;
-067import 
org.apache.hadoop.hbase.security.AuthMethod;
-068import 
org.apache.hadoop.hbase.security.HBasePolicyProvider;
-069import 
org.apache.hadoop.hbase.security.SaslStatus;
-070import 
org.apache.hadoop.hbase.security.SaslUtil;
-071import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService;
-072import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor;
-073import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.Message;
-074import 
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader;
-075import 
org.apache.hadoop.hbase.util.Bytes;
-076import 
org.apache.hadoop.hbase.util.JVM;
-077import 
org.apache.hadoop.hbase.util.Pair;
-078import 
org.apache.hadoop.io.IntWritable;
-079import 
org.apache.hadoop.security.authorize.ServiceAuthorizationManager;
-080import org.apache.htrace.TraceInfo;
-081
-082/**
-083 * An RPC server with Netty4 
implementation.
-084 *
-085 */
-086public class NettyRpcServer extends 
RpcServer {
-087
-088  public static final Log LOG = 
LogFactory.getLog(NettyRpcServer.class);
-089
-090  protected final InetSocketAddress 
bindAddress;
-091
-092  private final CountDownLatch closed = 
new CountDownLatch(1);
-093  private final Channel serverChannel;
-094  private final ChannelGroup allChannels 
= new DefaultChannelGroup(GlobalEventExecutor.INSTANCE);;
-095
-096  public NettyRpcServer(final Server 
server, final String name,
-097  final 
List services,
-098  final InetSocketAddress 
bindAddress, Configuration conf,
-099  RpcScheduler scheduler) throws 
IOException {
-100super(server, name, services, 
bindAddress, conf, scheduler);
-101this.bindAddress = bindAddress;
-102boolean useEpoll = useEpoll(conf);
-103int workerCount = 
conf.getInt("hbase.netty.rpc.server.worker.count",
-104
Runtime.getRuntime().availableProcessors() / 4);
-105EventLoopGroup bossGroup = null;
-106EventLoopGroup workerGroup = null;
-107if (useEpoll) {
-108  bossGroup = new 
EpollEventLoopGroup(1);
-109  workerGroup = new 
EpollEventLoopGroup(workerCount);
-110} else {
-111  bossGroup = new 
NioEventLoopGroup(1);
-112  workerGroup = new 
NioEventLoopGroup(workerCount);
-113}
-114ServerBootstrap bootstrap = new 
ServerBootstrap();
-115bootstrap.group(bossGroup, 
workerGroup);
-116if (useEpoll) {
-117  
bootstrap.channel(EpollServerSocketChannel.class);
-118} else {
-119  
bootstrap.channel(NioServerSocketChannel.class);
-120}
-121
bootstrap.childOption(ChannelOption.TCP_NODELAY, tcpNoDelay);
-122
bootstrap.childOption(ChannelOption.SO_KEEPALIVE, tcpKeepAlive);
-123
bootstrap.childOption(ChannelOption.ALLOCATOR,
-124
PooledByteBufAllocator.DEFAULT);
-125bootstrap.childHandler(new 
Initializer(maxRequestSize));
-126
-127try {
-128  serverChannel = 
bootstrap.bind(this.bindAddress).sync().channel();
-129  LOG.info("NettyRpcServer bind to 
address=" + serverChannel.localAddress()
-130  + ", 
hbase.netty.rpc.server.worker.count=" + workerCount
-131  + ", useEpoll=" + useEpoll);
-132  allChannels.add(serverChannel);
-133} catch (InterruptedException e) {
-134  throw new 
InterruptedIOException(e.getMessage());
-135}
-136initReconfigurable(conf);
-137this.scheduler.init(new 
RpcSchedulerContext(this));
-138  }
-139
-140  private static boolean 
useEpoll(Configuration conf) {
-141// Config to enable native 
transport.
-142boolean epollEnabled = 
conf.getBoolean("hbase.rpc.server.nativetransport",
-143true);
-144// Use the faster native epoll 
transport mechanism on linux if enabled
-145return epollEnabled && 
JVM.isLinux() && JVM.isAmd64();
-146  }
-147
-148  @Override
-149  public synchronized void start() {
-150if (started) {
-151  return;
-152}
-15

[27/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-05-09 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.Initializer.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.Initializer.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.Initializer.html
index 27e0dee..109b5f3 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.Initializer.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.Initializer.html
@@ -67,157 +67,157 @@
 059import 
org.apache.hadoop.hbase.CellScanner;
 060import 
org.apache.hadoop.hbase.HConstants;
 061import org.apache.hadoop.hbase.Server;
-062import 
org.apache.hadoop.hbase.classification.InterfaceStability;
-063import 
org.apache.hadoop.hbase.monitoring.MonitoredRPCHandler;
-064import 
org.apache.hadoop.hbase.nio.ByteBuff;
-065import 
org.apache.hadoop.hbase.nio.SingleByteBuff;
-066import 
org.apache.hadoop.hbase.security.AccessDeniedException;
-067import 
org.apache.hadoop.hbase.security.AuthMethod;
-068import 
org.apache.hadoop.hbase.security.HBasePolicyProvider;
-069import 
org.apache.hadoop.hbase.security.SaslStatus;
-070import 
org.apache.hadoop.hbase.security.SaslUtil;
-071import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService;
-072import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor;
-073import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.Message;
-074import 
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader;
-075import 
org.apache.hadoop.hbase.util.Bytes;
-076import 
org.apache.hadoop.hbase.util.JVM;
-077import 
org.apache.hadoop.hbase.util.Pair;
-078import 
org.apache.hadoop.io.IntWritable;
-079import 
org.apache.hadoop.security.authorize.ServiceAuthorizationManager;
-080import org.apache.htrace.TraceInfo;
-081
-082/**
-083 * An RPC server with Netty4 
implementation.
-084 *
-085 */
-086public class NettyRpcServer extends 
RpcServer {
-087
-088  public static final Log LOG = 
LogFactory.getLog(NettyRpcServer.class);
-089
-090  protected final InetSocketAddress 
bindAddress;
-091
-092  private final CountDownLatch closed = 
new CountDownLatch(1);
-093  private final Channel serverChannel;
-094  private final ChannelGroup allChannels 
= new DefaultChannelGroup(GlobalEventExecutor.INSTANCE);;
-095
-096  public NettyRpcServer(final Server 
server, final String name,
-097  final 
List services,
-098  final InetSocketAddress 
bindAddress, Configuration conf,
-099  RpcScheduler scheduler) throws 
IOException {
-100super(server, name, services, 
bindAddress, conf, scheduler);
-101this.bindAddress = bindAddress;
-102boolean useEpoll = useEpoll(conf);
-103int workerCount = 
conf.getInt("hbase.netty.rpc.server.worker.count",
-104
Runtime.getRuntime().availableProcessors() / 4);
-105EventLoopGroup bossGroup = null;
-106EventLoopGroup workerGroup = null;
-107if (useEpoll) {
-108  bossGroup = new 
EpollEventLoopGroup(1);
-109  workerGroup = new 
EpollEventLoopGroup(workerCount);
-110} else {
-111  bossGroup = new 
NioEventLoopGroup(1);
-112  workerGroup = new 
NioEventLoopGroup(workerCount);
-113}
-114ServerBootstrap bootstrap = new 
ServerBootstrap();
-115bootstrap.group(bossGroup, 
workerGroup);
-116if (useEpoll) {
-117  
bootstrap.channel(EpollServerSocketChannel.class);
-118} else {
-119  
bootstrap.channel(NioServerSocketChannel.class);
-120}
-121
bootstrap.childOption(ChannelOption.TCP_NODELAY, tcpNoDelay);
-122
bootstrap.childOption(ChannelOption.SO_KEEPALIVE, tcpKeepAlive);
-123
bootstrap.childOption(ChannelOption.ALLOCATOR,
-124
PooledByteBufAllocator.DEFAULT);
-125bootstrap.childHandler(new 
Initializer(maxRequestSize));
-126
-127try {
-128  serverChannel = 
bootstrap.bind(this.bindAddress).sync().channel();
-129  LOG.info("NettyRpcServer bind to 
address=" + serverChannel.localAddress()
-130  + ", 
hbase.netty.rpc.server.worker.count=" + workerCount
-131  + ", useEpoll=" + useEpoll);
-132  allChannels.add(serverChannel);
-133} catch (InterruptedException e) {
-134  throw new 
InterruptedIOException(e.getMessage());
-135}
-136initReconfigurable(conf);
-137this.scheduler.init(new 
RpcSchedulerContext(this));
-138  }
-139
-140  private static boolean 
useEpoll(Configuration conf) {
-141// Config to enable native 
transport.
-142boolean epollEnabled = 
conf.getBoolean("hbase.rpc.server.nativetransport",
-143true);
-144// Use the faster native epoll 
transport mechanism on linux if enabled
-145return epollEnabled && 
JVM.isLinux() && JVM.isAmd64();
-146  }
-147
-148  @Override
-149  public synchronized void start() {
-150if (started) {
-151  return;
-152}
-153authTokenS

[09/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-05-09 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.html 
b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.html
index 2987e7b..7f61b54 100644
--- a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.html
+++ b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.html
@@ -135,1305 +135,1251 @@
 127  private Listener listener = null;
 128  protected Responder responder = null;
 129
-130  /**
-131   * Datastructure that holds all 
necessary to a method invocation and then afterward, carries
-132   * the result.
-133   */
-134  @InterfaceStability.Evolving
-135  public class Call extends 
RpcServer.Call {
-136
-137protected Responder responder;
+130  /** Listens on the socket. Creates jobs 
for the handler threads*/
+131  private class Listener extends Thread 
{
+132
+133private ServerSocketChannel 
acceptChannel = null; //the accept channel
+134private Selector selector = null; 
//the selector that we use for the server
+135private Reader[] readers = null;
+136private int currentReader = 0;
+137private final int 
readerPendingConnectionQueueLength;
 138
-139
@edu.umd.cs.findbugs.annotations.SuppressWarnings(value="NP_NULL_ON_SOME_PATH",
-140justification="Can't figure why 
this complaint is happening... see below")
-141Call(int id, final BlockingService 
service, final MethodDescriptor md,
-142RequestHeader header, Message 
param, CellScanner cellScanner,
-143RpcServer.Connection connection, 
long size, TraceInfo tinfo,
-144final InetAddress remoteAddress, 
int timeout, CallCleanup reqCleanup,
-145Responder responder) {
-146  super(id, service, md, header, 
param, cellScanner, connection, size,
-147  tinfo, remoteAddress, timeout, 
reqCleanup);
-148  this.responder = responder;
-149}
+139private ExecutorService readPool;
+140
+141public Listener(final String name) 
throws IOException {
+142  super(name);
+143  // The backlog of requests that we 
will have the serversocket carry.
+144  int backlogLength = 
conf.getInt("hbase.ipc.server.listen.queue.size", 128);
+145  readerPendingConnectionQueueLength 
=
+146  
conf.getInt("hbase.ipc.server.read.connection-queue.size", 100);
+147  // Create a new server socket and 
set to non blocking mode
+148  acceptChannel = 
ServerSocketChannel.open();
+149  
acceptChannel.configureBlocking(false);
 150
-151/**
-152 * Call is done. Execution happened 
and we returned results to client. It is now safe to
-153 * cleanup.
-154 */
-155
@edu.umd.cs.findbugs.annotations.SuppressWarnings(value="IS2_INCONSISTENT_SYNC",
-156justification="Presume the lock 
on processing request held by caller is protection enough")
-157@Override
-158void done() {
-159  super.done();
-160  this.getConnection().decRpcCount(); 
// Say that we're done with this call.
-161}
-162
-163@Override
-164public long disconnectSince() {
-165  if 
(!getConnection().isConnectionOpen()) {
-166return System.currentTimeMillis() 
- timestamp;
-167  } else {
-168return -1L;
-169  }
-170}
-171
-172@Override
-173public synchronized void 
sendResponseIfReady() throws IOException {
-174  // set param null to reduce memory 
pressure
-175  this.param = null;
-176  this.responder.doRespond(this);
+151  // Bind the server socket to the 
binding addrees (can be different from the default interface)
+152  bind(acceptChannel.socket(), 
bindAddress, backlogLength);
+153  port = 
acceptChannel.socket().getLocalPort(); //Could be an ephemeral port
+154  address = 
(InetSocketAddress)acceptChannel.socket().getLocalSocketAddress();
+155  // create a selector;
+156  selector = Selector.open();
+157
+158  readers = new 
Reader[readThreads];
+159  // Why this executor thing? Why not 
like hadoop just start up all the threads? I suppose it
+160  // has an advantage in that it is 
easy to shutdown the pool.
+161  readPool = 
Executors.newFixedThreadPool(readThreads,
+162new 
ThreadFactoryBuilder().setNameFormat(
+163  
"RpcServer.reader=%d,bindAddress=" + bindAddress.getHostName() +
+164  ",port=" + 
port).setDaemon(true)
+165
.setUncaughtExceptionHandler(Threads.LOGGING_EXCEPTION_HANDLER).build());
+166  for (int i = 0; i < readThreads; 
++i) {
+167Reader reader = new Reader();
+168readers[i] = reader;
+169readPool.execute(reader);
+170  }
+171  LOG.info(getName() + ": started " + 
readThreads + " reader(s) listening on port=" + port);
+172
+173  // Register accepts on the server 
socket with the selector.
+174  acceptC

[19/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-05-09 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.CallCleanup.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.CallCleanup.html 
b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.CallCleanup.html
index 5cc356a..9e1c66c 100644
--- a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.CallCleanup.html
+++ b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.CallCleanup.html
@@ -70,2037 +70,1559 @@
 062import 
org.apache.hadoop.hbase.client.VersionInfoUtil;
 063import 
org.apache.hadoop.hbase.codec.Codec;
 064import 
org.apache.hadoop.hbase.conf.ConfigurationObserver;
-065import 
org.apache.hadoop.hbase.exceptions.RegionMovedException;
-066import 
org.apache.hadoop.hbase.exceptions.RequestTooBigException;
-067import 
org.apache.hadoop.hbase.io.ByteBufferListOutputStream;
-068import 
org.apache.hadoop.hbase.io.ByteBufferOutputStream;
-069import 
org.apache.hadoop.hbase.io.ByteBufferPool;
-070import 
org.apache.hadoop.hbase.io.crypto.aes.CryptoAES;
-071import 
org.apache.hadoop.hbase.monitoring.MonitoredRPCHandler;
-072import 
org.apache.hadoop.hbase.monitoring.TaskMonitor;
-073import 
org.apache.hadoop.hbase.nio.ByteBuff;
-074import 
org.apache.hadoop.hbase.nio.MultiByteBuff;
-075import 
org.apache.hadoop.hbase.nio.SingleByteBuff;
-076import 
org.apache.hadoop.hbase.regionserver.RSRpcServices;
-077import 
org.apache.hadoop.hbase.security.AccessDeniedException;
-078import 
org.apache.hadoop.hbase.security.AuthMethod;
-079import 
org.apache.hadoop.hbase.security.HBaseSaslRpcServer;
-080import 
org.apache.hadoop.hbase.security.HBaseSaslRpcServer.SaslDigestCallbackHandler;
-081import 
org.apache.hadoop.hbase.security.HBaseSaslRpcServer.SaslGssCallbackHandler;
-082import 
org.apache.hadoop.hbase.security.SaslStatus;
-083import 
org.apache.hadoop.hbase.security.SaslUtil;
-084import 
org.apache.hadoop.hbase.security.User;
-085import 
org.apache.hadoop.hbase.security.UserProvider;
-086import 
org.apache.hadoop.hbase.security.token.AuthenticationTokenSecretManager;
-087import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService;
-088import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteInput;
-089import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString;
-090import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream;
-091import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream;
-092import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor;
-093import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.Message;
-094import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException;
-095import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.TextFormat;
-096import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.UnsafeByteOperations;
-097import 
org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
-098import 
org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos;
-099import 
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.VersionInfo;
-100import 
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos;
-101import 
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.CellBlockMeta;
-102import 
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ConnectionHeader;
-103import 
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ExceptionResponse;
-104import 
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader;
-105import 
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ResponseHeader;
-106import 
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation;
-107import 
org.apache.hadoop.hbase.util.ByteBufferUtils;
-108import 
org.apache.hadoop.hbase.util.Bytes;
-109import 
org.apache.hadoop.hbase.util.Pair;
-110import 
org.apache.hadoop.io.BytesWritable;
-111import org.apache.hadoop.io.Writable;
-112import 
org.apache.hadoop.io.WritableUtils;
-113import 
org.apache.hadoop.io.compress.CompressionCodec;
-114import 
org.apache.hadoop.security.UserGroupInformation;
-115import 
org.apache.hadoop.security.UserGroupInformation.AuthenticationMethod;
-116import 
org.apache.hadoop.security.authorize.AuthorizationException;
-117import 
org.apache.hadoop.security.authorize.PolicyProvider;
-118import 
org.apache.hadoop.security.authorize.ProxyUsers;
-119import 
org.apache.hadoop.security.authorize.ServiceAuthorizationManager;
-120import 
org.apache.hadoop.security.token.SecretManager;
-121import 
org.apache.hadoop.security.token.SecretManager.InvalidToken;
-122import 
org.apache.hadoop.security.token.TokenIdentifier;
-123import 
org.apache.hadoop.util.StringUtils;
-124import org.apache.htrace.TraceInfo;
-125import 
org.codehaus.jackson.map.ObjectMapper;
-126
-127import 
com.google.common.annotations.VisibleForTesting;
-128
-129/**
-1

[29/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-05-09 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.CallWriteListener.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.CallWriteListener.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.CallWriteListener.html
index 27e0dee..109b5f3 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.CallWriteListener.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.CallWriteListener.html
@@ -67,157 +67,157 @@
 059import 
org.apache.hadoop.hbase.CellScanner;
 060import 
org.apache.hadoop.hbase.HConstants;
 061import org.apache.hadoop.hbase.Server;
-062import 
org.apache.hadoop.hbase.classification.InterfaceStability;
-063import 
org.apache.hadoop.hbase.monitoring.MonitoredRPCHandler;
-064import 
org.apache.hadoop.hbase.nio.ByteBuff;
-065import 
org.apache.hadoop.hbase.nio.SingleByteBuff;
-066import 
org.apache.hadoop.hbase.security.AccessDeniedException;
-067import 
org.apache.hadoop.hbase.security.AuthMethod;
-068import 
org.apache.hadoop.hbase.security.HBasePolicyProvider;
-069import 
org.apache.hadoop.hbase.security.SaslStatus;
-070import 
org.apache.hadoop.hbase.security.SaslUtil;
-071import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService;
-072import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor;
-073import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.Message;
-074import 
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader;
-075import 
org.apache.hadoop.hbase.util.Bytes;
-076import 
org.apache.hadoop.hbase.util.JVM;
-077import 
org.apache.hadoop.hbase.util.Pair;
-078import 
org.apache.hadoop.io.IntWritable;
-079import 
org.apache.hadoop.security.authorize.ServiceAuthorizationManager;
-080import org.apache.htrace.TraceInfo;
-081
-082/**
-083 * An RPC server with Netty4 
implementation.
-084 *
-085 */
-086public class NettyRpcServer extends 
RpcServer {
-087
-088  public static final Log LOG = 
LogFactory.getLog(NettyRpcServer.class);
-089
-090  protected final InetSocketAddress 
bindAddress;
-091
-092  private final CountDownLatch closed = 
new CountDownLatch(1);
-093  private final Channel serverChannel;
-094  private final ChannelGroup allChannels 
= new DefaultChannelGroup(GlobalEventExecutor.INSTANCE);;
-095
-096  public NettyRpcServer(final Server 
server, final String name,
-097  final 
List services,
-098  final InetSocketAddress 
bindAddress, Configuration conf,
-099  RpcScheduler scheduler) throws 
IOException {
-100super(server, name, services, 
bindAddress, conf, scheduler);
-101this.bindAddress = bindAddress;
-102boolean useEpoll = useEpoll(conf);
-103int workerCount = 
conf.getInt("hbase.netty.rpc.server.worker.count",
-104
Runtime.getRuntime().availableProcessors() / 4);
-105EventLoopGroup bossGroup = null;
-106EventLoopGroup workerGroup = null;
-107if (useEpoll) {
-108  bossGroup = new 
EpollEventLoopGroup(1);
-109  workerGroup = new 
EpollEventLoopGroup(workerCount);
-110} else {
-111  bossGroup = new 
NioEventLoopGroup(1);
-112  workerGroup = new 
NioEventLoopGroup(workerCount);
-113}
-114ServerBootstrap bootstrap = new 
ServerBootstrap();
-115bootstrap.group(bossGroup, 
workerGroup);
-116if (useEpoll) {
-117  
bootstrap.channel(EpollServerSocketChannel.class);
-118} else {
-119  
bootstrap.channel(NioServerSocketChannel.class);
-120}
-121
bootstrap.childOption(ChannelOption.TCP_NODELAY, tcpNoDelay);
-122
bootstrap.childOption(ChannelOption.SO_KEEPALIVE, tcpKeepAlive);
-123
bootstrap.childOption(ChannelOption.ALLOCATOR,
-124
PooledByteBufAllocator.DEFAULT);
-125bootstrap.childHandler(new 
Initializer(maxRequestSize));
-126
-127try {
-128  serverChannel = 
bootstrap.bind(this.bindAddress).sync().channel();
-129  LOG.info("NettyRpcServer bind to 
address=" + serverChannel.localAddress()
-130  + ", 
hbase.netty.rpc.server.worker.count=" + workerCount
-131  + ", useEpoll=" + useEpoll);
-132  allChannels.add(serverChannel);
-133} catch (InterruptedException e) {
-134  throw new 
InterruptedIOException(e.getMessage());
-135}
-136initReconfigurable(conf);
-137this.scheduler.init(new 
RpcSchedulerContext(this));
-138  }
-139
-140  private static boolean 
useEpoll(Configuration conf) {
-141// Config to enable native 
transport.
-142boolean epollEnabled = 
conf.getBoolean("hbase.rpc.server.nativetransport",
-143true);
-144// Use the faster native epoll 
transport mechanism on linux if enabled
-145return epollEnabled && 
JVM.isLinux() && JVM.isAmd64();
-146  }
-147
-148  @Override
-149  public synchronized void start() {
-150if (started) {
-151  return

[41/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-05-09 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/org/apache/hadoop/hbase/ipc/ServerCall.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/ipc/ServerCall.html 
b/devapidocs/org/apache/hadoop/hbase/ipc/ServerCall.html
new file mode 100644
index 000..434671a
--- /dev/null
+++ b/devapidocs/org/apache/hadoop/hbase/ipc/ServerCall.html
@@ -0,0 +1,1477 @@
+http://www.w3.org/TR/html4/loose.dtd";>
+
+
+
+
+
+ServerCall (Apache HBase 2.0.0-SNAPSHOT API)
+
+
+
+
+
+var methods = 
{"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":10,"i33":10,"i34":10,"i35":10,"i36":10,"i37":10,"i38":10,"i39":10};
+var tabs = {65535:["t0","All Methods"],2:["t2","Instance 
Methods"],8:["t4","Concrete Methods"]};
+var altColor = "altColor";
+var rowColor = "rowColor";
+var tableTab = "tableTab";
+var activeTableTab = "activeTableTab";
+
+
+JavaScript is disabled on your browser.
+
+
+
+
+
+Skip navigation links
+
+
+
+
+Overview
+Package
+Class
+Use
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+Prev Class
+Next Class
+
+
+Frames
+No Frames
+
+
+All Classes
+
+
+
+
+
+
+
+Summary: 
+Nested | 
+Field | 
+Constr | 
+Method
+
+
+Detail: 
+Field | 
+Constr | 
+Method
+
+
+
+
+
+
+
+
+org.apache.hadoop.hbase.ipc
+Class ServerCall
+
+
+
+http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true";
 title="class or interface in java.lang">java.lang.Object
+
+
+org.apache.hadoop.hbase.ipc.ServerCall
+
+
+
+
+
+
+
+All Implemented Interfaces:
+RpcCall, RpcCallContext
+
+
+Direct Known Subclasses:
+NettyServerCall, SimpleServerCall
+
+
+
+@InterfaceAudience.Private
+abstract class ServerCall
+extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true";
 title="class or interface in java.lang">Object
+implements RpcCall
+Datastructure that holds all necessary to a method 
invocation and then afterward, carries
+ the result.
+
+
+
+
+
+
+
+
+
+
+
+Field Summary
+
+Fields 
+
+Modifier and Type
+Field and Description
+
+
+protected CellBlockBuilder
+cellBlockBuilder 
+
+
+protected ByteBufferListOutputStream
+cellBlockStream 
+
+
+protected CellScanner
+cellScanner 
+
+
+protected RpcServer.Connection
+connection 
+
+
+protected long
+deadline 
+
+
+private long
+exceptionSize 
+
+
+protected 
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader
+header 
+
+
+protected int
+id 
+
+
+protected boolean
+isError 
+
+
+protected 
org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor
+md 
+
+
+protected 
org.apache.hadoop.hbase.shaded.com.google.protobuf.Message
+param 
+
+
+protected long
+receiveTime 
+
+
+protected http://docs.oracle.com/javase/8/docs/api/java/net/InetAddress.html?is-external=true";
 title="class or interface in java.net">InetAddress
+remoteAddress 
+
+
+protected RpcServer.CallCleanup
+reqCleanup 
+
+
+protected ByteBufferPool
+reservoir 
+
+
+protected BufferChain
+response
+Chain of buffers to send as response.
+
+
+
+private long
+responseBlockSize 
+
+
+private long
+responseCellSize 
+
+
+private boolean
+retryImmediatelySupported 
+
+
+protected RpcCallback
+rpcCallback 
+
+
+protected 
org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService
+service 
+
+
+protected long
+size 
+
+
+protected long
+startTime 
+
+
+protected int
+timeout 
+
+
+protected org.apache.htrace.TraceInfo
+tinfo 
+
+
+protected User
+user 
+
+
+
+
+
+
+
+
+
+Constructor Summary
+
+Constructors 
+
+Constructor and Description
+
+
+ServerCall(int id,
+  
org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService service,
+  
org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor md,
+  
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader header,
+  
org.apache.hadoop.hbase.shaded.com.google.protobuf.Message param,
+  CellScanner cellScanner,
+  RpcServer.Connection connection,
+  long size,
+  org.apache.htrace.TraceInfo tinfo,
+  http://docs.oracle.com/javase/8/docs/api/java/net/InetAddress.html?is-external=true";
 title="class or interface in java.net">InetAddress remoteAddress,
+  long receiveTime,
+  int timeout,
+  ByteBufferPool reservoir,
+  Cell

[01/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-05-09 Thread git-site-role
Repository: hbase-site
Updated Branches:
  refs/heads/asf-site baff48117 -> 7ef4c5a9a


http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/RSRpcServices.RegionScannerShippedCallBack.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/RSRpcServices.RegionScannerShippedCallBack.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/RSRpcServices.RegionScannerShippedCallBack.html
index 17d1bcb..6dfd1d4 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/RSRpcServices.RegionScannerShippedCallBack.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/RSRpcServices.RegionScannerShippedCallBack.html
@@ -35,863 +35,863 @@
 027import java.net.InetSocketAddress;
 028import java.net.UnknownHostException;
 029import java.nio.ByteBuffer;
-030import java.util.ArrayList;
-031import java.util.Collections;
-032import java.util.HashMap;
-033import java.util.Iterator;
-034import java.util.List;
-035import java.util.Map;
-036import java.util.Map.Entry;
-037import java.util.NavigableMap;
-038import java.util.Set;
-039import java.util.TreeSet;
-040import 
java.util.concurrent.ConcurrentHashMap;
-041import 
java.util.concurrent.ConcurrentMap;
-042import 
java.util.concurrent.atomic.AtomicLong;
-043import 
java.util.concurrent.atomic.LongAdder;
-044
-045import 
org.apache.commons.lang.mutable.MutableObject;
-046import org.apache.commons.logging.Log;
-047import 
org.apache.commons.logging.LogFactory;
-048import 
org.apache.hadoop.conf.Configuration;
-049import org.apache.hadoop.fs.Path;
-050import 
org.apache.hadoop.hbase.ByteBufferCell;
-051import org.apache.hadoop.hbase.Cell;
-052import 
org.apache.hadoop.hbase.CellScannable;
-053import 
org.apache.hadoop.hbase.CellScanner;
-054import 
org.apache.hadoop.hbase.CellUtil;
-055import 
org.apache.hadoop.hbase.DoNotRetryIOException;
-056import 
org.apache.hadoop.hbase.DroppedSnapshotException;
-057import 
org.apache.hadoop.hbase.HBaseIOException;
-058import 
org.apache.hadoop.hbase.HConstants;
-059import 
org.apache.hadoop.hbase.HRegionInfo;
-060import 
org.apache.hadoop.hbase.HTableDescriptor;
-061import 
org.apache.hadoop.hbase.MultiActionResultTooLarge;
-062import 
org.apache.hadoop.hbase.NotServingRegionException;
-063import 
org.apache.hadoop.hbase.ServerName;
-064import 
org.apache.hadoop.hbase.TableName;
-065import 
org.apache.hadoop.hbase.UnknownScannerException;
-066import 
org.apache.hadoop.hbase.classification.InterfaceAudience;
-067import 
org.apache.hadoop.hbase.client.Append;
-068import 
org.apache.hadoop.hbase.client.ConnectionUtils;
-069import 
org.apache.hadoop.hbase.client.Delete;
-070import 
org.apache.hadoop.hbase.client.Durability;
-071import 
org.apache.hadoop.hbase.client.Get;
-072import 
org.apache.hadoop.hbase.client.Increment;
-073import 
org.apache.hadoop.hbase.client.Mutation;
-074import 
org.apache.hadoop.hbase.client.Put;
-075import 
org.apache.hadoop.hbase.client.RegionReplicaUtil;
-076import 
org.apache.hadoop.hbase.client.Result;
-077import 
org.apache.hadoop.hbase.client.RowMutations;
-078import 
org.apache.hadoop.hbase.client.Scan;
-079import 
org.apache.hadoop.hbase.client.VersionInfoUtil;
-080import 
org.apache.hadoop.hbase.conf.ConfigurationObserver;
-081import 
org.apache.hadoop.hbase.exceptions.FailedSanityCheckException;
-082import 
org.apache.hadoop.hbase.exceptions.OutOfOrderScannerNextException;
-083import 
org.apache.hadoop.hbase.exceptions.ScannerResetException;
-084import 
org.apache.hadoop.hbase.filter.ByteArrayComparable;
-085import 
org.apache.hadoop.hbase.filter.CompareFilter.CompareOp;
-086import 
org.apache.hadoop.hbase.ipc.HBaseRPCErrorHandler;
-087import 
org.apache.hadoop.hbase.ipc.HBaseRpcController;
-088import 
org.apache.hadoop.hbase.ipc.PriorityFunction;
-089import 
org.apache.hadoop.hbase.ipc.QosPriority;
-090import 
org.apache.hadoop.hbase.ipc.RpcCallContext;
-091import 
org.apache.hadoop.hbase.ipc.RpcCallback;
-092import 
org.apache.hadoop.hbase.ipc.RpcServer;
-093import 
org.apache.hadoop.hbase.ipc.RpcServer.BlockingServiceAndInterface;
-094import 
org.apache.hadoop.hbase.ipc.RpcServerFactory;
-095import 
org.apache.hadoop.hbase.ipc.RpcServerInterface;
-096import 
org.apache.hadoop.hbase.ipc.ServerNotRunningYetException;
-097import 
org.apache.hadoop.hbase.ipc.ServerRpcController;
-098import 
org.apache.hadoop.hbase.master.MasterRpcServices;
-099import 
org.apache.hadoop.hbase.quotas.OperationQuota;
-100import 
org.apache.hadoop.hbase.quotas.RegionServerQuotaManager;
-101import 
org.apache.hadoop.hbase.regionserver.HRegion.RegionScannerImpl;
-102import 
org.apache.hadoop.hbase.regionserver.Leases.Lease;
-103import 
org.apache.hadoop.hbase.regionserver.Leases.LeaseStillHeldException;
-104import 
org.apache.hadoop.hbase.regionserver.Region.Operation;
-105import 
org.apache.hadoop.hbase.regionserver.ScannerContext.Limi

[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-05-09 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.ConnectionHeaderHandler.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.ConnectionHeaderHandler.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.ConnectionHeaderHandler.html
index 27e0dee..109b5f3 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.ConnectionHeaderHandler.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.ConnectionHeaderHandler.html
@@ -67,157 +67,157 @@
 059import 
org.apache.hadoop.hbase.CellScanner;
 060import 
org.apache.hadoop.hbase.HConstants;
 061import org.apache.hadoop.hbase.Server;
-062import 
org.apache.hadoop.hbase.classification.InterfaceStability;
-063import 
org.apache.hadoop.hbase.monitoring.MonitoredRPCHandler;
-064import 
org.apache.hadoop.hbase.nio.ByteBuff;
-065import 
org.apache.hadoop.hbase.nio.SingleByteBuff;
-066import 
org.apache.hadoop.hbase.security.AccessDeniedException;
-067import 
org.apache.hadoop.hbase.security.AuthMethod;
-068import 
org.apache.hadoop.hbase.security.HBasePolicyProvider;
-069import 
org.apache.hadoop.hbase.security.SaslStatus;
-070import 
org.apache.hadoop.hbase.security.SaslUtil;
-071import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService;
-072import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor;
-073import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.Message;
-074import 
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader;
-075import 
org.apache.hadoop.hbase.util.Bytes;
-076import 
org.apache.hadoop.hbase.util.JVM;
-077import 
org.apache.hadoop.hbase.util.Pair;
-078import 
org.apache.hadoop.io.IntWritable;
-079import 
org.apache.hadoop.security.authorize.ServiceAuthorizationManager;
-080import org.apache.htrace.TraceInfo;
-081
-082/**
-083 * An RPC server with Netty4 
implementation.
-084 *
-085 */
-086public class NettyRpcServer extends 
RpcServer {
-087
-088  public static final Log LOG = 
LogFactory.getLog(NettyRpcServer.class);
-089
-090  protected final InetSocketAddress 
bindAddress;
-091
-092  private final CountDownLatch closed = 
new CountDownLatch(1);
-093  private final Channel serverChannel;
-094  private final ChannelGroup allChannels 
= new DefaultChannelGroup(GlobalEventExecutor.INSTANCE);;
-095
-096  public NettyRpcServer(final Server 
server, final String name,
-097  final 
List services,
-098  final InetSocketAddress 
bindAddress, Configuration conf,
-099  RpcScheduler scheduler) throws 
IOException {
-100super(server, name, services, 
bindAddress, conf, scheduler);
-101this.bindAddress = bindAddress;
-102boolean useEpoll = useEpoll(conf);
-103int workerCount = 
conf.getInt("hbase.netty.rpc.server.worker.count",
-104
Runtime.getRuntime().availableProcessors() / 4);
-105EventLoopGroup bossGroup = null;
-106EventLoopGroup workerGroup = null;
-107if (useEpoll) {
-108  bossGroup = new 
EpollEventLoopGroup(1);
-109  workerGroup = new 
EpollEventLoopGroup(workerCount);
-110} else {
-111  bossGroup = new 
NioEventLoopGroup(1);
-112  workerGroup = new 
NioEventLoopGroup(workerCount);
-113}
-114ServerBootstrap bootstrap = new 
ServerBootstrap();
-115bootstrap.group(bossGroup, 
workerGroup);
-116if (useEpoll) {
-117  
bootstrap.channel(EpollServerSocketChannel.class);
-118} else {
-119  
bootstrap.channel(NioServerSocketChannel.class);
-120}
-121
bootstrap.childOption(ChannelOption.TCP_NODELAY, tcpNoDelay);
-122
bootstrap.childOption(ChannelOption.SO_KEEPALIVE, tcpKeepAlive);
-123
bootstrap.childOption(ChannelOption.ALLOCATOR,
-124
PooledByteBufAllocator.DEFAULT);
-125bootstrap.childHandler(new 
Initializer(maxRequestSize));
-126
-127try {
-128  serverChannel = 
bootstrap.bind(this.bindAddress).sync().channel();
-129  LOG.info("NettyRpcServer bind to 
address=" + serverChannel.localAddress()
-130  + ", 
hbase.netty.rpc.server.worker.count=" + workerCount
-131  + ", useEpoll=" + useEpoll);
-132  allChannels.add(serverChannel);
-133} catch (InterruptedException e) {
-134  throw new 
InterruptedIOException(e.getMessage());
-135}
-136initReconfigurable(conf);
-137this.scheduler.init(new 
RpcSchedulerContext(this));
-138  }
-139
-140  private static boolean 
useEpoll(Configuration conf) {
-141// Config to enable native 
transport.
-142boolean epollEnabled = 
conf.getBoolean("hbase.rpc.server.nativetransport",
-143true);
-144// Use the faster native epoll 
transport mechanism on linux if enabled
-145return epollEnabled && 
JVM.isLinux() && JVM.isAmd64();
-146  }
-147
-148  @Override
-149  public synchronized void start() {
-150i

[06/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-05-09 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.PeriodicDoMetrics.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.PeriodicDoMetrics.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.PeriodicDoMetrics.html
index 66fe094..17bc96c 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.PeriodicDoMetrics.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.PeriodicDoMetrics.html
@@ -57,3290 +57,3303 @@
 049import org.apache.commons.logging.Log;
 050import 
org.apache.commons.logging.LogFactory;
 051import 
org.apache.hadoop.conf.Configuration;
-052import org.apache.hadoop.fs.Path;
-053import 
org.apache.hadoop.hbase.ClusterStatus;
-054import 
org.apache.hadoop.hbase.CoordinatedStateException;
-055import 
org.apache.hadoop.hbase.CoordinatedStateManager;
-056import 
org.apache.hadoop.hbase.DoNotRetryIOException;
-057import 
org.apache.hadoop.hbase.HBaseIOException;
-058import 
org.apache.hadoop.hbase.HBaseInterfaceAudience;
-059import 
org.apache.hadoop.hbase.HColumnDescriptor;
-060import 
org.apache.hadoop.hbase.HConstants;
-061import 
org.apache.hadoop.hbase.HRegionInfo;
-062import 
org.apache.hadoop.hbase.HTableDescriptor;
-063import 
org.apache.hadoop.hbase.MasterNotRunningException;
-064import 
org.apache.hadoop.hbase.MetaTableAccessor;
-065import 
org.apache.hadoop.hbase.NamespaceDescriptor;
-066import 
org.apache.hadoop.hbase.PleaseHoldException;
-067import 
org.apache.hadoop.hbase.ProcedureInfo;
-068import 
org.apache.hadoop.hbase.ScheduledChore;
-069import 
org.apache.hadoop.hbase.ServerLoad;
-070import 
org.apache.hadoop.hbase.ServerName;
-071import 
org.apache.hadoop.hbase.TableDescriptors;
-072import 
org.apache.hadoop.hbase.TableName;
-073import 
org.apache.hadoop.hbase.TableNotDisabledException;
-074import 
org.apache.hadoop.hbase.TableNotFoundException;
-075import 
org.apache.hadoop.hbase.UnknownRegionException;
-076import 
org.apache.hadoop.hbase.classification.InterfaceAudience;
-077import 
org.apache.hadoop.hbase.client.MasterSwitchType;
-078import 
org.apache.hadoop.hbase.client.Result;
-079import 
org.apache.hadoop.hbase.client.TableState;
-080import 
org.apache.hadoop.hbase.coprocessor.BypassCoprocessorException;
-081import 
org.apache.hadoop.hbase.coprocessor.CoprocessorHost;
-082import 
org.apache.hadoop.hbase.exceptions.DeserializationException;
-083import 
org.apache.hadoop.hbase.exceptions.MergeRegionException;
-084import 
org.apache.hadoop.hbase.executor.ExecutorType;
-085import 
org.apache.hadoop.hbase.favored.FavoredNodesManager;
-086import 
org.apache.hadoop.hbase.favored.FavoredNodesPromoter;
-087import 
org.apache.hadoop.hbase.http.InfoServer;
-088import 
org.apache.hadoop.hbase.ipc.CoprocessorRpcUtils;
-089import 
org.apache.hadoop.hbase.ipc.RpcServer;
-090import 
org.apache.hadoop.hbase.ipc.ServerNotRunningYetException;
-091import 
org.apache.hadoop.hbase.master.MasterRpcServices.BalanceSwitchMode;
-092import 
org.apache.hadoop.hbase.master.balancer.BalancerChore;
-093import 
org.apache.hadoop.hbase.master.balancer.BaseLoadBalancer;
-094import 
org.apache.hadoop.hbase.master.balancer.ClusterStatusChore;
-095import 
org.apache.hadoop.hbase.master.balancer.LoadBalancerFactory;
-096import 
org.apache.hadoop.hbase.master.cleaner.HFileCleaner;
-097import 
org.apache.hadoop.hbase.master.cleaner.LogCleaner;
-098import 
org.apache.hadoop.hbase.master.cleaner.ReplicationMetaCleaner;
-099import 
org.apache.hadoop.hbase.master.cleaner.ReplicationZKNodeCleaner;
-100import 
org.apache.hadoop.hbase.master.cleaner.ReplicationZKNodeCleanerChore;
-101import 
org.apache.hadoop.hbase.master.locking.LockManager;
-102import 
org.apache.hadoop.hbase.master.normalizer.NormalizationPlan;
-103import 
org.apache.hadoop.hbase.master.normalizer.NormalizationPlan.PlanType;
-104import 
org.apache.hadoop.hbase.master.normalizer.RegionNormalizer;
-105import 
org.apache.hadoop.hbase.master.normalizer.RegionNormalizerChore;
-106import 
org.apache.hadoop.hbase.master.normalizer.RegionNormalizerFactory;
-107import 
org.apache.hadoop.hbase.master.procedure.AddColumnFamilyProcedure;
-108import 
org.apache.hadoop.hbase.master.procedure.CreateTableProcedure;
-109import 
org.apache.hadoop.hbase.master.procedure.DeleteColumnFamilyProcedure;
-110import 
org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure;
-111import 
org.apache.hadoop.hbase.master.procedure.DisableTableProcedure;
-112import 
org.apache.hadoop.hbase.master.procedure.EnableTableProcedure;
-113import 
org.apache.hadoop.hbase.master.procedure.MasterProcedureConstants;
-114import 
org.apache.hadoop.hbase.master.procedure.MasterProcedureEnv;
-115import 
org.apache.hadoop.hbase.master.procedure.MasterProcedureScheduler;
-116import 
org.apache.hadoop.hbase.master.procedure.MasterProcedureUtil;
-117import 
org.apache.hadoop.hbas

[13/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-05-09 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.ConnectionManager.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.ConnectionManager.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.ConnectionManager.html
index 2987e7b..7f61b54 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.ConnectionManager.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.ConnectionManager.html
@@ -135,1305 +135,1251 @@
 127  private Listener listener = null;
 128  protected Responder responder = null;
 129
-130  /**
-131   * Datastructure that holds all 
necessary to a method invocation and then afterward, carries
-132   * the result.
-133   */
-134  @InterfaceStability.Evolving
-135  public class Call extends 
RpcServer.Call {
-136
-137protected Responder responder;
+130  /** Listens on the socket. Creates jobs 
for the handler threads*/
+131  private class Listener extends Thread 
{
+132
+133private ServerSocketChannel 
acceptChannel = null; //the accept channel
+134private Selector selector = null; 
//the selector that we use for the server
+135private Reader[] readers = null;
+136private int currentReader = 0;
+137private final int 
readerPendingConnectionQueueLength;
 138
-139
@edu.umd.cs.findbugs.annotations.SuppressWarnings(value="NP_NULL_ON_SOME_PATH",
-140justification="Can't figure why 
this complaint is happening... see below")
-141Call(int id, final BlockingService 
service, final MethodDescriptor md,
-142RequestHeader header, Message 
param, CellScanner cellScanner,
-143RpcServer.Connection connection, 
long size, TraceInfo tinfo,
-144final InetAddress remoteAddress, 
int timeout, CallCleanup reqCleanup,
-145Responder responder) {
-146  super(id, service, md, header, 
param, cellScanner, connection, size,
-147  tinfo, remoteAddress, timeout, 
reqCleanup);
-148  this.responder = responder;
-149}
+139private ExecutorService readPool;
+140
+141public Listener(final String name) 
throws IOException {
+142  super(name);
+143  // The backlog of requests that we 
will have the serversocket carry.
+144  int backlogLength = 
conf.getInt("hbase.ipc.server.listen.queue.size", 128);
+145  readerPendingConnectionQueueLength 
=
+146  
conf.getInt("hbase.ipc.server.read.connection-queue.size", 100);
+147  // Create a new server socket and 
set to non blocking mode
+148  acceptChannel = 
ServerSocketChannel.open();
+149  
acceptChannel.configureBlocking(false);
 150
-151/**
-152 * Call is done. Execution happened 
and we returned results to client. It is now safe to
-153 * cleanup.
-154 */
-155
@edu.umd.cs.findbugs.annotations.SuppressWarnings(value="IS2_INCONSISTENT_SYNC",
-156justification="Presume the lock 
on processing request held by caller is protection enough")
-157@Override
-158void done() {
-159  super.done();
-160  this.getConnection().decRpcCount(); 
// Say that we're done with this call.
-161}
-162
-163@Override
-164public long disconnectSince() {
-165  if 
(!getConnection().isConnectionOpen()) {
-166return System.currentTimeMillis() 
- timestamp;
-167  } else {
-168return -1L;
-169  }
-170}
-171
-172@Override
-173public synchronized void 
sendResponseIfReady() throws IOException {
-174  // set param null to reduce memory 
pressure
-175  this.param = null;
-176  this.responder.doRespond(this);
+151  // Bind the server socket to the 
binding addrees (can be different from the default interface)
+152  bind(acceptChannel.socket(), 
bindAddress, backlogLength);
+153  port = 
acceptChannel.socket().getLocalPort(); //Could be an ephemeral port
+154  address = 
(InetSocketAddress)acceptChannel.socket().getLocalSocketAddress();
+155  // create a selector;
+156  selector = Selector.open();
+157
+158  readers = new 
Reader[readThreads];
+159  // Why this executor thing? Why not 
like hadoop just start up all the threads? I suppose it
+160  // has an advantage in that it is 
easy to shutdown the pool.
+161  readPool = 
Executors.newFixedThreadPool(readThreads,
+162new 
ThreadFactoryBuilder().setNameFormat(
+163  
"RpcServer.reader=%d,bindAddress=" + bindAddress.getHostName() +
+164  ",port=" + 
port).setDaemon(true)
+165
.setUncaughtExceptionHandler(Threads.LOGGING_EXCEPTION_HANDLER).build());
+166  for (int i = 0; i < readThreads; 
++i) {
+167Reader reader = new Reader();
+168readers[i] = reader;
+169readPool.execute(reader);
+170  }
+171  LOG.info(getName() + ": started " + 
readThreads + " reader(s) listening on port=" + port);

[04/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-05-09 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.html
--
diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.html 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.html
index 66fe094..17bc96c 100644
--- a/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.html
+++ b/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.html
@@ -57,3290 +57,3303 @@
 049import org.apache.commons.logging.Log;
 050import 
org.apache.commons.logging.LogFactory;
 051import 
org.apache.hadoop.conf.Configuration;
-052import org.apache.hadoop.fs.Path;
-053import 
org.apache.hadoop.hbase.ClusterStatus;
-054import 
org.apache.hadoop.hbase.CoordinatedStateException;
-055import 
org.apache.hadoop.hbase.CoordinatedStateManager;
-056import 
org.apache.hadoop.hbase.DoNotRetryIOException;
-057import 
org.apache.hadoop.hbase.HBaseIOException;
-058import 
org.apache.hadoop.hbase.HBaseInterfaceAudience;
-059import 
org.apache.hadoop.hbase.HColumnDescriptor;
-060import 
org.apache.hadoop.hbase.HConstants;
-061import 
org.apache.hadoop.hbase.HRegionInfo;
-062import 
org.apache.hadoop.hbase.HTableDescriptor;
-063import 
org.apache.hadoop.hbase.MasterNotRunningException;
-064import 
org.apache.hadoop.hbase.MetaTableAccessor;
-065import 
org.apache.hadoop.hbase.NamespaceDescriptor;
-066import 
org.apache.hadoop.hbase.PleaseHoldException;
-067import 
org.apache.hadoop.hbase.ProcedureInfo;
-068import 
org.apache.hadoop.hbase.ScheduledChore;
-069import 
org.apache.hadoop.hbase.ServerLoad;
-070import 
org.apache.hadoop.hbase.ServerName;
-071import 
org.apache.hadoop.hbase.TableDescriptors;
-072import 
org.apache.hadoop.hbase.TableName;
-073import 
org.apache.hadoop.hbase.TableNotDisabledException;
-074import 
org.apache.hadoop.hbase.TableNotFoundException;
-075import 
org.apache.hadoop.hbase.UnknownRegionException;
-076import 
org.apache.hadoop.hbase.classification.InterfaceAudience;
-077import 
org.apache.hadoop.hbase.client.MasterSwitchType;
-078import 
org.apache.hadoop.hbase.client.Result;
-079import 
org.apache.hadoop.hbase.client.TableState;
-080import 
org.apache.hadoop.hbase.coprocessor.BypassCoprocessorException;
-081import 
org.apache.hadoop.hbase.coprocessor.CoprocessorHost;
-082import 
org.apache.hadoop.hbase.exceptions.DeserializationException;
-083import 
org.apache.hadoop.hbase.exceptions.MergeRegionException;
-084import 
org.apache.hadoop.hbase.executor.ExecutorType;
-085import 
org.apache.hadoop.hbase.favored.FavoredNodesManager;
-086import 
org.apache.hadoop.hbase.favored.FavoredNodesPromoter;
-087import 
org.apache.hadoop.hbase.http.InfoServer;
-088import 
org.apache.hadoop.hbase.ipc.CoprocessorRpcUtils;
-089import 
org.apache.hadoop.hbase.ipc.RpcServer;
-090import 
org.apache.hadoop.hbase.ipc.ServerNotRunningYetException;
-091import 
org.apache.hadoop.hbase.master.MasterRpcServices.BalanceSwitchMode;
-092import 
org.apache.hadoop.hbase.master.balancer.BalancerChore;
-093import 
org.apache.hadoop.hbase.master.balancer.BaseLoadBalancer;
-094import 
org.apache.hadoop.hbase.master.balancer.ClusterStatusChore;
-095import 
org.apache.hadoop.hbase.master.balancer.LoadBalancerFactory;
-096import 
org.apache.hadoop.hbase.master.cleaner.HFileCleaner;
-097import 
org.apache.hadoop.hbase.master.cleaner.LogCleaner;
-098import 
org.apache.hadoop.hbase.master.cleaner.ReplicationMetaCleaner;
-099import 
org.apache.hadoop.hbase.master.cleaner.ReplicationZKNodeCleaner;
-100import 
org.apache.hadoop.hbase.master.cleaner.ReplicationZKNodeCleanerChore;
-101import 
org.apache.hadoop.hbase.master.locking.LockManager;
-102import 
org.apache.hadoop.hbase.master.normalizer.NormalizationPlan;
-103import 
org.apache.hadoop.hbase.master.normalizer.NormalizationPlan.PlanType;
-104import 
org.apache.hadoop.hbase.master.normalizer.RegionNormalizer;
-105import 
org.apache.hadoop.hbase.master.normalizer.RegionNormalizerChore;
-106import 
org.apache.hadoop.hbase.master.normalizer.RegionNormalizerFactory;
-107import 
org.apache.hadoop.hbase.master.procedure.AddColumnFamilyProcedure;
-108import 
org.apache.hadoop.hbase.master.procedure.CreateTableProcedure;
-109import 
org.apache.hadoop.hbase.master.procedure.DeleteColumnFamilyProcedure;
-110import 
org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure;
-111import 
org.apache.hadoop.hbase.master.procedure.DisableTableProcedure;
-112import 
org.apache.hadoop.hbase.master.procedure.EnableTableProcedure;
-113import 
org.apache.hadoop.hbase.master.procedure.MasterProcedureConstants;
-114import 
org.apache.hadoop.hbase.master.procedure.MasterProcedureEnv;
-115import 
org.apache.hadoop.hbase.master.procedure.MasterProcedureScheduler;
-116import 
org.apache.hadoop.hbase.master.procedure.MasterProcedureUtil;
-117import 
org.apache.hadoop.hbase.master.procedure.MergeTableRegionsProcedure;
-118import 
org.apache.hadoop.hbase.master.proc

[22/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-05-09 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.BlockingServiceAndInterface.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.BlockingServiceAndInterface.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.BlockingServiceAndInterface.html
index 5cc356a..9e1c66c 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.BlockingServiceAndInterface.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.BlockingServiceAndInterface.html
@@ -70,2037 +70,1559 @@
 062import 
org.apache.hadoop.hbase.client.VersionInfoUtil;
 063import 
org.apache.hadoop.hbase.codec.Codec;
 064import 
org.apache.hadoop.hbase.conf.ConfigurationObserver;
-065import 
org.apache.hadoop.hbase.exceptions.RegionMovedException;
-066import 
org.apache.hadoop.hbase.exceptions.RequestTooBigException;
-067import 
org.apache.hadoop.hbase.io.ByteBufferListOutputStream;
-068import 
org.apache.hadoop.hbase.io.ByteBufferOutputStream;
-069import 
org.apache.hadoop.hbase.io.ByteBufferPool;
-070import 
org.apache.hadoop.hbase.io.crypto.aes.CryptoAES;
-071import 
org.apache.hadoop.hbase.monitoring.MonitoredRPCHandler;
-072import 
org.apache.hadoop.hbase.monitoring.TaskMonitor;
-073import 
org.apache.hadoop.hbase.nio.ByteBuff;
-074import 
org.apache.hadoop.hbase.nio.MultiByteBuff;
-075import 
org.apache.hadoop.hbase.nio.SingleByteBuff;
-076import 
org.apache.hadoop.hbase.regionserver.RSRpcServices;
-077import 
org.apache.hadoop.hbase.security.AccessDeniedException;
-078import 
org.apache.hadoop.hbase.security.AuthMethod;
-079import 
org.apache.hadoop.hbase.security.HBaseSaslRpcServer;
-080import 
org.apache.hadoop.hbase.security.HBaseSaslRpcServer.SaslDigestCallbackHandler;
-081import 
org.apache.hadoop.hbase.security.HBaseSaslRpcServer.SaslGssCallbackHandler;
-082import 
org.apache.hadoop.hbase.security.SaslStatus;
-083import 
org.apache.hadoop.hbase.security.SaslUtil;
-084import 
org.apache.hadoop.hbase.security.User;
-085import 
org.apache.hadoop.hbase.security.UserProvider;
-086import 
org.apache.hadoop.hbase.security.token.AuthenticationTokenSecretManager;
-087import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService;
-088import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteInput;
-089import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString;
-090import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream;
-091import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream;
-092import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor;
-093import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.Message;
-094import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException;
-095import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.TextFormat;
-096import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.UnsafeByteOperations;
-097import 
org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
-098import 
org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos;
-099import 
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.VersionInfo;
-100import 
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos;
-101import 
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.CellBlockMeta;
-102import 
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ConnectionHeader;
-103import 
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ExceptionResponse;
-104import 
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader;
-105import 
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ResponseHeader;
-106import 
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation;
-107import 
org.apache.hadoop.hbase.util.ByteBufferUtils;
-108import 
org.apache.hadoop.hbase.util.Bytes;
-109import 
org.apache.hadoop.hbase.util.Pair;
-110import 
org.apache.hadoop.io.BytesWritable;
-111import org.apache.hadoop.io.Writable;
-112import 
org.apache.hadoop.io.WritableUtils;
-113import 
org.apache.hadoop.io.compress.CompressionCodec;
-114import 
org.apache.hadoop.security.UserGroupInformation;
-115import 
org.apache.hadoop.security.UserGroupInformation.AuthenticationMethod;
-116import 
org.apache.hadoop.security.authorize.AuthorizationException;
-117import 
org.apache.hadoop.security.authorize.PolicyProvider;
-118import 
org.apache.hadoop.security.authorize.ProxyUsers;
-119import 
org.apache.hadoop.security.authorize.ServiceAuthorizationManager;
-120import 
org.apache.hadoop.security.token.SecretManager;
-121import 
org.apache.hadoop.security.token.SecretManager.InvalidToken;
-122import 
org.apache.hadoop.security.token.TokenIdentifier;
-123import 
org.apache.hadoop.util.StringUtils;
-124import org.apache.htrace.TraceInfo;
-125import 
org.codehaus.jackson.map.ObjectMapper

[07/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-05-09 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.InitializationMonitor.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.InitializationMonitor.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.InitializationMonitor.html
index 66fe094..17bc96c 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.InitializationMonitor.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.InitializationMonitor.html
@@ -57,3290 +57,3303 @@
 049import org.apache.commons.logging.Log;
 050import 
org.apache.commons.logging.LogFactory;
 051import 
org.apache.hadoop.conf.Configuration;
-052import org.apache.hadoop.fs.Path;
-053import 
org.apache.hadoop.hbase.ClusterStatus;
-054import 
org.apache.hadoop.hbase.CoordinatedStateException;
-055import 
org.apache.hadoop.hbase.CoordinatedStateManager;
-056import 
org.apache.hadoop.hbase.DoNotRetryIOException;
-057import 
org.apache.hadoop.hbase.HBaseIOException;
-058import 
org.apache.hadoop.hbase.HBaseInterfaceAudience;
-059import 
org.apache.hadoop.hbase.HColumnDescriptor;
-060import 
org.apache.hadoop.hbase.HConstants;
-061import 
org.apache.hadoop.hbase.HRegionInfo;
-062import 
org.apache.hadoop.hbase.HTableDescriptor;
-063import 
org.apache.hadoop.hbase.MasterNotRunningException;
-064import 
org.apache.hadoop.hbase.MetaTableAccessor;
-065import 
org.apache.hadoop.hbase.NamespaceDescriptor;
-066import 
org.apache.hadoop.hbase.PleaseHoldException;
-067import 
org.apache.hadoop.hbase.ProcedureInfo;
-068import 
org.apache.hadoop.hbase.ScheduledChore;
-069import 
org.apache.hadoop.hbase.ServerLoad;
-070import 
org.apache.hadoop.hbase.ServerName;
-071import 
org.apache.hadoop.hbase.TableDescriptors;
-072import 
org.apache.hadoop.hbase.TableName;
-073import 
org.apache.hadoop.hbase.TableNotDisabledException;
-074import 
org.apache.hadoop.hbase.TableNotFoundException;
-075import 
org.apache.hadoop.hbase.UnknownRegionException;
-076import 
org.apache.hadoop.hbase.classification.InterfaceAudience;
-077import 
org.apache.hadoop.hbase.client.MasterSwitchType;
-078import 
org.apache.hadoop.hbase.client.Result;
-079import 
org.apache.hadoop.hbase.client.TableState;
-080import 
org.apache.hadoop.hbase.coprocessor.BypassCoprocessorException;
-081import 
org.apache.hadoop.hbase.coprocessor.CoprocessorHost;
-082import 
org.apache.hadoop.hbase.exceptions.DeserializationException;
-083import 
org.apache.hadoop.hbase.exceptions.MergeRegionException;
-084import 
org.apache.hadoop.hbase.executor.ExecutorType;
-085import 
org.apache.hadoop.hbase.favored.FavoredNodesManager;
-086import 
org.apache.hadoop.hbase.favored.FavoredNodesPromoter;
-087import 
org.apache.hadoop.hbase.http.InfoServer;
-088import 
org.apache.hadoop.hbase.ipc.CoprocessorRpcUtils;
-089import 
org.apache.hadoop.hbase.ipc.RpcServer;
-090import 
org.apache.hadoop.hbase.ipc.ServerNotRunningYetException;
-091import 
org.apache.hadoop.hbase.master.MasterRpcServices.BalanceSwitchMode;
-092import 
org.apache.hadoop.hbase.master.balancer.BalancerChore;
-093import 
org.apache.hadoop.hbase.master.balancer.BaseLoadBalancer;
-094import 
org.apache.hadoop.hbase.master.balancer.ClusterStatusChore;
-095import 
org.apache.hadoop.hbase.master.balancer.LoadBalancerFactory;
-096import 
org.apache.hadoop.hbase.master.cleaner.HFileCleaner;
-097import 
org.apache.hadoop.hbase.master.cleaner.LogCleaner;
-098import 
org.apache.hadoop.hbase.master.cleaner.ReplicationMetaCleaner;
-099import 
org.apache.hadoop.hbase.master.cleaner.ReplicationZKNodeCleaner;
-100import 
org.apache.hadoop.hbase.master.cleaner.ReplicationZKNodeCleanerChore;
-101import 
org.apache.hadoop.hbase.master.locking.LockManager;
-102import 
org.apache.hadoop.hbase.master.normalizer.NormalizationPlan;
-103import 
org.apache.hadoop.hbase.master.normalizer.NormalizationPlan.PlanType;
-104import 
org.apache.hadoop.hbase.master.normalizer.RegionNormalizer;
-105import 
org.apache.hadoop.hbase.master.normalizer.RegionNormalizerChore;
-106import 
org.apache.hadoop.hbase.master.normalizer.RegionNormalizerFactory;
-107import 
org.apache.hadoop.hbase.master.procedure.AddColumnFamilyProcedure;
-108import 
org.apache.hadoop.hbase.master.procedure.CreateTableProcedure;
-109import 
org.apache.hadoop.hbase.master.procedure.DeleteColumnFamilyProcedure;
-110import 
org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure;
-111import 
org.apache.hadoop.hbase.master.procedure.DisableTableProcedure;
-112import 
org.apache.hadoop.hbase.master.procedure.EnableTableProcedure;
-113import 
org.apache.hadoop.hbase.master.procedure.MasterProcedureConstants;
-114import 
org.apache.hadoop.hbase.master.procedure.MasterProcedureEnv;
-115import 
org.apache.hadoop.hbase.master.procedure.MasterProcedureScheduler;
-116import 
org.apache.hadoop.hbase.master.procedure.MasterProcedureUtil;
-117import 
or

hbase git commit: HBASE-16356 REST API scanner: row prefix filter and custom filter parameters are mutually exclusive (Ben Watson)

2017-05-09 Thread tedyu
Repository: hbase
Updated Branches:
  refs/heads/master 51d4c68b7 -> ac1024af2


HBASE-16356 REST API scanner: row prefix filter and custom filter parameters 
are mutually exclusive (Ben Watson)


Project: http://git-wip-us.apache.org/repos/asf/hbase/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/ac1024af
Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/ac1024af
Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/ac1024af

Branch: refs/heads/master
Commit: ac1024af213158d6528ffec964f2bf4aadd9ccd3
Parents: 51d4c68
Author: tedyu 
Authored: Tue May 9 07:58:20 2017 -0700
Committer: tedyu 
Committed: Tue May 9 07:58:20 2017 -0700

--
 .../apache/hadoop/hbase/filter/FilterList.java  |  2 +-
 .../apache/hadoop/hbase/rest/TableResource.java | 31 +-
 .../apache/hadoop/hbase/rest/TestTableScan.java | 33 ++--
 3 files changed, 39 insertions(+), 27 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/hbase/blob/ac1024af/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FilterList.java
--
diff --git 
a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FilterList.java 
b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FilterList.java
index d533026..0742b22 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FilterList.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FilterList.java
@@ -148,7 +148,7 @@ final public class FilterList extends FilterBase {
 return filters;
   }
 
-  private int size() {
+  public int size() {
 return filters.size();
   }
 

http://git-wip-us.apache.org/repos/asf/hbase/blob/ac1024af/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/TableResource.java
--
diff --git 
a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/TableResource.java 
b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/TableResource.java
index 9eb21ec..3019e40 100644
--- a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/TableResource.java
+++ b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/TableResource.java
@@ -135,14 +135,14 @@ public class TableResource extends ResourceBase {
   @DefaultValue(Long.MAX_VALUE + "") @QueryParam(Constants.SCAN_END_TIME) 
long endTime,
   @DefaultValue("true") @QueryParam(Constants.SCAN_BATCH_SIZE) boolean 
cacheBlocks,
   @DefaultValue("false") @QueryParam(Constants.SCAN_REVERSED) boolean 
reversed,
-  @DefaultValue("") @QueryParam(Constants.SCAN_FILTER) String filters) {
+  @DefaultValue("") @QueryParam(Constants.SCAN_FILTER) String paramFilter) 
{
 try {
-  Filter filter = null;
+  Filter prefixFilter = null;
   Scan tableScan = new Scan();
   if (scanSpec.indexOf('*') > 0) {
 String prefix = scanSpec.substring(0, scanSpec.indexOf('*'));
 byte[] prefixBytes = Bytes.toBytes(prefix);
-filter = new PrefixFilter(Bytes.toBytes(prefix));
+prefixFilter = new PrefixFilter(Bytes.toBytes(prefix));
 if (startRow.isEmpty()) {
   tableScan.setStartRow(prefixBytes);
 }
@@ -183,22 +183,21 @@ public class TableResource extends ResourceBase {
   tableScan.addFamily(Bytes.toBytes(familysplit[0]));
 }
   }
-  FilterList filterList = null;
-  if (StringUtils.isNotEmpty(filters)) {
-  ParseFilter pf = new ParseFilter();
-  Filter filterParam = pf.parseFilterString(filters);
-  if (filter != null) {
-filterList = new FilterList(filter, filterParam);
-  }
-  else {
-filter = filterParam;
-  }
+  FilterList filterList = new FilterList();
+  if (StringUtils.isNotEmpty(paramFilter)) {
+ParseFilter pf = new ParseFilter();
+Filter parsedParamFilter = pf.parseFilterString(paramFilter);
+if (parsedParamFilter != null) {
+  filterList.addFilter(parsedParamFilter);
+}
+if (prefixFilter != null) {
+  filterList.addFilter(prefixFilter);
+}
   }
-  if (filterList != null) {
+  if (filterList.size() > 0) {
 tableScan.setFilter(filterList);
-  } else if (filter != null) {
-tableScan.setFilter(filter);
   }
+
   int fetchSize = 
this.servlet.getConfiguration().getInt(Constants.SCAN_FETCH_SIZE, 10);
   tableScan.setCaching(fetchSize);
   tableScan.setReversed(reversed);

http://git-wip-us.apache.org/repos/asf/hbase/blob/ac1024af/hbase-rest/src/test/java/org/apache/hadoop/hbase/rest/TestTableScan.java
--
diff --git 
a/hbase-rest/src/test/java/org/apache/hadoop/hbase/rest/TestTableScan.java 
b/hbase-rest/src/test/java/org/apac

hbase git commit: HBASE-18009 Move RpcServer.Call to a separated file

2017-05-09 Thread zhangduo
Repository: hbase
Updated Branches:
  refs/heads/master 959deb0e5 -> 51d4c68b7


HBASE-18009 Move RpcServer.Call to a separated file


Project: http://git-wip-us.apache.org/repos/asf/hbase/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/51d4c68b
Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/51d4c68b
Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/51d4c68b

Branch: refs/heads/master
Commit: 51d4c68b7cce43af1190f9195bfb08963375bc27
Parents: 959deb0
Author: zhangduo 
Authored: Mon May 8 20:36:33 2017 +0800
Committer: zhangduo 
Committed: Tue May 9 14:56:03 2017 +0800

--
 .../org/apache/hadoop/hbase/ipc/CallRunner.java |   4 +-
 .../apache/hadoop/hbase/ipc/NettyRpcServer.java |  91 +---
 .../hadoop/hbase/ipc/NettyServerCall.java   |  67 +++
 .../org/apache/hadoop/hbase/ipc/RpcCall.java|   5 -
 .../org/apache/hadoop/hbase/ipc/RpcServer.java  | 498 +-
 .../org/apache/hadoop/hbase/ipc/ServerCall.java | 527 +++
 .../hadoop/hbase/ipc/SimpleRpcServer.java   | 122 ++---
 .../hadoop/hbase/ipc/SimpleServerCall.java  |  79 +++
 .../apache/hadoop/hbase/ipc/TestCallRunner.java |   3 +-
 .../hbase/ipc/TestSimpleRpcScheduler.java   |  60 +--
 .../hadoop/hbase/security/TestSecureIPC.java|   8 +-
 11 files changed, 774 insertions(+), 690 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/hbase/blob/51d4c68b/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/CallRunner.java
--
diff --git 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/CallRunner.java 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/CallRunner.java
index 0aabc10..f16fc50 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/CallRunner.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/CallRunner.java
@@ -75,8 +75,8 @@ public class CallRunner {
* @deprecated As of release 2.0, this will be removed in HBase 3.0
*/
   @Deprecated
-  public RpcServer.Call getCall() {
-return (RpcServer.Call) call;
+  public ServerCall getCall() {
+return (ServerCall) call;
   }
 
   public void setStatus(MonitoredRPCHandler status) {

http://git-wip-us.apache.org/repos/asf/hbase/blob/51d4c68b/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/NettyRpcServer.java
--
diff --git 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/NettyRpcServer.java 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/NettyRpcServer.java
index be55378..c18b894 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/NettyRpcServer.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/NettyRpcServer.java
@@ -59,7 +59,6 @@ import org.apache.hadoop.conf.Configuration;
 import org.apache.hadoop.hbase.CellScanner;
 import org.apache.hadoop.hbase.HConstants;
 import org.apache.hadoop.hbase.Server;
-import org.apache.hadoop.hbase.classification.InterfaceStability;
 import org.apache.hadoop.hbase.monitoring.MonitoredRPCHandler;
 import org.apache.hadoop.hbase.nio.ByteBuff;
 import org.apache.hadoop.hbase.nio.SingleByteBuff;
@@ -203,13 +202,14 @@ public class NettyRpcServer extends RpcServer {
 this.hostAddress = inetSocketAddress.getAddress().getHostAddress();
   }
   this.remotePort = inetSocketAddress.getPort();
-  this.saslCall = new Call(SASL_CALLID, null, null, null, null, null, this,
-  0, null, null, 0, null);
-  this.setConnectionHeaderResponseCall = new Call(
-  CONNECTION_HEADER_RESPONSE_CALLID, null, null, null, null, null,
-  this, 0, null, null, 0, null);
-  this.authFailedCall = new Call(AUTHORIZATION_FAILED_CALLID, null, null,
-  null, null, null, this, 0, null, null, 0, null);
+  this.saslCall = new NettyServerCall(SASL_CALLID, null, null, null, null, 
null, this, 0, null,
+  null, System.currentTimeMillis(), 0, reservoir, cellBlockBuilder, 
null);
+  this.setConnectionHeaderResponseCall =
+  new NettyServerCall(CONNECTION_HEADER_RESPONSE_CALLID, null, null, 
null, null, null, this,
+  0, null, null, System.currentTimeMillis(), 0, reservoir, 
cellBlockBuilder, null);
+  this.authFailedCall =
+  new NettyServerCall(AUTHORIZATION_FAILED_CALLID, null, null, null, 
null, null, this, 0,
+  null, null, System.currentTimeMillis(), 0, reservoir, 
cellBlockBuilder, null);
 }
 
 void readPreamble(ByteBuf buffer) throws IOException {
@@ -243,7 +243,7 @@ public class NettyRpcServer extends RpcServer {
   AccessDeniedException ae = new AccessDeniedException(
   "Authentication is required");
   setupResponse(authFailedResponse, authFailedCall, ae, 
ae.getMessage());
-  ((Call) a