hbase git commit: HBASE-17343 Changing the default type of CompactingMemStore to be BASIC
Repository: hbase Updated Branches: refs/heads/master 3bcc475ee -> c5cc81d8e HBASE-17343 Changing the default type of CompactingMemStore to be BASIC Project: http://git-wip-us.apache.org/repos/asf/hbase/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/c5cc81d8 Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/c5cc81d8 Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/c5cc81d8 Branch: refs/heads/master Commit: c5cc81d8e31ba76833adf25b6c357205745c23ad Parents: 3bcc475 Author: anastas Authored: Wed May 10 09:02:26 2017 +0300 Committer: anastas Committed: Wed May 10 09:02:26 2017 +0300 -- .../org/apache/hadoop/hbase/regionserver/CompactingMemStore.java | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) -- http://git-wip-us.apache.org/repos/asf/hbase/blob/c5cc81d8/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/CompactingMemStore.java -- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/CompactingMemStore.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/CompactingMemStore.java index b244997..aaf60e3 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/CompactingMemStore.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/CompactingMemStore.java @@ -56,7 +56,7 @@ public class CompactingMemStore extends AbstractMemStore { public static final String COMPACTING_MEMSTORE_TYPE_KEY = "hbase.hregion.compacting.memstore.type"; public static final String COMPACTING_MEMSTORE_TYPE_DEFAULT = - String.valueOf(MemoryCompactionPolicy.NONE); + String.valueOf(MemoryCompactionPolicy.BASIC); // Default fraction of in-memory-flush size w.r.t. flush-to-disk size public static final String IN_MEMORY_FLUSH_THRESHOLD_FACTOR_KEY = "hbase.memstore.inmemoryflush.threshold.factor";
hbase git commit: HBASE-8758 Error in RegionCoprocessorHost class preScanner method documentation
Repository: hbase Updated Branches: refs/heads/branch-1.3 6545b45b1 -> 286394ba6 HBASE-8758 Error in RegionCoprocessorHost class preScanner method documentation Signed-off-by: Chia-Ping Tsai Project: http://git-wip-us.apache.org/repos/asf/hbase/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/286394ba Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/286394ba Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/286394ba Branch: refs/heads/branch-1.3 Commit: 286394ba636f3be0f27ced00d58227b48ff290e8 Parents: 6545b45 Author: Roman Nikitchenko Authored: Wed May 10 11:30:16 2017 +0800 Committer: Chia-Ping Tsai Committed: Wed May 10 11:37:55 2017 +0800 -- .../apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java| 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) -- http://git-wip-us.apache.org/repos/asf/hbase/blob/286394ba/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java -- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java index 814370c..1ed866a 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java @@ -1288,7 +1288,7 @@ public class RegionCoprocessorHost /** * @param scan the Scan specification * @return scanner id to return to client if default operation should be - * bypassed, false otherwise + * bypassed, null otherwise * @exception IOException Exception */ public RegionScanner preScannerOpen(final Scan scan) throws IOException {
hbase git commit: HBASE-8758 Error in RegionCoprocessorHost class preScanner method documentation
Repository: hbase Updated Branches: refs/heads/branch-1.2 afb67a8d9 -> b6d1b19a3 HBASE-8758 Error in RegionCoprocessorHost class preScanner method documentation Signed-off-by: Chia-Ping Tsai Project: http://git-wip-us.apache.org/repos/asf/hbase/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/b6d1b19a Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/b6d1b19a Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/b6d1b19a Branch: refs/heads/branch-1.2 Commit: b6d1b19a3e5a645c7906fb00870775598f9d1514 Parents: afb67a8 Author: Roman Nikitchenko Authored: Wed May 10 11:30:16 2017 +0800 Committer: Chia-Ping Tsai Committed: Wed May 10 11:36:53 2017 +0800 -- .../apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java| 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) -- http://git-wip-us.apache.org/repos/asf/hbase/blob/b6d1b19a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java -- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java index 814370c..1ed866a 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java @@ -1288,7 +1288,7 @@ public class RegionCoprocessorHost /** * @param scan the Scan specification * @return scanner id to return to client if default operation should be - * bypassed, false otherwise + * bypassed, null otherwise * @exception IOException Exception */ public RegionScanner preScannerOpen(final Scan scan) throws IOException {
hbase git commit: HBASE-8758 Error in RegionCoprocessorHost class preScanner method documentation
Repository: hbase Updated Branches: refs/heads/master 815b0f853 -> 3bcc475ee HBASE-8758 Error in RegionCoprocessorHost class preScanner method documentation Signed-off-by: Chia-Ping Tsai Project: http://git-wip-us.apache.org/repos/asf/hbase/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/3bcc475e Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/3bcc475e Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/3bcc475e Branch: refs/heads/master Commit: 3bcc475ee245a083f39b346c18dfdd2892aae745 Parents: 815b0f8 Author: Roman Nikitchenko Authored: Wed May 10 11:30:16 2017 +0800 Committer: Chia-Ping Tsai Committed: Wed May 10 11:35:45 2017 +0800 -- .../apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java| 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) -- http://git-wip-us.apache.org/repos/asf/hbase/blob/3bcc475e/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java -- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java index 64823b6..0abc988 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java @@ -1254,7 +1254,7 @@ public class RegionCoprocessorHost /** * @param scan the Scan specification * @return scanner id to return to client if default operation should be - * bypassed, false otherwise + * bypassed, null otherwise * @exception IOException Exception */ public RegionScanner preScannerOpen(final Scan scan) throws IOException {
hbase git commit: HBASE-8758 Error in RegionCoprocessorHost class preScanner method documentation
Repository: hbase Updated Branches: refs/heads/branch-1 8d51c8371 -> ea89047ab HBASE-8758 Error in RegionCoprocessorHost class preScanner method documentation Signed-off-by: Chia-Ping Tsai Project: http://git-wip-us.apache.org/repos/asf/hbase/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/ea89047a Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/ea89047a Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/ea89047a Branch: refs/heads/branch-1 Commit: ea89047abf20b1dbf55dd1fa758a5545984157be Parents: 8d51c83 Author: Roman Nikitchenko Authored: Wed May 10 11:30:16 2017 +0800 Committer: Chia-Ping Tsai Committed: Wed May 10 11:34:18 2017 +0800 -- .../apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java| 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) -- http://git-wip-us.apache.org/repos/asf/hbase/blob/ea89047a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java -- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java index 8adc3a4..74ecbc8 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java @@ -1290,7 +1290,7 @@ public class RegionCoprocessorHost /** * @param scan the Scan specification * @return scanner id to return to client if default operation should be - * bypassed, false otherwise + * bypassed, null otherwise * @exception IOException Exception */ public RegionScanner preScannerOpen(final Scan scan) throws IOException {
[1/2] hbase git commit: HBASE-17928 Shell tool to clear compaction queues (Guangxu Cheng)
Repository: hbase Updated Branches: refs/heads/master b67f6fecc -> 815b0f853 http://git-wip-us.apache.org/repos/asf/hbase/blob/815b0f85/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/RegionServerObserver.java -- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/RegionServerObserver.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/RegionServerObserver.java index 9fbf4db..140bdbe 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/RegionServerObserver.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/RegionServerObserver.java @@ -171,4 +171,20 @@ public interface RegionServerObserver extends Coprocessor { default void postReplicateLogEntries( final ObserverContext ctx, List entries, CellScanner cells) throws IOException {} + + /** + * This will be called before clearing compaction queues + * @param ctx the environment to interact with the framework and region server. + */ + default void preClearCompactionQueues( + final ObserverContext ctx) + throws IOException {} + + /** + * This will be called after clearing compaction queues + * @param ctx the environment to interact with the framework and region server. + */ + default void postClearCompactionQueues( + final ObserverContext ctx) + throws IOException {} } http://git-wip-us.apache.org/repos/asf/hbase/blob/815b0f85/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/CompactSplitThread.java -- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/CompactSplitThread.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/CompactSplitThread.java index 5356ee1..a74c4cb 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/CompactSplitThread.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/CompactSplitThread.java @@ -672,4 +672,12 @@ public class CompactSplitThread implements CompactionRequestor, PropagatingConfi void shutdownLongCompactions(){ this.longCompactions.shutdown(); } + + public void clearLongCompactionsQueue() { +longCompactions.getQueue().clear(); + } + + public void clearShortCompactionsQueue() { +shortCompactions.getQueue().clear(); + } } http://git-wip-us.apache.org/repos/asf/hbase/blob/815b0f85/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RSRpcServices.java -- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RSRpcServices.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RSRpcServices.java index 5deb6ba..34401f2 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RSRpcServices.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RSRpcServices.java @@ -31,6 +31,7 @@ import java.util.*; import java.util.Map.Entry; import java.util.concurrent.ConcurrentHashMap; import java.util.concurrent.ConcurrentMap; +import java.util.concurrent.atomic.AtomicBoolean; import java.util.concurrent.atomic.AtomicLong; import java.util.concurrent.atomic.LongAdder; @@ -110,6 +111,8 @@ import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil; import org.apache.hadoop.hbase.shaded.protobuf.RequestConverter; import org.apache.hadoop.hbase.shaded.protobuf.ResponseConverter; import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.AdminService; +import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ClearCompactionQueuesRequest; +import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ClearCompactionQueuesResponse; import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CloseRegionForSplitOrMergeRequest; import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CloseRegionForSplitOrMergeResponse; import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CloseRegionRequest; @@ -268,6 +271,8 @@ public class RSRpcServices implements HBaseRPCErrorHandler, */ private final long minimumScanTimeLimitDelta; + final AtomicBoolean clearCompactionQueues = new AtomicBoolean(false); + /** * An Rpc callback for closing a RegionScanner. */ @@ -1611,6 +1616,44 @@ public class RSRpcServices implements HBaseRPCErrorHandler, return builder.build(); } + @Override + @QosPriority(priority=HConstants.ADMIN_QOS) + public ClearCompactionQueuesResponse clearCompactionQueues(RpcController controller, +ClearCompactionQueuesRequest request) throws ServiceException { +LOG.debug("Client=" + RpcServer.getRequestUserName() + "/" + RpcServer.getRemoteAddress() ++ " clear compactions queue"); +ClearCompactionQueuesResponse.Builder respB
[2/2] hbase git commit: HBASE-17928 Shell tool to clear compaction queues (Guangxu Cheng)
HBASE-17928 Shell tool to clear compaction queues (Guangxu Cheng) Project: http://git-wip-us.apache.org/repos/asf/hbase/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/815b0f85 Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/815b0f85 Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/815b0f85 Branch: refs/heads/master Commit: 815b0f853b8005535219d8e3381d5a44c06887f7 Parents: b67f6fe Author: tedyu Authored: Tue May 9 18:32:38 2017 -0700 Committer: tedyu Committed: Tue May 9 18:32:38 2017 -0700 -- .../org/apache/hadoop/hbase/client/Admin.java | 11 + .../apache/hadoop/hbase/client/HBaseAdmin.java | 22 + .../hbase/shaded/protobuf/RequestConverter.java | 10 + .../shaded/protobuf/generated/AdminProtos.java | 1356 +++--- .../src/main/protobuf/Admin.proto | 10 + .../hbase/coprocessor/RegionServerObserver.java | 16 + .../hbase/regionserver/CompactSplitThread.java |8 + .../hbase/regionserver/RSRpcServices.java | 43 + .../RegionServerCoprocessorHost.java| 20 + .../hbase/security/access/AccessController.java |6 + .../hadoop/hbase/master/MockRegionServer.java |8 + hbase-shell/src/main/ruby/hbase/admin.rb| 26 + hbase-shell/src/main/ruby/shell.rb |1 + .../shell/commands/clear_compaction_queues.rb | 41 + 14 files changed, 1418 insertions(+), 160 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hbase/blob/815b0f85/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Admin.java -- diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Admin.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Admin.java index 414c5ac..2efc9cb 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Admin.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Admin.java @@ -24,6 +24,7 @@ import java.util.ArrayList; import java.util.Collection; import java.util.List; import java.util.Map; +import java.util.Set; import java.util.concurrent.Future; import java.util.regex.Pattern; @@ -2026,4 +2027,14 @@ public interface Admin extends Abortable, Closeable { * @throws IOException if a remote or network exception occurs */ void disableTableReplication(final TableName tableName) throws IOException; + + /** + * Clear compacting queues on a regionserver. + * @param sn the region server name + * @param queues the set of queue name + * @throws IOException if a remote or network exception occurs + * @throws InterruptedException + */ + void clearCompactionQueues(final ServerName sn, final Set queues) +throws IOException, InterruptedException; } http://git-wip-us.apache.org/repos/asf/hbase/blob/815b0f85/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HBaseAdmin.java -- diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HBaseAdmin.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HBaseAdmin.java index ac5c239..9192b10 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HBaseAdmin.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HBaseAdmin.java @@ -92,6 +92,7 @@ import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil; import org.apache.hadoop.hbase.shaded.protobuf.RequestConverter; import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos; import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.AdminService; +import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ClearCompactionQueuesRequest; import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CloseRegionRequest; import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CloseRegionResponse; import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CompactRegionRequest; @@ -4227,4 +4228,25 @@ public class HBaseAdmin implements Admin { return otherConf; } + + @Override + public void clearCompactionQueues(final ServerName sn, final Set queues) +throws IOException, InterruptedException { +if (queues == null || queues.size() == 0) { + throw new IllegalArgumentException("queues cannot be null or empty"); +} +final AdminService.BlockingInterface admin = this.connection.getAdmin(sn); +Callable callable = new Callable() { + @Override + public Void call() throws Exception { +// TODO: There is no timeout on this controller. Set one! +HBaseRpcController controller = rpcControllerFactory.newController(); +ClearCompactionQueuesRequest request = +RequestConverter.buildClearCompactionQueuesRequest(queues); +admin.clearCompact
hbase git commit: HBASE-15199 (addendum) - When JRUBY_HOME is specified, update CLASSPATH and HBASE_OPTS when jruby is needed
Repository: hbase Updated Branches: refs/heads/master c38bf1244 -> b67f6fecc HBASE-15199 (addendum) - When JRUBY_HOME is specified, update CLASSPATH and HBASE_OPTS when jruby is needed Signed-off-by: Sean Busbey Signed-off-by: tedyu Signed-off-by: Jerry He Project: http://git-wip-us.apache.org/repos/asf/hbase/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/b67f6fec Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/b67f6fec Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/b67f6fec Branch: refs/heads/master Commit: b67f6fecc173ff1272284f3e47f95d493fab331d Parents: c38bf12 Author: Xiang Li Authored: Sat May 6 11:37:13 2017 +0800 Committer: Sean Busbey Committed: Tue May 9 16:12:45 2017 -0500 -- bin/hbase | 52 +++- bin/hbase.cmd | 44 2 files changed, 51 insertions(+), 45 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hbase/blob/b67f6fec/bin/hbase -- diff --git a/bin/hbase b/bin/hbase index 579c413..fa70f6e 100755 --- a/bin/hbase +++ b/bin/hbase @@ -294,32 +294,34 @@ if [ "$AUTH_AS_SERVER" == "true" ] || [ "$COMMAND" = "hbck" ]; then fi # for jruby -# (1) when JRUBY_HOME is specified explicitly, eg. export JRUBY_HOME=/usr/local/share/jruby -# for all commands, CLASSPATH and HBASE_OPTS are updated according to JRUBY_HOME specified -# (2) when JRUBY_HOME is not specified explicitly -# A. for jruby_cmds defined below, add JRUBY_PACKAGED_WITH_HBASE into CLASSPATH -# B. for other commands, do nothing - -if [ "$JRUBY_HOME" != "" ]; then # JRUBY_HOME is specified explicitly, eg. export JRUBY_HOME=/usr/local/share/jruby - # add jruby.jar into CLASSPATH - CLASSPATH="$JRUBY_HOME/lib/jruby.jar:$CLASSPATH" - - # add jruby to HBASE_OPTS - HBASE_OPTS="$HBASE_OPTS -Djruby.home=$JRUBY_HOME -Djruby.lib=$JRUBY_HOME/lib" - -else # JRUBY_HOME is not specified explicitly - # check if the commmand needs jruby - declare -a jruby_cmds=("shell" "org.jruby.Main") - for cmd in "${jruby_cmds[@]}"; do -if [[ $cmd == "$COMMAND" ]]; then - jruby_needed=true - break -fi - done +# (1) for the commands which need jruby (see jruby_cmds defined below) +# A. when JRUBY_HOME is specified explicitly, eg. export JRUBY_HOME=/usr/local/share/jruby +#CLASSPATH and HBASE_OPTS are updated according to JRUBY_HOME specified +# B. when JRUBY_HOME is not specified explicitly +#add jruby packaged with HBase to CLASSPATH +# (2) for other commands, do nothing + +# check if the commmand needs jruby +declare -a jruby_cmds=("shell" "org.jruby.Main") +for cmd in "${jruby_cmds[@]}"; do + if [[ $cmd == "$COMMAND" ]]; then +jruby_needed=true +break + fi +done + +# the command needs jruby +if [[ $jruby_needed ]]; then + if [ "$JRUBY_HOME" != "" ]; then # JRUBY_HOME is specified explicitly, eg. export JRUBY_HOME=/usr/local/share/jruby +# add jruby.jar into CLASSPATH +CLASSPATH="$JRUBY_HOME/lib/jruby.jar:$CLASSPATH" + +# add jruby to HBASE_OPTS +HBASE_OPTS="$HBASE_OPTS -Djruby.home=$JRUBY_HOME -Djruby.lib=$JRUBY_HOME/lib" - # add JRUBY_PACKAGED_WITH_HBASE to CLASSPATH when jruby is needed - JRUBY_PACKAGED_WITH_HBASE="$HBASE_HOME/lib/ruby/*.jar" - if [[ $jruby_needed ]]; then # the command needs jruby + else # JRUBY_HOME is not specified explicitly +# add jruby packaged with HBase to CLASSPATH +JRUBY_PACKAGED_WITH_HBASE="$HBASE_HOME/lib/ruby/*.jar" for jruby_jar in $JRUBY_PACKAGED_WITH_HBASE; do CLASSPATH=$jruby_jar:$CLASSPATH; done http://git-wip-us.apache.org/repos/asf/hbase/blob/b67f6fec/bin/hbase.cmd -- diff --git a/bin/hbase.cmd b/bin/hbase.cmd index 44cb9d3..67a3bc1 100644 --- a/bin/hbase.cmd +++ b/bin/hbase.cmd @@ -263,30 +263,34 @@ if defined service_entry ( ) @rem for jruby -@rem (1) when JRUBY_HOME is defined -@rem for all commands, CLASSPATH and HBASE_OPTS are updated according to JRUBY_HOME specified -@rem (2) when JRUBY_HOME is not defined -@rem A. for jruby-commands defined below, add JRUBY_PACKAGED_WITH_HBASE into CLASSPATH -@rem B. for other commands, do nothing - -@rem JRUBY_HOME is defined -if defined JRUBY_HOME ( - set CLASSPATH=%CLASSPATH%;%JRUBY_HOME%\lib\jruby.jar - set HBASE_OPTS=%HBASE_OPTS% -Djruby.home="%JRUBY_HOME%" -Djruby.lib="%JRUBY_HOME%\lib" +@rem (1) for the commands which need jruby (see jruby-commands defined below) +@rem A. when JRUBY_HOME is defined +@remCLASSPATH and HBASE_OPTS are updated according to JRUBY_HOME defined +@rem B. when JRUBY_HOME is not defined +@remadd jruby packaged with HBase to CLASSPATH +@rem (2) for other commands,
hbase git commit: HBASE-18017 Reduce frequency of setStoragePolicy failure warnings
Repository: hbase Updated Branches: refs/heads/master 951b23a44 -> c38bf1244 HBASE-18017 Reduce frequency of setStoragePolicy failure warnings Project: http://git-wip-us.apache.org/repos/asf/hbase/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/c38bf124 Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/c38bf124 Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/c38bf124 Branch: refs/heads/master Commit: c38bf12444aca77c7cb12637147c07dc711acbe9 Parents: 951b23a Author: Andrew Purtell Authored: Mon May 8 15:05:59 2017 -0700 Committer: Andrew Purtell Committed: Tue May 9 12:08:05 2017 -0700 -- .../org/apache/hadoop/hbase/util/FSUtils.java | 67 +--- 1 file changed, 44 insertions(+), 23 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hbase/blob/c38bf124/hbase-server/src/main/java/org/apache/hadoop/hbase/util/FSUtils.java -- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/util/FSUtils.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/util/FSUtils.java index c78ba06..284b786 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/util/FSUtils.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/util/FSUtils.java @@ -152,6 +152,9 @@ public abstract class FSUtils { setStoragePolicy(fs, path, storagePolicy); } + private static final Map warningMap = + new ConcurrentHashMap(); + /** * Sets storage policy for given path. * If the passed path is a directory, we'll set the storage policy for all files @@ -187,17 +190,20 @@ public abstract class FSUtils { try { distributed = isDistributedFileSystem(fs); } catch (IOException ioe) { - // This should NEVER happen. - LOG.warn("Failed setStoragePolicy=" + trimmedStoragePolicy + " on path=" + - path + "; failed isDFS test", ioe); + if (!warningMap.containsKey(fs)) { +warningMap.put(fs, true); +LOG.warn("FileSystem isn't an instance of DistributedFileSystem; presuming it doesn't " ++ "support setStoragePolicy. Unable to set storagePolicy=" + trimmedStoragePolicy ++ " on path=" + path); + } else if (LOG.isDebugEnabled()) { +LOG.debug("FileSystem isn't an instance of DistributedFileSystem; presuming it doesn't " ++ "support setStoragePolicy. Unable to set storagePolicy=" + trimmedStoragePolicy ++ " on path=" + path); + } return; } if (distributed) { invokeSetStoragePolicy(fs, path, trimmedStoragePolicy); -} else { - LOG.info("FileSystem isn't an instance of DistributedFileSystem; presuming it doesn't " + - "support setStoragePolicy. Unable to set storagePolicy=" + trimmedStoragePolicy + - " on path=" + path); } } @@ -209,41 +215,56 @@ public abstract class FSUtils { Method m = null; try { m = fs.getClass().getDeclaredMethod("setStoragePolicy", - new Class[] { Path.class, String.class }); +new Class[] { Path.class, String.class }); m.setAccessible(true); } catch (NoSuchMethodException e) { - LOG.info("FileSystem doesn't support setStoragePolicy; HDFS-6584 not available " - + "(hadoop-2.6.0+): " + e.getMessage()); + final String msg = "FileSystem doesn't support setStoragePolicy; HDFS-6584 not available"; + if (!warningMap.containsKey(fs)) { +warningMap.put(fs, true); +LOG.warn(msg, e); + } else if (LOG.isDebugEnabled()) { +LOG.debug(msg, e); + } + m = null; } catch (SecurityException e) { - LOG.info("Don't have access to setStoragePolicy on FileSystems; HDFS-6584 not available " - + "(hadoop-2.6.0+): ", e); + final String msg = "No access to setStoragePolicy on FileSystem; HDFS-6584 not available"; + if (!warningMap.containsKey(fs)) { +warningMap.put(fs, true); +LOG.warn(msg, e); + } else if (LOG.isDebugEnabled()) { +LOG.debug(msg, e); + } m = null; // could happen on setAccessible() } if (m != null) { try { m.invoke(fs, path, storagePolicy); -LOG.info("Set storagePolicy=" + storagePolicy + " for path=" + path); +if (LOG.isDebugEnabled()) { + LOG.debug("Set storagePolicy=" + storagePolicy + " for path=" + path); +} } catch (Exception e) { +// This swallows FNFE, should we be throwing it? seems more likely to indicate dev +// misuse than a runtime problem with HDFS. +if (!warningMap.containsKey(fs)) { + warningMap.put(fs, true); + LOG.warn("Unable to set storagePolicy=" + storagePolicy + " for path=" + path, e); +} else if (LOG.isDebugEnabled()) { + LOG.
hbase git commit: HBASE-11013: Clone Snapshots on Secure Cluster Should provide option to apply Retained User Permissions
Repository: hbase Updated Branches: refs/heads/master ac1024af2 -> 951b23a44 HBASE-11013: Clone Snapshots on Secure Cluster Should provide option to apply Retained User Permissions Signed-off-by: tedyu Project: http://git-wip-us.apache.org/repos/asf/hbase/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/951b23a4 Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/951b23a4 Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/951b23a4 Branch: refs/heads/master Commit: 951b23a44cd90ae4afed9b255de0e678fbfba946 Parents: ac1024a Author: huzheng Authored: Mon May 8 21:01:47 2017 +0800 Committer: tedyu Committed: Tue May 9 09:32:48 2017 -0700 -- .../org/apache/hadoop/hbase/client/Admin.java | 13 + .../apache/hadoop/hbase/client/HBaseAdmin.java | 25 +- .../hbase/security/access/TablePermission.java | 4 + .../shaded/protobuf/generated/MasterProtos.java | 597 +++ .../src/main/protobuf/Master.proto | 1 + .../org/apache/hadoop/hbase/master/HMaster.java | 5 +- .../hadoop/hbase/master/MasterRpcServices.java | 4 +- .../procedure/CloneSnapshotProcedure.java | 21 +- .../hbase/master/snapshot/SnapshotManager.java | 17 +- .../master/snapshot/TakeSnapshotHandler.java| 3 + .../security/access/AccessControlLists.java | 2 +- .../snapshot/SnapshotDescriptionUtils.java | 97 ++- .../hbase/client/TestSnapshotWithAcl.java | 203 +++ .../hbase/security/access/SecureTestUtil.java | 2 +- hbase-shell/src/main/ruby/hbase/admin.rb| 4 +- hbase-shell/src/main/ruby/hbase_constants.rb| 1 + .../main/ruby/shell/commands/clone_snapshot.rb | 11 +- .../hbase/client/TestReplicationShell.java | 1 - 18 files changed, 727 insertions(+), 284 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hbase/blob/951b23a4/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Admin.java -- diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Admin.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Admin.java index 8a9dc61..414c5ac 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Admin.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Admin.java @@ -1534,6 +1534,19 @@ public interface Admin extends Abortable, Closeable { /** * Create a new table by cloning the snapshot content. + * @param snapshotName name of the snapshot to be cloned + * @param tableName name of the table where the snapshot will be restored + * @param restoreAcl true to clone acl into newly created table + * @throws IOException if a remote or network exception occurs + * @throws TableExistsException if table to be created already exists + * @throws RestoreSnapshotException if snapshot failed to be cloned + * @throws IllegalArgumentException if the specified table has not a valid name + */ + void cloneSnapshot(final String snapshotName, final TableName tableName, final boolean restoreAcl) + throws IOException, TableExistsException, RestoreSnapshotException; + + /** + * Create a new table by cloning the snapshot content. * * @param snapshotName name of the snapshot to be cloned * @param tableName name of the table where the snapshot will be restored http://git-wip-us.apache.org/repos/asf/hbase/blob/951b23a4/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HBaseAdmin.java -- diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HBaseAdmin.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HBaseAdmin.java index 0c62688..ac5c239 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HBaseAdmin.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HBaseAdmin.java @@ -2572,7 +2572,7 @@ public class HBaseAdmin implements Admin { try { // Restore snapshot get( -internalRestoreSnapshotAsync(snapshotName, tableName), +internalRestoreSnapshotAsync(snapshotName, tableName, false), syncWaitTimeout, TimeUnit.MILLISECONDS); } catch (IOException e) { @@ -2581,7 +2581,7 @@ public class HBaseAdmin implements Admin { if (takeFailSafeSnapshot) { try { get( -internalRestoreSnapshotAsync(failSafeSnapshotSnapshotName, tableName), +internalRestoreSnapshotAsync(failSafeSnapshotSnapshotName, tableName, false), syncWaitTimeout, TimeUnit.MILLISECONDS); String msg = "Restore snapshot=" + snapshotName + @@ -2624,7 +2624,7 @@ public class HBaseAdmin implements Admin { throw new TableNotDisabledException(tableName); } -return internalRest
[50/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/checkstyle-aggregate.html -- diff --git a/checkstyle-aggregate.html b/checkstyle-aggregate.html index 153e8e6..27b828c 100644 --- a/checkstyle-aggregate.html +++ b/checkstyle-aggregate.html @@ -7,7 +7,7 @@ - + Apache HBase – Checkstyle Results @@ -286,10 +286,10 @@ Warnings Errors -2156 +2159 0 0 -14354 +14355 Files @@ -2457,7 +2457,7 @@ org/apache/hadoop/hbase/ipc/NettyRpcServer.java 0 0 -4 +3 org/apache/hadoop/hbase/ipc/PriorityFunction.java 0 @@ -2517,7 +2517,7 @@ org/apache/hadoop/hbase/ipc/RpcServer.java 0 0 -57 +53 org/apache/hadoop/hbase/ipc/RpcServerFactory.java 0 @@ -2529,4511 +2529,4516 @@ 0 4 +org/apache/hadoop/hbase/ipc/ServerCall.java +0 +0 +3 + org/apache/hadoop/hbase/ipc/SimpleRpcScheduler.java 0 0 8 - + org/apache/hadoop/hbase/ipc/SimpleRpcServer.java 0 0 -25 - +26 + org/apache/hadoop/hbase/mapred/Driver.java 0 0 4 - + org/apache/hadoop/hbase/mapred/GroupingTableMap.java 0 0 11 - + org/apache/hadoop/hbase/mapred/HRegionPartitioner.java 0 0 4 - + org/apache/hadoop/hbase/mapred/IdentityTableMap.java 0 0 8 - + org/apache/hadoop/hbase/mapred/IdentityTableReduce.java 0 0 7 - + org/apache/hadoop/hbase/mapred/MultiTableSnapshotInputFormat.java 0 0 5 - + org/apache/hadoop/hbase/mapred/RowCounter.java 0 0 7 - + org/apache/hadoop/hbase/mapred/TableInputFormat.java 0 0 1 - + org/apache/hadoop/hbase/mapred/TableInputFormatBase.java 0 0 6 - + org/apache/hadoop/hbase/mapred/TableMap.java 0 0 1 - + org/apache/hadoop/hbase/mapred/TableMapReduceUtil.java 0 0 11 - + org/apache/hadoop/hbase/mapred/TableOutputFormat.java 0 0 2 - + org/apache/hadoop/hbase/mapred/TableRecordReader.java 0 0 5 - + org/apache/hadoop/hbase/mapred/TableRecordReaderImpl.java 0 0 7 - + org/apache/hadoop/hbase/mapred/TableReduce.java 0 0 1 - + org/apache/hadoop/hbase/mapred/TableSnapshotInputFormat.java 0 0 4 - + org/apache/hadoop/hbase/mapred/TableSplit.java 0 0 17 - + org/apache/hadoop/hbase/mapreduce/CellCounter.java 0 0 10 - + org/apache/hadoop/hbase/mapreduce/CellCreator.java 0 0 6 - + org/apache/hadoop/hbase/mapreduce/CopyTable.java 0 0 12 - + org/apache/hadoop/hbase/mapreduce/DefaultVisibilityExpressionResolver.java 0 0 2 - + org/apache/hadoop/hbase/mapreduce/Driver.java 0 0 3 - + org/apache/hadoop/hbase/mapreduce/Export.java 0 0 7 - + org/apache/hadoop/hbase/mapreduce/GroupingTableMapper.java 0 0 3 - + org/apache/hadoop/hbase/mapreduce/HFileInputFormat.java 0 0 1 - + org/apache/hadoop/hbase/mapreduce/HFileOutputFormat2.java 0 0 5 - + org/apache/hadoop/hbase/mapreduce/HRegionPartitioner.java 0 0 4 - + org/apache/hadoop/hbase/mapreduce/HashTable.java 0 0 2 - + org/apache/hadoop/hbase/mapreduce/IdentityTableMapper.java 0 0 1 - + org/apache/hadoop/hbase/mapreduce/IdentityTableReducer.java 0 0 1 - + org/apache/hadoop/hbase/mapreduce/Import.java 0 0 9 - + org/apache/hadoop/hbase/mapreduce/ImportTsv.java 0 0 19 - + org/apache/hadoop/hbase/mapreduce/IndexBuilder.java 0 0 1 - + org/apache/hadoop/hbase/mapreduce/JarFinder.java 0 0 1 - + org/apache/hadoop/hbase/mapreduce/JobUtil.java 0 0 4 - + org/apache/hadoop/hbase/mapreduce/KeyValueSortReducer.java 0 0 4 - + org/apache/hadoop/hbase/mapreduce/LoadIncrementalHFiles.java 0 0 6 - + org/apache/hadoop/hbase/mapreduce/MultiHFileOutputFormat.java 0 0 7 - + org/apache/hadoop/hbase/mapreduce/MultiTableInputFormat.java 0 0 1 - + org/apache/hadoop/hbase/mapreduce/MultiTableInputFormatBase.java 0 0 5 - + org/apache/hadoop/hbase/mapreduce/MultiTableOutputFormat.java 0 0 3 - + org/apache/hadoop/hbase/mapreduce/MultiTableSnapshotInputFormat.java 0 0 1 - + org/apache/hadoop/hbase/mapreduce/MultiTableSnapshotInputFormatImpl.java 0 0 12 - + org/apache/hadoop/hbase/mapreduce/MultithreadedTableMapper.java 0 0 7 - + org/apache/hadoop/hbase/mapreduce/PutCombiner.java 0 0 3 - + org/apache/hadoop/hbase/mapreduce/PutSortReducer.java 0 0 2 - + org/apache/hadoop/hbase/mapreduce/ResultSerialization.java 0 0 1 - + org/apache/hadoop/hbase/mapreduce/RowCounter.java 0 0 6 - + org/apache/hadoop/hbase/mapreduce/SampleUploader.java 0 0 1 - + org/apache/hadoop/hbase/mapreduce/SimpleTotalOrderPartitioner.java 0 0 6 - + org/apache/hadoop/hbase/mapreduce/SyncTable.java 0 0 6 - + org/apache/hadoop/hbase/mapreduce/TableInputFormat.java 0 0 4 - + org/apache/hadoop/hbase/mapreduce/TableInputFormatBase.java 0 0 15 - + org/apache/hadoop/hbase/mapreduce/TableMapReduceUtil.java 0 0 54 - + org/apache/hadoop/hbase/mapreduce/TableMapper.java 0 0 1 - + org/apache/hadoop/hbase/mapreduce/TableOutputCommitter.java 0 0 1 - + org/apache/hadoop/hbase/mapreduce/TableOutputFormat.java 0 0 4 - + org/apache/hadoop/h
[38/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/org/apache/hadoop/hbase/ipc/class-use/BufferChain.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/ipc/class-use/BufferChain.html b/devapidocs/org/apache/hadoop/hbase/ipc/class-use/BufferChain.html index 627009f..f851c7e 100644 --- a/devapidocs/org/apache/hadoop/hbase/ipc/class-use/BufferChain.html +++ b/devapidocs/org/apache/hadoop/hbase/ipc/class-use/BufferChain.html @@ -106,7 +106,7 @@ protected BufferChain -RpcServer.Call.response +ServerCall.response Chain of buffers to send as response. @@ -121,7 +121,7 @@ protected BufferChain -RpcServer.Call.wrapWithSasl(BufferChain bc) +ServerCall.wrapWithSasl(BufferChain bc) @@ -141,7 +141,7 @@ protected BufferChain -RpcServer.Call.wrapWithSasl(BufferChain bc) +ServerCall.wrapWithSasl(BufferChain bc) http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/org/apache/hadoop/hbase/ipc/class-use/CellBlockBuilder.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/ipc/class-use/CellBlockBuilder.html b/devapidocs/org/apache/hadoop/hbase/ipc/class-use/CellBlockBuilder.html index d132558..1fb183e 100644 --- a/devapidocs/org/apache/hadoop/hbase/ipc/class-use/CellBlockBuilder.html +++ b/devapidocs/org/apache/hadoop/hbase/ipc/class-use/CellBlockBuilder.html @@ -114,6 +114,10 @@ protected CellBlockBuilder +ServerCall.cellBlockBuilder + + +protected CellBlockBuilder RpcServer.cellBlockBuilder @@ -130,6 +134,58 @@ Codec codec, org.apache.hadoop.io.compress.CompressionCodec compressor) + +NettyServerCall(int id, + org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService service, + org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor md, + org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader header, + org.apache.hadoop.hbase.shaded.com.google.protobuf.Message param, + CellScanner cellScanner, + RpcServer.Connection connection, + long size, + org.apache.htrace.TraceInfo tinfo, + http://docs.oracle.com/javase/8/docs/api/java/net/InetAddress.html?is-external=true"; title="class or interface in java.net">InetAddress remoteAddress, + long receiveTime, + int timeout, + ByteBufferPool reservoir, + CellBlockBuilder cellBlockBuilder, + RpcServer.CallCleanup reqCleanup) + + +ServerCall(int id, + org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService service, + org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor md, + org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader header, + org.apache.hadoop.hbase.shaded.com.google.protobuf.Message param, + CellScanner cellScanner, + RpcServer.Connection connection, + long size, + org.apache.htrace.TraceInfo tinfo, + http://docs.oracle.com/javase/8/docs/api/java/net/InetAddress.html?is-external=true"; title="class or interface in java.net">InetAddress remoteAddress, + long receiveTime, + int timeout, + ByteBufferPool reservoir, + CellBlockBuilder cellBlockBuilder, + RpcServer.CallCleanup reqCleanup) + + +SimpleServerCall(int id, + org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService service, + org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor md, + org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader header, + org.apache.hadoop.hbase.shaded.com.google.protobuf.Message param, +CellScanner cellScanner, +RpcServer.Connection connection, +long size, +org.apache.htrace.TraceInfo tinfo, +http://docs.oracle.com/javase/8/docs/api/java/net/InetAddress.html?is-external=true"; title="class or interface in java.net">InetAddress remoteAddress, +long receiveTime, +int timeout, +ByteBufferPool reservoir, +CellBlockBuilder cellBlockBuilder, +RpcServer.CallCleanup reqCleanup, +SimpleRpcServer.Responder responder) + http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/org/apache/hadoop/hbase/ipc/class-use/NettyRpcServer.Call.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/ipc/class-use/NettyRpcServer.Call.html b/devapidocs/org/apache/hadoop/hbase/ipc/class-use/NettyRpcServer.Call.ht
[30/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.Call.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.Call.html b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.Call.html deleted file mode 100644 index 27e0dee..000 --- a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.Call.html +++ /dev/null @@ -1,612 +0,0 @@ -http://www.w3.org/TR/html4/loose.dtd";> - - -Source code - - - - -001/** -002 * Licensed to the Apache Software Foundation (ASF) under one -003 * or more contributor license agreements. See the NOTICE file -004 * distributed with this work for additional information -005 * regarding copyright ownership. The ASF licenses this file -006 * to you under the Apache License, Version 2.0 (the -007 * "License"); you may not use this file except in compliance -008 * with the License. You may obtain a copy of the License at -009 * -010 * http://www.apache.org/licenses/LICENSE-2.0 -011 * -012 * Unless required by applicable law or agreed to in writing, software -013 * distributed under the License is distributed on an "AS IS" BASIS, -014 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -015 * See the License for the specific language governing permissions and -016 * limitations under the License. -017 */ -018 -019package org.apache.hadoop.hbase.ipc; -020 -021import io.netty.bootstrap.ServerBootstrap; -022import io.netty.buffer.ByteBuf; -023import io.netty.buffer.PooledByteBufAllocator; -024import io.netty.buffer.Unpooled; -025import io.netty.channel.Channel; -026import io.netty.channel.ChannelFuture; -027import io.netty.channel.ChannelFutureListener; -028import io.netty.channel.ChannelHandlerContext; -029import io.netty.channel.ChannelInboundHandlerAdapter; -030import io.netty.channel.ChannelInitializer; -031import io.netty.channel.ChannelOption; -032import io.netty.channel.ChannelOutboundHandlerAdapter; -033import io.netty.channel.ChannelPipeline; -034import io.netty.channel.ChannelPromise; -035import io.netty.channel.EventLoopGroup; -036import io.netty.channel.epoll.EpollEventLoopGroup; -037import io.netty.channel.epoll.EpollServerSocketChannel; -038import io.netty.channel.group.ChannelGroup; -039import io.netty.channel.group.DefaultChannelGroup; -040import io.netty.channel.nio.NioEventLoopGroup; -041import io.netty.channel.socket.SocketChannel; -042import io.netty.channel.socket.nio.NioServerSocketChannel; -043import io.netty.handler.codec.ByteToMessageDecoder; -044import io.netty.handler.codec.LengthFieldBasedFrameDecoder; -045import io.netty.util.concurrent.GlobalEventExecutor; -046 -047import java.io.IOException; -048import java.io.InterruptedIOException; -049import java.net.InetAddress; -050import java.net.InetSocketAddress; -051import java.nio.ByteBuffer; -052import java.util.Arrays; -053import java.util.List; -054import java.util.concurrent.CountDownLatch; -055 -056import org.apache.commons.logging.Log; -057import org.apache.commons.logging.LogFactory; -058import org.apache.hadoop.conf.Configuration; -059import org.apache.hadoop.hbase.CellScanner; -060import org.apache.hadoop.hbase.HConstants; -061import org.apache.hadoop.hbase.Server; -062import org.apache.hadoop.hbase.classification.InterfaceStability; -063import org.apache.hadoop.hbase.monitoring.MonitoredRPCHandler; -064import org.apache.hadoop.hbase.nio.ByteBuff; -065import org.apache.hadoop.hbase.nio.SingleByteBuff; -066import org.apache.hadoop.hbase.security.AccessDeniedException; -067import org.apache.hadoop.hbase.security.AuthMethod; -068import org.apache.hadoop.hbase.security.HBasePolicyProvider; -069import org.apache.hadoop.hbase.security.SaslStatus; -070import org.apache.hadoop.hbase.security.SaslUtil; -071import org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService; -072import org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor; -073import org.apache.hadoop.hbase.shaded.com.google.protobuf.Message; -074import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader; -075import org.apache.hadoop.hbase.util.Bytes; -076import org.apache.hadoop.hbase.util.JVM; -077import org.apache.hadoop.hbase.util.Pair; -078import org.apache.hadoop.io.IntWritable; -079import org.apache.hadoop.security.authorize.ServiceAuthorizationManager; -080import org.apache.htrace.TraceInfo; -081 -082/** -083 * An RPC server with Netty4 implementation. -084 * -085 */ -086public class NettyRpcServer extends RpcServer { -087 -088 public static final Log LOG = LogFactory.getLog(NettyRpcServer.class); -089 -090 protected final InetSocketAddress bindAddress; -091 -092 private final CountDownLatch closed = new CountDownLatch(1); -093 private final Channel serverChannel; -094 private final ChannelGroup allChanne
hbase-site git commit: INFRA-10751 Empty commit
Repository: hbase-site Updated Branches: refs/heads/asf-site 7ef4c5a9a -> 8678c699d INFRA-10751 Empty commit Project: http://git-wip-us.apache.org/repos/asf/hbase-site/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase-site/commit/8678c699 Tree: http://git-wip-us.apache.org/repos/asf/hbase-site/tree/8678c699 Diff: http://git-wip-us.apache.org/repos/asf/hbase-site/diff/8678c699 Branch: refs/heads/asf-site Commit: 8678c699dd903f277455fb29daf542d7d01e676e Parents: 7ef4c5a Author: jenkins Authored: Tue May 9 15:00:39 2017 + Committer: jenkins Committed: Tue May 9 15:00:39 2017 + -- --
[21/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.ByteBuffByteInput.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.ByteBuffByteInput.html b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.ByteBuffByteInput.html index 5cc356a..9e1c66c 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.ByteBuffByteInput.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.ByteBuffByteInput.html @@ -70,2037 +70,1559 @@ 062import org.apache.hadoop.hbase.client.VersionInfoUtil; 063import org.apache.hadoop.hbase.codec.Codec; 064import org.apache.hadoop.hbase.conf.ConfigurationObserver; -065import org.apache.hadoop.hbase.exceptions.RegionMovedException; -066import org.apache.hadoop.hbase.exceptions.RequestTooBigException; -067import org.apache.hadoop.hbase.io.ByteBufferListOutputStream; -068import org.apache.hadoop.hbase.io.ByteBufferOutputStream; -069import org.apache.hadoop.hbase.io.ByteBufferPool; -070import org.apache.hadoop.hbase.io.crypto.aes.CryptoAES; -071import org.apache.hadoop.hbase.monitoring.MonitoredRPCHandler; -072import org.apache.hadoop.hbase.monitoring.TaskMonitor; -073import org.apache.hadoop.hbase.nio.ByteBuff; -074import org.apache.hadoop.hbase.nio.MultiByteBuff; -075import org.apache.hadoop.hbase.nio.SingleByteBuff; -076import org.apache.hadoop.hbase.regionserver.RSRpcServices; -077import org.apache.hadoop.hbase.security.AccessDeniedException; -078import org.apache.hadoop.hbase.security.AuthMethod; -079import org.apache.hadoop.hbase.security.HBaseSaslRpcServer; -080import org.apache.hadoop.hbase.security.HBaseSaslRpcServer.SaslDigestCallbackHandler; -081import org.apache.hadoop.hbase.security.HBaseSaslRpcServer.SaslGssCallbackHandler; -082import org.apache.hadoop.hbase.security.SaslStatus; -083import org.apache.hadoop.hbase.security.SaslUtil; -084import org.apache.hadoop.hbase.security.User; -085import org.apache.hadoop.hbase.security.UserProvider; -086import org.apache.hadoop.hbase.security.token.AuthenticationTokenSecretManager; -087import org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService; -088import org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteInput; -089import org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString; -090import org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream; -091import org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream; -092import org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor; -093import org.apache.hadoop.hbase.shaded.com.google.protobuf.Message; -094import org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException; -095import org.apache.hadoop.hbase.shaded.com.google.protobuf.TextFormat; -096import org.apache.hadoop.hbase.shaded.com.google.protobuf.UnsafeByteOperations; -097import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil; -098import org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos; -099import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.VersionInfo; -100import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos; -101import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.CellBlockMeta; -102import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ConnectionHeader; -103import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ExceptionResponse; -104import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader; -105import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ResponseHeader; -106import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation; -107import org.apache.hadoop.hbase.util.ByteBufferUtils; -108import org.apache.hadoop.hbase.util.Bytes; -109import org.apache.hadoop.hbase.util.Pair; -110import org.apache.hadoop.io.BytesWritable; -111import org.apache.hadoop.io.Writable; -112import org.apache.hadoop.io.WritableUtils; -113import org.apache.hadoop.io.compress.CompressionCodec; -114import org.apache.hadoop.security.UserGroupInformation; -115import org.apache.hadoop.security.UserGroupInformation.AuthenticationMethod; -116import org.apache.hadoop.security.authorize.AuthorizationException; -117import org.apache.hadoop.security.authorize.PolicyProvider; -118import org.apache.hadoop.security.authorize.ProxyUsers; -119import org.apache.hadoop.security.authorize.ServiceAuthorizationManager; -120import org.apache.hadoop.security.token.SecretManager; -121import org.apache.hadoop.security.token.SecretManager.InvalidToken; -122import org.apache.hadoop.security.token.TokenIdentifier; -123import org.apache.hadoop.util.StringUtils; -124import org.apache.htrace.TraceInfo; -125import org.codehaus.jackson.map.ObjectMapper; -126 -127import com.google.common.annotations.V
[32/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html b/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html index 49a36a3..ff4d17b 100644 --- a/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html +++ b/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html @@ -683,20 +683,20 @@ java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true"; title="class or interface in java.lang">Enum(implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true"; title="class or interface in java.lang">Comparable , java.io.http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true"; title="class or interface in java.io">Serializable) -org.apache.hadoop.hbase.regionserver.ScanType -org.apache.hadoop.hbase.regionserver.FlushType -org.apache.hadoop.hbase.regionserver.StoreScanner.StoreScannerCompactionRace -org.apache.hadoop.hbase.regionserver.DefaultHeapMemoryTuner.StepDirection -org.apache.hadoop.hbase.regionserver.MemStoreCompactor.Action -org.apache.hadoop.hbase.regionserver.ScannerContext.NextState -org.apache.hadoop.hbase.regionserver.SplitLogWorker.TaskExecutor.Status +org.apache.hadoop.hbase.regionserver.Region.Operation +org.apache.hadoop.hbase.regionserver.ImmutableSegment.Type org.apache.hadoop.hbase.regionserver.RegionOpeningState org.apache.hadoop.hbase.regionserver.BloomType +org.apache.hadoop.hbase.regionserver.ScannerContext.NextState +org.apache.hadoop.hbase.regionserver.DefaultHeapMemoryTuner.StepDirection +org.apache.hadoop.hbase.regionserver.SplitLogWorker.TaskExecutor.Status +org.apache.hadoop.hbase.regionserver.ScannerContext.LimitScope +org.apache.hadoop.hbase.regionserver.FlushType org.apache.hadoop.hbase.regionserver.MetricsRegionServerSourceFactoryImpl.FactoryStorage -org.apache.hadoop.hbase.regionserver.Region.Operation +org.apache.hadoop.hbase.regionserver.ScanType +org.apache.hadoop.hbase.regionserver.StoreScanner.StoreScannerCompactionRace org.apache.hadoop.hbase.regionserver.Region.FlushResult.Result -org.apache.hadoop.hbase.regionserver.ImmutableSegment.Type -org.apache.hadoop.hbase.regionserver.ScannerContext.LimitScope +org.apache.hadoop.hbase.regionserver.MemStoreCompactor.Action http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/org/apache/hadoop/hbase/regionserver/querymatcher/package-tree.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/querymatcher/package-tree.html b/devapidocs/org/apache/hadoop/hbase/regionserver/querymatcher/package-tree.html index 80fd563..d4c8f41 100644 --- a/devapidocs/org/apache/hadoop/hbase/regionserver/querymatcher/package-tree.html +++ b/devapidocs/org/apache/hadoop/hbase/regionserver/querymatcher/package-tree.html @@ -125,10 +125,10 @@ java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true"; title="class or interface in java.lang">Enum (implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true"; title="class or interface in java.lang">Comparable , java.io.http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true"; title="class or interface in java.io">Serializable) -org.apache.hadoop.hbase.regionserver.querymatcher.StripeCompactionScanQueryMatcher.DropDeletesInOutput -org.apache.hadoop.hbase.regionserver.querymatcher.ScanQueryMatcher.MatchCode org.apache.hadoop.hbase.regionserver.querymatcher.DeleteTracker.DeleteCompare org.apache.hadoop.hbase.regionserver.querymatcher.DeleteTracker.DeleteResult +org.apache.hadoop.hbase.regionserver.querymatcher.ScanQueryMatcher.MatchCode +org.apache.hadoop.hbase.regionserver.querymatcher.StripeCompactionScanQueryMatcher.DropDeletesInOutput http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/org/apache/hadoop/hbase/rest/model/package-tree.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/rest/model/package-tree.html b/devapidocs/org/apache/hadoop/hbase/rest/model/package-tree.html index dbfbf42..d9efa05 100644 --- a/devapidocs/org/apache/hadoop/hbase/rest/model/package-tree.html +++ b/devapidocs/org/apache/hadoop/hbase/rest/model/package-tree.html @@ -110,8 +110,8 @@ java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true"; title="class or interface in java.lang">Enum (implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true"; title="class or interface in java.lang">Comparable , java.io.http://docs.oracle.com/javase/8/docs/
[11/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Listener.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Listener.html b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Listener.html index 2987e7b..7f61b54 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Listener.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Listener.html @@ -135,1305 +135,1251 @@ 127 private Listener listener = null; 128 protected Responder responder = null; 129 -130 /** -131 * Datastructure that holds all necessary to a method invocation and then afterward, carries -132 * the result. -133 */ -134 @InterfaceStability.Evolving -135 public class Call extends RpcServer.Call { -136 -137protected Responder responder; +130 /** Listens on the socket. Creates jobs for the handler threads*/ +131 private class Listener extends Thread { +132 +133private ServerSocketChannel acceptChannel = null; //the accept channel +134private Selector selector = null; //the selector that we use for the server +135private Reader[] readers = null; +136private int currentReader = 0; +137private final int readerPendingConnectionQueueLength; 138 -139 @edu.umd.cs.findbugs.annotations.SuppressWarnings(value="NP_NULL_ON_SOME_PATH", -140justification="Can't figure why this complaint is happening... see below") -141Call(int id, final BlockingService service, final MethodDescriptor md, -142RequestHeader header, Message param, CellScanner cellScanner, -143RpcServer.Connection connection, long size, TraceInfo tinfo, -144final InetAddress remoteAddress, int timeout, CallCleanup reqCleanup, -145Responder responder) { -146 super(id, service, md, header, param, cellScanner, connection, size, -147 tinfo, remoteAddress, timeout, reqCleanup); -148 this.responder = responder; -149} +139private ExecutorService readPool; +140 +141public Listener(final String name) throws IOException { +142 super(name); +143 // The backlog of requests that we will have the serversocket carry. +144 int backlogLength = conf.getInt("hbase.ipc.server.listen.queue.size", 128); +145 readerPendingConnectionQueueLength = +146 conf.getInt("hbase.ipc.server.read.connection-queue.size", 100); +147 // Create a new server socket and set to non blocking mode +148 acceptChannel = ServerSocketChannel.open(); +149 acceptChannel.configureBlocking(false); 150 -151/** -152 * Call is done. Execution happened and we returned results to client. It is now safe to -153 * cleanup. -154 */ -155 @edu.umd.cs.findbugs.annotations.SuppressWarnings(value="IS2_INCONSISTENT_SYNC", -156justification="Presume the lock on processing request held by caller is protection enough") -157@Override -158void done() { -159 super.done(); -160 this.getConnection().decRpcCount(); // Say that we're done with this call. -161} -162 -163@Override -164public long disconnectSince() { -165 if (!getConnection().isConnectionOpen()) { -166return System.currentTimeMillis() - timestamp; -167 } else { -168return -1L; -169 } -170} -171 -172@Override -173public synchronized void sendResponseIfReady() throws IOException { -174 // set param null to reduce memory pressure -175 this.param = null; -176 this.responder.doRespond(this); +151 // Bind the server socket to the binding addrees (can be different from the default interface) +152 bind(acceptChannel.socket(), bindAddress, backlogLength); +153 port = acceptChannel.socket().getLocalPort(); //Could be an ephemeral port +154 address = (InetSocketAddress)acceptChannel.socket().getLocalSocketAddress(); +155 // create a selector; +156 selector = Selector.open(); +157 +158 readers = new Reader[readThreads]; +159 // Why this executor thing? Why not like hadoop just start up all the threads? I suppose it +160 // has an advantage in that it is easy to shutdown the pool. +161 readPool = Executors.newFixedThreadPool(readThreads, +162new ThreadFactoryBuilder().setNameFormat( +163 "RpcServer.reader=%d,bindAddress=" + bindAddress.getHostName() + +164 ",port=" + port).setDaemon(true) +165 .setUncaughtExceptionHandler(Threads.LOGGING_EXCEPTION_HANDLER).build()); +166 for (int i = 0; i < readThreads; ++i) { +167Reader reader = new Reader(); +168readers[i] = reader; +169readPool.execute(reader); +170 } +171 LOG.info(getName() + ": started " + readThreads + " reader(s) listening on port=" + port); +172 +173 // Register accepts on the serv
[08/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleServerCall.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleServerCall.html b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleServerCall.html new file mode 100644 index 000..31df5ee --- /dev/null +++ b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleServerCall.html @@ -0,0 +1,151 @@ +http://www.w3.org/TR/html4/loose.dtd";> + + +Source code + + + + +001/** +002 * Licensed to the Apache Software Foundation (ASF) under one +003 * or more contributor license agreements. See the NOTICE file +004 * distributed with this work for additional information +005 * regarding copyright ownership. The ASF licenses this file +006 * to you under the Apache License, Version 2.0 (the +007 * "License"); you may not use this file except in compliance +008 * with the License. You may obtain a copy of the License at +009 * +010 * http://www.apache.org/licenses/LICENSE-2.0 +011 * +012 * Unless required by applicable law or agreed to in writing, software +013 * distributed under the License is distributed on an "AS IS" BASIS, +014 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +015 * See the License for the specific language governing permissions and +016 * limitations under the License. +017 */ +018package org.apache.hadoop.hbase.ipc; +019 +020import java.io.IOException; +021import java.net.InetAddress; +022 +023import org.apache.hadoop.hbase.CellScanner; +024import org.apache.hadoop.hbase.classification.InterfaceAudience; +025import org.apache.hadoop.hbase.io.ByteBufferPool; +026import org.apache.hadoop.hbase.ipc.RpcServer.CallCleanup; +027import org.apache.hadoop.hbase.ipc.SimpleRpcServer.Connection; +028import org.apache.hadoop.hbase.ipc.SimpleRpcServer.Responder; +029import org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService; +030import org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor; +031import org.apache.hadoop.hbase.shaded.com.google.protobuf.Message; +032import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader; +033import org.apache.htrace.TraceInfo; +034 +035/** +036 * Datastructure that holds all necessary to a method invocation and then afterward, carries the +037 * result. +038 */ +039@InterfaceAudience.Private +040class SimpleServerCall extends ServerCall { +041 +042 long lastSentTime; +043 +044 final Responder responder; +045 +046 @edu.umd.cs.findbugs.annotations.SuppressWarnings(value = "NP_NULL_ON_SOME_PATH", +047 justification = "Can't figure why this complaint is happening... see below") +048 SimpleServerCall(int id, final BlockingService service, final MethodDescriptor md, +049 RequestHeader header, Message param, CellScanner cellScanner, RpcServer.Connection connection, +050 long size, TraceInfo tinfo, final InetAddress remoteAddress, long receiveTime, int timeout, +051 ByteBufferPool reservoir, CellBlockBuilder cellBlockBuilder, CallCleanup reqCleanup, +052 Responder responder) { +053super(id, service, md, header, param, cellScanner, connection, size, tinfo, remoteAddress, +054receiveTime, timeout, reservoir, cellBlockBuilder, reqCleanup); +055this.responder = responder; +056 } +057 +058 /** +059 * Call is done. Execution happened and we returned results to client. It is now safe to cleanup. +060 */ +061 @edu.umd.cs.findbugs.annotations.SuppressWarnings(value = "IS2_INCONSISTENT_SYNC", +062 justification = "Presume the lock on processing request held by caller is protection enough") +063 @Override +064 void done() { +065super.done(); +066this.getConnection().decRpcCount(); // Say that we're done with this call. +067 } +068 +069 @Override +070 public synchronized void sendResponseIfReady() throws IOException { +071// set param null to reduce memory pressure +072this.param = null; +073this.responder.doRespond(this); +074 } +075 +076 Connection getConnection() { +077return (Connection) this.connection; +078 } +079} + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
[47/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/org/apache/hadoop/hbase/backup/package-tree.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/backup/package-tree.html b/devapidocs/org/apache/hadoop/hbase/backup/package-tree.html index 1f603c7..72dbc96 100644 --- a/devapidocs/org/apache/hadoop/hbase/backup/package-tree.html +++ b/devapidocs/org/apache/hadoop/hbase/backup/package-tree.html @@ -165,10 +165,10 @@ java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true"; title="class or interface in java.lang">Enum(implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true"; title="class or interface in java.lang">Comparable , java.io.http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true"; title="class or interface in java.io">Serializable) -org.apache.hadoop.hbase.backup.BackupRestoreConstants.BackupCommand +org.apache.hadoop.hbase.backup.BackupInfo.BackupPhase org.apache.hadoop.hbase.backup.BackupInfo.BackupState org.apache.hadoop.hbase.backup.BackupType -org.apache.hadoop.hbase.backup.BackupInfo.BackupPhase +org.apache.hadoop.hbase.backup.BackupRestoreConstants.BackupCommand http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/org/apache/hadoop/hbase/class-use/CellScanner.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/class-use/CellScanner.html b/devapidocs/org/apache/hadoop/hbase/class-use/CellScanner.html index 20110fb..3fa62ce 100644 --- a/devapidocs/org/apache/hadoop/hbase/class-use/CellScanner.html +++ b/devapidocs/org/apache/hadoop/hbase/class-use/CellScanner.html @@ -441,7 +441,7 @@ protected CellScanner -RpcServer.Call.cellScanner +ServerCall.cellScanner @@ -474,7 +474,7 @@ CellScanner -RpcServer.Call.getCellScanner() +ServerCall.getCellScanner() CellScanner @@ -680,7 +680,7 @@ int timeout) -abstract RpcServer.Call +abstract ServerCall RpcServer.Connection.createCall(int id, org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService service, org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor md, @@ -695,7 +695,7 @@ RpcServer.CallCleanup reqCleanup) -RpcServer.Call +ServerCall SimpleRpcServer.Connection.createCall(int id, org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService service, org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor md, @@ -710,7 +710,7 @@ RpcServer.CallCleanup reqCleanup) -RpcServer.Call +ServerCall NettyRpcServer.NettyConnection.createCall(int id, org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService service, org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor md, @@ -774,7 +774,7 @@ void -RpcServer.Call.setResponse(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message m, +ServerCall.setResponse(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message m, CellScanner cells, http://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true"; title="class or interface in java.lang">Throwable t, http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true"; title="class or interface in java.lang">String errorMsg) @@ -797,49 +797,6 @@ -Call(int id, - org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService service, - org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor md, - org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader header, -org.apache.hadoop.hbase.shaded.com.google.protobuf.Message param, -CellScanner cellScanner, -RpcServer.Connection connection, -long size, -org.apache.htrace.TraceInfo tinfo, -http://docs.oracle.com/javase/8/docs/api/java/net/InetAddress.html?is-external=true"; title="class or interface in java.net">InetAddress remoteAddress, -int timeout, -RpcServer.CallCleanup reqCleanup) - - -Call(int id, - org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService service, - org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor md, - org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader header, -org.apache.hadoop.hbase.shaded.com.google.protobuf.Message param, -CellScanner cellScanner, -RpcServer.Connection connection, -long size, -org.apache.htrace.TraceInfo tinfo, -http://docs.oracle.com/javase/8/docs/api/java/net/InetAddress.html?is-external=true"; title="class or interface in java.net">InetAddress remoteAddress, -int timeout, -RpcServer.CallCleanup reqCleanup) - - -Call(int id, - org.a
[48/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/index-all.html -- diff --git a/devapidocs/index-all.html b/devapidocs/index-all.html index 3a10992..6d90f3b 100644 --- a/devapidocs/index-all.html +++ b/devapidocs/index-all.html @@ -8235,16 +8235,12 @@ call(BlockingService, Descriptors.MethodDescriptor, Message, CellScanner, long, MonitoredRPCHandler, long, int) - Method in class org.apache.hadoop.hbase.ipc.NettyRpcServer -Call(int, BlockingService, Descriptors.MethodDescriptor, RPCProtos.RequestHeader, Message, CellScanner, RpcServer.Connection, long, TraceInfo, InetAddress, int, RpcServer.CallCleanup) - Constructor for class org.apache.hadoop.hbase.ipc.NettyRpcServer.Call - call - Variable in class org.apache.hadoop.hbase.ipc.NettyRpcServer.CallWriteListener call(RpcCall, MonitoredRPCHandler) - Method in class org.apache.hadoop.hbase.ipc.RpcServer This is a server side method, which is invoked over RPC. -Call(int, BlockingService, Descriptors.MethodDescriptor, RPCProtos.RequestHeader, Message, CellScanner, RpcServer.Connection, long, TraceInfo, InetAddress, int, RpcServer.CallCleanup) - Constructor for class org.apache.hadoop.hbase.ipc.RpcServer.Call - call(BlockingService, Descriptors.MethodDescriptor, Message, CellScanner, long, MonitoredRPCHandler) - Method in interface org.apache.hadoop.hbase.ipc.RpcServerInterface Deprecated. @@ -8263,8 +8259,6 @@ call(BlockingService, Descriptors.MethodDescriptor, Message, CellScanner, long, MonitoredRPCHandler, long, int) - Method in class org.apache.hadoop.hbase.ipc.SimpleRpcServer -Call(int, BlockingService, Descriptors.MethodDescriptor, RPCProtos.RequestHeader, Message, CellScanner, RpcServer.Connection, long, TraceInfo, InetAddress, int, RpcServer.CallCleanup, SimpleRpcServer.Responder) - Constructor for class org.apache.hadoop.hbase.ipc.SimpleRpcServer.Call - call() - Method in class org.apache.hadoop.hbase.master.AssignCallable call(MasterObserver, ObserverContext) - Method in class org.apache.hadoop.hbase.master.MasterCoprocessorHost.CoprocessorOperation @@ -8580,7 +8574,7 @@ callWithRetries(RetryingCallable , int) - Method in class org.apache.hadoop.hbase.client.RpcRetryingCallerImpl -CallWriteListener(NettyRpcServer.Call) - Constructor for class org.apache.hadoop.hbase.ipc.NettyRpcServer.CallWriteListener +CallWriteListener(NettyServerCall) - Constructor for class org.apache.hadoop.hbase.ipc.NettyRpcServer.CallWriteListener calSleepTimeMs() - Method in class org.apache.hadoop.hbase.io.hadoopbackport.ThrottledInputStream @@ -9005,6 +8999,8 @@ cellBlockBuilder - Variable in class org.apache.hadoop.hbase.ipc.RpcServer +cellBlockBuilder - Variable in class org.apache.hadoop.hbase.ipc.ServerCall + CellBlockBuilder.ByteBufferOutputStreamSupplier - Class in org.apache.hadoop.hbase.ipc CellBlockBuilder.ByteBufOutputStreamSupplier - Class in org.apache.hadoop.hbase.ipc @@ -9017,7 +9013,7 @@ How much we think the decompressor will expand the original compressed content. -cellBlockStream - Variable in class org.apache.hadoop.hbase.ipc.RpcServer.Call +cellBlockStream - Variable in class org.apache.hadoop.hbase.ipc.ServerCall CellCodec - Class in org.apache.hadoop.hbase.codec @@ -9220,7 +9216,7 @@ cellScanner() - Method in class org.apache.hadoop.hbase.ipc.HBaseRpcControllerImpl -cellScanner - Variable in class org.apache.hadoop.hbase.ipc.RpcServer.Call +cellScanner - Variable in class org.apache.hadoop.hbase.ipc.ServerCall CellScanner(Iterator ) - Constructor for class org.apache.hadoop.hbase.mapreduce.SyncTable.SyncMapper.CellScanner @@ -10942,7 +10938,7 @@ Do the necessary cleanup after the call if needed. -cleanup() - Method in class org.apache.hadoop.hbase.ipc.RpcServer.Call +cleanup() - Method in class org.apache.hadoop.hbase.ipc.ServerCall cleanup(Mapper .Context) - Method in class org.apache.hadoop.hbase.mapreduce.HashTable.HashMapper @@ -14498,6 +14494,8 @@ compareTo(WALKey) - Method in class org.apache.hadoop.hbase.wal.WALKey +compareTo(WALSplitter.MutationReplay) - Method in class org.apache.hadoop.hbase.wal.WALSplitter.MutationReplay + compareToCurrentToken(Cell) - Method in class org.apache.hadoop.hbase.codec.prefixtree.decode.PrefixTreeArraySearcher Compare only the bytes within the window of the current token @@ -15763,10 +15761,10 @@ connection - Variable in class org.apache.hadoop.hbase.ipc.NettyRpcServer.MessageDecoder -connection - Variable in class org.apache.hadoop.hbase.ipc.RpcServer.Call - Connection() - Constructor for class org.apache.hadoop.hbase.ipc.RpcServer.Connection +connection - Variable in class org.apache.hadoop.hbase.ipc.ServerCall + Connection(SocketCh
[45/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/org/apache/hadoop/hbase/ipc/NettyServerCall.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/ipc/NettyServerCall.html b/devapidocs/org/apache/hadoop/hbase/ipc/NettyServerCall.html new file mode 100644 index 000..43d126f --- /dev/null +++ b/devapidocs/org/apache/hadoop/hbase/ipc/NettyServerCall.html @@ -0,0 +1,374 @@ +http://www.w3.org/TR/html4/loose.dtd";> + + + + + +NettyServerCall (Apache HBase 2.0.0-SNAPSHOT API) + + + + + +var methods = {"i0":10,"i1":10,"i2":10}; +var tabs = {65535:["t0","All Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]}; +var altColor = "altColor"; +var rowColor = "rowColor"; +var tableTab = "tableTab"; +var activeTableTab = "activeTableTab"; + + +JavaScript is disabled on your browser. + + + + + +Skip navigation links + + + + +Overview +Package +Class +Use +Tree +Deprecated +Index +Help + + + + +Prev Class +Next Class + + +Frames +No Frames + + +All Classes + + + + + + + +Summary: +Nested | +Field | +Constr | +Method + + +Detail: +Field | +Constr | +Method + + + + + + + + +org.apache.hadoop.hbase.ipc +Class NettyServerCall + + + +http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true"; title="class or interface in java.lang">java.lang.Object + + +org.apache.hadoop.hbase.ipc.ServerCall + + +org.apache.hadoop.hbase.ipc.NettyServerCall + + + + + + + + + +All Implemented Interfaces: +RpcCall, RpcCallContext + + + +@InterfaceAudience.Private +class NettyServerCall +extends ServerCall +Datastructure that holds all necessary to a method invocation and then afterward, carries the + result. + + + + + + + + + + + +Field Summary + + + + +Fields inherited from class org.apache.hadoop.hbase.ipc.ServerCall +cellBlockBuilder, cellBlockStream, cellScanner, connection, deadline, header, id, isError, md, param, receiveTime, remoteAddress, reqCleanup, reservoir, response, rpcCallback, service, size, startTime, timeout, tinfo, user + + + + + + + + +Constructor Summary + +Constructors + +Constructor and Description + + +NettyServerCall(int id, + org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService service, + org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor md, + org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader header, + org.apache.hadoop.hbase.shaded.com.google.protobuf.Message param, + CellScanner cellScanner, + RpcServer.Connection connection, + long size, + org.apache.htrace.TraceInfo tinfo, + http://docs.oracle.com/javase/8/docs/api/java/net/InetAddress.html?is-external=true"; title="class or interface in java.net">InetAddress remoteAddress, + long receiveTime, + int timeout, + ByteBufferPool reservoir, + CellBlockBuilder cellBlockBuilder, + RpcServer.CallCleanup reqCleanup) + + + + + + + + + +Method Summary + +All Methods Instance Methods Concrete Methods + +Modifier and Type +Method and Description + + +(package private) NettyRpcServer.NettyConnection +getConnection() + + +void +sendResponseIfReady() +If we have a response, and delay is not set, then respond immediately. + + + +void +sendResponseIfReady(io.netty.channel.ChannelFutureListener listener) + + + + + + +Methods inherited from class org.apache.hadoop.hbase.ipc.ServerCall +cleanup, createHeaderAndMessageBytes, disconnectSince, done, getCellScanner, getClientVersionInfo, getDeadline, getHeader, getMethod, getParam, getPriority, getReceiveTime, getRemoteAddress, getRemotePort, getRequestUser, getRequestUserName, getResponseBlockSize, getResponseCellSize, getResponseExceptionSize, getService, getSize, getStartTime, getTimeout, getTraceInfo, incrementResponseBlockSize, incrementResponseCellSize, incrementResponseExceptionSize, isClientCellBlockSupported, isRetryImmediatelySupported, setCallBack, setConnectionHeaderResponse, setExceptionResponse, setResponse, setSaslTokenResponse, setStartTime, toShortString, toString, wrapWithSasl + + + + + +Methods inherited from class java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true"; title="class or interface in java.lang">Object +http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#cl
[49/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/checkstyle.rss -- diff --git a/checkstyle.rss b/checkstyle.rss index b77bb56..cda80bb 100644 --- a/checkstyle.rss +++ b/checkstyle.rss @@ -25,8 +25,8 @@ under the License. en-us ©2007 - 2017 The Apache Software Foundation - File: 2156, - Errors: 14354, + File: 2159, + Errors: 14355, Warnings: 0, Infos: 0 @@ -11787,7 +11787,7 @@ under the License. 0 - 25 + 26 @@ -15824,6 +15824,20 @@ under the License. + http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.ipc.SimpleServerCall.java";>org/apache/hadoop/hbase/ipc/SimpleServerCall.java + + + 0 + + + 0 + + + 0 + + + + http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.shaded.com.google.protobuf.Any.java";>org/apache/hadoop/hbase/shaded/com/google/protobuf/Any.java @@ -16967,7 +16981,7 @@ under the License. 0 - 57 + 53 @@ -20271,7 +20285,7 @@ under the License. 0 - 4 + 3 @@ -23183,7 +23197,7 @@ under the License. 0 - 117 + 119 @@ -23454,6 +23468,20 @@ under the License. + http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.ipc.ServerCall.java";>org/apache/hadoop/hbase/ipc/ServerCall.java + + + 0 + + + 0 + + + 3 + + + + http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.regionserver.LeaseListener.java";>org/apache/hadoop/hbase/regionserver/LeaseListener.java @@ -30160,6 +30188,20 @@ under the License. + http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.ipc.NettyServerCall.java";>org/apache/hadoop/hbase/ipc/NettyServerCall.java + + + 0 + + + 0 + + + 0 + + + + http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.ZooKeeperConnectionException.java";>org/apache/hadoop/hbase/ZooKeeperConnectionException.java http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/coc.html -- diff --git a/coc.html b/coc.html index 1cdd55d..c9d7435 100644 --- a/coc.html +++ b/coc.html @@ -7,7 +7,7 @@ - + Apache HBase – Code of Conduct Policy @@ -380,7 +380,7 @@ email to mailto:priv...@hbase.apache.org";>the priv https://www.apache.org/";>The Apache Software Foundation. All rights reserved. - Last Published: 2017-05-06 + Last Published: 2017-05-09 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/cygwin.html -- diff --git a/cygwin.html b/cygwin.html index ad6e1eb..83a8ea1 100644 --- a/cygwin.html +++ b/cygwin.html @@ -7,7 +7,7 @@ - + Apache HBase – Installing Apache HBase (TM) on Windows using Cygwin @@ -679,7 +679,7 @@ Now your HBase server is running, start coding and build that next https://www.apache.org/";>The Apache Software Foundation. All rights reserved. - Last Published: 2017-05-06 + Last Published: 2017-05-09 http://git-wip-us.apache.org/repos/asf/hb
[43/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/org/apache/hadoop/hbase/ipc/RpcServer.CallCleanup.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/ipc/RpcServer.CallCleanup.html b/devapidocs/org/apache/hadoop/hbase/ipc/RpcServer.CallCleanup.html index 2d1c621..09e77fb 100644 --- a/devapidocs/org/apache/hadoop/hbase/ipc/RpcServer.CallCleanup.html +++ b/devapidocs/org/apache/hadoop/hbase/ipc/RpcServer.CallCleanup.html @@ -49,7 +49,7 @@ var activeTableTab = "activeTableTab"; -Prev Class +Prev Class Next Class @@ -110,7 +110,7 @@ var activeTableTab = "activeTableTab"; http://docs.oracle.com/javase/8/docs/api/java/lang/FunctionalInterface.html?is-external=true"; title="class or interface in java.lang">@FunctionalInterface -protected static interface RpcServer.CallCleanup +protected static interface RpcServer.CallCleanup @@ -154,7 +154,7 @@ protected static interface run -void run() +void run() @@ -185,7 +185,7 @@ protected static interface -Prev Class +Prev Class Next Class http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/org/apache/hadoop/hbase/ipc/RpcServer.Connection.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/ipc/RpcServer.Connection.html b/devapidocs/org/apache/hadoop/hbase/ipc/RpcServer.Connection.html index 2319d90..9239da0 100644 --- a/devapidocs/org/apache/hadoop/hbase/ipc/RpcServer.Connection.html +++ b/devapidocs/org/apache/hadoop/hbase/ipc/RpcServer.Connection.html @@ -121,7 +121,7 @@ var activeTableTab = "activeTableTab"; -public abstract class RpcServer.Connection +public abstract class RpcServer.Connection extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true"; title="class or interface in java.lang">Object implements http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.html?is-external=true"; title="class or interface in java.io">Closeable Reads calls from a connection and queues them for handling. @@ -156,7 +156,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.h authenticatedWithFallback -protected RpcServer.Call +protected ServerCall authFailedCall @@ -224,7 +224,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.h SASL_CALLID -protected RpcServer.Call +protected ServerCall saslCall @@ -240,7 +240,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.h service -protected RpcServer.Call +protected ServerCall setConnectionHeaderResponseCall @@ -312,7 +312,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.h authorizeConnection() -abstract RpcServer.Call +abstract ServerCall createCall(int id, org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService service, org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor md, @@ -458,7 +458,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.h connectionPreambleRead -protected boolean connectionPreambleRead +protected boolean connectionPreambleRead @@ -467,7 +467,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.h connectionHeaderRead -protected boolean connectionHeaderRead +protected boolean connectionHeaderRead @@ -476,7 +476,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.h callCleanup -protected RpcServer.CallCleanup callCleanup +protected RpcServer.CallCleanup callCleanup @@ -485,7 +485,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.h hostAddress -protected http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true"; title="class or interface in java.lang">String hostAddress +protected http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true"; title="class or interface in java.lang">String hostAddress @@ -494,7 +494,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.h remotePort -protected int remotePort +protected int remotePort @@ -503,7 +503,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.h addr -protected http://docs.oracle.com/javase/8/docs/api/java/net/InetAddress.html?is-external=true"; title="class or interface in java.net">InetAddress addr +protected http://docs.oracle.com/javase/8/docs/api/java/net/InetAddress.html?is-external=true"; title="class or interface in java.net">InetAddress addr @@ -512,7 +512,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.h connectionHeader -protected org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ConnectionHeader connectionHeader +protected org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ConnectionHea
[16/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/src-html/org/apache/hadoop/hbase/ipc/ServerCall.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/ServerCall.html b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/ServerCall.html new file mode 100644 index 000..6ad1e1b --- /dev/null +++ b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/ServerCall.html @@ -0,0 +1,599 @@ +http://www.w3.org/TR/html4/loose.dtd";> + + +Source code + + + + +001/** +002 * Licensed to the Apache Software Foundation (ASF) under one +003 * or more contributor license agreements. See the NOTICE file +004 * distributed with this work for additional information +005 * regarding copyright ownership. The ASF licenses this file +006 * to you under the Apache License, Version 2.0 (the +007 * "License"); you may not use this file except in compliance +008 * with the License. You may obtain a copy of the License at +009 * +010 * http://www.apache.org/licenses/LICENSE-2.0 +011 * +012 * Unless required by applicable law or agreed to in writing, software +013 * distributed under the License is distributed on an "AS IS" BASIS, +014 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +015 * See the License for the specific language governing permissions and +016 * limitations under the License. +017 */ +018package org.apache.hadoop.hbase.ipc; +019 +020import java.io.IOException; +021import java.net.InetAddress; +022import java.nio.ByteBuffer; +023import java.util.ArrayList; +024import java.util.List; +025 +026import org.apache.hadoop.hbase.CellScanner; +027import org.apache.hadoop.hbase.DoNotRetryIOException; +028import org.apache.hadoop.hbase.classification.InterfaceAudience; +029import org.apache.hadoop.hbase.exceptions.RegionMovedException; +030import org.apache.hadoop.hbase.io.ByteBufferListOutputStream; +031import org.apache.hadoop.hbase.io.ByteBufferPool; +032import org.apache.hadoop.hbase.ipc.RpcServer.CallCleanup; +033import org.apache.hadoop.hbase.ipc.RpcServer.Connection; +034import org.apache.hadoop.hbase.security.User; +035import org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService; +036import org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream; +037import org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor; +038import org.apache.hadoop.hbase.shaded.com.google.protobuf.Message; +039import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil; +040import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.VersionInfo; +041import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.CellBlockMeta; +042import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ExceptionResponse; +043import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader; +044import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ResponseHeader; +045import org.apache.hadoop.hbase.util.ByteBufferUtils; +046import org.apache.hadoop.hbase.util.Bytes; +047import org.apache.hadoop.util.StringUtils; +048import org.apache.htrace.TraceInfo; +049 +050/** +051 * Datastructure that holds all necessary to a method invocation and then afterward, carries +052 * the result. +053 */ +054@InterfaceAudience.Private +055abstract class ServerCall implements RpcCall { +056 +057 protected final int id; // the client's call id +058 protected final BlockingService service; +059 protected final MethodDescriptor md; +060 protected final RequestHeader header; +061 protected Message param; // the parameter passed +062 // Optional cell data passed outside of protobufs. +063 protected final CellScanner cellScanner; +064 protected final Connection connection; // connection to client +065 protected final long receiveTime; // the time received when response is null +066 // the time served when response is not null +067 protected final int timeout; +068 protected long startTime; +069 protected final long deadline;// the deadline to handle this call, if exceed we can drop it. +070 +071 protected final ByteBufferPool reservoir; +072 +073 protected final CellBlockBuilder cellBlockBuilder; +074 +075 /** +076 * Chain of buffers to send as response. +077 */ +078 protected BufferChain response; +079 +080 protected final long size; // size of current call +081 protected boolean isError; +082 protected final TraceInfo tinfo; +083 protected ByteBufferListOutputStream cellBlockStream = null; +084 protected CallCleanup reqCleanup = null; +085 +086 protected User user; +087 protected final InetAddress remoteAddress; +088 protected RpcCallback rpcCallback; +089 +090 private long responseCellSize = 0; +091 private long responseBlockSize = 0; +092 // cumul
[20/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.Call.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.Call.html b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.Call.html deleted file mode 100644 index 5cc356a..000 --- a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.Call.html +++ /dev/null @@ -1,2167 +0,0 @@ -http://www.w3.org/TR/html4/loose.dtd";> - - -Source code - - - - -001/** -002 * Licensed to the Apache Software Foundation (ASF) under one -003 * or more contributor license agreements. See the NOTICE file -004 * distributed with this work for additional information -005 * regarding copyright ownership. The ASF licenses this file -006 * to you under the Apache License, Version 2.0 (the -007 * "License"); you may not use this file except in compliance -008 * with the License. You may obtain a copy of the License at -009 * -010 * http://www.apache.org/licenses/LICENSE-2.0 -011 * -012 * Unless required by applicable law or agreed to in writing, software -013 * distributed under the License is distributed on an "AS IS" BASIS, -014 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -015 * See the License for the specific language governing permissions and -016 * limitations under the License. -017 */ -018 -019package org.apache.hadoop.hbase.ipc; -020 -021import static org.apache.hadoop.fs.CommonConfigurationKeysPublic.HADOOP_SECURITY_AUTHORIZATION; -022 -023import java.io.ByteArrayInputStream; -024import java.io.ByteArrayOutputStream; -025import java.io.Closeable; -026import java.io.DataOutputStream; -027import java.io.IOException; -028import java.net.InetAddress; -029import java.net.InetSocketAddress; -030import java.nio.ByteBuffer; -031import java.nio.channels.Channels; -032import java.nio.channels.GatheringByteChannel; -033import java.nio.channels.ReadableByteChannel; -034import java.nio.channels.WritableByteChannel; -035import java.security.GeneralSecurityException; -036import java.security.PrivilegedExceptionAction; -037import java.util.ArrayList; -038import java.util.HashMap; -039import java.util.List; -040import java.util.Map; -041import java.util.Properties; -042import java.util.concurrent.atomic.LongAdder; -043 -044import javax.security.sasl.Sasl; -045import javax.security.sasl.SaslException; -046import javax.security.sasl.SaslServer; -047 -048import org.apache.commons.crypto.cipher.CryptoCipherFactory; -049import org.apache.commons.crypto.random.CryptoRandom; -050import org.apache.commons.crypto.random.CryptoRandomFactory; -051import org.apache.commons.logging.Log; -052import org.apache.commons.logging.LogFactory; -053import org.apache.hadoop.conf.Configuration; -054import org.apache.hadoop.hbase.CallQueueTooBigException; -055import org.apache.hadoop.hbase.CellScanner; -056import org.apache.hadoop.hbase.DoNotRetryIOException; -057import org.apache.hadoop.hbase.HBaseInterfaceAudience; -058import org.apache.hadoop.hbase.HConstants; -059import org.apache.hadoop.hbase.Server; -060import org.apache.hadoop.hbase.classification.InterfaceAudience; -061import org.apache.hadoop.hbase.classification.InterfaceStability; -062import org.apache.hadoop.hbase.client.VersionInfoUtil; -063import org.apache.hadoop.hbase.codec.Codec; -064import org.apache.hadoop.hbase.conf.ConfigurationObserver; -065import org.apache.hadoop.hbase.exceptions.RegionMovedException; -066import org.apache.hadoop.hbase.exceptions.RequestTooBigException; -067import org.apache.hadoop.hbase.io.ByteBufferListOutputStream; -068import org.apache.hadoop.hbase.io.ByteBufferOutputStream; -069import org.apache.hadoop.hbase.io.ByteBufferPool; -070import org.apache.hadoop.hbase.io.crypto.aes.CryptoAES; -071import org.apache.hadoop.hbase.monitoring.MonitoredRPCHandler; -072import org.apache.hadoop.hbase.monitoring.TaskMonitor; -073import org.apache.hadoop.hbase.nio.ByteBuff; -074import org.apache.hadoop.hbase.nio.MultiByteBuff; -075import org.apache.hadoop.hbase.nio.SingleByteBuff; -076import org.apache.hadoop.hbase.regionserver.RSRpcServices; -077import org.apache.hadoop.hbase.security.AccessDeniedException; -078import org.apache.hadoop.hbase.security.AuthMethod; -079import org.apache.hadoop.hbase.security.HBaseSaslRpcServer; -080import org.apache.hadoop.hbase.security.HBaseSaslRpcServer.SaslDigestCallbackHandler; -081import org.apache.hadoop.hbase.security.HBaseSaslRpcServer.SaslGssCallbackHandler; -082import org.apache.hadoop.hbase.security.SaslStatus; -083import org.apache.hadoop.hbase.security.SaslUtil; -084import org.apache.hadoop.hbase.security.User; -085import org.apache.hadoop.hbase.security.UserProvider; -086import org.apache.hadoop.hbase.security.token.AuthenticationTokenSecretManager; -087import org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingSe
[42/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/org/apache/hadoop/hbase/ipc/RpcServer.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/ipc/RpcServer.html b/devapidocs/org/apache/hadoop/hbase/ipc/RpcServer.html index 737b9b3..307b468 100644 --- a/devapidocs/org/apache/hadoop/hbase/ipc/RpcServer.html +++ b/devapidocs/org/apache/hadoop/hbase/ipc/RpcServer.html @@ -119,7 +119,7 @@ var activeTableTab = "activeTableTab"; @InterfaceAudience.LimitedPrivate(value={"Coprocesssor","Phoenix"}) @InterfaceStability.Evolving -public abstract class RpcServer +public abstract class RpcServer extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true"; title="class or interface in java.lang">Object implements RpcServerInterface, ConfigurationObserver An RPC server that hosts protobuf described Services. @@ -153,17 +153,10 @@ implements RpcServer.ByteBuffByteInput -class -RpcServer.Call -Datastructure that holds all necessary to a method invocation and then afterward, carries - the result. - - - protected static interface RpcServer.CallCleanup - + class RpcServer.Connection Reads calls from a connection and queues them for handling. @@ -639,8 +632,8 @@ implements protected void -setupResponse(http://docs.oracle.com/javase/8/docs/api/java/io/ByteArrayOutputStream.html?is-external=true"; title="class or interface in java.io">ByteArrayOutputStream response, - RpcServer.Call call, +setupResponse(http://docs.oracle.com/javase/8/docs/api/java/io/ByteArrayOutputStream.html?is-external=true"; title="class or interface in java.io">ByteArrayOutputStream response, + ServerCall call, http://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true"; title="class or interface in java.lang">Throwable t, http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true"; title="class or interface in java.lang">String error) Setup response for the RPC Call. @@ -681,7 +674,7 @@ implements LOG -public static final org.apache.commons.logging.Log LOG +public static final org.apache.commons.logging.Log LOG @@ -690,7 +683,7 @@ implements CALL_QUEUE_TOO_BIG_EXCEPTION -protected static final CallQueueTooBigException CALL_QUEUE_TOO_BIG_EXCEPTION +protected static final CallQueueTooBigException CALL_QUEUE_TOO_BIG_EXCEPTION @@ -699,7 +692,7 @@ implements authorize -private final boolean authorize +private final boolean authorize @@ -708,7 +701,7 @@ implements isSecurityEnabled -protected boolean isSecurityEnabled +protected boolean isSecurityEnabled @@ -717,7 +710,7 @@ implements CURRENT_VERSION -public static final byte CURRENT_VERSION +public static final byte CURRENT_VERSION See Also: Constant Field Values @@ -730,7 +723,7 @@ implements FALLBACK_TO_INSECURE_CLIENT_AUTH -public static final http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true"; title="class or interface in java.lang">String FALLBACK_TO_INSECURE_CLIENT_AUTH +public static final http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true"; title="class or interface in java.lang">String FALLBACK_TO_INSECURE_CLIENT_AUTH Whether we allow a fallback to SIMPLE auth for insecure clients when security is enabled. See Also: @@ -744,7 +737,7 @@ implements DEFAULT_MAX_CALLQUEUE_LENGTH_PER_HANDLER -protected static final int DEFAULT_MAX_CALLQUEUE_LENGTH_PER_HANDLER +protected static final int DEFAULT_MAX_CALLQUEUE_LENGTH_PER_HANDLER How many calls/handler are allowed in the queue. See Also: @@ -758,7 +751,7 @@ implements cellBlockBuilder -protected final CellBlockBuilder cellBlockBuilder +protected final CellBlockBuilder cellBlockBuilder @@ -767,7 +760,7 @@ implements AUTH_FAILED_FOR -protected static final http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true"; title="class or interface in java.lang">String AUTH_FAILED_FOR +protected static final http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true"; title="class or interface in java.lang">String AUTH_FAILED_FOR See Also: Constant Field Values @@ -780,7 +773,7 @@ implements AUTH_SUCCESSFUL_FOR -protected static final http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true"; title="class or interface in java.lang">String AUTH_SUCCESSFUL_FOR +protected static final http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true"; title="class or interface in java.lang">String AUTH_SUCCESSFUL_FOR See Also: Constant Field Values @@ -793,7 +786,7 @@ implements AUDITLOG -protected static final org.apache.commons.logging.Log AUDITLOG +protected static final org.apache.commons.logging.Log AUDITLOG @@ -802,7 +795,7 @@ implements secretManager -protected org.a
[17/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.html b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.html index 5cc356a..9e1c66c 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.html @@ -70,2037 +70,1559 @@ 062import org.apache.hadoop.hbase.client.VersionInfoUtil; 063import org.apache.hadoop.hbase.codec.Codec; 064import org.apache.hadoop.hbase.conf.ConfigurationObserver; -065import org.apache.hadoop.hbase.exceptions.RegionMovedException; -066import org.apache.hadoop.hbase.exceptions.RequestTooBigException; -067import org.apache.hadoop.hbase.io.ByteBufferListOutputStream; -068import org.apache.hadoop.hbase.io.ByteBufferOutputStream; -069import org.apache.hadoop.hbase.io.ByteBufferPool; -070import org.apache.hadoop.hbase.io.crypto.aes.CryptoAES; -071import org.apache.hadoop.hbase.monitoring.MonitoredRPCHandler; -072import org.apache.hadoop.hbase.monitoring.TaskMonitor; -073import org.apache.hadoop.hbase.nio.ByteBuff; -074import org.apache.hadoop.hbase.nio.MultiByteBuff; -075import org.apache.hadoop.hbase.nio.SingleByteBuff; -076import org.apache.hadoop.hbase.regionserver.RSRpcServices; -077import org.apache.hadoop.hbase.security.AccessDeniedException; -078import org.apache.hadoop.hbase.security.AuthMethod; -079import org.apache.hadoop.hbase.security.HBaseSaslRpcServer; -080import org.apache.hadoop.hbase.security.HBaseSaslRpcServer.SaslDigestCallbackHandler; -081import org.apache.hadoop.hbase.security.HBaseSaslRpcServer.SaslGssCallbackHandler; -082import org.apache.hadoop.hbase.security.SaslStatus; -083import org.apache.hadoop.hbase.security.SaslUtil; -084import org.apache.hadoop.hbase.security.User; -085import org.apache.hadoop.hbase.security.UserProvider; -086import org.apache.hadoop.hbase.security.token.AuthenticationTokenSecretManager; -087import org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService; -088import org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteInput; -089import org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString; -090import org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream; -091import org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream; -092import org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor; -093import org.apache.hadoop.hbase.shaded.com.google.protobuf.Message; -094import org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException; -095import org.apache.hadoop.hbase.shaded.com.google.protobuf.TextFormat; -096import org.apache.hadoop.hbase.shaded.com.google.protobuf.UnsafeByteOperations; -097import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil; -098import org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos; -099import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.VersionInfo; -100import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos; -101import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.CellBlockMeta; -102import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ConnectionHeader; -103import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ExceptionResponse; -104import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader; -105import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ResponseHeader; -106import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation; -107import org.apache.hadoop.hbase.util.ByteBufferUtils; -108import org.apache.hadoop.hbase.util.Bytes; -109import org.apache.hadoop.hbase.util.Pair; -110import org.apache.hadoop.io.BytesWritable; -111import org.apache.hadoop.io.Writable; -112import org.apache.hadoop.io.WritableUtils; -113import org.apache.hadoop.io.compress.CompressionCodec; -114import org.apache.hadoop.security.UserGroupInformation; -115import org.apache.hadoop.security.UserGroupInformation.AuthenticationMethod; -116import org.apache.hadoop.security.authorize.AuthorizationException; -117import org.apache.hadoop.security.authorize.PolicyProvider; -118import org.apache.hadoop.security.authorize.ProxyUsers; -119import org.apache.hadoop.security.authorize.ServiceAuthorizationManager; -120import org.apache.hadoop.security.token.SecretManager; -121import org.apache.hadoop.security.token.SecretManager.InvalidToken; -122import org.apache.hadoop.security.token.TokenIdentifier; -123import org.apache.hadoop.util.StringUtils; -124import org.apache.htrace.TraceInfo; -125import org.codehaus.jackson.map.ObjectMapper; -126 -127import com.google.common.annotations.VisibleForTesting; -128 -129/** -130 * An RPC server that hosts protobuf described Services. -
[35/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/org/apache/hadoop/hbase/master/HMaster.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/master/HMaster.html b/devapidocs/org/apache/hadoop/hbase/master/HMaster.html index bb60410..8967608 100644 --- a/devapidocs/org/apache/hadoop/hbase/master/HMaster.html +++ b/devapidocs/org/apache/hadoop/hbase/master/HMaster.html @@ -128,7 +128,7 @@ var activeTableTab = "activeTableTab"; @InterfaceAudience.LimitedPrivate(value="Tools") -public class HMaster +public class HMaster extends HRegionServer implements MasterServices HMaster is the "master server" for HBase. An HBase cluster has one active @@ -1382,7 +1382,7 @@ implements LOG -private static final org.apache.commons.logging.Log LOG +private static final org.apache.commons.logging.Log LOG @@ -1391,7 +1391,7 @@ implements MASTER -public static final http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true"; title="class or interface in java.lang">String MASTER +public static final http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true"; title="class or interface in java.lang">String MASTER See Also: Constant Field Values @@ -1404,7 +1404,7 @@ implements activeMasterManager -private final ActiveMasterManager activeMasterManager +private final ActiveMasterManager activeMasterManager @@ -1413,7 +1413,7 @@ implements regionServerTracker -RegionServerTracker regionServerTracker +RegionServerTracker regionServerTracker @@ -1422,7 +1422,7 @@ implements drainingServerTracker -private DrainingServerTracker drainingServerTracker +private DrainingServerTracker drainingServerTracker @@ -1431,7 +1431,7 @@ implements loadBalancerTracker -LoadBalancerTracker loadBalancerTracker +LoadBalancerTracker loadBalancerTracker @@ -1440,7 +1440,7 @@ implements splitOrMergeTracker -private SplitOrMergeTracker splitOrMergeTracker +private SplitOrMergeTracker splitOrMergeTracker @@ -1449,7 +1449,7 @@ implements regionNormalizerTracker -private RegionNormalizerTracker regionNormalizerTracker +private RegionNormalizerTracker regionNormalizerTracker @@ -1458,7 +1458,7 @@ implements maintenanceModeTracker -private MasterMaintenanceModeTracker maintenanceModeTracker +private MasterMaintenanceModeTracker maintenanceModeTracker @@ -1467,7 +1467,7 @@ implements clusterSchemaService -private ClusterSchemaService clusterSchemaService +private ClusterSchemaService clusterSchemaService @@ -1476,7 +1476,7 @@ implements metricsMaster -final MetricsMaster metricsMaster +final MetricsMaster metricsMaster @@ -1485,7 +1485,7 @@ implements fileSystemManager -private MasterFileSystem fileSystemManager +private MasterFileSystem fileSystemManager @@ -1494,7 +1494,7 @@ implements walManager -private MasterWalManager walManager +private MasterWalManager walManager @@ -1503,7 +1503,7 @@ implements serverManager -private volatile ServerManager serverManager +private volatile ServerManager serverManager @@ -1512,7 +1512,7 @@ implements assignmentManager -private AssignmentManager assignmentManager +private AssignmentManager assignmentManager @@ -1521,7 +1521,7 @@ implements replicationManager -private ReplicationManager replicationManager +private ReplicationManager replicationManager @@ -1530,7 +1530,7 @@ implements rsFatals -MemoryBoundedLogMessageBuffer rsFatals +MemoryBoundedLogMessageBuffer rsFatals @@ -1539,7 +1539,7 @@ implements activeMaster -private volatile boolean activeMaster +private volatile boolean activeMaster @@ -1548,7 +1548,7 @@ implements initialized -private final ProcedureEvent initialized +private final ProcedureEvent initialized @@ -1557,7 +1557,7 @@ implements serviceStarted -volatile boolean serviceStarted +volatile boolean serviceStarted @@ -1566,7 +1566,7 @@ implements serverCrashProcessingEnabled -private final ProcedureEvent serverCrashProcessingEnabled +private final ProcedureEvent serverCrashProcessingEnabled @@ -1575,7 +1575,7 @@ implements maxBlancingTime -private final int maxBlancingTime +private final int maxBlancingTime @@ -1584,7 +1584,7 @@ implements maxRitPercent -private final double maxRitPercent +private final double maxRitPercent @@ -1593,7 +1593,7 @@ implements lockManager -private final LockManager lockManager +private final LockManager lockManager @@ -1602,7 +1602,7 @@ implements balancer -private LoadBalancer balancer +private LoadBalancer balancer @@ -1611,7 +1611,7 @@ implements normalizer -private RegionNormalizer normalizer +private RegionNormalizer normalizer @@ -1620,7 +1620,7 @@ implements balancerChore -private BalancerChore balancerChore +private BalancerChore balancerChore @@ -1629,7 +1
[36/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/org/apache/hadoop/hbase/master/HMaster.InitializationMonitor.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/master/HMaster.InitializationMonitor.html b/devapidocs/org/apache/hadoop/hbase/master/HMaster.InitializationMonitor.html index 0791ec1..6476d55 100644 --- a/devapidocs/org/apache/hadoop/hbase/master/HMaster.InitializationMonitor.html +++ b/devapidocs/org/apache/hadoop/hbase/master/HMaster.InitializationMonitor.html @@ -122,7 +122,7 @@ var activeTableTab = "activeTableTab"; -private static class HMaster.InitializationMonitor +private static class HMaster.InitializationMonitor extends HasThread Protection against zombie master. Started once Master accepts active responsibility and starts taking over responsibilities. Allows a finite time window before giving up ownership. @@ -250,7 +250,7 @@ extends TIMEOUT_KEY -public static final http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true"; title="class or interface in java.lang">String TIMEOUT_KEY +public static final http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true"; title="class or interface in java.lang">String TIMEOUT_KEY The amount of time in milliseconds to sleep before checking initialization status. See Also: @@ -264,7 +264,7 @@ extends TIMEOUT_DEFAULT -public static final long TIMEOUT_DEFAULT +public static final long TIMEOUT_DEFAULT @@ -273,7 +273,7 @@ extends HALT_KEY -public static final http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true"; title="class or interface in java.lang">String HALT_KEY +public static final http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true"; title="class or interface in java.lang">String HALT_KEY When timeout expired and initialization has not complete, call http://docs.oracle.com/javase/8/docs/api/java/lang/System.html?is-external=true#exit-int-"; title="class or interface in java.lang">System.exit(int) when true, do nothing otherwise. @@ -288,7 +288,7 @@ extends HALT_DEFAULT -public static final boolean HALT_DEFAULT +public static final boolean HALT_DEFAULT See Also: Constant Field Values @@ -301,7 +301,7 @@ extends master -private final HMaster master +private final HMaster master @@ -310,7 +310,7 @@ extends timeout -private final long timeout +private final long timeout @@ -319,7 +319,7 @@ extends haltOnTimeout -private final boolean haltOnTimeout +private final boolean haltOnTimeout @@ -336,7 +336,7 @@ extends InitializationMonitor -InitializationMonitor(HMaster master) +InitializationMonitor(HMaster master) Creates a Thread that monitors the HMaster.isInitialized() state. @@ -354,7 +354,7 @@ extends run -public void run() +public void run() Specified by: http://docs.oracle.com/javase/8/docs/api/java/lang/Runnable.html?is-external=true#run--"; title="class or interface in java.lang">run in interface http://docs.oracle.com/javase/8/docs/api/java/lang/Runnable.html?is-external=true"; title="class or interface in java.lang">Runnable http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/org/apache/hadoop/hbase/master/HMaster.PeriodicDoMetrics.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/master/HMaster.PeriodicDoMetrics.html b/devapidocs/org/apache/hadoop/hbase/master/HMaster.PeriodicDoMetrics.html index 6537ea5..c1296d6 100644 --- a/devapidocs/org/apache/hadoop/hbase/master/HMaster.PeriodicDoMetrics.html +++ b/devapidocs/org/apache/hadoop/hbase/master/HMaster.PeriodicDoMetrics.html @@ -122,7 +122,7 @@ var activeTableTab = "activeTableTab"; -private static class HMaster.PeriodicDoMetrics +private static class HMaster.PeriodicDoMetrics extends ScheduledChore @@ -220,7 +220,7 @@ extends server -private final HMaster server +private final HMaster server @@ -237,7 +237,7 @@ extends PeriodicDoMetrics -public PeriodicDoMetrics(int doMetricsInterval, +public PeriodicDoMetrics(int doMetricsInterval, HMaster server) @@ -255,7 +255,7 @@ extends chore -protected void chore() +protected void chore() Description copied from class: ScheduledChore The task to execute on each scheduled execution of the Chore http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/org/apache/hadoop/hbase/master/HMaster.RedirectServlet.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/master/HMaster.RedirectServlet.html b/devapidocs/org/apache/hadoop/hbase/master/HMaster.RedirectServlet.html index a619b44..6ec0266 100644 --- a/devapidocs/org/apache/hadoop/hbase/master/HMaster.RedirectServlet.html +++ b/devapidocs/org/apache
[37/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/org/apache/hadoop/hbase/ipc/class-use/RpcServer.Connection.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/ipc/class-use/RpcServer.Connection.html b/devapidocs/org/apache/hadoop/hbase/ipc/class-use/RpcServer.Connection.html index 3856a94..b7d3911 100644 --- a/devapidocs/org/apache/hadoop/hbase/ipc/class-use/RpcServer.Connection.html +++ b/devapidocs/org/apache/hadoop/hbase/ipc/class-use/RpcServer.Connection.html @@ -129,7 +129,7 @@ protected RpcServer.Connection -RpcServer.Call.connection +ServerCall.connection @@ -141,7 +141,7 @@ -abstract RpcServer.Call +abstract ServerCall RpcServer.Connection.createCall(int id, org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService service, org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor md, @@ -156,7 +156,7 @@ RpcServer.CallCleanup reqCleanup) -RpcServer.Call +ServerCall SimpleRpcServer.Connection.createCall(int id, org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService service, org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor md, @@ -171,7 +171,7 @@ RpcServer.CallCleanup reqCleanup) -RpcServer.Call +ServerCall NettyRpcServer.NettyConnection.createCall(int id, org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService service, org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor md, @@ -194,47 +194,56 @@ -Call(int id, - org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService service, - org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor md, - org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader header, -org.apache.hadoop.hbase.shaded.com.google.protobuf.Message param, -CellScanner cellScanner, -RpcServer.Connection connection, -long size, -org.apache.htrace.TraceInfo tinfo, -http://docs.oracle.com/javase/8/docs/api/java/net/InetAddress.html?is-external=true"; title="class or interface in java.net">InetAddress remoteAddress, -int timeout, -RpcServer.CallCleanup reqCleanup) +NettyServerCall(int id, + org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService service, + org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor md, + org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader header, + org.apache.hadoop.hbase.shaded.com.google.protobuf.Message param, + CellScanner cellScanner, + RpcServer.Connection connection, + long size, + org.apache.htrace.TraceInfo tinfo, + http://docs.oracle.com/javase/8/docs/api/java/net/InetAddress.html?is-external=true"; title="class or interface in java.net">InetAddress remoteAddress, + long receiveTime, + int timeout, + ByteBufferPool reservoir, + CellBlockBuilder cellBlockBuilder, + RpcServer.CallCleanup reqCleanup) -Call(int id, - org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService service, - org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor md, - org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader header, -org.apache.hadoop.hbase.shaded.com.google.protobuf.Message param, -CellScanner cellScanner, -RpcServer.Connection connection, -long size, -org.apache.htrace.TraceInfo tinfo, -http://docs.oracle.com/javase/8/docs/api/java/net/InetAddress.html?is-external=true"; title="class or interface in java.net">InetAddress remoteAddress, -int timeout, -RpcServer.CallCleanup reqCleanup) +ServerCall(int id, + org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService service, + org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor md, + org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader header, + org.apache.hadoop.hbase.shaded.com.google.protobuf.Message param, + CellScanner cellScanner, + RpcServer.Connection connection, + long size, + org.apache.htrace.TraceInfo tinfo, + http://docs.oracle.com/javase/8/docs/api/java/net/InetAddress.html?is-external=true"; title="class or interface in java.net">InetAddress remoteAddress, + long receiveTime, + int timeout, + ByteBufferPool reservoir, + CellBlockBuilder cellBlockBuilder, + RpcServer.CallCleanup reqCleanup) -Call(int id, - org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService service, - org.apache.hadoop.hbase.shaded.com
[03/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/RSRpcServices.RegionScannerCloseCallBack.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/RSRpcServices.RegionScannerCloseCallBack.html b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/RSRpcServices.RegionScannerCloseCallBack.html index 17d1bcb..6dfd1d4 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/RSRpcServices.RegionScannerCloseCallBack.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/RSRpcServices.RegionScannerCloseCallBack.html @@ -35,863 +35,863 @@ 027import java.net.InetSocketAddress; 028import java.net.UnknownHostException; 029import java.nio.ByteBuffer; -030import java.util.ArrayList; -031import java.util.Collections; -032import java.util.HashMap; -033import java.util.Iterator; -034import java.util.List; -035import java.util.Map; -036import java.util.Map.Entry; -037import java.util.NavigableMap; -038import java.util.Set; -039import java.util.TreeSet; -040import java.util.concurrent.ConcurrentHashMap; -041import java.util.concurrent.ConcurrentMap; -042import java.util.concurrent.atomic.AtomicLong; -043import java.util.concurrent.atomic.LongAdder; -044 -045import org.apache.commons.lang.mutable.MutableObject; -046import org.apache.commons.logging.Log; -047import org.apache.commons.logging.LogFactory; -048import org.apache.hadoop.conf.Configuration; -049import org.apache.hadoop.fs.Path; -050import org.apache.hadoop.hbase.ByteBufferCell; -051import org.apache.hadoop.hbase.Cell; -052import org.apache.hadoop.hbase.CellScannable; -053import org.apache.hadoop.hbase.CellScanner; -054import org.apache.hadoop.hbase.CellUtil; -055import org.apache.hadoop.hbase.DoNotRetryIOException; -056import org.apache.hadoop.hbase.DroppedSnapshotException; -057import org.apache.hadoop.hbase.HBaseIOException; -058import org.apache.hadoop.hbase.HConstants; -059import org.apache.hadoop.hbase.HRegionInfo; -060import org.apache.hadoop.hbase.HTableDescriptor; -061import org.apache.hadoop.hbase.MultiActionResultTooLarge; -062import org.apache.hadoop.hbase.NotServingRegionException; -063import org.apache.hadoop.hbase.ServerName; -064import org.apache.hadoop.hbase.TableName; -065import org.apache.hadoop.hbase.UnknownScannerException; -066import org.apache.hadoop.hbase.classification.InterfaceAudience; -067import org.apache.hadoop.hbase.client.Append; -068import org.apache.hadoop.hbase.client.ConnectionUtils; -069import org.apache.hadoop.hbase.client.Delete; -070import org.apache.hadoop.hbase.client.Durability; -071import org.apache.hadoop.hbase.client.Get; -072import org.apache.hadoop.hbase.client.Increment; -073import org.apache.hadoop.hbase.client.Mutation; -074import org.apache.hadoop.hbase.client.Put; -075import org.apache.hadoop.hbase.client.RegionReplicaUtil; -076import org.apache.hadoop.hbase.client.Result; -077import org.apache.hadoop.hbase.client.RowMutations; -078import org.apache.hadoop.hbase.client.Scan; -079import org.apache.hadoop.hbase.client.VersionInfoUtil; -080import org.apache.hadoop.hbase.conf.ConfigurationObserver; -081import org.apache.hadoop.hbase.exceptions.FailedSanityCheckException; -082import org.apache.hadoop.hbase.exceptions.OutOfOrderScannerNextException; -083import org.apache.hadoop.hbase.exceptions.ScannerResetException; -084import org.apache.hadoop.hbase.filter.ByteArrayComparable; -085import org.apache.hadoop.hbase.filter.CompareFilter.CompareOp; -086import org.apache.hadoop.hbase.ipc.HBaseRPCErrorHandler; -087import org.apache.hadoop.hbase.ipc.HBaseRpcController; -088import org.apache.hadoop.hbase.ipc.PriorityFunction; -089import org.apache.hadoop.hbase.ipc.QosPriority; -090import org.apache.hadoop.hbase.ipc.RpcCallContext; -091import org.apache.hadoop.hbase.ipc.RpcCallback; -092import org.apache.hadoop.hbase.ipc.RpcServer; -093import org.apache.hadoop.hbase.ipc.RpcServer.BlockingServiceAndInterface; -094import org.apache.hadoop.hbase.ipc.RpcServerFactory; -095import org.apache.hadoop.hbase.ipc.RpcServerInterface; -096import org.apache.hadoop.hbase.ipc.ServerNotRunningYetException; -097import org.apache.hadoop.hbase.ipc.ServerRpcController; -098import org.apache.hadoop.hbase.master.MasterRpcServices; -099import org.apache.hadoop.hbase.quotas.OperationQuota; -100import org.apache.hadoop.hbase.quotas.RegionServerQuotaManager; -101import org.apache.hadoop.hbase.regionserver.HRegion.RegionScannerImpl; -102import org.apache.hadoop.hbase.regionserver.Leases.Lease; -103import org.apache.hadoop.hbase.regionserver.Leases.LeaseStillHeldException; -104import org.apache.hadoop.hbase.regionserver.Region.Operation; -105import org.apache.hadoop.hbase.regionserver.ScannerContext.LimitScope; -106import org.apache.hadoop.hbase.regionserver.handler.OpenMetaHandler; -107import org.
[05/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.RedirectServlet.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.RedirectServlet.html b/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.RedirectServlet.html index 66fe094..17bc96c 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.RedirectServlet.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.RedirectServlet.html @@ -57,3290 +57,3303 @@ 049import org.apache.commons.logging.Log; 050import org.apache.commons.logging.LogFactory; 051import org.apache.hadoop.conf.Configuration; -052import org.apache.hadoop.fs.Path; -053import org.apache.hadoop.hbase.ClusterStatus; -054import org.apache.hadoop.hbase.CoordinatedStateException; -055import org.apache.hadoop.hbase.CoordinatedStateManager; -056import org.apache.hadoop.hbase.DoNotRetryIOException; -057import org.apache.hadoop.hbase.HBaseIOException; -058import org.apache.hadoop.hbase.HBaseInterfaceAudience; -059import org.apache.hadoop.hbase.HColumnDescriptor; -060import org.apache.hadoop.hbase.HConstants; -061import org.apache.hadoop.hbase.HRegionInfo; -062import org.apache.hadoop.hbase.HTableDescriptor; -063import org.apache.hadoop.hbase.MasterNotRunningException; -064import org.apache.hadoop.hbase.MetaTableAccessor; -065import org.apache.hadoop.hbase.NamespaceDescriptor; -066import org.apache.hadoop.hbase.PleaseHoldException; -067import org.apache.hadoop.hbase.ProcedureInfo; -068import org.apache.hadoop.hbase.ScheduledChore; -069import org.apache.hadoop.hbase.ServerLoad; -070import org.apache.hadoop.hbase.ServerName; -071import org.apache.hadoop.hbase.TableDescriptors; -072import org.apache.hadoop.hbase.TableName; -073import org.apache.hadoop.hbase.TableNotDisabledException; -074import org.apache.hadoop.hbase.TableNotFoundException; -075import org.apache.hadoop.hbase.UnknownRegionException; -076import org.apache.hadoop.hbase.classification.InterfaceAudience; -077import org.apache.hadoop.hbase.client.MasterSwitchType; -078import org.apache.hadoop.hbase.client.Result; -079import org.apache.hadoop.hbase.client.TableState; -080import org.apache.hadoop.hbase.coprocessor.BypassCoprocessorException; -081import org.apache.hadoop.hbase.coprocessor.CoprocessorHost; -082import org.apache.hadoop.hbase.exceptions.DeserializationException; -083import org.apache.hadoop.hbase.exceptions.MergeRegionException; -084import org.apache.hadoop.hbase.executor.ExecutorType; -085import org.apache.hadoop.hbase.favored.FavoredNodesManager; -086import org.apache.hadoop.hbase.favored.FavoredNodesPromoter; -087import org.apache.hadoop.hbase.http.InfoServer; -088import org.apache.hadoop.hbase.ipc.CoprocessorRpcUtils; -089import org.apache.hadoop.hbase.ipc.RpcServer; -090import org.apache.hadoop.hbase.ipc.ServerNotRunningYetException; -091import org.apache.hadoop.hbase.master.MasterRpcServices.BalanceSwitchMode; -092import org.apache.hadoop.hbase.master.balancer.BalancerChore; -093import org.apache.hadoop.hbase.master.balancer.BaseLoadBalancer; -094import org.apache.hadoop.hbase.master.balancer.ClusterStatusChore; -095import org.apache.hadoop.hbase.master.balancer.LoadBalancerFactory; -096import org.apache.hadoop.hbase.master.cleaner.HFileCleaner; -097import org.apache.hadoop.hbase.master.cleaner.LogCleaner; -098import org.apache.hadoop.hbase.master.cleaner.ReplicationMetaCleaner; -099import org.apache.hadoop.hbase.master.cleaner.ReplicationZKNodeCleaner; -100import org.apache.hadoop.hbase.master.cleaner.ReplicationZKNodeCleanerChore; -101import org.apache.hadoop.hbase.master.locking.LockManager; -102import org.apache.hadoop.hbase.master.normalizer.NormalizationPlan; -103import org.apache.hadoop.hbase.master.normalizer.NormalizationPlan.PlanType; -104import org.apache.hadoop.hbase.master.normalizer.RegionNormalizer; -105import org.apache.hadoop.hbase.master.normalizer.RegionNormalizerChore; -106import org.apache.hadoop.hbase.master.normalizer.RegionNormalizerFactory; -107import org.apache.hadoop.hbase.master.procedure.AddColumnFamilyProcedure; -108import org.apache.hadoop.hbase.master.procedure.CreateTableProcedure; -109import org.apache.hadoop.hbase.master.procedure.DeleteColumnFamilyProcedure; -110import org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure; -111import org.apache.hadoop.hbase.master.procedure.DisableTableProcedure; -112import org.apache.hadoop.hbase.master.procedure.EnableTableProcedure; -113import org.apache.hadoop.hbase.master.procedure.MasterProcedureConstants; -114import org.apache.hadoop.hbase.master.procedure.MasterProcedureEnv; -115import org.apache.hadoop.hbase.master.procedure.MasterProcedureScheduler; -116import org.apache.hadoop.hbase.master.procedure.MasterProcedureUtil; -117import org.apache.hadoop.hbase.master.p
[51/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
Published site at 82d554e3783372cc6b05489452c815b57c06f6cd. Project: http://git-wip-us.apache.org/repos/asf/hbase-site/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase-site/commit/7ef4c5a9 Tree: http://git-wip-us.apache.org/repos/asf/hbase-site/tree/7ef4c5a9 Diff: http://git-wip-us.apache.org/repos/asf/hbase-site/diff/7ef4c5a9 Branch: refs/heads/asf-site Commit: 7ef4c5a9a2c4f8fd9b2456b7dfd0b36ab6e45ad4 Parents: baff481 Author: jenkins Authored: Tue May 9 15:00:02 2017 + Committer: jenkins Committed: Tue May 9 15:00:02 2017 + -- acid-semantics.html | 4 +- apache_hbase_reference_guide.pdf| 4 +- apache_hbase_reference_guide.pdfmarks | 4 +- bulk-loads.html | 4 +- checkstyle-aggregate.html | 34006 + checkstyle.rss |54 +- coc.html| 4 +- cygwin.html | 4 +- dependencies.html | 4 +- dependency-convergence.html | 4 +- dependency-info.html| 4 +- dependency-management.html | 4 +- devapidocs/allclasses-frame.html| 3 + devapidocs/allclasses-noframe.html | 3 + devapidocs/constant-values.html | 6 +- devapidocs/index-all.html | 238 +- .../hadoop/hbase/backup/package-tree.html | 4 +- .../hadoop/hbase/class-use/CellScanner.html | 109 +- .../class-use/InterfaceAudience.Private.html|42 +- .../class-use/InterfaceStability.Evolving.html |27 +- .../hbase/classification/package-tree.html | 4 +- .../hadoop/hbase/client/package-tree.html |18 +- .../hadoop/hbase/filter/package-tree.html | 8 +- .../class-use/ByteBufferListOutputStream.html | 2 +- .../hbase/io/class-use/ByteBufferPool.html |64 + .../hadoop/hbase/io/hfile/package-tree.html | 6 +- .../org/apache/hadoop/hbase/ipc/CallRunner.html | 4 +- .../hadoop/hbase/ipc/NettyRpcServer.Call.html | 395 - .../ipc/NettyRpcServer.CallWriteListener.html |18 +- .../NettyRpcServer.ConnectionHeaderHandler.html | 8 +- .../hbase/ipc/NettyRpcServer.Initializer.html | 8 +- .../ipc/NettyRpcServer.MessageDecoder.html |16 +- .../ipc/NettyRpcServer.MessageEncoder.html | 6 +- .../ipc/NettyRpcServer.NettyConnection.html |36 +- .../apache/hadoop/hbase/ipc/NettyRpcServer.html |55 +- .../hadoop/hbase/ipc/NettyServerCall.html | 374 + .../hadoop/hbase/ipc/PriorityFunction.html | 4 +- .../hadoop/hbase/ipc/RWQueueRpcExecutor.html| 4 +- .../org/apache/hadoop/hbase/ipc/RpcCall.html|50 +- .../apache/hadoop/hbase/ipc/RpcCallContext.html | 2 +- .../RpcServer.BlockingServiceAndInterface.html |12 +- .../hbase/ipc/RpcServer.ByteBuffByteInput.html |24 +- .../apache/hadoop/hbase/ipc/RpcServer.Call.html | 1453 - .../hadoop/hbase/ipc/RpcServer.CallCleanup.html | 8 +- .../hadoop/hbase/ipc/RpcServer.Connection.html | 144 +- .../org/apache/hadoop/hbase/ipc/RpcServer.html | 189 +- .../org/apache/hadoop/hbase/ipc/ServerCall.html | 1477 + .../hbase/ipc/ServerNotRunningYetException.html | 4 +- .../hadoop/hbase/ipc/SimpleRpcServer.Call.html | 429 - .../hbase/ipc/SimpleRpcServer.Connection.html |82 +- .../ipc/SimpleRpcServer.ConnectionManager.html |40 +- .../ipc/SimpleRpcServer.Listener.Reader.html|14 +- .../hbase/ipc/SimpleRpcServer.Listener.html |30 +- .../hbase/ipc/SimpleRpcServer.Responder.html|38 +- .../hadoop/hbase/ipc/SimpleRpcServer.html |45 +- .../hadoop/hbase/ipc/SimpleServerCall.html | 420 + .../hbase/ipc/StoppedRpcClientException.html| 4 +- .../hadoop/hbase/ipc/class-use/BufferChain.html | 6 +- .../hbase/ipc/class-use/CellBlockBuilder.html |56 + .../ipc/class-use/NettyRpcServer.Call.html | 178 - .../NettyRpcServer.NettyConnection.html | 2 +- .../hbase/ipc/class-use/NettyServerCall.html| 178 + .../hadoop/hbase/ipc/class-use/RpcCall.html |16 +- .../hbase/ipc/class-use/RpcCallContext.html |16 +- .../hadoop/hbase/ipc/class-use/RpcCallback.html | 4 +- .../hbase/ipc/class-use/RpcServer.Call.html | 278 - .../ipc/class-use/RpcServer.CallCleanup.html|91 +- .../ipc/class-use/RpcServer.Connection.html |91 +- .../hadoop/hbase/ipc/class-use/ServerCall.html | 278 + .../ipc/class-use/SimpleRpcServer.Call.html | 186 - .../class-use/SimpleRpcServer.Connection.html | 2 +- .../class-use/SimpleRpcServer.Responder.html|35 +- .../hbase/ipc/class-use/SimpleServerCall.html | 1
[46/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/org/apache/hadoop/hbase/ipc/NettyRpcServer.Call.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/ipc/NettyRpcServer.Call.html b/devapidocs/org/apache/hadoop/hbase/ipc/NettyRpcServer.Call.html deleted file mode 100644 index 8e2f829..000 --- a/devapidocs/org/apache/hadoop/hbase/ipc/NettyRpcServer.Call.html +++ /dev/null @@ -1,395 +0,0 @@ -http://www.w3.org/TR/html4/loose.dtd";> - - - - - -NettyRpcServer.Call (Apache HBase 2.0.0-SNAPSHOT API) - - - - - -var methods = {"i0":10,"i1":10,"i2":10,"i3":10}; -var tabs = {65535:["t0","All Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]}; -var altColor = "altColor"; -var rowColor = "rowColor"; -var tableTab = "tableTab"; -var activeTableTab = "activeTableTab"; - - -JavaScript is disabled on your browser. - - - - - -Skip navigation links - - - - -Overview -Package -Class -Use -Tree -Deprecated -Index -Help - - - - -Prev Class -Next Class - - -Frames -No Frames - - -All Classes - - - - - - - -Summary: -Nested | -Field | -Constr | -Method - - -Detail: -Field | -Constr | -Method - - - - - - - - -org.apache.hadoop.hbase.ipc -Class NettyRpcServer.Call - - - -http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true"; title="class or interface in java.lang">java.lang.Object - - -org.apache.hadoop.hbase.ipc.RpcServer.Call - - -org.apache.hadoop.hbase.ipc.NettyRpcServer.Call - - - - - - - - - -All Implemented Interfaces: -RpcCall, RpcCallContext - - -Enclosing class: -NettyRpcServer - - - -@InterfaceStability.Evolving -public class NettyRpcServer.Call -extends RpcServer.Call -Datastructure that holds all necessary to a method invocation and then afterward, carries the - result. - - - - - - - - - - - -Field Summary - - - - -Fields inherited from class org.apache.hadoop.hbase.ipc.RpcServer.Call -cellBlockStream, cellScanner, connection, deadline, header, id, isError, md, param, remoteAddress, reqCleanup, response, rpcCallback, service, size, startTime, timeout, timestamp, tinfo, user - - - - - - - - -Constructor Summary - -Constructors - -Constructor and Description - - -Call(int id, - org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService service, - org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor md, - org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader header, -org.apache.hadoop.hbase.shaded.com.google.protobuf.Message param, -CellScanner cellScanner, -RpcServer.Connection connection, -long size, -org.apache.htrace.TraceInfo tinfo, -http://docs.oracle.com/javase/8/docs/api/java/net/InetAddress.html?is-external=true"; title="class or interface in java.net">InetAddress remoteAddress, -int timeout, -RpcServer.CallCleanup reqCleanup) - - - - - - - - - -Method Summary - -All Methods Instance Methods Concrete Methods - -Modifier and Type -Method and Description - - -long -disconnectSince() -Check if the caller who made this IPC call has disconnected. - - - -(package private) NettyRpcServer.NettyConnection -getConnection() - - -void -sendResponseIfReady() -If we have a response, and delay is not set, then respond immediately. - - - -void -sendResponseIfReady(io.netty.channel.ChannelFutureListener listener) - - - - - - -Methods inherited from class org.apache.hadoop.hbase.ipc.RpcServer.Call -cleanup, createHeaderAndMessageBytes, done, getCellScanner, getClientVersionInfo, getDeadline, getHeader, getMethod, href="../../../../../org/apache/hadoop/hbase/ipc/RpcServer.Call.html#getParam--">getParam, > href="../../../../../org/apache/hadoop/hbase/ipc/RpcServer.Call.html#getPriority--">getPriority, > href="../../../../../org/apache/hadoop/hbase/ipc/RpcServer.Call.html#getReceiveTime--">getReceiveTime, > href="../../../../../org/apache/hadoop/hbase/ipc/RpcServer.Call.html#getRemoteAddress--">getRemoteAddress, > href="../../../../../org/apache/hadoop/hbase/ipc/RpcServer.Call.html#getRemotePort--">getRemotePort, > href="../../../../../org/apache/hadoop/hbase/ipc/RpcServer.Call.html#getRequestUser--">getRequestUser, > href="../../../../../org/apache/hadoop/hbase/ipc/RpcServer.Call.html#getRequestUserName--">getRequestUserName, > href="../../../../../org/apache/hadoop/hbase/ipc/RpcServer.Call.html#getResponseBlockSize--">getResponseBlockSize, > href="../../../../../org/apache/hadoop/hbase/ipc/RpcServer.Call.html#getRes
[15/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Call.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Call.html b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Call.html deleted file mode 100644 index 2987e7b..000 --- a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Call.html +++ /dev/null @@ -1,1500 +0,0 @@ -http://www.w3.org/TR/html4/loose.dtd";> - - -Source code - - - - -001/** -002 * Licensed to the Apache Software Foundation (ASF) under one -003 * or more contributor license agreements. See the NOTICE file -004 * distributed with this work for additional information -005 * regarding copyright ownership. The ASF licenses this file -006 * to you under the Apache License, Version 2.0 (the -007 * "License"); you may not use this file except in compliance -008 * with the License. You may obtain a copy of the License at -009 * -010 * http://www.apache.org/licenses/LICENSE-2.0 -011 * -012 * Unless required by applicable law or agreed to in writing, software -013 * distributed under the License is distributed on an "AS IS" BASIS, -014 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -015 * See the License for the specific language governing permissions and -016 * limitations under the License. -017 */ -018 -019package org.apache.hadoop.hbase.ipc; -020 -021import java.io.IOException; -022import java.io.InputStream; -023import java.net.BindException; -024import java.net.InetAddress; -025import java.net.InetSocketAddress; -026import java.net.ServerSocket; -027import java.net.Socket; -028import java.net.SocketException; -029import java.net.UnknownHostException; -030import java.nio.ByteBuffer; -031import java.nio.channels.CancelledKeyException; -032import java.nio.channels.ClosedChannelException; -033import java.nio.channels.GatheringByteChannel; -034import java.nio.channels.ReadableByteChannel; -035import java.nio.channels.SelectionKey; -036import java.nio.channels.Selector; -037import java.nio.channels.ServerSocketChannel; -038import java.nio.channels.SocketChannel; -039import java.util.ArrayList; -040import java.util.Arrays; -041import java.util.Collections; -042import java.util.Iterator; -043import java.util.List; -044import java.util.Set; -045import java.util.Timer; -046import java.util.TimerTask; -047import java.util.concurrent.ConcurrentHashMap; -048import java.util.concurrent.ConcurrentLinkedDeque; -049import java.util.concurrent.ExecutorService; -050import java.util.concurrent.Executors; -051import java.util.concurrent.LinkedBlockingQueue; -052import java.util.concurrent.atomic.AtomicInteger; -053import java.util.concurrent.atomic.LongAdder; -054import java.util.concurrent.locks.Lock; -055import java.util.concurrent.locks.ReentrantLock; -056 -057import org.apache.hadoop.conf.Configuration; -058import org.apache.hadoop.hbase.CellScanner; -059import org.apache.hadoop.hbase.DoNotRetryIOException; -060import org.apache.hadoop.hbase.HBaseIOException; -061import org.apache.hadoop.hbase.HConstants; -062import org.apache.hadoop.hbase.Server; -063import org.apache.hadoop.hbase.classification.InterfaceAudience; -064import org.apache.hadoop.hbase.classification.InterfaceStability; -065import org.apache.hadoop.hbase.client.VersionInfoUtil; -066import org.apache.hadoop.hbase.exceptions.RequestTooBigException; -067import org.apache.hadoop.hbase.monitoring.MonitoredRPCHandler; -068import org.apache.hadoop.hbase.nio.ByteBuff; -069import org.apache.hadoop.hbase.nio.SingleByteBuff; -070import org.apache.hadoop.hbase.security.AccessDeniedException; -071import org.apache.hadoop.hbase.security.AuthMethod; -072import org.apache.hadoop.hbase.security.HBasePolicyProvider; -073import org.apache.hadoop.hbase.security.SaslStatus; -074import org.apache.hadoop.hbase.security.SaslUtil; -075import org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService; -076import org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream; -077import org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor; -078import org.apache.hadoop.hbase.shaded.com.google.protobuf.Message; -079import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil; -080import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader; -081import org.apache.hadoop.hbase.util.Bytes; -082import org.apache.hadoop.hbase.util.Pair; -083import org.apache.hadoop.hbase.util.Threads; -084import org.apache.hadoop.io.IOUtils; -085import org.apache.hadoop.io.IntWritable; -086import org.apache.hadoop.security.authorize.ServiceAuthorizationManager; -087import org.apache.hadoop.util.StringUtils; -088import org.apache.htrace.TraceInfo; -089 -090import com.google.common.util.concurrent.ThreadFactoryBuilder; -091 -092/** -093 * The
[24/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.NettyConnection.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.NettyConnection.html b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.NettyConnection.html index 27e0dee..109b5f3 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.NettyConnection.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.NettyConnection.html @@ -67,157 +67,157 @@ 059import org.apache.hadoop.hbase.CellScanner; 060import org.apache.hadoop.hbase.HConstants; 061import org.apache.hadoop.hbase.Server; -062import org.apache.hadoop.hbase.classification.InterfaceStability; -063import org.apache.hadoop.hbase.monitoring.MonitoredRPCHandler; -064import org.apache.hadoop.hbase.nio.ByteBuff; -065import org.apache.hadoop.hbase.nio.SingleByteBuff; -066import org.apache.hadoop.hbase.security.AccessDeniedException; -067import org.apache.hadoop.hbase.security.AuthMethod; -068import org.apache.hadoop.hbase.security.HBasePolicyProvider; -069import org.apache.hadoop.hbase.security.SaslStatus; -070import org.apache.hadoop.hbase.security.SaslUtil; -071import org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService; -072import org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor; -073import org.apache.hadoop.hbase.shaded.com.google.protobuf.Message; -074import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader; -075import org.apache.hadoop.hbase.util.Bytes; -076import org.apache.hadoop.hbase.util.JVM; -077import org.apache.hadoop.hbase.util.Pair; -078import org.apache.hadoop.io.IntWritable; -079import org.apache.hadoop.security.authorize.ServiceAuthorizationManager; -080import org.apache.htrace.TraceInfo; -081 -082/** -083 * An RPC server with Netty4 implementation. -084 * -085 */ -086public class NettyRpcServer extends RpcServer { -087 -088 public static final Log LOG = LogFactory.getLog(NettyRpcServer.class); -089 -090 protected final InetSocketAddress bindAddress; -091 -092 private final CountDownLatch closed = new CountDownLatch(1); -093 private final Channel serverChannel; -094 private final ChannelGroup allChannels = new DefaultChannelGroup(GlobalEventExecutor.INSTANCE);; -095 -096 public NettyRpcServer(final Server server, final String name, -097 final Listservices, -098 final InetSocketAddress bindAddress, Configuration conf, -099 RpcScheduler scheduler) throws IOException { -100super(server, name, services, bindAddress, conf, scheduler); -101this.bindAddress = bindAddress; -102boolean useEpoll = useEpoll(conf); -103int workerCount = conf.getInt("hbase.netty.rpc.server.worker.count", -104 Runtime.getRuntime().availableProcessors() / 4); -105EventLoopGroup bossGroup = null; -106EventLoopGroup workerGroup = null; -107if (useEpoll) { -108 bossGroup = new EpollEventLoopGroup(1); -109 workerGroup = new EpollEventLoopGroup(workerCount); -110} else { -111 bossGroup = new NioEventLoopGroup(1); -112 workerGroup = new NioEventLoopGroup(workerCount); -113} -114ServerBootstrap bootstrap = new ServerBootstrap(); -115bootstrap.group(bossGroup, workerGroup); -116if (useEpoll) { -117 bootstrap.channel(EpollServerSocketChannel.class); -118} else { -119 bootstrap.channel(NioServerSocketChannel.class); -120} -121 bootstrap.childOption(ChannelOption.TCP_NODELAY, tcpNoDelay); -122 bootstrap.childOption(ChannelOption.SO_KEEPALIVE, tcpKeepAlive); -123 bootstrap.childOption(ChannelOption.ALLOCATOR, -124 PooledByteBufAllocator.DEFAULT); -125bootstrap.childHandler(new Initializer(maxRequestSize)); -126 -127try { -128 serverChannel = bootstrap.bind(this.bindAddress).sync().channel(); -129 LOG.info("NettyRpcServer bind to address=" + serverChannel.localAddress() -130 + ", hbase.netty.rpc.server.worker.count=" + workerCount -131 + ", useEpoll=" + useEpoll); -132 allChannels.add(serverChannel); -133} catch (InterruptedException e) { -134 throw new InterruptedIOException(e.getMessage()); -135} -136initReconfigurable(conf); -137this.scheduler.init(new RpcSchedulerContext(this)); -138 } -139 -140 private static boolean useEpoll(Configuration conf) { -141// Config to enable native transport. -142boolean epollEnabled = conf.getBoolean("hbase.rpc.server.nativetransport", -143true); -144// Use the faster native epoll transport mechanism on linux if enabled -145return epollEnabled && JVM.isLinux() && JVM.isAmd64(); -146 } -147 -148 @Override -149 public synchronized void start() { -150if (started) { -151 return; -152
[10/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Responder.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Responder.html b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Responder.html index 2987e7b..7f61b54 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Responder.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Responder.html @@ -135,1305 +135,1251 @@ 127 private Listener listener = null; 128 protected Responder responder = null; 129 -130 /** -131 * Datastructure that holds all necessary to a method invocation and then afterward, carries -132 * the result. -133 */ -134 @InterfaceStability.Evolving -135 public class Call extends RpcServer.Call { -136 -137protected Responder responder; +130 /** Listens on the socket. Creates jobs for the handler threads*/ +131 private class Listener extends Thread { +132 +133private ServerSocketChannel acceptChannel = null; //the accept channel +134private Selector selector = null; //the selector that we use for the server +135private Reader[] readers = null; +136private int currentReader = 0; +137private final int readerPendingConnectionQueueLength; 138 -139 @edu.umd.cs.findbugs.annotations.SuppressWarnings(value="NP_NULL_ON_SOME_PATH", -140justification="Can't figure why this complaint is happening... see below") -141Call(int id, final BlockingService service, final MethodDescriptor md, -142RequestHeader header, Message param, CellScanner cellScanner, -143RpcServer.Connection connection, long size, TraceInfo tinfo, -144final InetAddress remoteAddress, int timeout, CallCleanup reqCleanup, -145Responder responder) { -146 super(id, service, md, header, param, cellScanner, connection, size, -147 tinfo, remoteAddress, timeout, reqCleanup); -148 this.responder = responder; -149} +139private ExecutorService readPool; +140 +141public Listener(final String name) throws IOException { +142 super(name); +143 // The backlog of requests that we will have the serversocket carry. +144 int backlogLength = conf.getInt("hbase.ipc.server.listen.queue.size", 128); +145 readerPendingConnectionQueueLength = +146 conf.getInt("hbase.ipc.server.read.connection-queue.size", 100); +147 // Create a new server socket and set to non blocking mode +148 acceptChannel = ServerSocketChannel.open(); +149 acceptChannel.configureBlocking(false); 150 -151/** -152 * Call is done. Execution happened and we returned results to client. It is now safe to -153 * cleanup. -154 */ -155 @edu.umd.cs.findbugs.annotations.SuppressWarnings(value="IS2_INCONSISTENT_SYNC", -156justification="Presume the lock on processing request held by caller is protection enough") -157@Override -158void done() { -159 super.done(); -160 this.getConnection().decRpcCount(); // Say that we're done with this call. -161} -162 -163@Override -164public long disconnectSince() { -165 if (!getConnection().isConnectionOpen()) { -166return System.currentTimeMillis() - timestamp; -167 } else { -168return -1L; -169 } -170} -171 -172@Override -173public synchronized void sendResponseIfReady() throws IOException { -174 // set param null to reduce memory pressure -175 this.param = null; -176 this.responder.doRespond(this); +151 // Bind the server socket to the binding addrees (can be different from the default interface) +152 bind(acceptChannel.socket(), bindAddress, backlogLength); +153 port = acceptChannel.socket().getLocalPort(); //Could be an ephemeral port +154 address = (InetSocketAddress)acceptChannel.socket().getLocalSocketAddress(); +155 // create a selector; +156 selector = Selector.open(); +157 +158 readers = new Reader[readThreads]; +159 // Why this executor thing? Why not like hadoop just start up all the threads? I suppose it +160 // has an advantage in that it is easy to shutdown the pool. +161 readPool = Executors.newFixedThreadPool(readThreads, +162new ThreadFactoryBuilder().setNameFormat( +163 "RpcServer.reader=%d,bindAddress=" + bindAddress.getHostName() + +164 ",port=" + port).setDaemon(true) +165 .setUncaughtExceptionHandler(Threads.LOGGING_EXCEPTION_HANDLER).build()); +166 for (int i = 0; i < readThreads; ++i) { +167Reader reader = new Reader(); +168readers[i] = reader; +169readPool.execute(reader); +170 } +171 LOG.info(getName() + ": started " + readThreads + " reader(s) listening on port=" + port); +172 +173 // Register accepts on th
[12/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Listener.Reader.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Listener.Reader.html b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Listener.Reader.html index 2987e7b..7f61b54 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Listener.Reader.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Listener.Reader.html @@ -135,1305 +135,1251 @@ 127 private Listener listener = null; 128 protected Responder responder = null; 129 -130 /** -131 * Datastructure that holds all necessary to a method invocation and then afterward, carries -132 * the result. -133 */ -134 @InterfaceStability.Evolving -135 public class Call extends RpcServer.Call { -136 -137protected Responder responder; +130 /** Listens on the socket. Creates jobs for the handler threads*/ +131 private class Listener extends Thread { +132 +133private ServerSocketChannel acceptChannel = null; //the accept channel +134private Selector selector = null; //the selector that we use for the server +135private Reader[] readers = null; +136private int currentReader = 0; +137private final int readerPendingConnectionQueueLength; 138 -139 @edu.umd.cs.findbugs.annotations.SuppressWarnings(value="NP_NULL_ON_SOME_PATH", -140justification="Can't figure why this complaint is happening... see below") -141Call(int id, final BlockingService service, final MethodDescriptor md, -142RequestHeader header, Message param, CellScanner cellScanner, -143RpcServer.Connection connection, long size, TraceInfo tinfo, -144final InetAddress remoteAddress, int timeout, CallCleanup reqCleanup, -145Responder responder) { -146 super(id, service, md, header, param, cellScanner, connection, size, -147 tinfo, remoteAddress, timeout, reqCleanup); -148 this.responder = responder; -149} +139private ExecutorService readPool; +140 +141public Listener(final String name) throws IOException { +142 super(name); +143 // The backlog of requests that we will have the serversocket carry. +144 int backlogLength = conf.getInt("hbase.ipc.server.listen.queue.size", 128); +145 readerPendingConnectionQueueLength = +146 conf.getInt("hbase.ipc.server.read.connection-queue.size", 100); +147 // Create a new server socket and set to non blocking mode +148 acceptChannel = ServerSocketChannel.open(); +149 acceptChannel.configureBlocking(false); 150 -151/** -152 * Call is done. Execution happened and we returned results to client. It is now safe to -153 * cleanup. -154 */ -155 @edu.umd.cs.findbugs.annotations.SuppressWarnings(value="IS2_INCONSISTENT_SYNC", -156justification="Presume the lock on processing request held by caller is protection enough") -157@Override -158void done() { -159 super.done(); -160 this.getConnection().decRpcCount(); // Say that we're done with this call. -161} -162 -163@Override -164public long disconnectSince() { -165 if (!getConnection().isConnectionOpen()) { -166return System.currentTimeMillis() - timestamp; -167 } else { -168return -1L; -169 } -170} -171 -172@Override -173public synchronized void sendResponseIfReady() throws IOException { -174 // set param null to reduce memory pressure -175 this.param = null; -176 this.responder.doRespond(this); +151 // Bind the server socket to the binding addrees (can be different from the default interface) +152 bind(acceptChannel.socket(), bindAddress, backlogLength); +153 port = acceptChannel.socket().getLocalPort(); //Could be an ephemeral port +154 address = (InetSocketAddress)acceptChannel.socket().getLocalSocketAddress(); +155 // create a selector; +156 selector = Selector.open(); +157 +158 readers = new Reader[readThreads]; +159 // Why this executor thing? Why not like hadoop just start up all the threads? I suppose it +160 // has an advantage in that it is easy to shutdown the pool. +161 readPool = Executors.newFixedThreadPool(readThreads, +162new ThreadFactoryBuilder().setNameFormat( +163 "RpcServer.reader=%d,bindAddress=" + bindAddress.getHostName() + +164 ",port=" + port).setDaemon(true) +165 .setUncaughtExceptionHandler(Threads.LOGGING_EXCEPTION_HANDLER).build()); +166 for (int i = 0; i < readThreads; ++i) { +167Reader reader = new Reader(); +168readers[i] = reader; +169readPool.execute(reader); +170 } +171 LOG.info(getName() + ": started " + readThreads + " reader(s) listening on port=" + port); +172 +173
[31/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/org/apache/hadoop/hbase/tmpl/regionserver/RSStatusTmpl.ImplData.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/tmpl/regionserver/RSStatusTmpl.ImplData.html b/devapidocs/org/apache/hadoop/hbase/tmpl/regionserver/RSStatusTmpl.ImplData.html index 19716c1..6034ffe 100644 --- a/devapidocs/org/apache/hadoop/hbase/tmpl/regionserver/RSStatusTmpl.ImplData.html +++ b/devapidocs/org/apache/hadoop/hbase/tmpl/regionserver/RSStatusTmpl.ImplData.html @@ -293,76 +293,76 @@ extends org.jamon.AbstractTemplateProxy.ImplData private HRegionServer m_regionServer - + -m_format -private http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true"; title="class or interface in java.lang">String m_format +m_filter +private http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true"; title="class or interface in java.lang">String m_filter - + -m_format__IsNotDefault -private boolean m_format__IsNotDefault +m_filter__IsNotDefault +private boolean m_filter__IsNotDefault - + -m_bcv -private http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true"; title="class or interface in java.lang">String m_bcv +m_bcn +private http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true"; title="class or interface in java.lang">String m_bcn - + -m_bcv__IsNotDefault -private boolean m_bcv__IsNotDefault +m_bcn__IsNotDefault +private boolean m_bcn__IsNotDefault - + -m_bcn -private http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true"; title="class or interface in java.lang">String m_bcn +m_format +private http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true"; title="class or interface in java.lang">String m_format - + -m_bcn__IsNotDefault -private boolean m_bcn__IsNotDefault +m_format__IsNotDefault +private boolean m_format__IsNotDefault - + -m_filter -private http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true"; title="class or interface in java.lang">String m_filter +m_bcv +private http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true"; title="class or interface in java.lang">String m_bcv - + -m_filter__IsNotDefault -private boolean m_filter__IsNotDefault +m_bcv__IsNotDefault +private boolean m_bcv__IsNotDefault @@ -408,112 +408,112 @@ extends org.jamon.AbstractTemplateProxy.ImplData public HRegionServer getRegionServer() - + -setFormat -public void setFormat(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true"; title="class or interface in java.lang">String format) +setFilter +public void setFilter(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true"; title="class or interface in java.lang">String filter) - + -getFormat -public http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true"; title="class or interface in java.lang">String getFormat() +getFilter +public http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true"; title="class or interface in java.lang">String getFilter() - + -getFormat__IsNotDefault -public boolean getFormat__IsNotDefault() +getFilter__IsNotDefault +public boolean getFilter__IsNotDefault() - + -setBcv -public void setBcv(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true"; title="class or interface in java.lang">String bcv) +setBcn +public void setBcn(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true"; title="class or interface in java.lang">String bcn) - + -getBcv -public http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true"; title="class or interface in java.lang">String getBcv() +getBcn +public http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true"; title="class or interface in java.lang">String getBcn() - + -getBcv__IsNotDefault -public boolean getBcv__IsNotDefault() +getBcn__IsNotDefault +public boolean getBcn__IsNotDefault() - + -setBcn -public void setBcn(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true"; title="class or interface in java.lang">String bcn) +setFormat +public void setFormat(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true"; title="class or interface in java.lang">String format) - + -getBcn -public http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true"; title="class or interface in java.lang">String getBcn() +getFormat +public http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true"; title="class or interface in java.lang">
[23/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.html b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.html index 27e0dee..109b5f3 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.html @@ -67,157 +67,157 @@ 059import org.apache.hadoop.hbase.CellScanner; 060import org.apache.hadoop.hbase.HConstants; 061import org.apache.hadoop.hbase.Server; -062import org.apache.hadoop.hbase.classification.InterfaceStability; -063import org.apache.hadoop.hbase.monitoring.MonitoredRPCHandler; -064import org.apache.hadoop.hbase.nio.ByteBuff; -065import org.apache.hadoop.hbase.nio.SingleByteBuff; -066import org.apache.hadoop.hbase.security.AccessDeniedException; -067import org.apache.hadoop.hbase.security.AuthMethod; -068import org.apache.hadoop.hbase.security.HBasePolicyProvider; -069import org.apache.hadoop.hbase.security.SaslStatus; -070import org.apache.hadoop.hbase.security.SaslUtil; -071import org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService; -072import org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor; -073import org.apache.hadoop.hbase.shaded.com.google.protobuf.Message; -074import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader; -075import org.apache.hadoop.hbase.util.Bytes; -076import org.apache.hadoop.hbase.util.JVM; -077import org.apache.hadoop.hbase.util.Pair; -078import org.apache.hadoop.io.IntWritable; -079import org.apache.hadoop.security.authorize.ServiceAuthorizationManager; -080import org.apache.htrace.TraceInfo; -081 -082/** -083 * An RPC server with Netty4 implementation. -084 * -085 */ -086public class NettyRpcServer extends RpcServer { -087 -088 public static final Log LOG = LogFactory.getLog(NettyRpcServer.class); -089 -090 protected final InetSocketAddress bindAddress; -091 -092 private final CountDownLatch closed = new CountDownLatch(1); -093 private final Channel serverChannel; -094 private final ChannelGroup allChannels = new DefaultChannelGroup(GlobalEventExecutor.INSTANCE);; -095 -096 public NettyRpcServer(final Server server, final String name, -097 final Listservices, -098 final InetSocketAddress bindAddress, Configuration conf, -099 RpcScheduler scheduler) throws IOException { -100super(server, name, services, bindAddress, conf, scheduler); -101this.bindAddress = bindAddress; -102boolean useEpoll = useEpoll(conf); -103int workerCount = conf.getInt("hbase.netty.rpc.server.worker.count", -104 Runtime.getRuntime().availableProcessors() / 4); -105EventLoopGroup bossGroup = null; -106EventLoopGroup workerGroup = null; -107if (useEpoll) { -108 bossGroup = new EpollEventLoopGroup(1); -109 workerGroup = new EpollEventLoopGroup(workerCount); -110} else { -111 bossGroup = new NioEventLoopGroup(1); -112 workerGroup = new NioEventLoopGroup(workerCount); -113} -114ServerBootstrap bootstrap = new ServerBootstrap(); -115bootstrap.group(bossGroup, workerGroup); -116if (useEpoll) { -117 bootstrap.channel(EpollServerSocketChannel.class); -118} else { -119 bootstrap.channel(NioServerSocketChannel.class); -120} -121 bootstrap.childOption(ChannelOption.TCP_NODELAY, tcpNoDelay); -122 bootstrap.childOption(ChannelOption.SO_KEEPALIVE, tcpKeepAlive); -123 bootstrap.childOption(ChannelOption.ALLOCATOR, -124 PooledByteBufAllocator.DEFAULT); -125bootstrap.childHandler(new Initializer(maxRequestSize)); -126 -127try { -128 serverChannel = bootstrap.bind(this.bindAddress).sync().channel(); -129 LOG.info("NettyRpcServer bind to address=" + serverChannel.localAddress() -130 + ", hbase.netty.rpc.server.worker.count=" + workerCount -131 + ", useEpoll=" + useEpoll); -132 allChannels.add(serverChannel); -133} catch (InterruptedException e) { -134 throw new InterruptedIOException(e.getMessage()); -135} -136initReconfigurable(conf); -137this.scheduler.init(new RpcSchedulerContext(this)); -138 } -139 -140 private static boolean useEpoll(Configuration conf) { -141// Config to enable native transport. -142boolean epollEnabled = conf.getBoolean("hbase.rpc.server.nativetransport", -143true); -144// Use the faster native epoll transport mechanism on linux if enabled -145return epollEnabled && JVM.isLinux() && JVM.isAmd64(); -146 } -147 -148 @Override -149 public synchronized void start() { -150if (started) { -151 return; -152} -153authTokenSecretMgr = createSecretManager(); -154if (authTokenSecretM
[26/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.MessageDecoder.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.MessageDecoder.html b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.MessageDecoder.html index 27e0dee..109b5f3 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.MessageDecoder.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.MessageDecoder.html @@ -67,157 +67,157 @@ 059import org.apache.hadoop.hbase.CellScanner; 060import org.apache.hadoop.hbase.HConstants; 061import org.apache.hadoop.hbase.Server; -062import org.apache.hadoop.hbase.classification.InterfaceStability; -063import org.apache.hadoop.hbase.monitoring.MonitoredRPCHandler; -064import org.apache.hadoop.hbase.nio.ByteBuff; -065import org.apache.hadoop.hbase.nio.SingleByteBuff; -066import org.apache.hadoop.hbase.security.AccessDeniedException; -067import org.apache.hadoop.hbase.security.AuthMethod; -068import org.apache.hadoop.hbase.security.HBasePolicyProvider; -069import org.apache.hadoop.hbase.security.SaslStatus; -070import org.apache.hadoop.hbase.security.SaslUtil; -071import org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService; -072import org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor; -073import org.apache.hadoop.hbase.shaded.com.google.protobuf.Message; -074import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader; -075import org.apache.hadoop.hbase.util.Bytes; -076import org.apache.hadoop.hbase.util.JVM; -077import org.apache.hadoop.hbase.util.Pair; -078import org.apache.hadoop.io.IntWritable; -079import org.apache.hadoop.security.authorize.ServiceAuthorizationManager; -080import org.apache.htrace.TraceInfo; -081 -082/** -083 * An RPC server with Netty4 implementation. -084 * -085 */ -086public class NettyRpcServer extends RpcServer { -087 -088 public static final Log LOG = LogFactory.getLog(NettyRpcServer.class); -089 -090 protected final InetSocketAddress bindAddress; -091 -092 private final CountDownLatch closed = new CountDownLatch(1); -093 private final Channel serverChannel; -094 private final ChannelGroup allChannels = new DefaultChannelGroup(GlobalEventExecutor.INSTANCE);; -095 -096 public NettyRpcServer(final Server server, final String name, -097 final Listservices, -098 final InetSocketAddress bindAddress, Configuration conf, -099 RpcScheduler scheduler) throws IOException { -100super(server, name, services, bindAddress, conf, scheduler); -101this.bindAddress = bindAddress; -102boolean useEpoll = useEpoll(conf); -103int workerCount = conf.getInt("hbase.netty.rpc.server.worker.count", -104 Runtime.getRuntime().availableProcessors() / 4); -105EventLoopGroup bossGroup = null; -106EventLoopGroup workerGroup = null; -107if (useEpoll) { -108 bossGroup = new EpollEventLoopGroup(1); -109 workerGroup = new EpollEventLoopGroup(workerCount); -110} else { -111 bossGroup = new NioEventLoopGroup(1); -112 workerGroup = new NioEventLoopGroup(workerCount); -113} -114ServerBootstrap bootstrap = new ServerBootstrap(); -115bootstrap.group(bossGroup, workerGroup); -116if (useEpoll) { -117 bootstrap.channel(EpollServerSocketChannel.class); -118} else { -119 bootstrap.channel(NioServerSocketChannel.class); -120} -121 bootstrap.childOption(ChannelOption.TCP_NODELAY, tcpNoDelay); -122 bootstrap.childOption(ChannelOption.SO_KEEPALIVE, tcpKeepAlive); -123 bootstrap.childOption(ChannelOption.ALLOCATOR, -124 PooledByteBufAllocator.DEFAULT); -125bootstrap.childHandler(new Initializer(maxRequestSize)); -126 -127try { -128 serverChannel = bootstrap.bind(this.bindAddress).sync().channel(); -129 LOG.info("NettyRpcServer bind to address=" + serverChannel.localAddress() -130 + ", hbase.netty.rpc.server.worker.count=" + workerCount -131 + ", useEpoll=" + useEpoll); -132 allChannels.add(serverChannel); -133} catch (InterruptedException e) { -134 throw new InterruptedIOException(e.getMessage()); -135} -136initReconfigurable(conf); -137this.scheduler.init(new RpcSchedulerContext(this)); -138 } -139 -140 private static boolean useEpoll(Configuration conf) { -141// Config to enable native transport. -142boolean epollEnabled = conf.getBoolean("hbase.rpc.server.nativetransport", -143true); -144// Use the faster native epoll transport mechanism on linux if enabled -145return epollEnabled && JVM.isLinux() && JVM.isAmd64(); -146 } -147 -148 @Override -149 public synchronized void start() { -150if (started) { -151 return; -152} -15
[33/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/org/apache/hadoop/hbase/regionserver/RSRpcServices.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/RSRpcServices.html b/devapidocs/org/apache/hadoop/hbase/regionserver/RSRpcServices.html index d98892a..c1b5dd1 100644 --- a/devapidocs/org/apache/hadoop/hbase/regionserver/RSRpcServices.html +++ b/devapidocs/org/apache/hadoop/hbase/regionserver/RSRpcServices.html @@ -118,7 +118,7 @@ var activeTableTab = "activeTableTab"; @InterfaceAudience.Private -public class RSRpcServices +public class RSRpcServices extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true"; title="class or interface in java.lang">Object implements HBaseRPCErrorHandler, org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.AdminService.BlockingInterface, org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.ClientService.BlockingInterface, PriorityFunction, ConfigurationObserver Implements the regionserver RPC services. @@ -826,7 +826,7 @@ implements LOG -protected static final org.apache.commons.logging.Log LOG +protected static final org.apache.commons.logging.Log LOG @@ -835,7 +835,7 @@ implements REGION_SERVER_RPC_SCHEDULER_FACTORY_CLASS -public static final http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true"; title="class or interface in java.lang">String REGION_SERVER_RPC_SCHEDULER_FACTORY_CLASS +public static final http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true"; title="class or interface in java.lang">String REGION_SERVER_RPC_SCHEDULER_FACTORY_CLASS RPC scheduler to use for the region server. See Also: @@ -849,7 +849,7 @@ implements REGION_SERVER_RPC_MINIMUM_SCAN_TIME_LIMIT_DELTA -private static final http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true"; title="class or interface in java.lang">String REGION_SERVER_RPC_MINIMUM_SCAN_TIME_LIMIT_DELTA +private static final http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true"; title="class or interface in java.lang">String REGION_SERVER_RPC_MINIMUM_SCAN_TIME_LIMIT_DELTA Minimum allowable time limit delta (in milliseconds) that can be enforced during scans. This configuration exists to prevent the scenario where a time limit is specified to be so restrictive that the time limit is reached immediately (before any cells are scanned). @@ -865,7 +865,7 @@ implements DEFAULT_REGION_SERVER_RPC_MINIMUM_SCAN_TIME_LIMIT_DELTA -private static final long DEFAULT_REGION_SERVER_RPC_MINIMUM_SCAN_TIME_LIMIT_DELTA +private static final long DEFAULT_REGION_SERVER_RPC_MINIMUM_SCAN_TIME_LIMIT_DELTA Default value of REGION_SERVER_RPC_MINIMUM_SCAN_TIME_LIMIT_DELTA See Also: @@ -879,7 +879,7 @@ implements requestCount -final http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/atomic/LongAdder.html?is-external=true"; title="class or interface in java.util.concurrent.atomic">LongAdder requestCount +final http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/atomic/LongAdder.html?is-external=true"; title="class or interface in java.util.concurrent.atomic">LongAdder requestCount @@ -888,7 +888,7 @@ implements rpcGetRequestCount -final http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/atomic/LongAdder.html?is-external=true"; title="class or interface in java.util.concurrent.atomic">LongAdder rpcGetRequestCount +final http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/atomic/LongAdder.html?is-external=true"; title="class or interface in java.util.concurrent.atomic">LongAdder rpcGetRequestCount @@ -897,7 +897,7 @@ implements rpcScanRequestCount -final http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/atomic/LongAdder.html?is-external=true"; title="class or interface in java.util.concurrent.atomic">LongAdder rpcScanRequestCount +final http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/atomic/LongAdder.html?is-external=true"; title="class or interface in java.util.concurrent.atomic">LongAdder rpcScanRequestCount @@ -906,7 +906,7 @@ implements rpcMultiRequestCount -final http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/atomic/LongAdder.html?is-external=true"; title="class or interface in java.util.concurrent.atomic">LongAdder rpcMultiRequestCount +final http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/atomic/LongAdder.html?is-external=true"; title="class or interface in java.util.concurrent.atomic">LongAdder rpcMultiRequestCount @@ -915,7 +915,7 @@ implements rpcMutateRequestCount -final http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/atomic/LongAdder.html?is-external=true"; title="class or interface in java.util.concurrent.atomic">LongAdder rpcMutateRequestCount +final http://docs.oracl
[14/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Connection.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Connection.html b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Connection.html index 2987e7b..7f61b54 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Connection.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Connection.html @@ -135,1305 +135,1251 @@ 127 private Listener listener = null; 128 protected Responder responder = null; 129 -130 /** -131 * Datastructure that holds all necessary to a method invocation and then afterward, carries -132 * the result. -133 */ -134 @InterfaceStability.Evolving -135 public class Call extends RpcServer.Call { -136 -137protected Responder responder; +130 /** Listens on the socket. Creates jobs for the handler threads*/ +131 private class Listener extends Thread { +132 +133private ServerSocketChannel acceptChannel = null; //the accept channel +134private Selector selector = null; //the selector that we use for the server +135private Reader[] readers = null; +136private int currentReader = 0; +137private final int readerPendingConnectionQueueLength; 138 -139 @edu.umd.cs.findbugs.annotations.SuppressWarnings(value="NP_NULL_ON_SOME_PATH", -140justification="Can't figure why this complaint is happening... see below") -141Call(int id, final BlockingService service, final MethodDescriptor md, -142RequestHeader header, Message param, CellScanner cellScanner, -143RpcServer.Connection connection, long size, TraceInfo tinfo, -144final InetAddress remoteAddress, int timeout, CallCleanup reqCleanup, -145Responder responder) { -146 super(id, service, md, header, param, cellScanner, connection, size, -147 tinfo, remoteAddress, timeout, reqCleanup); -148 this.responder = responder; -149} +139private ExecutorService readPool; +140 +141public Listener(final String name) throws IOException { +142 super(name); +143 // The backlog of requests that we will have the serversocket carry. +144 int backlogLength = conf.getInt("hbase.ipc.server.listen.queue.size", 128); +145 readerPendingConnectionQueueLength = +146 conf.getInt("hbase.ipc.server.read.connection-queue.size", 100); +147 // Create a new server socket and set to non blocking mode +148 acceptChannel = ServerSocketChannel.open(); +149 acceptChannel.configureBlocking(false); 150 -151/** -152 * Call is done. Execution happened and we returned results to client. It is now safe to -153 * cleanup. -154 */ -155 @edu.umd.cs.findbugs.annotations.SuppressWarnings(value="IS2_INCONSISTENT_SYNC", -156justification="Presume the lock on processing request held by caller is protection enough") -157@Override -158void done() { -159 super.done(); -160 this.getConnection().decRpcCount(); // Say that we're done with this call. -161} -162 -163@Override -164public long disconnectSince() { -165 if (!getConnection().isConnectionOpen()) { -166return System.currentTimeMillis() - timestamp; -167 } else { -168return -1L; -169 } -170} -171 -172@Override -173public synchronized void sendResponseIfReady() throws IOException { -174 // set param null to reduce memory pressure -175 this.param = null; -176 this.responder.doRespond(this); +151 // Bind the server socket to the binding addrees (can be different from the default interface) +152 bind(acceptChannel.socket(), bindAddress, backlogLength); +153 port = acceptChannel.socket().getLocalPort(); //Could be an ephemeral port +154 address = (InetSocketAddress)acceptChannel.socket().getLocalSocketAddress(); +155 // create a selector; +156 selector = Selector.open(); +157 +158 readers = new Reader[readThreads]; +159 // Why this executor thing? Why not like hadoop just start up all the threads? I suppose it +160 // has an advantage in that it is easy to shutdown the pool. +161 readPool = Executors.newFixedThreadPool(readThreads, +162new ThreadFactoryBuilder().setNameFormat( +163 "RpcServer.reader=%d,bindAddress=" + bindAddress.getHostName() + +164 ",port=" + port).setDaemon(true) +165 .setUncaughtExceptionHandler(Threads.LOGGING_EXCEPTION_HANDLER).build()); +166 for (int i = 0; i < readThreads; ++i) { +167Reader reader = new Reader(); +168readers[i] = reader; +169readPool.execute(reader); +170 } +171 LOG.info(getName() + ": started " + readThreads + " reader(s) listening on port=" + port); +172 +173 // Register accepts
[39/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Listener.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Listener.html b/devapidocs/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Listener.html index 85f0adc..5c85b2d 100644 --- a/devapidocs/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Listener.html +++ b/devapidocs/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Listener.html @@ -122,7 +122,7 @@ var activeTableTab = "activeTableTab"; -private class SimpleRpcServer.Listener +private class SimpleRpcServer.Listener extends http://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html?is-external=true"; title="class or interface in java.lang">Thread Listens on the socket. Creates jobs for the handler threads @@ -296,7 +296,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html? acceptChannel -private http://docs.oracle.com/javase/8/docs/api/java/nio/channels/ServerSocketChannel.html?is-external=true"; title="class or interface in java.nio.channels">ServerSocketChannel acceptChannel +private http://docs.oracle.com/javase/8/docs/api/java/nio/channels/ServerSocketChannel.html?is-external=true"; title="class or interface in java.nio.channels">ServerSocketChannel acceptChannel @@ -305,7 +305,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html? selector -private http://docs.oracle.com/javase/8/docs/api/java/nio/channels/Selector.html?is-external=true"; title="class or interface in java.nio.channels">Selector selector +private http://docs.oracle.com/javase/8/docs/api/java/nio/channels/Selector.html?is-external=true"; title="class or interface in java.nio.channels">Selector selector @@ -314,7 +314,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html? readers -private SimpleRpcServer.Listener.Reader[] readers +private SimpleRpcServer.Listener.Reader[] readers @@ -323,7 +323,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html? currentReader -private int currentReader +private int currentReader @@ -332,7 +332,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html? readerPendingConnectionQueueLength -private final int readerPendingConnectionQueueLength +private final int readerPendingConnectionQueueLength @@ -341,7 +341,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html? readPool -private http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/ExecutorService.html?is-external=true"; title="class or interface in java.util.concurrent">ExecutorService readPool +private http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/ExecutorService.html?is-external=true"; title="class or interface in java.util.concurrent">ExecutorService readPool @@ -358,7 +358,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html? Listener -public Listener(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true"; title="class or interface in java.lang">String name) +public Listener(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true"; title="class or interface in java.lang">String name) throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true"; title="class or interface in java.io">IOException Throws: @@ -380,7 +380,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html? run -public void run() +public void run() Specified by: http://docs.oracle.com/javase/8/docs/api/java/lang/Runnable.html?is-external=true#run--"; title="class or interface in java.lang">run in interface http://docs.oracle.com/javase/8/docs/api/java/lang/Runnable.html?is-external=true"; title="class or interface in java.lang">Runnable @@ -395,7 +395,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html? closeCurrentConnection -private void closeCurrentConnection(http://docs.oracle.com/javase/8/docs/api/java/nio/channels/SelectionKey.html?is-external=true"; title="class or interface in java.nio.channels">SelectionKey key, +private void closeCurrentConnection(http://docs.oracle.com/javase/8/docs/api/java/nio/channels/SelectionKey.html?is-external=true"; title="class or interface in java.nio.channels">SelectionKey key, http://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true"; title="class or interface in java.lang">Throwable e) @@ -405,7 +405,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html? getAddress -http://docs.oracle.com/javase/8/docs/api/java/net/InetSocketAddress.html?is-external=true"; title="class or interface in java.net">InetSocketAddress getAddress() +http://docs.oracle.com/javase/8/docs/api/java/net/
[02/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/RSRpcServices.RegionScannerHolder.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/RSRpcServices.RegionScannerHolder.html b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/RSRpcServices.RegionScannerHolder.html index 17d1bcb..6dfd1d4 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/RSRpcServices.RegionScannerHolder.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/RSRpcServices.RegionScannerHolder.html @@ -35,863 +35,863 @@ 027import java.net.InetSocketAddress; 028import java.net.UnknownHostException; 029import java.nio.ByteBuffer; -030import java.util.ArrayList; -031import java.util.Collections; -032import java.util.HashMap; -033import java.util.Iterator; -034import java.util.List; -035import java.util.Map; -036import java.util.Map.Entry; -037import java.util.NavigableMap; -038import java.util.Set; -039import java.util.TreeSet; -040import java.util.concurrent.ConcurrentHashMap; -041import java.util.concurrent.ConcurrentMap; -042import java.util.concurrent.atomic.AtomicLong; -043import java.util.concurrent.atomic.LongAdder; -044 -045import org.apache.commons.lang.mutable.MutableObject; -046import org.apache.commons.logging.Log; -047import org.apache.commons.logging.LogFactory; -048import org.apache.hadoop.conf.Configuration; -049import org.apache.hadoop.fs.Path; -050import org.apache.hadoop.hbase.ByteBufferCell; -051import org.apache.hadoop.hbase.Cell; -052import org.apache.hadoop.hbase.CellScannable; -053import org.apache.hadoop.hbase.CellScanner; -054import org.apache.hadoop.hbase.CellUtil; -055import org.apache.hadoop.hbase.DoNotRetryIOException; -056import org.apache.hadoop.hbase.DroppedSnapshotException; -057import org.apache.hadoop.hbase.HBaseIOException; -058import org.apache.hadoop.hbase.HConstants; -059import org.apache.hadoop.hbase.HRegionInfo; -060import org.apache.hadoop.hbase.HTableDescriptor; -061import org.apache.hadoop.hbase.MultiActionResultTooLarge; -062import org.apache.hadoop.hbase.NotServingRegionException; -063import org.apache.hadoop.hbase.ServerName; -064import org.apache.hadoop.hbase.TableName; -065import org.apache.hadoop.hbase.UnknownScannerException; -066import org.apache.hadoop.hbase.classification.InterfaceAudience; -067import org.apache.hadoop.hbase.client.Append; -068import org.apache.hadoop.hbase.client.ConnectionUtils; -069import org.apache.hadoop.hbase.client.Delete; -070import org.apache.hadoop.hbase.client.Durability; -071import org.apache.hadoop.hbase.client.Get; -072import org.apache.hadoop.hbase.client.Increment; -073import org.apache.hadoop.hbase.client.Mutation; -074import org.apache.hadoop.hbase.client.Put; -075import org.apache.hadoop.hbase.client.RegionReplicaUtil; -076import org.apache.hadoop.hbase.client.Result; -077import org.apache.hadoop.hbase.client.RowMutations; -078import org.apache.hadoop.hbase.client.Scan; -079import org.apache.hadoop.hbase.client.VersionInfoUtil; -080import org.apache.hadoop.hbase.conf.ConfigurationObserver; -081import org.apache.hadoop.hbase.exceptions.FailedSanityCheckException; -082import org.apache.hadoop.hbase.exceptions.OutOfOrderScannerNextException; -083import org.apache.hadoop.hbase.exceptions.ScannerResetException; -084import org.apache.hadoop.hbase.filter.ByteArrayComparable; -085import org.apache.hadoop.hbase.filter.CompareFilter.CompareOp; -086import org.apache.hadoop.hbase.ipc.HBaseRPCErrorHandler; -087import org.apache.hadoop.hbase.ipc.HBaseRpcController; -088import org.apache.hadoop.hbase.ipc.PriorityFunction; -089import org.apache.hadoop.hbase.ipc.QosPriority; -090import org.apache.hadoop.hbase.ipc.RpcCallContext; -091import org.apache.hadoop.hbase.ipc.RpcCallback; -092import org.apache.hadoop.hbase.ipc.RpcServer; -093import org.apache.hadoop.hbase.ipc.RpcServer.BlockingServiceAndInterface; -094import org.apache.hadoop.hbase.ipc.RpcServerFactory; -095import org.apache.hadoop.hbase.ipc.RpcServerInterface; -096import org.apache.hadoop.hbase.ipc.ServerNotRunningYetException; -097import org.apache.hadoop.hbase.ipc.ServerRpcController; -098import org.apache.hadoop.hbase.master.MasterRpcServices; -099import org.apache.hadoop.hbase.quotas.OperationQuota; -100import org.apache.hadoop.hbase.quotas.RegionServerQuotaManager; -101import org.apache.hadoop.hbase.regionserver.HRegion.RegionScannerImpl; -102import org.apache.hadoop.hbase.regionserver.Leases.Lease; -103import org.apache.hadoop.hbase.regionserver.Leases.LeaseStillHeldException; -104import org.apache.hadoop.hbase.regionserver.Region.Operation; -105import org.apache.hadoop.hbase.regionserver.ScannerContext.LimitScope; -106import org.apache.hadoop.hbase.regionserver.handler.OpenMetaHandler; -107import org.apache.hadoop.hbase.regionserver.ha
[44/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/org/apache/hadoop/hbase/ipc/RpcServer.Call.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/ipc/RpcServer.Call.html b/devapidocs/org/apache/hadoop/hbase/ipc/RpcServer.Call.html deleted file mode 100644 index eff339f..000 --- a/devapidocs/org/apache/hadoop/hbase/ipc/RpcServer.Call.html +++ /dev/null @@ -1,1453 +0,0 @@ -http://www.w3.org/TR/html4/loose.dtd";> - - - - - -RpcServer.Call (Apache HBase 2.0.0-SNAPSHOT API) - - - - - -var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":10,"i33":10,"i34":10,"i35":10,"i36":10,"i37":10,"i38":10,"i39":10}; -var tabs = {65535:["t0","All Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]}; -var altColor = "altColor"; -var rowColor = "rowColor"; -var tableTab = "tableTab"; -var activeTableTab = "activeTableTab"; - - -JavaScript is disabled on your browser. - - - - - -Skip navigation links - - - - -Overview -Package -Class -Use -Tree -Deprecated -Index -Help - - - - -Prev Class -Next Class - - -Frames -No Frames - - -All Classes - - - - - - - -Summary: -Nested | -Field | -Constr | -Method - - -Detail: -Field | -Constr | -Method - - - - - - - - -org.apache.hadoop.hbase.ipc -Class RpcServer.Call - - - -http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true"; title="class or interface in java.lang">java.lang.Object - - -org.apache.hadoop.hbase.ipc.RpcServer.Call - - - - - - - -All Implemented Interfaces: -RpcCall, RpcCallContext - - -Direct Known Subclasses: -NettyRpcServer.Call, SimpleRpcServer.Call - - -Enclosing class: -RpcServer - - - -@InterfaceStability.Evolving - @InterfaceAudience.Private -public abstract class RpcServer.Call -extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true"; title="class or interface in java.lang">Object -implements RpcCall -Datastructure that holds all necessary to a method invocation and then afterward, carries - the result. - - - - - - - - - - - -Field Summary - -Fields - -Modifier and Type -Field and Description - - -protected ByteBufferListOutputStream -cellBlockStream - - -protected CellScanner -cellScanner - - -protected RpcServer.Connection -connection - - -protected long -deadline - - -private long -exceptionSize - - -protected org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader -header - - -protected int -id - - -protected boolean -isError - - -protected org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor -md - - -protected org.apache.hadoop.hbase.shaded.com.google.protobuf.Message -param - - -protected http://docs.oracle.com/javase/8/docs/api/java/net/InetAddress.html?is-external=true"; title="class or interface in java.net">InetAddress -remoteAddress - - -protected RpcServer.CallCleanup -reqCleanup - - -protected BufferChain -response -Chain of buffers to send as response. - - - -private long -responseBlockSize - - -private long -responseCellSize - - -private boolean -retryImmediatelySupported - - -protected RpcCallback -rpcCallback - - -protected org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService -service - - -protected long -size - - -protected long -startTime - - -protected int -timeout - - -protected long -timestamp - - -protected org.apache.htrace.TraceInfo -tinfo - - -protected User -user - - - - - - - - - -Constructor Summary - -Constructors - -Constructor and Description - - -Call(int id, - org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService service, - org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor md, - org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader header, -org.apache.hadoop.hbase.shaded.com.google.protobuf.Message param, -CellScanner cellScanner, -RpcServer.Connection connection, -long size, -org.apache.htrace.TraceInfo tinfo, -http://docs.oracle.com/javase/8/docs/api/java/net/InetAddress.html?is-external=true"; title="class or interface in java.net">InetAddress remoteAddress, -int timeout, -RpcServer.CallCleanup reqCleanup) - - - - - - - - - -Method Summary - -All Methods Instance Methods Concrete Methods
[34/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/org/apache/hadoop/hbase/master/package-tree.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/master/package-tree.html b/devapidocs/org/apache/hadoop/hbase/master/package-tree.html index a96a542..c18daae 100644 --- a/devapidocs/org/apache/hadoop/hbase/master/package-tree.html +++ b/devapidocs/org/apache/hadoop/hbase/master/package-tree.html @@ -322,11 +322,11 @@ java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true"; title="class or interface in java.lang">Enum(implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true"; title="class or interface in java.lang">Comparable , java.io.http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true"; title="class or interface in java.io">Serializable) -org.apache.hadoop.hbase.master.RegionState.State -org.apache.hadoop.hbase.master.MetricsMasterSourceFactoryImpl.FactoryStorage -org.apache.hadoop.hbase.master.MasterRpcServices.BalanceSwitchMode -org.apache.hadoop.hbase.master.SplitLogManager.TerminationStatus org.apache.hadoop.hbase.master.SplitLogManager.ResubmitDirective +org.apache.hadoop.hbase.master.SplitLogManager.TerminationStatus +org.apache.hadoop.hbase.master.MasterRpcServices.BalanceSwitchMode +org.apache.hadoop.hbase.master.MetricsMasterSourceFactoryImpl.FactoryStorage +org.apache.hadoop.hbase.master.RegionState.State http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/org/apache/hadoop/hbase/master/procedure/package-tree.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/master/procedure/package-tree.html b/devapidocs/org/apache/hadoop/hbase/master/procedure/package-tree.html index 4460f73..ff790b4 100644 --- a/devapidocs/org/apache/hadoop/hbase/master/procedure/package-tree.html +++ b/devapidocs/org/apache/hadoop/hbase/master/procedure/package-tree.html @@ -174,8 +174,8 @@ java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true"; title="class or interface in java.lang">Enum (implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true"; title="class or interface in java.lang">Comparable , java.io.http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true"; title="class or interface in java.io">Serializable) -org.apache.hadoop.hbase.master.procedure.TableProcedureInterface.TableOperationType org.apache.hadoop.hbase.master.procedure.ServerProcedureInterface.ServerOperationType +org.apache.hadoop.hbase.master.procedure.TableProcedureInterface.TableOperationType org.apache.hadoop.hbase.master.procedure.DisableTableProcedure.MarkRegionOfflineOpResult http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/org/apache/hadoop/hbase/package-tree.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/package-tree.html b/devapidocs/org/apache/hadoop/hbase/package-tree.html index 2cc1064..d7279ca 100644 --- a/devapidocs/org/apache/hadoop/hbase/package-tree.html +++ b/devapidocs/org/apache/hadoop/hbase/package-tree.html @@ -422,16 +422,16 @@ java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true"; title="class or interface in java.lang">Enum (implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true"; title="class or interface in java.lang">Comparable , java.io.http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true"; title="class or interface in java.io">Serializable) -org.apache.hadoop.hbase.KeyValue.Type +org.apache.hadoop.hbase.HConstants.Modify +org.apache.hadoop.hbase.CompatibilitySingletonFactory.SingletonStorage org.apache.hadoop.hbase.KeepDeletedCells +org.apache.hadoop.hbase.HConstants.OperationStatusCode +org.apache.hadoop.hbase.Coprocessor.State org.apache.hadoop.hbase.MetaTableAccessor.QueryType org.apache.hadoop.hbase.MemoryCompactionPolicy -org.apache.hadoop.hbase.HConstants.OperationStatusCode org.apache.hadoop.hbase.HealthChecker.HealthCheckerExitStatus -org.apache.hadoop.hbase.HConstants.Modify org.apache.hadoop.hbase.ProcedureState -org.apache.hadoop.hbase.CompatibilitySingletonFactory.SingletonStorage -org.apache.hadoop.hbase.Coprocessor.State +org.apache.hadoop.hbase.KeyValue.Type http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/org/apache/hadoop/hbase/procedure2/package-tree.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/procedure2/package-tree.html b/devapidocs/org/apache/hadoop/hbase/procedure2/package-tree.html index cbe5806..2992239 1
[40/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/org/apache/hadoop/hbase/ipc/ServerNotRunningYetException.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/ipc/ServerNotRunningYetException.html b/devapidocs/org/apache/hadoop/hbase/ipc/ServerNotRunningYetException.html index b113a5a..63453fa 100644 --- a/devapidocs/org/apache/hadoop/hbase/ipc/ServerNotRunningYetException.html +++ b/devapidocs/org/apache/hadoop/hbase/ipc/ServerNotRunningYetException.html @@ -43,7 +43,7 @@ -Prev Class +Prev Class Next Class @@ -223,7 +223,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/io/IOException.ht -Prev Class +Prev Class Next Class http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Call.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Call.html b/devapidocs/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Call.html deleted file mode 100644 index 59c6329..000 --- a/devapidocs/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Call.html +++ /dev/null @@ -1,429 +0,0 @@ -http://www.w3.org/TR/html4/loose.dtd";> - - - - - -SimpleRpcServer.Call (Apache HBase 2.0.0-SNAPSHOT API) - - - - - -var methods = {"i0":10,"i1":10,"i2":10,"i3":10}; -var tabs = {65535:["t0","All Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]}; -var altColor = "altColor"; -var rowColor = "rowColor"; -var tableTab = "tableTab"; -var activeTableTab = "activeTableTab"; - - -JavaScript is disabled on your browser. - - - - - -Skip navigation links - - - - -Overview -Package -Class -Use -Tree -Deprecated -Index -Help - - - - -Prev Class -Next Class - - -Frames -No Frames - - -All Classes - - - - - - - -Summary: -Nested | -Field | -Constr | -Method - - -Detail: -Field | -Constr | -Method - - - - - - - - -org.apache.hadoop.hbase.ipc -Class SimpleRpcServer.Call - - - -http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true"; title="class or interface in java.lang">java.lang.Object - - -org.apache.hadoop.hbase.ipc.RpcServer.Call - - -org.apache.hadoop.hbase.ipc.SimpleRpcServer.Call - - - - - - - - - -All Implemented Interfaces: -RpcCall, RpcCallContext - - -Enclosing class: -SimpleRpcServer - - - -@InterfaceStability.Evolving -public class SimpleRpcServer.Call -extends RpcServer.Call -Datastructure that holds all necessary to a method invocation and then afterward, carries - the result. - - - - - - - - - - - -Field Summary - -Fields - -Modifier and Type -Field and Description - - -protected SimpleRpcServer.Responder -responder - - - - - - -Fields inherited from class org.apache.hadoop.hbase.ipc.RpcServer.Call -cellBlockStream, cellScanner, connection, deadline, header, id, isError, md, param, remoteAddress, reqCleanup, response, rpcCallback, service, size, startTime, timeout, timestamp, tinfo, user - - - - - - - - -Constructor Summary - -Constructors - -Constructor and Description - - -Call(int id, - org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService service, - org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor md, - org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader header, -org.apache.hadoop.hbase.shaded.com.google.protobuf.Message param, -CellScanner cellScanner, -RpcServer.Connection connection, -long size, -org.apache.htrace.TraceInfo tinfo, -http://docs.oracle.com/javase/8/docs/api/java/net/InetAddress.html?is-external=true"; title="class or interface in java.net">InetAddress remoteAddress, -int timeout, -RpcServer.CallCleanup reqCleanup, -SimpleRpcServer.Responder responder) - - - - - - - - - -Method Summary - -All Methods Instance Methods Concrete Methods - -Modifier and Type -Method and Description - - -long -disconnectSince() -Check if the caller who made this IPC call has disconnected. - - - -(package private) void -done() -Call is done. - - - -(package private) SimpleRpcServer.Connection -getConnection() - - -void -sendResponseIfReady() -Send the response of this RPC call. - - - - - - - -Methods inherited from class org.apache.hadoop.hbase.ipc.RpcServer.Call -cleanup, createHeaderAndMessageBytes, getCellScanner, getClientVersionInfo, getDeadline, getHeader, getMethod, get Param, getPriority, getReceiveTime, getRemoteAddress, getRemotePort
[18/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.Connection.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.Connection.html b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.Connection.html index 5cc356a..9e1c66c 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.Connection.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.Connection.html @@ -70,2037 +70,1559 @@ 062import org.apache.hadoop.hbase.client.VersionInfoUtil; 063import org.apache.hadoop.hbase.codec.Codec; 064import org.apache.hadoop.hbase.conf.ConfigurationObserver; -065import org.apache.hadoop.hbase.exceptions.RegionMovedException; -066import org.apache.hadoop.hbase.exceptions.RequestTooBigException; -067import org.apache.hadoop.hbase.io.ByteBufferListOutputStream; -068import org.apache.hadoop.hbase.io.ByteBufferOutputStream; -069import org.apache.hadoop.hbase.io.ByteBufferPool; -070import org.apache.hadoop.hbase.io.crypto.aes.CryptoAES; -071import org.apache.hadoop.hbase.monitoring.MonitoredRPCHandler; -072import org.apache.hadoop.hbase.monitoring.TaskMonitor; -073import org.apache.hadoop.hbase.nio.ByteBuff; -074import org.apache.hadoop.hbase.nio.MultiByteBuff; -075import org.apache.hadoop.hbase.nio.SingleByteBuff; -076import org.apache.hadoop.hbase.regionserver.RSRpcServices; -077import org.apache.hadoop.hbase.security.AccessDeniedException; -078import org.apache.hadoop.hbase.security.AuthMethod; -079import org.apache.hadoop.hbase.security.HBaseSaslRpcServer; -080import org.apache.hadoop.hbase.security.HBaseSaslRpcServer.SaslDigestCallbackHandler; -081import org.apache.hadoop.hbase.security.HBaseSaslRpcServer.SaslGssCallbackHandler; -082import org.apache.hadoop.hbase.security.SaslStatus; -083import org.apache.hadoop.hbase.security.SaslUtil; -084import org.apache.hadoop.hbase.security.User; -085import org.apache.hadoop.hbase.security.UserProvider; -086import org.apache.hadoop.hbase.security.token.AuthenticationTokenSecretManager; -087import org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService; -088import org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteInput; -089import org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString; -090import org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream; -091import org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream; -092import org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor; -093import org.apache.hadoop.hbase.shaded.com.google.protobuf.Message; -094import org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException; -095import org.apache.hadoop.hbase.shaded.com.google.protobuf.TextFormat; -096import org.apache.hadoop.hbase.shaded.com.google.protobuf.UnsafeByteOperations; -097import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil; -098import org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos; -099import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.VersionInfo; -100import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos; -101import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.CellBlockMeta; -102import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ConnectionHeader; -103import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ExceptionResponse; -104import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader; -105import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ResponseHeader; -106import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation; -107import org.apache.hadoop.hbase.util.ByteBufferUtils; -108import org.apache.hadoop.hbase.util.Bytes; -109import org.apache.hadoop.hbase.util.Pair; -110import org.apache.hadoop.io.BytesWritable; -111import org.apache.hadoop.io.Writable; -112import org.apache.hadoop.io.WritableUtils; -113import org.apache.hadoop.io.compress.CompressionCodec; -114import org.apache.hadoop.security.UserGroupInformation; -115import org.apache.hadoop.security.UserGroupInformation.AuthenticationMethod; -116import org.apache.hadoop.security.authorize.AuthorizationException; -117import org.apache.hadoop.security.authorize.PolicyProvider; -118import org.apache.hadoop.security.authorize.ProxyUsers; -119import org.apache.hadoop.security.authorize.ServiceAuthorizationManager; -120import org.apache.hadoop.security.token.SecretManager; -121import org.apache.hadoop.security.token.SecretManager.InvalidToken; -122import org.apache.hadoop.security.token.TokenIdentifier; -123import org.apache.hadoop.util.StringUtils; -124import org.apache.htrace.TraceInfo; -125import org.codehaus.jackson.map.ObjectMapper; -126 -127import com.google.common.annotations.VisibleForTesting; -128 -129/** -130 *
[25/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.MessageEncoder.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.MessageEncoder.html b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.MessageEncoder.html index 27e0dee..109b5f3 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.MessageEncoder.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.MessageEncoder.html @@ -67,157 +67,157 @@ 059import org.apache.hadoop.hbase.CellScanner; 060import org.apache.hadoop.hbase.HConstants; 061import org.apache.hadoop.hbase.Server; -062import org.apache.hadoop.hbase.classification.InterfaceStability; -063import org.apache.hadoop.hbase.monitoring.MonitoredRPCHandler; -064import org.apache.hadoop.hbase.nio.ByteBuff; -065import org.apache.hadoop.hbase.nio.SingleByteBuff; -066import org.apache.hadoop.hbase.security.AccessDeniedException; -067import org.apache.hadoop.hbase.security.AuthMethod; -068import org.apache.hadoop.hbase.security.HBasePolicyProvider; -069import org.apache.hadoop.hbase.security.SaslStatus; -070import org.apache.hadoop.hbase.security.SaslUtil; -071import org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService; -072import org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor; -073import org.apache.hadoop.hbase.shaded.com.google.protobuf.Message; -074import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader; -075import org.apache.hadoop.hbase.util.Bytes; -076import org.apache.hadoop.hbase.util.JVM; -077import org.apache.hadoop.hbase.util.Pair; -078import org.apache.hadoop.io.IntWritable; -079import org.apache.hadoop.security.authorize.ServiceAuthorizationManager; -080import org.apache.htrace.TraceInfo; -081 -082/** -083 * An RPC server with Netty4 implementation. -084 * -085 */ -086public class NettyRpcServer extends RpcServer { -087 -088 public static final Log LOG = LogFactory.getLog(NettyRpcServer.class); -089 -090 protected final InetSocketAddress bindAddress; -091 -092 private final CountDownLatch closed = new CountDownLatch(1); -093 private final Channel serverChannel; -094 private final ChannelGroup allChannels = new DefaultChannelGroup(GlobalEventExecutor.INSTANCE);; -095 -096 public NettyRpcServer(final Server server, final String name, -097 final Listservices, -098 final InetSocketAddress bindAddress, Configuration conf, -099 RpcScheduler scheduler) throws IOException { -100super(server, name, services, bindAddress, conf, scheduler); -101this.bindAddress = bindAddress; -102boolean useEpoll = useEpoll(conf); -103int workerCount = conf.getInt("hbase.netty.rpc.server.worker.count", -104 Runtime.getRuntime().availableProcessors() / 4); -105EventLoopGroup bossGroup = null; -106EventLoopGroup workerGroup = null; -107if (useEpoll) { -108 bossGroup = new EpollEventLoopGroup(1); -109 workerGroup = new EpollEventLoopGroup(workerCount); -110} else { -111 bossGroup = new NioEventLoopGroup(1); -112 workerGroup = new NioEventLoopGroup(workerCount); -113} -114ServerBootstrap bootstrap = new ServerBootstrap(); -115bootstrap.group(bossGroup, workerGroup); -116if (useEpoll) { -117 bootstrap.channel(EpollServerSocketChannel.class); -118} else { -119 bootstrap.channel(NioServerSocketChannel.class); -120} -121 bootstrap.childOption(ChannelOption.TCP_NODELAY, tcpNoDelay); -122 bootstrap.childOption(ChannelOption.SO_KEEPALIVE, tcpKeepAlive); -123 bootstrap.childOption(ChannelOption.ALLOCATOR, -124 PooledByteBufAllocator.DEFAULT); -125bootstrap.childHandler(new Initializer(maxRequestSize)); -126 -127try { -128 serverChannel = bootstrap.bind(this.bindAddress).sync().channel(); -129 LOG.info("NettyRpcServer bind to address=" + serverChannel.localAddress() -130 + ", hbase.netty.rpc.server.worker.count=" + workerCount -131 + ", useEpoll=" + useEpoll); -132 allChannels.add(serverChannel); -133} catch (InterruptedException e) { -134 throw new InterruptedIOException(e.getMessage()); -135} -136initReconfigurable(conf); -137this.scheduler.init(new RpcSchedulerContext(this)); -138 } -139 -140 private static boolean useEpoll(Configuration conf) { -141// Config to enable native transport. -142boolean epollEnabled = conf.getBoolean("hbase.rpc.server.nativetransport", -143true); -144// Use the faster native epoll transport mechanism on linux if enabled -145return epollEnabled && JVM.isLinux() && JVM.isAmd64(); -146 } -147 -148 @Override -149 public synchronized void start() { -150if (started) { -151 return; -152} -15
[27/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.Initializer.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.Initializer.html b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.Initializer.html index 27e0dee..109b5f3 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.Initializer.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.Initializer.html @@ -67,157 +67,157 @@ 059import org.apache.hadoop.hbase.CellScanner; 060import org.apache.hadoop.hbase.HConstants; 061import org.apache.hadoop.hbase.Server; -062import org.apache.hadoop.hbase.classification.InterfaceStability; -063import org.apache.hadoop.hbase.monitoring.MonitoredRPCHandler; -064import org.apache.hadoop.hbase.nio.ByteBuff; -065import org.apache.hadoop.hbase.nio.SingleByteBuff; -066import org.apache.hadoop.hbase.security.AccessDeniedException; -067import org.apache.hadoop.hbase.security.AuthMethod; -068import org.apache.hadoop.hbase.security.HBasePolicyProvider; -069import org.apache.hadoop.hbase.security.SaslStatus; -070import org.apache.hadoop.hbase.security.SaslUtil; -071import org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService; -072import org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor; -073import org.apache.hadoop.hbase.shaded.com.google.protobuf.Message; -074import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader; -075import org.apache.hadoop.hbase.util.Bytes; -076import org.apache.hadoop.hbase.util.JVM; -077import org.apache.hadoop.hbase.util.Pair; -078import org.apache.hadoop.io.IntWritable; -079import org.apache.hadoop.security.authorize.ServiceAuthorizationManager; -080import org.apache.htrace.TraceInfo; -081 -082/** -083 * An RPC server with Netty4 implementation. -084 * -085 */ -086public class NettyRpcServer extends RpcServer { -087 -088 public static final Log LOG = LogFactory.getLog(NettyRpcServer.class); -089 -090 protected final InetSocketAddress bindAddress; -091 -092 private final CountDownLatch closed = new CountDownLatch(1); -093 private final Channel serverChannel; -094 private final ChannelGroup allChannels = new DefaultChannelGroup(GlobalEventExecutor.INSTANCE);; -095 -096 public NettyRpcServer(final Server server, final String name, -097 final Listservices, -098 final InetSocketAddress bindAddress, Configuration conf, -099 RpcScheduler scheduler) throws IOException { -100super(server, name, services, bindAddress, conf, scheduler); -101this.bindAddress = bindAddress; -102boolean useEpoll = useEpoll(conf); -103int workerCount = conf.getInt("hbase.netty.rpc.server.worker.count", -104 Runtime.getRuntime().availableProcessors() / 4); -105EventLoopGroup bossGroup = null; -106EventLoopGroup workerGroup = null; -107if (useEpoll) { -108 bossGroup = new EpollEventLoopGroup(1); -109 workerGroup = new EpollEventLoopGroup(workerCount); -110} else { -111 bossGroup = new NioEventLoopGroup(1); -112 workerGroup = new NioEventLoopGroup(workerCount); -113} -114ServerBootstrap bootstrap = new ServerBootstrap(); -115bootstrap.group(bossGroup, workerGroup); -116if (useEpoll) { -117 bootstrap.channel(EpollServerSocketChannel.class); -118} else { -119 bootstrap.channel(NioServerSocketChannel.class); -120} -121 bootstrap.childOption(ChannelOption.TCP_NODELAY, tcpNoDelay); -122 bootstrap.childOption(ChannelOption.SO_KEEPALIVE, tcpKeepAlive); -123 bootstrap.childOption(ChannelOption.ALLOCATOR, -124 PooledByteBufAllocator.DEFAULT); -125bootstrap.childHandler(new Initializer(maxRequestSize)); -126 -127try { -128 serverChannel = bootstrap.bind(this.bindAddress).sync().channel(); -129 LOG.info("NettyRpcServer bind to address=" + serverChannel.localAddress() -130 + ", hbase.netty.rpc.server.worker.count=" + workerCount -131 + ", useEpoll=" + useEpoll); -132 allChannels.add(serverChannel); -133} catch (InterruptedException e) { -134 throw new InterruptedIOException(e.getMessage()); -135} -136initReconfigurable(conf); -137this.scheduler.init(new RpcSchedulerContext(this)); -138 } -139 -140 private static boolean useEpoll(Configuration conf) { -141// Config to enable native transport. -142boolean epollEnabled = conf.getBoolean("hbase.rpc.server.nativetransport", -143true); -144// Use the faster native epoll transport mechanism on linux if enabled -145return epollEnabled && JVM.isLinux() && JVM.isAmd64(); -146 } -147 -148 @Override -149 public synchronized void start() { -150if (started) { -151 return; -152} -153authTokenS
[09/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.html b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.html index 2987e7b..7f61b54 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.html @@ -135,1305 +135,1251 @@ 127 private Listener listener = null; 128 protected Responder responder = null; 129 -130 /** -131 * Datastructure that holds all necessary to a method invocation and then afterward, carries -132 * the result. -133 */ -134 @InterfaceStability.Evolving -135 public class Call extends RpcServer.Call { -136 -137protected Responder responder; +130 /** Listens on the socket. Creates jobs for the handler threads*/ +131 private class Listener extends Thread { +132 +133private ServerSocketChannel acceptChannel = null; //the accept channel +134private Selector selector = null; //the selector that we use for the server +135private Reader[] readers = null; +136private int currentReader = 0; +137private final int readerPendingConnectionQueueLength; 138 -139 @edu.umd.cs.findbugs.annotations.SuppressWarnings(value="NP_NULL_ON_SOME_PATH", -140justification="Can't figure why this complaint is happening... see below") -141Call(int id, final BlockingService service, final MethodDescriptor md, -142RequestHeader header, Message param, CellScanner cellScanner, -143RpcServer.Connection connection, long size, TraceInfo tinfo, -144final InetAddress remoteAddress, int timeout, CallCleanup reqCleanup, -145Responder responder) { -146 super(id, service, md, header, param, cellScanner, connection, size, -147 tinfo, remoteAddress, timeout, reqCleanup); -148 this.responder = responder; -149} +139private ExecutorService readPool; +140 +141public Listener(final String name) throws IOException { +142 super(name); +143 // The backlog of requests that we will have the serversocket carry. +144 int backlogLength = conf.getInt("hbase.ipc.server.listen.queue.size", 128); +145 readerPendingConnectionQueueLength = +146 conf.getInt("hbase.ipc.server.read.connection-queue.size", 100); +147 // Create a new server socket and set to non blocking mode +148 acceptChannel = ServerSocketChannel.open(); +149 acceptChannel.configureBlocking(false); 150 -151/** -152 * Call is done. Execution happened and we returned results to client. It is now safe to -153 * cleanup. -154 */ -155 @edu.umd.cs.findbugs.annotations.SuppressWarnings(value="IS2_INCONSISTENT_SYNC", -156justification="Presume the lock on processing request held by caller is protection enough") -157@Override -158void done() { -159 super.done(); -160 this.getConnection().decRpcCount(); // Say that we're done with this call. -161} -162 -163@Override -164public long disconnectSince() { -165 if (!getConnection().isConnectionOpen()) { -166return System.currentTimeMillis() - timestamp; -167 } else { -168return -1L; -169 } -170} -171 -172@Override -173public synchronized void sendResponseIfReady() throws IOException { -174 // set param null to reduce memory pressure -175 this.param = null; -176 this.responder.doRespond(this); +151 // Bind the server socket to the binding addrees (can be different from the default interface) +152 bind(acceptChannel.socket(), bindAddress, backlogLength); +153 port = acceptChannel.socket().getLocalPort(); //Could be an ephemeral port +154 address = (InetSocketAddress)acceptChannel.socket().getLocalSocketAddress(); +155 // create a selector; +156 selector = Selector.open(); +157 +158 readers = new Reader[readThreads]; +159 // Why this executor thing? Why not like hadoop just start up all the threads? I suppose it +160 // has an advantage in that it is easy to shutdown the pool. +161 readPool = Executors.newFixedThreadPool(readThreads, +162new ThreadFactoryBuilder().setNameFormat( +163 "RpcServer.reader=%d,bindAddress=" + bindAddress.getHostName() + +164 ",port=" + port).setDaemon(true) +165 .setUncaughtExceptionHandler(Threads.LOGGING_EXCEPTION_HANDLER).build()); +166 for (int i = 0; i < readThreads; ++i) { +167Reader reader = new Reader(); +168readers[i] = reader; +169readPool.execute(reader); +170 } +171 LOG.info(getName() + ": started " + readThreads + " reader(s) listening on port=" + port); +172 +173 // Register accepts on the server socket with the selector. +174 acceptC
[19/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.CallCleanup.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.CallCleanup.html b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.CallCleanup.html index 5cc356a..9e1c66c 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.CallCleanup.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.CallCleanup.html @@ -70,2037 +70,1559 @@ 062import org.apache.hadoop.hbase.client.VersionInfoUtil; 063import org.apache.hadoop.hbase.codec.Codec; 064import org.apache.hadoop.hbase.conf.ConfigurationObserver; -065import org.apache.hadoop.hbase.exceptions.RegionMovedException; -066import org.apache.hadoop.hbase.exceptions.RequestTooBigException; -067import org.apache.hadoop.hbase.io.ByteBufferListOutputStream; -068import org.apache.hadoop.hbase.io.ByteBufferOutputStream; -069import org.apache.hadoop.hbase.io.ByteBufferPool; -070import org.apache.hadoop.hbase.io.crypto.aes.CryptoAES; -071import org.apache.hadoop.hbase.monitoring.MonitoredRPCHandler; -072import org.apache.hadoop.hbase.monitoring.TaskMonitor; -073import org.apache.hadoop.hbase.nio.ByteBuff; -074import org.apache.hadoop.hbase.nio.MultiByteBuff; -075import org.apache.hadoop.hbase.nio.SingleByteBuff; -076import org.apache.hadoop.hbase.regionserver.RSRpcServices; -077import org.apache.hadoop.hbase.security.AccessDeniedException; -078import org.apache.hadoop.hbase.security.AuthMethod; -079import org.apache.hadoop.hbase.security.HBaseSaslRpcServer; -080import org.apache.hadoop.hbase.security.HBaseSaslRpcServer.SaslDigestCallbackHandler; -081import org.apache.hadoop.hbase.security.HBaseSaslRpcServer.SaslGssCallbackHandler; -082import org.apache.hadoop.hbase.security.SaslStatus; -083import org.apache.hadoop.hbase.security.SaslUtil; -084import org.apache.hadoop.hbase.security.User; -085import org.apache.hadoop.hbase.security.UserProvider; -086import org.apache.hadoop.hbase.security.token.AuthenticationTokenSecretManager; -087import org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService; -088import org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteInput; -089import org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString; -090import org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream; -091import org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream; -092import org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor; -093import org.apache.hadoop.hbase.shaded.com.google.protobuf.Message; -094import org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException; -095import org.apache.hadoop.hbase.shaded.com.google.protobuf.TextFormat; -096import org.apache.hadoop.hbase.shaded.com.google.protobuf.UnsafeByteOperations; -097import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil; -098import org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos; -099import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.VersionInfo; -100import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos; -101import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.CellBlockMeta; -102import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ConnectionHeader; -103import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ExceptionResponse; -104import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader; -105import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ResponseHeader; -106import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation; -107import org.apache.hadoop.hbase.util.ByteBufferUtils; -108import org.apache.hadoop.hbase.util.Bytes; -109import org.apache.hadoop.hbase.util.Pair; -110import org.apache.hadoop.io.BytesWritable; -111import org.apache.hadoop.io.Writable; -112import org.apache.hadoop.io.WritableUtils; -113import org.apache.hadoop.io.compress.CompressionCodec; -114import org.apache.hadoop.security.UserGroupInformation; -115import org.apache.hadoop.security.UserGroupInformation.AuthenticationMethod; -116import org.apache.hadoop.security.authorize.AuthorizationException; -117import org.apache.hadoop.security.authorize.PolicyProvider; -118import org.apache.hadoop.security.authorize.ProxyUsers; -119import org.apache.hadoop.security.authorize.ServiceAuthorizationManager; -120import org.apache.hadoop.security.token.SecretManager; -121import org.apache.hadoop.security.token.SecretManager.InvalidToken; -122import org.apache.hadoop.security.token.TokenIdentifier; -123import org.apache.hadoop.util.StringUtils; -124import org.apache.htrace.TraceInfo; -125import org.codehaus.jackson.map.ObjectMapper; -126 -127import com.google.common.annotations.VisibleForTesting; -128 -129/** -1
[29/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.CallWriteListener.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.CallWriteListener.html b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.CallWriteListener.html index 27e0dee..109b5f3 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.CallWriteListener.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.CallWriteListener.html @@ -67,157 +67,157 @@ 059import org.apache.hadoop.hbase.CellScanner; 060import org.apache.hadoop.hbase.HConstants; 061import org.apache.hadoop.hbase.Server; -062import org.apache.hadoop.hbase.classification.InterfaceStability; -063import org.apache.hadoop.hbase.monitoring.MonitoredRPCHandler; -064import org.apache.hadoop.hbase.nio.ByteBuff; -065import org.apache.hadoop.hbase.nio.SingleByteBuff; -066import org.apache.hadoop.hbase.security.AccessDeniedException; -067import org.apache.hadoop.hbase.security.AuthMethod; -068import org.apache.hadoop.hbase.security.HBasePolicyProvider; -069import org.apache.hadoop.hbase.security.SaslStatus; -070import org.apache.hadoop.hbase.security.SaslUtil; -071import org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService; -072import org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor; -073import org.apache.hadoop.hbase.shaded.com.google.protobuf.Message; -074import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader; -075import org.apache.hadoop.hbase.util.Bytes; -076import org.apache.hadoop.hbase.util.JVM; -077import org.apache.hadoop.hbase.util.Pair; -078import org.apache.hadoop.io.IntWritable; -079import org.apache.hadoop.security.authorize.ServiceAuthorizationManager; -080import org.apache.htrace.TraceInfo; -081 -082/** -083 * An RPC server with Netty4 implementation. -084 * -085 */ -086public class NettyRpcServer extends RpcServer { -087 -088 public static final Log LOG = LogFactory.getLog(NettyRpcServer.class); -089 -090 protected final InetSocketAddress bindAddress; -091 -092 private final CountDownLatch closed = new CountDownLatch(1); -093 private final Channel serverChannel; -094 private final ChannelGroup allChannels = new DefaultChannelGroup(GlobalEventExecutor.INSTANCE);; -095 -096 public NettyRpcServer(final Server server, final String name, -097 final Listservices, -098 final InetSocketAddress bindAddress, Configuration conf, -099 RpcScheduler scheduler) throws IOException { -100super(server, name, services, bindAddress, conf, scheduler); -101this.bindAddress = bindAddress; -102boolean useEpoll = useEpoll(conf); -103int workerCount = conf.getInt("hbase.netty.rpc.server.worker.count", -104 Runtime.getRuntime().availableProcessors() / 4); -105EventLoopGroup bossGroup = null; -106EventLoopGroup workerGroup = null; -107if (useEpoll) { -108 bossGroup = new EpollEventLoopGroup(1); -109 workerGroup = new EpollEventLoopGroup(workerCount); -110} else { -111 bossGroup = new NioEventLoopGroup(1); -112 workerGroup = new NioEventLoopGroup(workerCount); -113} -114ServerBootstrap bootstrap = new ServerBootstrap(); -115bootstrap.group(bossGroup, workerGroup); -116if (useEpoll) { -117 bootstrap.channel(EpollServerSocketChannel.class); -118} else { -119 bootstrap.channel(NioServerSocketChannel.class); -120} -121 bootstrap.childOption(ChannelOption.TCP_NODELAY, tcpNoDelay); -122 bootstrap.childOption(ChannelOption.SO_KEEPALIVE, tcpKeepAlive); -123 bootstrap.childOption(ChannelOption.ALLOCATOR, -124 PooledByteBufAllocator.DEFAULT); -125bootstrap.childHandler(new Initializer(maxRequestSize)); -126 -127try { -128 serverChannel = bootstrap.bind(this.bindAddress).sync().channel(); -129 LOG.info("NettyRpcServer bind to address=" + serverChannel.localAddress() -130 + ", hbase.netty.rpc.server.worker.count=" + workerCount -131 + ", useEpoll=" + useEpoll); -132 allChannels.add(serverChannel); -133} catch (InterruptedException e) { -134 throw new InterruptedIOException(e.getMessage()); -135} -136initReconfigurable(conf); -137this.scheduler.init(new RpcSchedulerContext(this)); -138 } -139 -140 private static boolean useEpoll(Configuration conf) { -141// Config to enable native transport. -142boolean epollEnabled = conf.getBoolean("hbase.rpc.server.nativetransport", -143true); -144// Use the faster native epoll transport mechanism on linux if enabled -145return epollEnabled && JVM.isLinux() && JVM.isAmd64(); -146 } -147 -148 @Override -149 public synchronized void start() { -150if (started) { -151 return
[41/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/org/apache/hadoop/hbase/ipc/ServerCall.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/ipc/ServerCall.html b/devapidocs/org/apache/hadoop/hbase/ipc/ServerCall.html new file mode 100644 index 000..434671a --- /dev/null +++ b/devapidocs/org/apache/hadoop/hbase/ipc/ServerCall.html @@ -0,0 +1,1477 @@ +http://www.w3.org/TR/html4/loose.dtd";> + + + + + +ServerCall (Apache HBase 2.0.0-SNAPSHOT API) + + + + + +var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":10,"i33":10,"i34":10,"i35":10,"i36":10,"i37":10,"i38":10,"i39":10}; +var tabs = {65535:["t0","All Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]}; +var altColor = "altColor"; +var rowColor = "rowColor"; +var tableTab = "tableTab"; +var activeTableTab = "activeTableTab"; + + +JavaScript is disabled on your browser. + + + + + +Skip navigation links + + + + +Overview +Package +Class +Use +Tree +Deprecated +Index +Help + + + + +Prev Class +Next Class + + +Frames +No Frames + + +All Classes + + + + + + + +Summary: +Nested | +Field | +Constr | +Method + + +Detail: +Field | +Constr | +Method + + + + + + + + +org.apache.hadoop.hbase.ipc +Class ServerCall + + + +http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true"; title="class or interface in java.lang">java.lang.Object + + +org.apache.hadoop.hbase.ipc.ServerCall + + + + + + + +All Implemented Interfaces: +RpcCall, RpcCallContext + + +Direct Known Subclasses: +NettyServerCall, SimpleServerCall + + + +@InterfaceAudience.Private +abstract class ServerCall +extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true"; title="class or interface in java.lang">Object +implements RpcCall +Datastructure that holds all necessary to a method invocation and then afterward, carries + the result. + + + + + + + + + + + +Field Summary + +Fields + +Modifier and Type +Field and Description + + +protected CellBlockBuilder +cellBlockBuilder + + +protected ByteBufferListOutputStream +cellBlockStream + + +protected CellScanner +cellScanner + + +protected RpcServer.Connection +connection + + +protected long +deadline + + +private long +exceptionSize + + +protected org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader +header + + +protected int +id + + +protected boolean +isError + + +protected org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor +md + + +protected org.apache.hadoop.hbase.shaded.com.google.protobuf.Message +param + + +protected long +receiveTime + + +protected http://docs.oracle.com/javase/8/docs/api/java/net/InetAddress.html?is-external=true"; title="class or interface in java.net">InetAddress +remoteAddress + + +protected RpcServer.CallCleanup +reqCleanup + + +protected ByteBufferPool +reservoir + + +protected BufferChain +response +Chain of buffers to send as response. + + + +private long +responseBlockSize + + +private long +responseCellSize + + +private boolean +retryImmediatelySupported + + +protected RpcCallback +rpcCallback + + +protected org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService +service + + +protected long +size + + +protected long +startTime + + +protected int +timeout + + +protected org.apache.htrace.TraceInfo +tinfo + + +protected User +user + + + + + + + + + +Constructor Summary + +Constructors + +Constructor and Description + + +ServerCall(int id, + org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService service, + org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor md, + org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader header, + org.apache.hadoop.hbase.shaded.com.google.protobuf.Message param, + CellScanner cellScanner, + RpcServer.Connection connection, + long size, + org.apache.htrace.TraceInfo tinfo, + http://docs.oracle.com/javase/8/docs/api/java/net/InetAddress.html?is-external=true"; title="class or interface in java.net">InetAddress remoteAddress, + long receiveTime, + int timeout, + ByteBufferPool reservoir, + Cell
[01/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
Repository: hbase-site Updated Branches: refs/heads/asf-site baff48117 -> 7ef4c5a9a http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/RSRpcServices.RegionScannerShippedCallBack.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/RSRpcServices.RegionScannerShippedCallBack.html b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/RSRpcServices.RegionScannerShippedCallBack.html index 17d1bcb..6dfd1d4 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/RSRpcServices.RegionScannerShippedCallBack.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/RSRpcServices.RegionScannerShippedCallBack.html @@ -35,863 +35,863 @@ 027import java.net.InetSocketAddress; 028import java.net.UnknownHostException; 029import java.nio.ByteBuffer; -030import java.util.ArrayList; -031import java.util.Collections; -032import java.util.HashMap; -033import java.util.Iterator; -034import java.util.List; -035import java.util.Map; -036import java.util.Map.Entry; -037import java.util.NavigableMap; -038import java.util.Set; -039import java.util.TreeSet; -040import java.util.concurrent.ConcurrentHashMap; -041import java.util.concurrent.ConcurrentMap; -042import java.util.concurrent.atomic.AtomicLong; -043import java.util.concurrent.atomic.LongAdder; -044 -045import org.apache.commons.lang.mutable.MutableObject; -046import org.apache.commons.logging.Log; -047import org.apache.commons.logging.LogFactory; -048import org.apache.hadoop.conf.Configuration; -049import org.apache.hadoop.fs.Path; -050import org.apache.hadoop.hbase.ByteBufferCell; -051import org.apache.hadoop.hbase.Cell; -052import org.apache.hadoop.hbase.CellScannable; -053import org.apache.hadoop.hbase.CellScanner; -054import org.apache.hadoop.hbase.CellUtil; -055import org.apache.hadoop.hbase.DoNotRetryIOException; -056import org.apache.hadoop.hbase.DroppedSnapshotException; -057import org.apache.hadoop.hbase.HBaseIOException; -058import org.apache.hadoop.hbase.HConstants; -059import org.apache.hadoop.hbase.HRegionInfo; -060import org.apache.hadoop.hbase.HTableDescriptor; -061import org.apache.hadoop.hbase.MultiActionResultTooLarge; -062import org.apache.hadoop.hbase.NotServingRegionException; -063import org.apache.hadoop.hbase.ServerName; -064import org.apache.hadoop.hbase.TableName; -065import org.apache.hadoop.hbase.UnknownScannerException; -066import org.apache.hadoop.hbase.classification.InterfaceAudience; -067import org.apache.hadoop.hbase.client.Append; -068import org.apache.hadoop.hbase.client.ConnectionUtils; -069import org.apache.hadoop.hbase.client.Delete; -070import org.apache.hadoop.hbase.client.Durability; -071import org.apache.hadoop.hbase.client.Get; -072import org.apache.hadoop.hbase.client.Increment; -073import org.apache.hadoop.hbase.client.Mutation; -074import org.apache.hadoop.hbase.client.Put; -075import org.apache.hadoop.hbase.client.RegionReplicaUtil; -076import org.apache.hadoop.hbase.client.Result; -077import org.apache.hadoop.hbase.client.RowMutations; -078import org.apache.hadoop.hbase.client.Scan; -079import org.apache.hadoop.hbase.client.VersionInfoUtil; -080import org.apache.hadoop.hbase.conf.ConfigurationObserver; -081import org.apache.hadoop.hbase.exceptions.FailedSanityCheckException; -082import org.apache.hadoop.hbase.exceptions.OutOfOrderScannerNextException; -083import org.apache.hadoop.hbase.exceptions.ScannerResetException; -084import org.apache.hadoop.hbase.filter.ByteArrayComparable; -085import org.apache.hadoop.hbase.filter.CompareFilter.CompareOp; -086import org.apache.hadoop.hbase.ipc.HBaseRPCErrorHandler; -087import org.apache.hadoop.hbase.ipc.HBaseRpcController; -088import org.apache.hadoop.hbase.ipc.PriorityFunction; -089import org.apache.hadoop.hbase.ipc.QosPriority; -090import org.apache.hadoop.hbase.ipc.RpcCallContext; -091import org.apache.hadoop.hbase.ipc.RpcCallback; -092import org.apache.hadoop.hbase.ipc.RpcServer; -093import org.apache.hadoop.hbase.ipc.RpcServer.BlockingServiceAndInterface; -094import org.apache.hadoop.hbase.ipc.RpcServerFactory; -095import org.apache.hadoop.hbase.ipc.RpcServerInterface; -096import org.apache.hadoop.hbase.ipc.ServerNotRunningYetException; -097import org.apache.hadoop.hbase.ipc.ServerRpcController; -098import org.apache.hadoop.hbase.master.MasterRpcServices; -099import org.apache.hadoop.hbase.quotas.OperationQuota; -100import org.apache.hadoop.hbase.quotas.RegionServerQuotaManager; -101import org.apache.hadoop.hbase.regionserver.HRegion.RegionScannerImpl; -102import org.apache.hadoop.hbase.regionserver.Leases.Lease; -103import org.apache.hadoop.hbase.regionserver.Leases.LeaseStillHeldException; -104import org.apache.hadoop.hbase.regionserver.Region.Operation; -105import org.apache.hadoop.hbase.regionserver.ScannerContext.Limi
[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.ConnectionHeaderHandler.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.ConnectionHeaderHandler.html b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.ConnectionHeaderHandler.html index 27e0dee..109b5f3 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.ConnectionHeaderHandler.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.ConnectionHeaderHandler.html @@ -67,157 +67,157 @@ 059import org.apache.hadoop.hbase.CellScanner; 060import org.apache.hadoop.hbase.HConstants; 061import org.apache.hadoop.hbase.Server; -062import org.apache.hadoop.hbase.classification.InterfaceStability; -063import org.apache.hadoop.hbase.monitoring.MonitoredRPCHandler; -064import org.apache.hadoop.hbase.nio.ByteBuff; -065import org.apache.hadoop.hbase.nio.SingleByteBuff; -066import org.apache.hadoop.hbase.security.AccessDeniedException; -067import org.apache.hadoop.hbase.security.AuthMethod; -068import org.apache.hadoop.hbase.security.HBasePolicyProvider; -069import org.apache.hadoop.hbase.security.SaslStatus; -070import org.apache.hadoop.hbase.security.SaslUtil; -071import org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService; -072import org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor; -073import org.apache.hadoop.hbase.shaded.com.google.protobuf.Message; -074import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader; -075import org.apache.hadoop.hbase.util.Bytes; -076import org.apache.hadoop.hbase.util.JVM; -077import org.apache.hadoop.hbase.util.Pair; -078import org.apache.hadoop.io.IntWritable; -079import org.apache.hadoop.security.authorize.ServiceAuthorizationManager; -080import org.apache.htrace.TraceInfo; -081 -082/** -083 * An RPC server with Netty4 implementation. -084 * -085 */ -086public class NettyRpcServer extends RpcServer { -087 -088 public static final Log LOG = LogFactory.getLog(NettyRpcServer.class); -089 -090 protected final InetSocketAddress bindAddress; -091 -092 private final CountDownLatch closed = new CountDownLatch(1); -093 private final Channel serverChannel; -094 private final ChannelGroup allChannels = new DefaultChannelGroup(GlobalEventExecutor.INSTANCE);; -095 -096 public NettyRpcServer(final Server server, final String name, -097 final Listservices, -098 final InetSocketAddress bindAddress, Configuration conf, -099 RpcScheduler scheduler) throws IOException { -100super(server, name, services, bindAddress, conf, scheduler); -101this.bindAddress = bindAddress; -102boolean useEpoll = useEpoll(conf); -103int workerCount = conf.getInt("hbase.netty.rpc.server.worker.count", -104 Runtime.getRuntime().availableProcessors() / 4); -105EventLoopGroup bossGroup = null; -106EventLoopGroup workerGroup = null; -107if (useEpoll) { -108 bossGroup = new EpollEventLoopGroup(1); -109 workerGroup = new EpollEventLoopGroup(workerCount); -110} else { -111 bossGroup = new NioEventLoopGroup(1); -112 workerGroup = new NioEventLoopGroup(workerCount); -113} -114ServerBootstrap bootstrap = new ServerBootstrap(); -115bootstrap.group(bossGroup, workerGroup); -116if (useEpoll) { -117 bootstrap.channel(EpollServerSocketChannel.class); -118} else { -119 bootstrap.channel(NioServerSocketChannel.class); -120} -121 bootstrap.childOption(ChannelOption.TCP_NODELAY, tcpNoDelay); -122 bootstrap.childOption(ChannelOption.SO_KEEPALIVE, tcpKeepAlive); -123 bootstrap.childOption(ChannelOption.ALLOCATOR, -124 PooledByteBufAllocator.DEFAULT); -125bootstrap.childHandler(new Initializer(maxRequestSize)); -126 -127try { -128 serverChannel = bootstrap.bind(this.bindAddress).sync().channel(); -129 LOG.info("NettyRpcServer bind to address=" + serverChannel.localAddress() -130 + ", hbase.netty.rpc.server.worker.count=" + workerCount -131 + ", useEpoll=" + useEpoll); -132 allChannels.add(serverChannel); -133} catch (InterruptedException e) { -134 throw new InterruptedIOException(e.getMessage()); -135} -136initReconfigurable(conf); -137this.scheduler.init(new RpcSchedulerContext(this)); -138 } -139 -140 private static boolean useEpoll(Configuration conf) { -141// Config to enable native transport. -142boolean epollEnabled = conf.getBoolean("hbase.rpc.server.nativetransport", -143true); -144// Use the faster native epoll transport mechanism on linux if enabled -145return epollEnabled && JVM.isLinux() && JVM.isAmd64(); -146 } -147 -148 @Override -149 public synchronized void start() { -150i
[06/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.PeriodicDoMetrics.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.PeriodicDoMetrics.html b/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.PeriodicDoMetrics.html index 66fe094..17bc96c 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.PeriodicDoMetrics.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.PeriodicDoMetrics.html @@ -57,3290 +57,3303 @@ 049import org.apache.commons.logging.Log; 050import org.apache.commons.logging.LogFactory; 051import org.apache.hadoop.conf.Configuration; -052import org.apache.hadoop.fs.Path; -053import org.apache.hadoop.hbase.ClusterStatus; -054import org.apache.hadoop.hbase.CoordinatedStateException; -055import org.apache.hadoop.hbase.CoordinatedStateManager; -056import org.apache.hadoop.hbase.DoNotRetryIOException; -057import org.apache.hadoop.hbase.HBaseIOException; -058import org.apache.hadoop.hbase.HBaseInterfaceAudience; -059import org.apache.hadoop.hbase.HColumnDescriptor; -060import org.apache.hadoop.hbase.HConstants; -061import org.apache.hadoop.hbase.HRegionInfo; -062import org.apache.hadoop.hbase.HTableDescriptor; -063import org.apache.hadoop.hbase.MasterNotRunningException; -064import org.apache.hadoop.hbase.MetaTableAccessor; -065import org.apache.hadoop.hbase.NamespaceDescriptor; -066import org.apache.hadoop.hbase.PleaseHoldException; -067import org.apache.hadoop.hbase.ProcedureInfo; -068import org.apache.hadoop.hbase.ScheduledChore; -069import org.apache.hadoop.hbase.ServerLoad; -070import org.apache.hadoop.hbase.ServerName; -071import org.apache.hadoop.hbase.TableDescriptors; -072import org.apache.hadoop.hbase.TableName; -073import org.apache.hadoop.hbase.TableNotDisabledException; -074import org.apache.hadoop.hbase.TableNotFoundException; -075import org.apache.hadoop.hbase.UnknownRegionException; -076import org.apache.hadoop.hbase.classification.InterfaceAudience; -077import org.apache.hadoop.hbase.client.MasterSwitchType; -078import org.apache.hadoop.hbase.client.Result; -079import org.apache.hadoop.hbase.client.TableState; -080import org.apache.hadoop.hbase.coprocessor.BypassCoprocessorException; -081import org.apache.hadoop.hbase.coprocessor.CoprocessorHost; -082import org.apache.hadoop.hbase.exceptions.DeserializationException; -083import org.apache.hadoop.hbase.exceptions.MergeRegionException; -084import org.apache.hadoop.hbase.executor.ExecutorType; -085import org.apache.hadoop.hbase.favored.FavoredNodesManager; -086import org.apache.hadoop.hbase.favored.FavoredNodesPromoter; -087import org.apache.hadoop.hbase.http.InfoServer; -088import org.apache.hadoop.hbase.ipc.CoprocessorRpcUtils; -089import org.apache.hadoop.hbase.ipc.RpcServer; -090import org.apache.hadoop.hbase.ipc.ServerNotRunningYetException; -091import org.apache.hadoop.hbase.master.MasterRpcServices.BalanceSwitchMode; -092import org.apache.hadoop.hbase.master.balancer.BalancerChore; -093import org.apache.hadoop.hbase.master.balancer.BaseLoadBalancer; -094import org.apache.hadoop.hbase.master.balancer.ClusterStatusChore; -095import org.apache.hadoop.hbase.master.balancer.LoadBalancerFactory; -096import org.apache.hadoop.hbase.master.cleaner.HFileCleaner; -097import org.apache.hadoop.hbase.master.cleaner.LogCleaner; -098import org.apache.hadoop.hbase.master.cleaner.ReplicationMetaCleaner; -099import org.apache.hadoop.hbase.master.cleaner.ReplicationZKNodeCleaner; -100import org.apache.hadoop.hbase.master.cleaner.ReplicationZKNodeCleanerChore; -101import org.apache.hadoop.hbase.master.locking.LockManager; -102import org.apache.hadoop.hbase.master.normalizer.NormalizationPlan; -103import org.apache.hadoop.hbase.master.normalizer.NormalizationPlan.PlanType; -104import org.apache.hadoop.hbase.master.normalizer.RegionNormalizer; -105import org.apache.hadoop.hbase.master.normalizer.RegionNormalizerChore; -106import org.apache.hadoop.hbase.master.normalizer.RegionNormalizerFactory; -107import org.apache.hadoop.hbase.master.procedure.AddColumnFamilyProcedure; -108import org.apache.hadoop.hbase.master.procedure.CreateTableProcedure; -109import org.apache.hadoop.hbase.master.procedure.DeleteColumnFamilyProcedure; -110import org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure; -111import org.apache.hadoop.hbase.master.procedure.DisableTableProcedure; -112import org.apache.hadoop.hbase.master.procedure.EnableTableProcedure; -113import org.apache.hadoop.hbase.master.procedure.MasterProcedureConstants; -114import org.apache.hadoop.hbase.master.procedure.MasterProcedureEnv; -115import org.apache.hadoop.hbase.master.procedure.MasterProcedureScheduler; -116import org.apache.hadoop.hbase.master.procedure.MasterProcedureUtil; -117import org.apache.hadoop.hbas
[13/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.ConnectionManager.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.ConnectionManager.html b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.ConnectionManager.html index 2987e7b..7f61b54 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.ConnectionManager.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.ConnectionManager.html @@ -135,1305 +135,1251 @@ 127 private Listener listener = null; 128 protected Responder responder = null; 129 -130 /** -131 * Datastructure that holds all necessary to a method invocation and then afterward, carries -132 * the result. -133 */ -134 @InterfaceStability.Evolving -135 public class Call extends RpcServer.Call { -136 -137protected Responder responder; +130 /** Listens on the socket. Creates jobs for the handler threads*/ +131 private class Listener extends Thread { +132 +133private ServerSocketChannel acceptChannel = null; //the accept channel +134private Selector selector = null; //the selector that we use for the server +135private Reader[] readers = null; +136private int currentReader = 0; +137private final int readerPendingConnectionQueueLength; 138 -139 @edu.umd.cs.findbugs.annotations.SuppressWarnings(value="NP_NULL_ON_SOME_PATH", -140justification="Can't figure why this complaint is happening... see below") -141Call(int id, final BlockingService service, final MethodDescriptor md, -142RequestHeader header, Message param, CellScanner cellScanner, -143RpcServer.Connection connection, long size, TraceInfo tinfo, -144final InetAddress remoteAddress, int timeout, CallCleanup reqCleanup, -145Responder responder) { -146 super(id, service, md, header, param, cellScanner, connection, size, -147 tinfo, remoteAddress, timeout, reqCleanup); -148 this.responder = responder; -149} +139private ExecutorService readPool; +140 +141public Listener(final String name) throws IOException { +142 super(name); +143 // The backlog of requests that we will have the serversocket carry. +144 int backlogLength = conf.getInt("hbase.ipc.server.listen.queue.size", 128); +145 readerPendingConnectionQueueLength = +146 conf.getInt("hbase.ipc.server.read.connection-queue.size", 100); +147 // Create a new server socket and set to non blocking mode +148 acceptChannel = ServerSocketChannel.open(); +149 acceptChannel.configureBlocking(false); 150 -151/** -152 * Call is done. Execution happened and we returned results to client. It is now safe to -153 * cleanup. -154 */ -155 @edu.umd.cs.findbugs.annotations.SuppressWarnings(value="IS2_INCONSISTENT_SYNC", -156justification="Presume the lock on processing request held by caller is protection enough") -157@Override -158void done() { -159 super.done(); -160 this.getConnection().decRpcCount(); // Say that we're done with this call. -161} -162 -163@Override -164public long disconnectSince() { -165 if (!getConnection().isConnectionOpen()) { -166return System.currentTimeMillis() - timestamp; -167 } else { -168return -1L; -169 } -170} -171 -172@Override -173public synchronized void sendResponseIfReady() throws IOException { -174 // set param null to reduce memory pressure -175 this.param = null; -176 this.responder.doRespond(this); +151 // Bind the server socket to the binding addrees (can be different from the default interface) +152 bind(acceptChannel.socket(), bindAddress, backlogLength); +153 port = acceptChannel.socket().getLocalPort(); //Could be an ephemeral port +154 address = (InetSocketAddress)acceptChannel.socket().getLocalSocketAddress(); +155 // create a selector; +156 selector = Selector.open(); +157 +158 readers = new Reader[readThreads]; +159 // Why this executor thing? Why not like hadoop just start up all the threads? I suppose it +160 // has an advantage in that it is easy to shutdown the pool. +161 readPool = Executors.newFixedThreadPool(readThreads, +162new ThreadFactoryBuilder().setNameFormat( +163 "RpcServer.reader=%d,bindAddress=" + bindAddress.getHostName() + +164 ",port=" + port).setDaemon(true) +165 .setUncaughtExceptionHandler(Threads.LOGGING_EXCEPTION_HANDLER).build()); +166 for (int i = 0; i < readThreads; ++i) { +167Reader reader = new Reader(); +168readers[i] = reader; +169readPool.execute(reader); +170 } +171 LOG.info(getName() + ": started " + readThreads + " reader(s) listening on port=" + port);
[04/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.html b/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.html index 66fe094..17bc96c 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.html @@ -57,3290 +57,3303 @@ 049import org.apache.commons.logging.Log; 050import org.apache.commons.logging.LogFactory; 051import org.apache.hadoop.conf.Configuration; -052import org.apache.hadoop.fs.Path; -053import org.apache.hadoop.hbase.ClusterStatus; -054import org.apache.hadoop.hbase.CoordinatedStateException; -055import org.apache.hadoop.hbase.CoordinatedStateManager; -056import org.apache.hadoop.hbase.DoNotRetryIOException; -057import org.apache.hadoop.hbase.HBaseIOException; -058import org.apache.hadoop.hbase.HBaseInterfaceAudience; -059import org.apache.hadoop.hbase.HColumnDescriptor; -060import org.apache.hadoop.hbase.HConstants; -061import org.apache.hadoop.hbase.HRegionInfo; -062import org.apache.hadoop.hbase.HTableDescriptor; -063import org.apache.hadoop.hbase.MasterNotRunningException; -064import org.apache.hadoop.hbase.MetaTableAccessor; -065import org.apache.hadoop.hbase.NamespaceDescriptor; -066import org.apache.hadoop.hbase.PleaseHoldException; -067import org.apache.hadoop.hbase.ProcedureInfo; -068import org.apache.hadoop.hbase.ScheduledChore; -069import org.apache.hadoop.hbase.ServerLoad; -070import org.apache.hadoop.hbase.ServerName; -071import org.apache.hadoop.hbase.TableDescriptors; -072import org.apache.hadoop.hbase.TableName; -073import org.apache.hadoop.hbase.TableNotDisabledException; -074import org.apache.hadoop.hbase.TableNotFoundException; -075import org.apache.hadoop.hbase.UnknownRegionException; -076import org.apache.hadoop.hbase.classification.InterfaceAudience; -077import org.apache.hadoop.hbase.client.MasterSwitchType; -078import org.apache.hadoop.hbase.client.Result; -079import org.apache.hadoop.hbase.client.TableState; -080import org.apache.hadoop.hbase.coprocessor.BypassCoprocessorException; -081import org.apache.hadoop.hbase.coprocessor.CoprocessorHost; -082import org.apache.hadoop.hbase.exceptions.DeserializationException; -083import org.apache.hadoop.hbase.exceptions.MergeRegionException; -084import org.apache.hadoop.hbase.executor.ExecutorType; -085import org.apache.hadoop.hbase.favored.FavoredNodesManager; -086import org.apache.hadoop.hbase.favored.FavoredNodesPromoter; -087import org.apache.hadoop.hbase.http.InfoServer; -088import org.apache.hadoop.hbase.ipc.CoprocessorRpcUtils; -089import org.apache.hadoop.hbase.ipc.RpcServer; -090import org.apache.hadoop.hbase.ipc.ServerNotRunningYetException; -091import org.apache.hadoop.hbase.master.MasterRpcServices.BalanceSwitchMode; -092import org.apache.hadoop.hbase.master.balancer.BalancerChore; -093import org.apache.hadoop.hbase.master.balancer.BaseLoadBalancer; -094import org.apache.hadoop.hbase.master.balancer.ClusterStatusChore; -095import org.apache.hadoop.hbase.master.balancer.LoadBalancerFactory; -096import org.apache.hadoop.hbase.master.cleaner.HFileCleaner; -097import org.apache.hadoop.hbase.master.cleaner.LogCleaner; -098import org.apache.hadoop.hbase.master.cleaner.ReplicationMetaCleaner; -099import org.apache.hadoop.hbase.master.cleaner.ReplicationZKNodeCleaner; -100import org.apache.hadoop.hbase.master.cleaner.ReplicationZKNodeCleanerChore; -101import org.apache.hadoop.hbase.master.locking.LockManager; -102import org.apache.hadoop.hbase.master.normalizer.NormalizationPlan; -103import org.apache.hadoop.hbase.master.normalizer.NormalizationPlan.PlanType; -104import org.apache.hadoop.hbase.master.normalizer.RegionNormalizer; -105import org.apache.hadoop.hbase.master.normalizer.RegionNormalizerChore; -106import org.apache.hadoop.hbase.master.normalizer.RegionNormalizerFactory; -107import org.apache.hadoop.hbase.master.procedure.AddColumnFamilyProcedure; -108import org.apache.hadoop.hbase.master.procedure.CreateTableProcedure; -109import org.apache.hadoop.hbase.master.procedure.DeleteColumnFamilyProcedure; -110import org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure; -111import org.apache.hadoop.hbase.master.procedure.DisableTableProcedure; -112import org.apache.hadoop.hbase.master.procedure.EnableTableProcedure; -113import org.apache.hadoop.hbase.master.procedure.MasterProcedureConstants; -114import org.apache.hadoop.hbase.master.procedure.MasterProcedureEnv; -115import org.apache.hadoop.hbase.master.procedure.MasterProcedureScheduler; -116import org.apache.hadoop.hbase.master.procedure.MasterProcedureUtil; -117import org.apache.hadoop.hbase.master.procedure.MergeTableRegionsProcedure; -118import org.apache.hadoop.hbase.master.proc
[22/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.BlockingServiceAndInterface.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.BlockingServiceAndInterface.html b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.BlockingServiceAndInterface.html index 5cc356a..9e1c66c 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.BlockingServiceAndInterface.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.BlockingServiceAndInterface.html @@ -70,2037 +70,1559 @@ 062import org.apache.hadoop.hbase.client.VersionInfoUtil; 063import org.apache.hadoop.hbase.codec.Codec; 064import org.apache.hadoop.hbase.conf.ConfigurationObserver; -065import org.apache.hadoop.hbase.exceptions.RegionMovedException; -066import org.apache.hadoop.hbase.exceptions.RequestTooBigException; -067import org.apache.hadoop.hbase.io.ByteBufferListOutputStream; -068import org.apache.hadoop.hbase.io.ByteBufferOutputStream; -069import org.apache.hadoop.hbase.io.ByteBufferPool; -070import org.apache.hadoop.hbase.io.crypto.aes.CryptoAES; -071import org.apache.hadoop.hbase.monitoring.MonitoredRPCHandler; -072import org.apache.hadoop.hbase.monitoring.TaskMonitor; -073import org.apache.hadoop.hbase.nio.ByteBuff; -074import org.apache.hadoop.hbase.nio.MultiByteBuff; -075import org.apache.hadoop.hbase.nio.SingleByteBuff; -076import org.apache.hadoop.hbase.regionserver.RSRpcServices; -077import org.apache.hadoop.hbase.security.AccessDeniedException; -078import org.apache.hadoop.hbase.security.AuthMethod; -079import org.apache.hadoop.hbase.security.HBaseSaslRpcServer; -080import org.apache.hadoop.hbase.security.HBaseSaslRpcServer.SaslDigestCallbackHandler; -081import org.apache.hadoop.hbase.security.HBaseSaslRpcServer.SaslGssCallbackHandler; -082import org.apache.hadoop.hbase.security.SaslStatus; -083import org.apache.hadoop.hbase.security.SaslUtil; -084import org.apache.hadoop.hbase.security.User; -085import org.apache.hadoop.hbase.security.UserProvider; -086import org.apache.hadoop.hbase.security.token.AuthenticationTokenSecretManager; -087import org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService; -088import org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteInput; -089import org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString; -090import org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream; -091import org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream; -092import org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor; -093import org.apache.hadoop.hbase.shaded.com.google.protobuf.Message; -094import org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException; -095import org.apache.hadoop.hbase.shaded.com.google.protobuf.TextFormat; -096import org.apache.hadoop.hbase.shaded.com.google.protobuf.UnsafeByteOperations; -097import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil; -098import org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos; -099import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.VersionInfo; -100import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos; -101import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.CellBlockMeta; -102import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ConnectionHeader; -103import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ExceptionResponse; -104import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader; -105import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ResponseHeader; -106import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation; -107import org.apache.hadoop.hbase.util.ByteBufferUtils; -108import org.apache.hadoop.hbase.util.Bytes; -109import org.apache.hadoop.hbase.util.Pair; -110import org.apache.hadoop.io.BytesWritable; -111import org.apache.hadoop.io.Writable; -112import org.apache.hadoop.io.WritableUtils; -113import org.apache.hadoop.io.compress.CompressionCodec; -114import org.apache.hadoop.security.UserGroupInformation; -115import org.apache.hadoop.security.UserGroupInformation.AuthenticationMethod; -116import org.apache.hadoop.security.authorize.AuthorizationException; -117import org.apache.hadoop.security.authorize.PolicyProvider; -118import org.apache.hadoop.security.authorize.ProxyUsers; -119import org.apache.hadoop.security.authorize.ServiceAuthorizationManager; -120import org.apache.hadoop.security.token.SecretManager; -121import org.apache.hadoop.security.token.SecretManager.InvalidToken; -122import org.apache.hadoop.security.token.TokenIdentifier; -123import org.apache.hadoop.util.StringUtils; -124import org.apache.htrace.TraceInfo; -125import org.codehaus.jackson.map.ObjectMapper
[07/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.InitializationMonitor.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.InitializationMonitor.html b/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.InitializationMonitor.html index 66fe094..17bc96c 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.InitializationMonitor.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.InitializationMonitor.html @@ -57,3290 +57,3303 @@ 049import org.apache.commons.logging.Log; 050import org.apache.commons.logging.LogFactory; 051import org.apache.hadoop.conf.Configuration; -052import org.apache.hadoop.fs.Path; -053import org.apache.hadoop.hbase.ClusterStatus; -054import org.apache.hadoop.hbase.CoordinatedStateException; -055import org.apache.hadoop.hbase.CoordinatedStateManager; -056import org.apache.hadoop.hbase.DoNotRetryIOException; -057import org.apache.hadoop.hbase.HBaseIOException; -058import org.apache.hadoop.hbase.HBaseInterfaceAudience; -059import org.apache.hadoop.hbase.HColumnDescriptor; -060import org.apache.hadoop.hbase.HConstants; -061import org.apache.hadoop.hbase.HRegionInfo; -062import org.apache.hadoop.hbase.HTableDescriptor; -063import org.apache.hadoop.hbase.MasterNotRunningException; -064import org.apache.hadoop.hbase.MetaTableAccessor; -065import org.apache.hadoop.hbase.NamespaceDescriptor; -066import org.apache.hadoop.hbase.PleaseHoldException; -067import org.apache.hadoop.hbase.ProcedureInfo; -068import org.apache.hadoop.hbase.ScheduledChore; -069import org.apache.hadoop.hbase.ServerLoad; -070import org.apache.hadoop.hbase.ServerName; -071import org.apache.hadoop.hbase.TableDescriptors; -072import org.apache.hadoop.hbase.TableName; -073import org.apache.hadoop.hbase.TableNotDisabledException; -074import org.apache.hadoop.hbase.TableNotFoundException; -075import org.apache.hadoop.hbase.UnknownRegionException; -076import org.apache.hadoop.hbase.classification.InterfaceAudience; -077import org.apache.hadoop.hbase.client.MasterSwitchType; -078import org.apache.hadoop.hbase.client.Result; -079import org.apache.hadoop.hbase.client.TableState; -080import org.apache.hadoop.hbase.coprocessor.BypassCoprocessorException; -081import org.apache.hadoop.hbase.coprocessor.CoprocessorHost; -082import org.apache.hadoop.hbase.exceptions.DeserializationException; -083import org.apache.hadoop.hbase.exceptions.MergeRegionException; -084import org.apache.hadoop.hbase.executor.ExecutorType; -085import org.apache.hadoop.hbase.favored.FavoredNodesManager; -086import org.apache.hadoop.hbase.favored.FavoredNodesPromoter; -087import org.apache.hadoop.hbase.http.InfoServer; -088import org.apache.hadoop.hbase.ipc.CoprocessorRpcUtils; -089import org.apache.hadoop.hbase.ipc.RpcServer; -090import org.apache.hadoop.hbase.ipc.ServerNotRunningYetException; -091import org.apache.hadoop.hbase.master.MasterRpcServices.BalanceSwitchMode; -092import org.apache.hadoop.hbase.master.balancer.BalancerChore; -093import org.apache.hadoop.hbase.master.balancer.BaseLoadBalancer; -094import org.apache.hadoop.hbase.master.balancer.ClusterStatusChore; -095import org.apache.hadoop.hbase.master.balancer.LoadBalancerFactory; -096import org.apache.hadoop.hbase.master.cleaner.HFileCleaner; -097import org.apache.hadoop.hbase.master.cleaner.LogCleaner; -098import org.apache.hadoop.hbase.master.cleaner.ReplicationMetaCleaner; -099import org.apache.hadoop.hbase.master.cleaner.ReplicationZKNodeCleaner; -100import org.apache.hadoop.hbase.master.cleaner.ReplicationZKNodeCleanerChore; -101import org.apache.hadoop.hbase.master.locking.LockManager; -102import org.apache.hadoop.hbase.master.normalizer.NormalizationPlan; -103import org.apache.hadoop.hbase.master.normalizer.NormalizationPlan.PlanType; -104import org.apache.hadoop.hbase.master.normalizer.RegionNormalizer; -105import org.apache.hadoop.hbase.master.normalizer.RegionNormalizerChore; -106import org.apache.hadoop.hbase.master.normalizer.RegionNormalizerFactory; -107import org.apache.hadoop.hbase.master.procedure.AddColumnFamilyProcedure; -108import org.apache.hadoop.hbase.master.procedure.CreateTableProcedure; -109import org.apache.hadoop.hbase.master.procedure.DeleteColumnFamilyProcedure; -110import org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure; -111import org.apache.hadoop.hbase.master.procedure.DisableTableProcedure; -112import org.apache.hadoop.hbase.master.procedure.EnableTableProcedure; -113import org.apache.hadoop.hbase.master.procedure.MasterProcedureConstants; -114import org.apache.hadoop.hbase.master.procedure.MasterProcedureEnv; -115import org.apache.hadoop.hbase.master.procedure.MasterProcedureScheduler; -116import org.apache.hadoop.hbase.master.procedure.MasterProcedureUtil; -117import or
hbase git commit: HBASE-16356 REST API scanner: row prefix filter and custom filter parameters are mutually exclusive (Ben Watson)
Repository: hbase Updated Branches: refs/heads/master 51d4c68b7 -> ac1024af2 HBASE-16356 REST API scanner: row prefix filter and custom filter parameters are mutually exclusive (Ben Watson) Project: http://git-wip-us.apache.org/repos/asf/hbase/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/ac1024af Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/ac1024af Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/ac1024af Branch: refs/heads/master Commit: ac1024af213158d6528ffec964f2bf4aadd9ccd3 Parents: 51d4c68 Author: tedyu Authored: Tue May 9 07:58:20 2017 -0700 Committer: tedyu Committed: Tue May 9 07:58:20 2017 -0700 -- .../apache/hadoop/hbase/filter/FilterList.java | 2 +- .../apache/hadoop/hbase/rest/TableResource.java | 31 +- .../apache/hadoop/hbase/rest/TestTableScan.java | 33 ++-- 3 files changed, 39 insertions(+), 27 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hbase/blob/ac1024af/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FilterList.java -- diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FilterList.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FilterList.java index d533026..0742b22 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FilterList.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FilterList.java @@ -148,7 +148,7 @@ final public class FilterList extends FilterBase { return filters; } - private int size() { + public int size() { return filters.size(); } http://git-wip-us.apache.org/repos/asf/hbase/blob/ac1024af/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/TableResource.java -- diff --git a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/TableResource.java b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/TableResource.java index 9eb21ec..3019e40 100644 --- a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/TableResource.java +++ b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/TableResource.java @@ -135,14 +135,14 @@ public class TableResource extends ResourceBase { @DefaultValue(Long.MAX_VALUE + "") @QueryParam(Constants.SCAN_END_TIME) long endTime, @DefaultValue("true") @QueryParam(Constants.SCAN_BATCH_SIZE) boolean cacheBlocks, @DefaultValue("false") @QueryParam(Constants.SCAN_REVERSED) boolean reversed, - @DefaultValue("") @QueryParam(Constants.SCAN_FILTER) String filters) { + @DefaultValue("") @QueryParam(Constants.SCAN_FILTER) String paramFilter) { try { - Filter filter = null; + Filter prefixFilter = null; Scan tableScan = new Scan(); if (scanSpec.indexOf('*') > 0) { String prefix = scanSpec.substring(0, scanSpec.indexOf('*')); byte[] prefixBytes = Bytes.toBytes(prefix); -filter = new PrefixFilter(Bytes.toBytes(prefix)); +prefixFilter = new PrefixFilter(Bytes.toBytes(prefix)); if (startRow.isEmpty()) { tableScan.setStartRow(prefixBytes); } @@ -183,22 +183,21 @@ public class TableResource extends ResourceBase { tableScan.addFamily(Bytes.toBytes(familysplit[0])); } } - FilterList filterList = null; - if (StringUtils.isNotEmpty(filters)) { - ParseFilter pf = new ParseFilter(); - Filter filterParam = pf.parseFilterString(filters); - if (filter != null) { -filterList = new FilterList(filter, filterParam); - } - else { -filter = filterParam; - } + FilterList filterList = new FilterList(); + if (StringUtils.isNotEmpty(paramFilter)) { +ParseFilter pf = new ParseFilter(); +Filter parsedParamFilter = pf.parseFilterString(paramFilter); +if (parsedParamFilter != null) { + filterList.addFilter(parsedParamFilter); +} +if (prefixFilter != null) { + filterList.addFilter(prefixFilter); +} } - if (filterList != null) { + if (filterList.size() > 0) { tableScan.setFilter(filterList); - } else if (filter != null) { -tableScan.setFilter(filter); } + int fetchSize = this.servlet.getConfiguration().getInt(Constants.SCAN_FETCH_SIZE, 10); tableScan.setCaching(fetchSize); tableScan.setReversed(reversed); http://git-wip-us.apache.org/repos/asf/hbase/blob/ac1024af/hbase-rest/src/test/java/org/apache/hadoop/hbase/rest/TestTableScan.java -- diff --git a/hbase-rest/src/test/java/org/apache/hadoop/hbase/rest/TestTableScan.java b/hbase-rest/src/test/java/org/apac
hbase git commit: HBASE-18009 Move RpcServer.Call to a separated file
Repository: hbase Updated Branches: refs/heads/master 959deb0e5 -> 51d4c68b7 HBASE-18009 Move RpcServer.Call to a separated file Project: http://git-wip-us.apache.org/repos/asf/hbase/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/51d4c68b Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/51d4c68b Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/51d4c68b Branch: refs/heads/master Commit: 51d4c68b7cce43af1190f9195bfb08963375bc27 Parents: 959deb0 Author: zhangduo Authored: Mon May 8 20:36:33 2017 +0800 Committer: zhangduo Committed: Tue May 9 14:56:03 2017 +0800 -- .../org/apache/hadoop/hbase/ipc/CallRunner.java | 4 +- .../apache/hadoop/hbase/ipc/NettyRpcServer.java | 91 +--- .../hadoop/hbase/ipc/NettyServerCall.java | 67 +++ .../org/apache/hadoop/hbase/ipc/RpcCall.java| 5 - .../org/apache/hadoop/hbase/ipc/RpcServer.java | 498 +- .../org/apache/hadoop/hbase/ipc/ServerCall.java | 527 +++ .../hadoop/hbase/ipc/SimpleRpcServer.java | 122 ++--- .../hadoop/hbase/ipc/SimpleServerCall.java | 79 +++ .../apache/hadoop/hbase/ipc/TestCallRunner.java | 3 +- .../hbase/ipc/TestSimpleRpcScheduler.java | 60 +-- .../hadoop/hbase/security/TestSecureIPC.java| 8 +- 11 files changed, 774 insertions(+), 690 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hbase/blob/51d4c68b/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/CallRunner.java -- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/CallRunner.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/CallRunner.java index 0aabc10..f16fc50 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/CallRunner.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/CallRunner.java @@ -75,8 +75,8 @@ public class CallRunner { * @deprecated As of release 2.0, this will be removed in HBase 3.0 */ @Deprecated - public RpcServer.Call getCall() { -return (RpcServer.Call) call; + public ServerCall getCall() { +return (ServerCall) call; } public void setStatus(MonitoredRPCHandler status) { http://git-wip-us.apache.org/repos/asf/hbase/blob/51d4c68b/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/NettyRpcServer.java -- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/NettyRpcServer.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/NettyRpcServer.java index be55378..c18b894 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/NettyRpcServer.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/NettyRpcServer.java @@ -59,7 +59,6 @@ import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.CellScanner; import org.apache.hadoop.hbase.HConstants; import org.apache.hadoop.hbase.Server; -import org.apache.hadoop.hbase.classification.InterfaceStability; import org.apache.hadoop.hbase.monitoring.MonitoredRPCHandler; import org.apache.hadoop.hbase.nio.ByteBuff; import org.apache.hadoop.hbase.nio.SingleByteBuff; @@ -203,13 +202,14 @@ public class NettyRpcServer extends RpcServer { this.hostAddress = inetSocketAddress.getAddress().getHostAddress(); } this.remotePort = inetSocketAddress.getPort(); - this.saslCall = new Call(SASL_CALLID, null, null, null, null, null, this, - 0, null, null, 0, null); - this.setConnectionHeaderResponseCall = new Call( - CONNECTION_HEADER_RESPONSE_CALLID, null, null, null, null, null, - this, 0, null, null, 0, null); - this.authFailedCall = new Call(AUTHORIZATION_FAILED_CALLID, null, null, - null, null, null, this, 0, null, null, 0, null); + this.saslCall = new NettyServerCall(SASL_CALLID, null, null, null, null, null, this, 0, null, + null, System.currentTimeMillis(), 0, reservoir, cellBlockBuilder, null); + this.setConnectionHeaderResponseCall = + new NettyServerCall(CONNECTION_HEADER_RESPONSE_CALLID, null, null, null, null, null, this, + 0, null, null, System.currentTimeMillis(), 0, reservoir, cellBlockBuilder, null); + this.authFailedCall = + new NettyServerCall(AUTHORIZATION_FAILED_CALLID, null, null, null, null, null, this, 0, + null, null, System.currentTimeMillis(), 0, reservoir, cellBlockBuilder, null); } void readPreamble(ByteBuf buffer) throws IOException { @@ -243,7 +243,7 @@ public class NettyRpcServer extends RpcServer { AccessDeniedException ae = new AccessDeniedException( "Authentication is required"); setupResponse(authFailedResponse, authFailedCall, ae, ae.getMessage()); - ((Call) a