hbase git commit: HBASE-20459 Majority of scan CPU time in HBase-1 spent in size estimation

2018-04-21 Thread stack
Repository: hbase
Updated Branches:
  refs/heads/branch-2.0 51069e6e8 -> 9d0a2e850


HBASE-20459 Majority of scan CPU time in HBase-1 spent in size estimation


Project: http://git-wip-us.apache.org/repos/asf/hbase/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/9d0a2e85
Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/9d0a2e85
Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/9d0a2e85

Branch: refs/heads/branch-2.0
Commit: 9d0a2e850ea0f685435009894adc8f6250342169
Parents: 51069e6
Author: Lars Hofhansl 
Authored: Sat Apr 21 08:09:18 2018 -0700
Committer: Michael Stack 
Committed: Sat Apr 21 20:06:51 2018 -0700

--
 .../src/main/java/org/apache/hadoop/hbase/util/ClassSize.java| 4 ++--
 1 file changed, 2 insertions(+), 2 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/hbase/blob/9d0a2e85/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ClassSize.java
--
diff --git 
a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ClassSize.java 
b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ClassSize.java
index 82f6e60..6f88c00 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ClassSize.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ClassSize.java
@@ -172,7 +172,7 @@ public class ClassSize {
 }
 
 long sizeOfByteArray(int len) {
-  return align(arrayHeaderSize() + len);
+  return align(ARRAY + len);
 }
   }
 
@@ -217,7 +217,7 @@ public class ClassSize {
 @Override
 @SuppressWarnings("static-access")
 long sizeOfByteArray(int len) {
-  return align(arrayHeaderSize() + len * 
UnsafeAccess.theUnsafe.ARRAY_BYTE_INDEX_SCALE);
+  return align(ARRAY + len * 
UnsafeAccess.theUnsafe.ARRAY_BYTE_INDEX_SCALE);
 }
   }
 



hbase git commit: HBASE-20459 Majority of scan CPU time in HBase-1 spent in size estimation

2018-04-21 Thread stack
Repository: hbase
Updated Branches:
  refs/heads/branch-2 cb1aaa683 -> 42d2445cf


HBASE-20459 Majority of scan CPU time in HBase-1 spent in size estimation


Project: http://git-wip-us.apache.org/repos/asf/hbase/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/42d2445c
Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/42d2445c
Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/42d2445c

Branch: refs/heads/branch-2
Commit: 42d2445cf527f717d6f6af17ac56c77dc8d44d05
Parents: cb1aaa6
Author: Lars Hofhansl 
Authored: Sat Apr 21 08:09:18 2018 -0700
Committer: Michael Stack 
Committed: Sat Apr 21 20:07:22 2018 -0700

--
 .../src/main/java/org/apache/hadoop/hbase/util/ClassSize.java| 4 ++--
 1 file changed, 2 insertions(+), 2 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/hbase/blob/42d2445c/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ClassSize.java
--
diff --git 
a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ClassSize.java 
b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ClassSize.java
index 82f6e60..6f88c00 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ClassSize.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ClassSize.java
@@ -172,7 +172,7 @@ public class ClassSize {
 }
 
 long sizeOfByteArray(int len) {
-  return align(arrayHeaderSize() + len);
+  return align(ARRAY + len);
 }
   }
 
@@ -217,7 +217,7 @@ public class ClassSize {
 @Override
 @SuppressWarnings("static-access")
 long sizeOfByteArray(int len) {
-  return align(arrayHeaderSize() + len * 
UnsafeAccess.theUnsafe.ARRAY_BYTE_INDEX_SCALE);
+  return align(ARRAY + len * 
UnsafeAccess.theUnsafe.ARRAY_BYTE_INDEX_SCALE);
 }
   }
 



[hbase] Git Push Summary

2018-04-21 Thread busbey
Repository: hbase
Updated Branches:
  refs/heads/HBASE-20388 [deleted] 6e4efa3b1


hbase git commit: HBASE-20388 nightly tests running on a feature branch should only comment on that feature branch's jira

2018-04-21 Thread busbey
Repository: hbase
Updated Branches:
  refs/heads/master 00fca5a4c -> f4a39043e


HBASE-20388 nightly tests running on a feature branch should only comment on 
that feature branch's jira

Signed-off-by: Mike Drob 


Project: http://git-wip-us.apache.org/repos/asf/hbase/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/f4a39043
Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/f4a39043
Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/f4a39043

Branch: refs/heads/master
Commit: f4a39043e2fb3da69e378b4e1cbcb4bd7e2a6662
Parents: 00fca5a
Author: Sean Busbey 
Authored: Thu Apr 12 21:10:53 2018 -0500
Committer: Sean Busbey 
Committed: Sat Apr 21 13:50:40 2018 -0500

--
 dev-support/Jenkinsfile | 35 +++
 1 file changed, 23 insertions(+), 12 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/hbase/blob/f4a39043/dev-support/Jenkinsfile
--
diff --git a/dev-support/Jenkinsfile b/dev-support/Jenkinsfile
index 6e320ca..821d20e 100644
--- a/dev-support/Jenkinsfile
+++ b/dev-support/Jenkinsfile
@@ -528,8 +528,14 @@ curl -L  -o personality.sh "${env.PROJECT_PERSONALITY}"
echo "[INFO] Comment:"
echo comment
echo ""
-   echo "[INFO] There are ${currentBuild.changeSets.size()} change 
sets."
-   getJirasToComment(currentBuild).each { currentIssue ->
+   echo "[DEBUG] checking to see if feature branch"
+   def jiras = getJirasToComment(env.BRANCH_NAME, [])
+   if (jiras.isEmpty()) {
+ echo "[DEBUG] non-feature branch, checking change messages for 
jira keys."
+ echo "[INFO] There are ${currentBuild.changeSets.size()} change 
sets."
+ jiras = getJirasToCommentFromChangesets(currentBuild)
+   }
+   jiras.each { currentIssue ->
  jiraComment issueKey: currentIssue, body: comment
}
 } catch (Exception exception) {
@@ -542,7 +548,7 @@ curl -L  -o personality.sh "${env.PROJECT_PERSONALITY}"
 }
 import org.jenkinsci.plugins.workflow.support.steps.build.RunWrapper
 @NonCPS
-List getJirasToComment(RunWrapper thisBuild) {
+List getJirasToCommentFromChangesets(RunWrapper thisBuild) {
   def seenJiras = []
   thisBuild.changeSets.each { cs ->
 cs.getItems().each { change ->
@@ -552,16 +558,21 @@ List getJirasToComment(RunWrapper thisBuild) {
   echo " ${change.commitId}"
   echo " ${change.author}"
   echo ""
-  msg.eachMatch("HBASE-[0-9]+") { currentIssue ->
-echo "[DEBUG] found jira key: ${currentIssue}"
-if (currentIssue in seenJiras) {
-  echo "[DEBUG] already commented on ${currentIssue}."
-} else {
-  echo "[INFO] commenting on ${currentIssue}."
-  seenJiras << currentIssue
-}
-  }
+  seenJiras = getJirasToComment(msg, seenJiras)
 }
   }
   return seenJiras
 }
+@NonCPS
+List getJirasToComment(CharSequence source, List seen) {
+  source.eachMatch("HBASE-[0-9]+") { currentIssue ->
+echo "[DEBUG] found jira key: ${currentIssue}"
+if (currentIssue in seen) {
+  echo "[DEBUG] already commented on ${currentIssue}."
+} else {
+  echo "[INFO] commenting on ${currentIssue}."
+  seen << currentIssue
+}
+  }
+  return seen
+}



hbase git commit: HBASE-20459 Majority of scan CPU time in HBase-1 spent in size estimation.

2018-04-21 Thread larsh
Repository: hbase
Updated Branches:
  refs/heads/master 46cb5dfa2 -> 00fca5a4c


HBASE-20459 Majority of scan CPU time in HBase-1 spent in size estimation.


Project: http://git-wip-us.apache.org/repos/asf/hbase/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/00fca5a4
Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/00fca5a4
Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/00fca5a4

Branch: refs/heads/master
Commit: 00fca5a4c9bdc73b4e0c962af0cbfcc7db2029fc
Parents: 46cb5df
Author: Lars Hofhansl 
Authored: Sat Apr 21 10:12:24 2018 -0700
Committer: Lars Hofhansl 
Committed: Sat Apr 21 10:12:24 2018 -0700

--
 .../src/main/java/org/apache/hadoop/hbase/util/ClassSize.java| 4 ++--
 1 file changed, 2 insertions(+), 2 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/hbase/blob/00fca5a4/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ClassSize.java
--
diff --git 
a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ClassSize.java 
b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ClassSize.java
index 82f6e60..6f88c00 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ClassSize.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ClassSize.java
@@ -172,7 +172,7 @@ public class ClassSize {
 }
 
 long sizeOfByteArray(int len) {
-  return align(arrayHeaderSize() + len);
+  return align(ARRAY + len);
 }
   }
 
@@ -217,7 +217,7 @@ public class ClassSize {
 @Override
 @SuppressWarnings("static-access")
 long sizeOfByteArray(int len) {
-  return align(arrayHeaderSize() + len * 
UnsafeAccess.theUnsafe.ARRAY_BYTE_INDEX_SCALE);
+  return align(ARRAY + len * 
UnsafeAccess.theUnsafe.ARRAY_BYTE_INDEX_SCALE);
 }
   }
 



hbase git commit: HBASE-20459 Majority of scan CPU time in HBase-1 spent in size estimation.

2018-04-21 Thread larsh
Repository: hbase
Updated Branches:
  refs/heads/branch-1.4 1be417a99 -> e30023071


HBASE-20459 Majority of scan CPU time in HBase-1 spent in size estimation.


Project: http://git-wip-us.apache.org/repos/asf/hbase/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/e3002307
Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/e3002307
Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/e3002307

Branch: refs/heads/branch-1.4
Commit: e30023071a012d8332925b65a508392cbf738df7
Parents: 1be417a
Author: Lars Hofhansl 
Authored: Sat Apr 21 10:04:15 2018 -0700
Committer: Lars Hofhansl 
Committed: Sat Apr 21 10:04:15 2018 -0700

--
 .../src/main/java/org/apache/hadoop/hbase/util/ClassSize.java| 4 ++--
 1 file changed, 2 insertions(+), 2 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/hbase/blob/e3002307/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ClassSize.java
--
diff --git 
a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ClassSize.java 
b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ClassSize.java
index edb5b2d..5e6627f 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ClassSize.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ClassSize.java
@@ -141,7 +141,7 @@ public class ClassSize {
 }
 
 long sizeOf(byte[] b, int len) {
-  return align(arrayHeaderSize() + len);
+  return align(ARRAY + len);
 }
   }
 
@@ -186,7 +186,7 @@ public class ClassSize {
 @Override
 @SuppressWarnings("static-access")
 long sizeOf(byte[] b, int len) {
-  return align(arrayHeaderSize() + len * 
UnsafeAccess.theUnsafe.ARRAY_BYTE_INDEX_SCALE);
+  return align(ARRAY + len * 
UnsafeAccess.theUnsafe.ARRAY_BYTE_INDEX_SCALE);
 }
   }
 



hbase git commit: HBASE-20459 Majority of scan CPU time in HBase-1 spent in size estimation.

2018-04-21 Thread larsh
Repository: hbase
Updated Branches:
  refs/heads/branch-1 e60189035 -> 6685b62ab


HBASE-20459 Majority of scan CPU time in HBase-1 spent in size estimation.


Project: http://git-wip-us.apache.org/repos/asf/hbase/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/6685b62a
Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/6685b62a
Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/6685b62a

Branch: refs/heads/branch-1
Commit: 6685b62abe44227ee98b4e88386c12a02d718ce3
Parents: e601890
Author: Lars Hofhansl 
Authored: Sat Apr 21 10:01:14 2018 -0700
Committer: Lars Hofhansl 
Committed: Sat Apr 21 10:02:07 2018 -0700

--
 .../src/main/java/org/apache/hadoop/hbase/util/ClassSize.java| 4 ++--
 1 file changed, 2 insertions(+), 2 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/hbase/blob/6685b62a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ClassSize.java
--
diff --git 
a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ClassSize.java 
b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ClassSize.java
index edb5b2d..5e6627f 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ClassSize.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ClassSize.java
@@ -141,7 +141,7 @@ public class ClassSize {
 }
 
 long sizeOf(byte[] b, int len) {
-  return align(arrayHeaderSize() + len);
+  return align(ARRAY + len);
 }
   }
 
@@ -186,7 +186,7 @@ public class ClassSize {
 @Override
 @SuppressWarnings("static-access")
 long sizeOf(byte[] b, int len) {
-  return align(arrayHeaderSize() + len * 
UnsafeAccess.theUnsafe.ARRAY_BYTE_INDEX_SCALE);
+  return align(ARRAY + len * 
UnsafeAccess.theUnsafe.ARRAY_BYTE_INDEX_SCALE);
 }
   }
 



hbase-site git commit: INFRA-10751 Empty commit

2018-04-21 Thread git-site-role
Repository: hbase-site
Updated Branches:
  refs/heads/asf-site 8bc57a1a2 -> 1a85017e7


INFRA-10751 Empty commit


Project: http://git-wip-us.apache.org/repos/asf/hbase-site/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase-site/commit/1a85017e
Tree: http://git-wip-us.apache.org/repos/asf/hbase-site/tree/1a85017e
Diff: http://git-wip-us.apache.org/repos/asf/hbase-site/diff/1a85017e

Branch: refs/heads/asf-site
Commit: 1a85017e791e22693b6121ddacba1cd02a42fa03
Parents: 8bc57a1
Author: jenkins 
Authored: Sat Apr 21 14:47:37 2018 +
Committer: jenkins 
Committed: Sat Apr 21 14:47:37 2018 +

--

--




[21/22] hbase-site git commit: Published site at 46cb5dfa226892fd2580f26ce9ce77225bd7e67c.

2018-04-21 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8bc57a1a/apache_hbase_reference_guide.pdf
--
diff --git a/apache_hbase_reference_guide.pdf b/apache_hbase_reference_guide.pdf
index 7ac23b6..396a172 100644
--- a/apache_hbase_reference_guide.pdf
+++ b/apache_hbase_reference_guide.pdf
@@ -5,8 +5,8 @@
 /Author (Apache HBase Team)
 /Creator (Asciidoctor PDF 1.5.0.alpha.15, based on Prawn 2.2.2)
 /Producer (Apache HBase Team)
-/ModDate (D:20180420144439+00'00')
-/CreationDate (D:20180420144439+00'00')
+/ModDate (D:20180421142950+00'00')
+/CreationDate (D:20180421144439+00'00')
 >>
 endobj
 2 0 obj
@@ -32028,7 +32028,7 @@ endobj
 endobj
 157 0 obj
 << /Limits [(__anchor-top) (adding.new.node)]
-/Names [(__anchor-top) 25 0 R (__indexterm-7613324) 3545 0 R 
(__indexterm-7615728) 3547 0 R (__indexterm-7617276) 3549 0 R 
(__indexterm-7619796) 3552 0 R (acid) 936 0 R (acl) 3356 0 R 
(add-metric-name-and-function-to-hadoop-compat-interface) 3645 0 R 
(add-the-implementation-to-both-hadoop-1-and-hadoop-2-compat-modules) 3647 0 R 
(add.metrics) 3643 0 R (adding-a-new-chapter-to-the-hbase-reference-guide) 3887 
0 R (adding.new.node) 3102 0 R]
+/Names [(__anchor-top) 25 0 R (__indexterm-2002) 3545 0 R (__indexterm-2004) 
3547 0 R (__indexterm-2006) 3549 0 R (__indexterm-2008) 3552 0 R (acid) 936 0 R 
(acl) 3356 0 R (add-metric-name-and-function-to-hadoop-compat-interface) 3645 0 
R (add-the-implementation-to-both-hadoop-1-and-hadoop-2-compat-modules) 3647 0 
R (add.metrics) 3643 0 R (adding-a-new-chapter-to-the-hbase-reference-guide) 
3887 0 R (adding.new.node) 3102 0 R]
 >>
 endobj
 158 0 obj
@@ -800821,4762 +800821,4762 @@ xref
 519751 0 n 
 520162 0 n 
 520209 0 n 
-520717 0 n 
-521058 0 n 
-521207 0 n 
-521427 0 n 
-521474 0 n 
-536662 0 n 
-537081 0 n 
-537127 0 n 
-537252 0 n 
-537378 0 n 
-537424 0 n 
-537548 0 n 
-537594 0 n 
-537640 0 n 
-548886 0 n 
-549337 0 n 
-549521 0 n 
-549753 0 n 
-549983 0 n 
-550029 0 n 
-550157 0 n 
-550291 0 n 
-550338 0 n 
-550484 0 n 
-550609 0 n 
-569270 0 n 
-569655 0 n 
-569785 0 n 
-569929 0 n 
-574618 0 n 
-574977 0 n 
-585005 0 n 
-585395 0 n 
-585441 0 n 
-585572 0 n 
-596438 0 n 
-596841 0 n 
-596887 0 n 
-596933 0 n 
-597075 0 n 
-597121 0 n 
-597167 0 n 
-597214 0 n 
-597261 0 n 
-605049 0 n 
-605408 0 n 
-605454 0 n 
-605501 0 n 
-605548 0 n 
-606152 0 n 
-606199 0 n 
-606246 0 n 
-606292 0 n 
-615083 0 n 
-615442 0 n 
-615488 0 n 
-615535 0 n 
-615582 0 n 
-615629 0 n 
-615676 0 n 
-615722 0 n 
-623349 0 n 
-623708 0 n 
-623756 0 n 
-624442 0 n 
-624490 0 n 
-624538 0 n 
-624586 0 n 
-624634 0 n 
-624681 0 n 
-635763 0 n 
-636122 0 n 
-636169 0 n 
-636216 0 n 
-636262 0 n 
-644211 0 n 
-644570 0 n 
-644617 0 n 
-644664 0 n 
-644711 0 n 
-644758 0 n 
-644805 0 n 
-645578 0 n 
-645624 0 n 
-654093 0 n 
-654452 0 n 
-654499 0 n 
-654546 0 n 
-654593 0 n 
-654640 0 n 
-654687 0 n 
-654733 0 n 
-663824 0 n 
-664209 0 n 
-664256 0 n 
-664489 0 n 
-664721 0 n 
-664768 0 n 
-664815 0 n 
-664862 0 n 
-664907 0 n 
-672944 0 n 
-673345 0 n 
-673393 0 n 
-673639 0 n 
-673884 0 n 
-673932 0 n 
-674178 0 n 
-674423 0 n 
-674471 0 n 
-674519 0 n 
-675052 0 n 
-675100 0 n 
-675148 0 n 
-684161 0 n 
-684533 0 n 
-684581 0 n 
-684629 0 n 
-684677 0 n 
-684725 0 n 
-684773 0 n 
-694088 0 n 
-694473 0 n 
-694521 0 n 
-694569 0 n 
-694617 0 n 
-694665 0 n 
-695180 0 n 
-695377 0 n 
-695572 0 n 
-695620 0 n 
-703955 0 n 
-704314 0 n 
-704361 0 n 
-704408 0 n 
-704455 0 n 
-704502 0 n 
-704549 0 n 
-704595 0 n 
-711508 0 n 
-711867 0 n 
-711914 0 n 
-711961 0 n 
-712008 0 n 
-712055 0 n 
-712102 0 n 
-712149 0 n 
-721699 0 n 
-722058 0 n 
-722104 0 n 
-722151 0 n 
-722198 0 n 
-722245 0 n 

[12/22] hbase-site git commit: Published site at 46cb5dfa226892fd2580f26ce9ce77225bd7e67c.

2018-04-21 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8bc57a1a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/FifoRpcScheduler.FifoCallRunner.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/FifoRpcScheduler.FifoCallRunner.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/FifoRpcScheduler.FifoCallRunner.html
index 9014fce..231ba93 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/FifoRpcScheduler.FifoCallRunner.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/FifoRpcScheduler.FifoCallRunner.html
@@ -142,78 +142,92 @@
 134  }
 135
 136  @Override
-137  public long getNumGeneralCallsDropped() 
{
-138return 0;
+137  public int 
getActiveGeneralRpcHandlerCount() {
+138return getActiveRpcHandlerCount();
 139  }
 140
 141  @Override
-142  public long getNumLifoModeSwitches() 
{
+142  public int 
getActivePriorityRpcHandlerCount() {
 143return 0;
 144  }
 145
 146  @Override
-147  public int getWriteQueueLength() {
+147  public int 
getActiveReplicationRpcHandlerCount() {
 148return 0;
 149  }
 150
 151  @Override
-152  public int getReadQueueLength() {
+152  public long getNumGeneralCallsDropped() 
{
 153return 0;
 154  }
 155
 156  @Override
-157  public int getScanQueueLength() {
+157  public long getNumLifoModeSwitches() 
{
 158return 0;
 159  }
 160
 161  @Override
-162  public int 
getActiveWriteRpcHandlerCount() {
+162  public int getWriteQueueLength() {
 163return 0;
 164  }
 165
 166  @Override
-167  public int 
getActiveReadRpcHandlerCount() {
+167  public int getReadQueueLength() {
 168return 0;
 169  }
 170
 171  @Override
-172  public int 
getActiveScanRpcHandlerCount() {
+172  public int getScanQueueLength() {
 173return 0;
 174  }
 175
 176  @Override
-177  public CallQueueInfo getCallQueueInfo() 
{
-178String queueName = "Fifo Queue";
-179
-180HashMapString, Long 
methodCount = new HashMap();
-181HashMapString, Long 
methodSize = new HashMap();
-182
-183CallQueueInfo callQueueInfo = new 
CallQueueInfo();
-184
callQueueInfo.setCallMethodCount(queueName, methodCount);
-185
callQueueInfo.setCallMethodSize(queueName, methodSize);
-186
-187
-188for (Runnable r:executor.getQueue()) 
{
-189  FifoCallRunner mcr = 
(FifoCallRunner) r;
-190  RpcCall rpcCall = 
mcr.getCallRunner().getRpcCall();
-191
-192  String method;
-193
-194  if (null==rpcCall.getMethod() ||
-195
StringUtil.isNullOrEmpty(method = rpcCall.getMethod().getName())) {
-196method = "Unknown";
-197  }
-198
-199  long size = rpcCall.getSize();
-200
-201  methodCount.put(method, 1 + 
methodCount.getOrDefault(method, 0L));
-202  methodSize.put(method, size + 
methodSize.getOrDefault(method, 0L));
-203}
-204
-205return callQueueInfo;
-206  }
-207
-208}
+177  public int 
getActiveWriteRpcHandlerCount() {
+178return 0;
+179  }
+180
+181  @Override
+182  public int 
getActiveReadRpcHandlerCount() {
+183return 0;
+184  }
+185
+186  @Override
+187  public int 
getActiveScanRpcHandlerCount() {
+188return 0;
+189  }
+190
+191  @Override
+192  public CallQueueInfo getCallQueueInfo() 
{
+193String queueName = "Fifo Queue";
+194
+195HashMapString, Long 
methodCount = new HashMap();
+196HashMapString, Long 
methodSize = new HashMap();
+197
+198CallQueueInfo callQueueInfo = new 
CallQueueInfo();
+199
callQueueInfo.setCallMethodCount(queueName, methodCount);
+200
callQueueInfo.setCallMethodSize(queueName, methodSize);
+201
+202
+203for (Runnable r:executor.getQueue()) 
{
+204  FifoCallRunner mcr = 
(FifoCallRunner) r;
+205  RpcCall rpcCall = 
mcr.getCallRunner().getRpcCall();
+206
+207  String method;
+208
+209  if (null==rpcCall.getMethod() ||
+210
StringUtil.isNullOrEmpty(method = rpcCall.getMethod().getName())) {
+211method = "Unknown";
+212  }
+213
+214  long size = rpcCall.getSize();
+215
+216  methodCount.put(method, 1 + 
methodCount.getOrDefault(method, 0L));
+217  methodSize.put(method, size + 
methodSize.getOrDefault(method, 0L));
+218}
+219
+220return callQueueInfo;
+221  }
+222}
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8bc57a1a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/FifoRpcScheduler.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/FifoRpcScheduler.html 
b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/FifoRpcScheduler.html
index 9014fce..231ba93 100644
--- a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/FifoRpcScheduler.html
+++ b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/FifoRpcScheduler.html
@@ -142,78 +142,92 @@
 134  }
 135
 136  @Override
-137  public long getNumGeneralCallsDropped() 
{
-138return 0;
+137  public int 
getActiveGeneralRpcHandlerCount() {
+138return 

[20/22] hbase-site git commit: Published site at 46cb5dfa226892fd2580f26ce9ce77225bd7e67c.

2018-04-21 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8bc57a1a/book.html
--
diff --git a/book.html b/book.html
index 8de5bcb..d82af18 100644
--- a/book.html
+++ b/book.html
@@ -37824,7 +37824,7 @@ The server will return cellblocks compressed using this 
same compressor as long
 
 
 Version 3.0.0-SNAPSHOT
-Last updated 2018-04-20 14:29:52 UTC
+Last updated 2018-04-21 14:29:50 UTC
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8bc57a1a/bulk-loads.html
--
diff --git a/bulk-loads.html b/bulk-loads.html
index 2c97d08..111fe28 100644
--- a/bulk-loads.html
+++ b/bulk-loads.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase   
   Bulk Loads in Apache HBase (TM)
@@ -306,7 +306,7 @@ under the License. -->
 https://www.apache.org/;>The Apache Software 
Foundation.
 All rights reserved.  
 
-  Last Published: 
2018-04-20
+  Last Published: 
2018-04-21
 
 
 



[01/22] hbase-site git commit: Published site at 46cb5dfa226892fd2580f26ce9ce77225bd7e67c.

2018-04-21 Thread git-site-role
Repository: hbase-site
Updated Branches:
  refs/heads/asf-site 84a296a43 -> 8bc57a1a2


http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8bc57a1a/testdevapidocs/src-html/org/apache/hadoop/hbase/thrift/TestThriftHttpServer.html
--
diff --git 
a/testdevapidocs/src-html/org/apache/hadoop/hbase/thrift/TestThriftHttpServer.html
 
b/testdevapidocs/src-html/org/apache/hadoop/hbase/thrift/TestThriftHttpServer.html
index 2d059c0..0e3c33c 100644
--- 
a/testdevapidocs/src-html/org/apache/hadoop/hbase/thrift/TestThriftHttpServer.html
+++ 
b/testdevapidocs/src-html/org/apache/hadoop/hbase/thrift/TestThriftHttpServer.html
@@ -29,214 +29,227 @@
 021import static 
org.junit.Assert.assertNull;
 022import static org.junit.Assert.fail;
 023
-024import java.util.ArrayList;
-025import java.util.List;
-026import 
org.apache.hadoop.conf.Configuration;
-027import 
org.apache.hadoop.hbase.HBaseClassTestRule;
-028import 
org.apache.hadoop.hbase.HBaseTestingUtility;
-029import 
org.apache.hadoop.hbase.HConstants;
-030import 
org.apache.hadoop.hbase.testclassification.ClientTests;
-031import 
org.apache.hadoop.hbase.testclassification.LargeTests;
-032import 
org.apache.hadoop.hbase.thrift.generated.Hbase;
-033import 
org.apache.hadoop.hbase.util.EnvironmentEdgeManager;
-034import 
org.apache.hadoop.hbase.util.EnvironmentEdgeManagerTestHelper;
-035import 
org.apache.hadoop.hbase.util.IncrementingEnvironmentEdge;
-036import 
org.apache.thrift.protocol.TBinaryProtocol;
-037import 
org.apache.thrift.protocol.TProtocol;
-038import 
org.apache.thrift.transport.THttpClient;
-039import 
org.apache.thrift.transport.TTransportException;
-040import org.junit.AfterClass;
-041import org.junit.BeforeClass;
-042import org.junit.ClassRule;
-043import org.junit.Rule;
-044import org.junit.Test;
-045import 
org.junit.experimental.categories.Category;
-046import 
org.junit.rules.ExpectedException;
-047import org.slf4j.Logger;
-048import org.slf4j.LoggerFactory;
-049
-050import 
org.apache.hbase.thirdparty.com.google.common.base.Joiner;
-051
-052/**
-053 * Start the HBase Thrift HTTP server on 
a random port through the command-line
-054 * interface and talk to it from client 
side.
-055 */
-056@Category({ClientTests.class, 
LargeTests.class})
-057
-058public class TestThriftHttpServer {
-059
-060  @ClassRule
-061  public static final HBaseClassTestRule 
CLASS_RULE =
-062  
HBaseClassTestRule.forClass(TestThriftHttpServer.class);
-063
-064  private static final Logger LOG =
-065  
LoggerFactory.getLogger(TestThriftHttpServer.class);
+024import java.net.HttpURLConnection;
+025import java.net.URL;
+026import java.util.ArrayList;
+027import java.util.List;
+028import 
org.apache.hadoop.conf.Configuration;
+029import 
org.apache.hadoop.hbase.HBaseClassTestRule;
+030import 
org.apache.hadoop.hbase.HBaseTestingUtility;
+031import 
org.apache.hadoop.hbase.HConstants;
+032import 
org.apache.hadoop.hbase.testclassification.ClientTests;
+033import 
org.apache.hadoop.hbase.testclassification.LargeTests;
+034import 
org.apache.hadoop.hbase.thrift.generated.Hbase;
+035import 
org.apache.hadoop.hbase.util.EnvironmentEdgeManager;
+036import 
org.apache.hadoop.hbase.util.EnvironmentEdgeManagerTestHelper;
+037import 
org.apache.hadoop.hbase.util.IncrementingEnvironmentEdge;
+038import 
org.apache.thrift.protocol.TBinaryProtocol;
+039import 
org.apache.thrift.protocol.TProtocol;
+040import 
org.apache.thrift.transport.THttpClient;
+041import 
org.apache.thrift.transport.TTransportException;
+042import org.junit.AfterClass;
+043import org.junit.Assert;
+044import org.junit.BeforeClass;
+045import org.junit.ClassRule;
+046import org.junit.Rule;
+047import org.junit.Test;
+048import 
org.junit.experimental.categories.Category;
+049import 
org.junit.rules.ExpectedException;
+050import org.slf4j.Logger;
+051import org.slf4j.LoggerFactory;
+052
+053import 
org.apache.hbase.thirdparty.com.google.common.base.Joiner;
+054
+055/**
+056 * Start the HBase Thrift HTTP server on 
a random port through the command-line
+057 * interface and talk to it from client 
side.
+058 */
+059@Category({ClientTests.class, 
LargeTests.class})
+060
+061public class TestThriftHttpServer {
+062
+063  @ClassRule
+064  public static final HBaseClassTestRule 
CLASS_RULE =
+065  
HBaseClassTestRule.forClass(TestThriftHttpServer.class);
 066
-067  private static final 
HBaseTestingUtility TEST_UTIL =
-068  new HBaseTestingUtility();
+067  private static final Logger LOG =
+068  
LoggerFactory.getLogger(TestThriftHttpServer.class);
 069
-070  private Thread httpServerThread;
-071  private volatile Exception 
httpServerException;
+070  private static final 
HBaseTestingUtility TEST_UTIL =
+071  new HBaseTestingUtility();
 072
-073  private Exception 
clientSideException;
-074
-075  private ThriftServer thriftServer;
-076  private int port;
+073  private Thread httpServerThread;
+074  private volatile Exception 

[17/22] hbase-site git commit: Published site at 46cb5dfa226892fd2580f26ce9ce77225bd7e67c.

2018-04-21 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8bc57a1a/devapidocs/org/apache/hadoop/hbase/ipc/MetricsHBaseServerSource.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/ipc/MetricsHBaseServerSource.html 
b/devapidocs/org/apache/hadoop/hbase/ipc/MetricsHBaseServerSource.html
index 5db80d5..b3fddf3 100644
--- a/devapidocs/org/apache/hadoop/hbase/ipc/MetricsHBaseServerSource.html
+++ b/devapidocs/org/apache/hadoop/hbase/ipc/MetricsHBaseServerSource.html
@@ -180,6 +180,14 @@ extends 
 static https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String
+NUM_ACTIVE_GENERAL_HANDLER_DESC
+
+
+static https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String
+NUM_ACTIVE_GENERAL_HANDLER_NAME
+
+
+static https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String
 NUM_ACTIVE_HANDLER_DESC
 
 
@@ -188,6 +196,14 @@ extends 
 static https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String
+NUM_ACTIVE_PRIORITY_HANDLER_DESC
+
+
+static https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String
+NUM_ACTIVE_PRIORITY_HANDLER_NAME
+
+
+static https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String
 NUM_ACTIVE_READ_HANDLER_DESC
 
 
@@ -196,6 +212,14 @@ extends 
 static https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String
+NUM_ACTIVE_REPLICATION_HANDLER_DESC
+
+
+static https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String
+NUM_ACTIVE_REPLICATION_HANDLER_NAME
+
+
+static https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String
 NUM_ACTIVE_SCAN_HANDLER_DESC
 
 
@@ -990,13 +1014,91 @@ extends 
+
+
+
+
+NUM_ACTIVE_GENERAL_HANDLER_NAME
+static finalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String NUM_ACTIVE_GENERAL_HANDLER_NAME
+
+See Also:
+Constant
 Field Values
+
+
+
+
+
+
+
+
+NUM_ACTIVE_GENERAL_HANDLER_DESC
+static finalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String NUM_ACTIVE_GENERAL_HANDLER_DESC
+
+See Also:
+Constant
 Field Values
+
+
+
+
+
+
+
+
+NUM_ACTIVE_PRIORITY_HANDLER_NAME
+static finalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String NUM_ACTIVE_PRIORITY_HANDLER_NAME
+
+See Also:
+Constant
 Field Values
+
+
+
+
+
+
+
+
+NUM_ACTIVE_PRIORITY_HANDLER_DESC
+static finalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String NUM_ACTIVE_PRIORITY_HANDLER_DESC
+
+See Also:
+Constant
 Field Values
+
+
+
+
+
+
+
+
+NUM_ACTIVE_REPLICATION_HANDLER_NAME
+static finalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String NUM_ACTIVE_REPLICATION_HANDLER_NAME
+
+See Also:
+Constant
 Field Values
+
+
+
+
+
+
+
+
+NUM_ACTIVE_REPLICATION_HANDLER_DESC
+static finalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String NUM_ACTIVE_REPLICATION_HANDLER_DESC
+
+See Also:
+Constant
 Field Values
+
+
+
 
 
 
 
 
 NUM_ACTIVE_WRITE_HANDLER_NAME
-static finalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String NUM_ACTIVE_WRITE_HANDLER_NAME
+static finalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String NUM_ACTIVE_WRITE_HANDLER_NAME
 
 See Also:
 Constant
 Field Values
@@ -1009,7 +,7 @@ extends 
 
 NUM_ACTIVE_WRITE_HANDLER_DESC
-static finalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String NUM_ACTIVE_WRITE_HANDLER_DESC
+static finalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String NUM_ACTIVE_WRITE_HANDLER_DESC
 
 See Also:
 Constant
 Field Values
@@ -1022,7 +1124,7 @@ extends 
 
 NUM_ACTIVE_READ_HANDLER_NAME
-static finalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String NUM_ACTIVE_READ_HANDLER_NAME
+static 

[19/22] hbase-site git commit: Published site at 46cb5dfa226892fd2580f26ce9ce77225bd7e67c.

2018-04-21 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8bc57a1a/checkstyle-aggregate.html
--
diff --git a/checkstyle-aggregate.html b/checkstyle-aggregate.html
index 67a62c1..88750a0 100644
--- a/checkstyle-aggregate.html
+++ b/checkstyle-aggregate.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase  Checkstyle Results
 
@@ -284,7 +284,7 @@
 3604
 0
 0
-15864
+15859
 
 Files
 
@@ -6947,7 +6947,7 @@
 org/apache/hadoop/hbase/regionserver/StoreFileInfo.java
 0
 0
-23
+18
 
 org/apache/hadoop/hbase/regionserver/StoreFileManager.java
 0
@@ -10332,7 +10332,7 @@
 
 
 http://checkstyle.sourceforge.net/config_javadoc.html#NonEmptyAtclauseDescription;>NonEmptyAtclauseDescription
-3831
+3826
 Error
 
 misc
@@ -52880,7 +52880,7 @@
 coding
 InnerAssignment
 Inner assignments should be avoided.
-195
+210
 
 org/apache/hadoop/hbase/ipc/HBaseRpcController.java
 
@@ -54608,37 +54608,37 @@
 sizes
 LineLength
 Line is longer than 100 characters (found 101).
-90
+93
 
 Error
 indentation
 Indentation
 'array initialization' child have incorrect indentation level 8, expected 
level should be one of the following: 6, 35.
-153
+156
 
 Error
 indentation
 Indentation
 'array initialization' child have incorrect indentation level 8, expected 
level should be one of the following: 6, 35.
-154
+157
 
 Error
 indentation
 Indentation
 'array initialization' child have incorrect indentation level 8, expected 
level should be one of the following: 6, 35.
-157
+160
 
 Error
 indentation
 Indentation
 'array initialization' child have incorrect indentation level 8, expected 
level should be one of the following: 6, 35.
-158
+161
 
 Error
 indentation
 Indentation
 'array initialization' child have incorrect indentation level 8, expected 
level should be one of the following: 6, 35.
-159
+162
 
 org/apache/hadoop/hbase/ipc/TestSimpleRpcScheduler.java
 
@@ -78137,37 +78137,37 @@
 blocks
 NeedBraces
 'if' construct must use '{}'s.
-189
+192
 
 Error
 blocks
 NeedBraces
 'if' construct must use '{}'s.
-190
+193
 
 Error
 blocks
 NeedBraces
 'if' construct must use '{}'s.
-191
+194
 
 Error
 blocks
 NeedBraces
 'if' construct must use '{}'s.
-192
+195
 
 Error
 blocks
 NeedBraces
 'if' construct must use '{}'s.
-193
+196
 
 Error
 blocks
 NeedBraces
 'if' construct must use '{}'s.
-194
+197
 
 org/apache/hadoop/hbase/quotas/UserQuotaState.java
 
@@ -85844,137 +85844,107 @@
 javadoc
 NonEmptyAtclauseDescription
 At-clause should have a non-empty description.
-178
-
-Error
-javadoc
-NonEmptyAtclauseDescription
-At-clause should have a non-empty description.
-179
-
-Error
-javadoc
-NonEmptyAtclauseDescription
-At-clause should have a non-empty description.
-180
-
-Error
-javadoc
-NonEmptyAtclauseDescription
-At-clause should have a non-empty description.
-181
-
-Error
-javadoc
-NonEmptyAtclauseDescription
-At-clause should have a non-empty description.
-182
+212
 
 Error
-javadoc
-NonEmptyAtclauseDescription
-At-clause should have a non-empty description.
-197
+blocks
+NeedBraces
+'if' construct must use '{}'s.
+486
 
 Error
 blocks
 NeedBraces
 'if' construct must use '{}'s.
-471
+500
 
 Error
 blocks
 NeedBraces
 'if' construct must use '{}'s.
-485
+549
 
 Error
 blocks
 NeedBraces
 'if' construct must use '{}'s.
-534
+550
 
 Error
 blocks
 NeedBraces
 'if' construct must use '{}'s.
-535
+552
 
 Error
 blocks
 NeedBraces
 'if' construct must use '{}'s.
-537
+555
 
 Error
 blocks
 NeedBraces
 'if' construct must use '{}'s.
-540
+556
 
 Error
 blocks
 NeedBraces
 'if' construct must use '{}'s.
-541
+557
 
 Error
 blocks
 NeedBraces
 'if' construct must use '{}'s.
-542
+560
 
 Error
 blocks
 NeedBraces
 'if' construct must use '{}'s.
-545
+561
 
 Error
 blocks
 NeedBraces
 'if' construct must use '{}'s.
-546
+562
 
 Error
 blocks
 NeedBraces
 'if' construct must use '{}'s.
-547
+565
 
 Error
 blocks
 NeedBraces
 'if' construct must use '{}'s.
-550
+566
 
 Error
 blocks
 NeedBraces
 'if' construct must use '{}'s.
-551
-
-Error
-blocks
-NeedBraces
-'if' construct must use '{}'s.
-552
+567
 
 org/apache/hadoop/hbase/regionserver/StoreFileManager.java
 
-
+
 Severity
 Category
 Rule
 Message
 Line
-
+
 Error
 javadoc
 NonEmptyAtclauseDescription
 At-clause should have a non-empty description.
 68
-
+
 Error
 javadoc
 NonEmptyAtclauseDescription
@@ -85983,31 +85953,31 @@
 
 org/apache/hadoop/hbase/regionserver/StoreFileReader.java
 
-
+
 Severity
 Category
 Rule
 Message
 Line
-
+
 Error
 imports
 ImportOrder
 Wrong order for 'org.apache.hadoop.hbase.KeyValue' import.
 39
-
+
 Error
 blocks
 NeedBraces
 'if' construct must use '{}'s.
 469
-
+
 Error
 blocks
 NeedBraces
 'if' construct must use '{}'s.
 489
-
+
 Error
 blocks
 NeedBraces
@@ -86016,67 +85986,67 @@
 
 org/apache/hadoop/hbase/regionserver/StoreFileScanner.java
 
-
+
 Severity
 Category
 Rule
 Message
 Line
-
+
 Error
 imports
 ImportOrder
 Wrong order for 'org.apache.hadoop.hbase.client.Scan' 

[09/22] hbase-site git commit: Published site at 46cb5dfa226892fd2580f26ce9ce77225bd7e67c.

2018-04-21 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8bc57a1a/devapidocs/src-html/org/apache/hadoop/hbase/thrift/ThriftServerRunner.HBaseHandler.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/thrift/ThriftServerRunner.HBaseHandler.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/thrift/ThriftServerRunner.HBaseHandler.html
index d6f672a..fd7ab7a 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/thrift/ThriftServerRunner.HBaseHandler.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/thrift/ThriftServerRunner.HBaseHandler.html
@@ -87,1949 +87,1951 @@
 079import 
org.apache.hadoop.hbase.filter.ParseFilter;
 080import 
org.apache.hadoop.hbase.filter.PrefixFilter;
 081import 
org.apache.hadoop.hbase.filter.WhileMatchFilter;
-082import 
org.apache.hadoop.hbase.log.HBaseMarkers;
-083import 
org.apache.hadoop.hbase.security.SaslUtil;
-084import 
org.apache.hadoop.hbase.security.SaslUtil.QualityOfProtection;
-085import 
org.apache.hadoop.hbase.security.SecurityUtil;
-086import 
org.apache.hadoop.hbase.security.UserProvider;
-087import 
org.apache.hadoop.hbase.thrift.generated.AlreadyExists;
-088import 
org.apache.hadoop.hbase.thrift.generated.BatchMutation;
-089import 
org.apache.hadoop.hbase.thrift.generated.ColumnDescriptor;
-090import 
org.apache.hadoop.hbase.thrift.generated.Hbase;
-091import 
org.apache.hadoop.hbase.thrift.generated.IOError;
-092import 
org.apache.hadoop.hbase.thrift.generated.IllegalArgument;
-093import 
org.apache.hadoop.hbase.thrift.generated.Mutation;
-094import 
org.apache.hadoop.hbase.thrift.generated.TAppend;
-095import 
org.apache.hadoop.hbase.thrift.generated.TCell;
-096import 
org.apache.hadoop.hbase.thrift.generated.TIncrement;
-097import 
org.apache.hadoop.hbase.thrift.generated.TRegionInfo;
-098import 
org.apache.hadoop.hbase.thrift.generated.TRowResult;
-099import 
org.apache.hadoop.hbase.thrift.generated.TScan;
-100import 
org.apache.hadoop.hbase.util.Bytes;
-101import 
org.apache.hadoop.hbase.util.ConnectionCache;
-102import 
org.apache.hadoop.hbase.util.DNS;
-103import 
org.apache.hadoop.hbase.util.JvmPauseMonitor;
-104import 
org.apache.hadoop.hbase.util.Strings;
-105import 
org.apache.hadoop.security.SaslRpcServer.SaslGssCallbackHandler;
-106import 
org.apache.hadoop.security.UserGroupInformation;
-107import 
org.apache.hadoop.security.authorize.ProxyUsers;
-108import org.apache.thrift.TException;
-109import org.apache.thrift.TProcessor;
-110import 
org.apache.thrift.protocol.TBinaryProtocol;
-111import 
org.apache.thrift.protocol.TCompactProtocol;
-112import 
org.apache.thrift.protocol.TProtocol;
-113import 
org.apache.thrift.protocol.TProtocolFactory;
-114import 
org.apache.thrift.server.THsHaServer;
-115import 
org.apache.thrift.server.TNonblockingServer;
-116import 
org.apache.thrift.server.TServer;
-117import 
org.apache.thrift.server.TServlet;
-118import 
org.apache.thrift.server.TThreadedSelectorServer;
-119import 
org.apache.thrift.transport.TFramedTransport;
-120import 
org.apache.thrift.transport.TNonblockingServerSocket;
-121import 
org.apache.thrift.transport.TNonblockingServerTransport;
-122import 
org.apache.thrift.transport.TSaslServerTransport;
-123import 
org.apache.thrift.transport.TServerSocket;
-124import 
org.apache.thrift.transport.TServerTransport;
-125import 
org.apache.thrift.transport.TTransportFactory;
-126import 
org.apache.yetus.audience.InterfaceAudience;
-127import 
org.eclipse.jetty.http.HttpVersion;
-128import 
org.eclipse.jetty.server.HttpConfiguration;
-129import 
org.eclipse.jetty.server.HttpConnectionFactory;
-130import 
org.eclipse.jetty.server.SecureRequestCustomizer;
-131import org.eclipse.jetty.server.Server;
-132import 
org.eclipse.jetty.server.ServerConnector;
-133import 
org.eclipse.jetty.server.SslConnectionFactory;
-134import 
org.eclipse.jetty.servlet.ServletContextHandler;
-135import 
org.eclipse.jetty.servlet.ServletHolder;
-136import 
org.eclipse.jetty.util.ssl.SslContextFactory;
-137import 
org.eclipse.jetty.util.thread.QueuedThreadPool;
-138import org.slf4j.Logger;
-139import org.slf4j.LoggerFactory;
-140import 
org.apache.hbase.thirdparty.com.google.common.base.Joiner;
-141import 
org.apache.hbase.thirdparty.com.google.common.base.Throwables;
-142import 
org.apache.hbase.thirdparty.com.google.common.util.concurrent.ThreadFactoryBuilder;
-143import 
org.apache.hbase.thirdparty.org.apache.commons.cli.CommandLine;
-144import 
org.apache.hbase.thirdparty.org.apache.commons.cli.Option;
-145import 
org.apache.hbase.thirdparty.org.apache.commons.cli.OptionGroup;
-146
-147/**
-148 * ThriftServerRunner - this class starts 
up a Thrift server which implements
-149 * the Hbase API specified in the 
Hbase.thrift IDL file.
-150 */
-151@InterfaceAudience.Private
-152public class ThriftServerRunner 
implements Runnable {
-153
-154  private static final Logger LOG = 
LoggerFactory.getLogger(ThriftServerRunner.class);
-155
-156  

[15/22] hbase-site git commit: Published site at 46cb5dfa226892fd2580f26ce9ce77225bd7e67c.

2018-04-21 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8bc57a1a/devapidocs/org/apache/hadoop/hbase/quotas/class-use/RpcThrottlingException.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/quotas/class-use/RpcThrottlingException.html
 
b/devapidocs/org/apache/hadoop/hbase/quotas/class-use/RpcThrottlingException.html
index f7bd99d..f5a70f3 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/quotas/class-use/RpcThrottlingException.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/quotas/class-use/RpcThrottlingException.html
@@ -124,19 +124,25 @@
 
 
 void
-TimeBasedLimiter.checkQuota(longwriteSize,
-  longreadSize)
+TimeBasedLimiter.checkQuota(longwriteReqs,
+  longestimateWriteSize,
+  longreadReqs,
+  longestimateReadSize)
 
 
 void
-QuotaLimiter.checkQuota(longestimateWriteSize,
+QuotaLimiter.checkQuota(longwriteReqs,
+  longestimateWriteSize,
+  longreadReqs,
   longestimateReadSize)
 Checks if it is possible to execute the specified 
operation.
 
 
 
 void
-NoopQuotaLimiter.checkQuota(longestimateWriteSize,
+NoopQuotaLimiter.checkQuota(longwriteReqs,
+  longestimateWriteSize,
+  longreadReqs,
   longestimateReadSize)
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8bc57a1a/devapidocs/org/apache/hadoop/hbase/quotas/package-tree.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/quotas/package-tree.html 
b/devapidocs/org/apache/hadoop/hbase/quotas/package-tree.html
index 4e25e3e..152e2b7 100644
--- a/devapidocs/org/apache/hadoop/hbase/quotas/package-tree.html
+++ b/devapidocs/org/apache/hadoop/hbase/quotas/package-tree.html
@@ -229,13 +229,13 @@
 
 java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true;
 title="class or interface in java.lang">EnumE (implements java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true;
 title="class or interface in java.lang">ComparableT, java.io.https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true;
 title="class or interface in java.io">Serializable)
 
-org.apache.hadoop.hbase.quotas.SpaceViolationPolicy
-org.apache.hadoop.hbase.quotas.OperationQuota.OperationType
+org.apache.hadoop.hbase.quotas.QuotaType
 org.apache.hadoop.hbase.quotas.ThrottleType
-org.apache.hadoop.hbase.quotas.QuotaScope
 org.apache.hadoop.hbase.quotas.RpcThrottlingException.Type
+org.apache.hadoop.hbase.quotas.OperationQuota.OperationType
+org.apache.hadoop.hbase.quotas.SpaceViolationPolicy
 org.apache.hadoop.hbase.quotas.ThrottlingException.Type
-org.apache.hadoop.hbase.quotas.QuotaType
+org.apache.hadoop.hbase.quotas.QuotaScope
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8bc57a1a/devapidocs/org/apache/hadoop/hbase/regionserver/StoreFileInfo.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/StoreFileInfo.html 
b/devapidocs/org/apache/hadoop/hbase/regionserver/StoreFileInfo.html
index 1158ff7..df2be42 100644
--- a/devapidocs/org/apache/hadoop/hbase/regionserver/StoreFileInfo.html
+++ b/devapidocs/org/apache/hadoop/hbase/regionserver/StoreFileInfo.html
@@ -237,6 +237,15 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 
+StoreFileInfo(org.apache.hadoop.conf.Configurationconf,
+ org.apache.hadoop.fs.FileSystemfs,
+ org.apache.hadoop.fs.FileStatusfileStatus,
+ Referencereference,
+ HFileLinklink)
+Create a Store File Info from an HFileLink and a 
Reference
+
+
+
 StoreFileInfo(org.apache.hadoop.conf.Configurationconf,
  org.apache.hadoop.fs.FileSystemfs,
  org.apache.hadoop.fs.PathinitialPath)
@@ -622,39 +631,55 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 publicStoreFileInfo(org.apache.hadoop.conf.Configurationconf,
  org.apache.hadoop.fs.FileSystemfs,
  org.apache.hadoop.fs.FileStatusfileStatus,
- HFileLinklink)
-  throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
 title="class or interface in java.io">IOException
+ HFileLinklink)
 Create a Store File Info from an HFileLink
 
 Parameters:
-conf - the Configuration to use
-fs - The current file system to use.
+conf - The Configuration to use
+fs - The current file system to use
 fileStatus - The FileStatus of the file
-Throws:
-https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
 title="class or interface in java.io">IOException
 
 
 
 
 
 
-
+
 
 StoreFileInfo
-publicStoreFileInfo(org.apache.hadoop.conf.Configurationconf,
+publicStoreFileInfo(org.apache.hadoop.conf.Configurationconf,
  

[11/22] hbase-site git commit: Published site at 46cb5dfa226892fd2580f26ce9ce77225bd7e67c.

2018-04-21 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8bc57a1a/devapidocs/src-html/org/apache/hadoop/hbase/quotas/QuotaLimiter.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/quotas/QuotaLimiter.html 
b/devapidocs/src-html/org/apache/hadoop/hbase/quotas/QuotaLimiter.html
index 80bd305..ee8178a 100644
--- a/devapidocs/src-html/org/apache/hadoop/hbase/quotas/QuotaLimiter.html
+++ b/devapidocs/src-html/org/apache/hadoop/hbase/quotas/QuotaLimiter.html
@@ -39,44 +39,48 @@
 031  /**
 032   * Checks if it is possible to execute 
the specified operation.
 033   *
-034   * @param estimateWriteSize the write 
size that will be checked against the available quota
-035   * @param estimateReadSize the read 
size that will be checked against the available quota
-036   * @throws RpcThrottlingException 
thrown if not enough avialable resources to perform operation.
-037   */
-038  void checkQuota(long estimateWriteSize, 
long estimateReadSize)
-039throws RpcThrottlingException;
-040
-041  /**
-042   * Removes the specified write and read 
amount from the quota.
-043   * At this point the write and read 
amount will be an estimate,
-044   * that will be later adjusted with a 
consumeWrite()/consumeRead() call.
-045   *
-046   * @param writeSize the write size that 
will be removed from the current quota
-047   * @param readSize the read size that 
will be removed from the current quota
-048   */
-049  void grabQuota(long writeSize, long 
readSize);
-050
-051  /**
-052   * Removes or add back some write 
amount to the quota.
-053   * (called at the end of an operation 
in case the estimate quota was off)
-054   */
-055  void consumeWrite(long size);
-056
-057  /**
-058   * Removes or add back some read amount 
to the quota.
-059   * (called at the end of an operation 
in case the estimate quota was off)
-060   */
-061  void consumeRead(long size);
-062
-063  /** @return true if the limiter is a 
noop */
-064  boolean isBypass();
-065
-066/** @return the number of bytes 
available to read to avoid exceeding the quota */
-067  long getReadAvailable();
-068
-069  /** @return the number of bytes 
available to write to avoid exceeding the quota */
-070  long getWriteAvailable();
-071}
+034   * @param writeReqs the write requests 
that will be checked against the available quota
+035   * @param estimateWriteSize the write 
size that will be checked against the available quota
+036   * @param readReqs the read requests 
that will be checked against the available quota
+037   * @param estimateReadSize the read 
size that will be checked against the available quota
+038   * @throws RpcThrottlingException 
thrown if not enough available resources to perform operation.
+039   */
+040  void checkQuota(long writeReqs, long 
estimateWriteSize, long readReqs, long estimateReadSize)
+041  throws RpcThrottlingException;
+042
+043  /**
+044   * Removes the specified write and read 
amount from the quota.
+045   * At this point the write and read 
amount will be an estimate,
+046   * that will be later adjusted with a 
consumeWrite()/consumeRead() call.
+047   *
+048   * @param writeReqs the write requests 
that will be removed from the current quota
+049   * @param writeSize the write size that 
will be removed from the current quota
+050   * @param readReqs the read requests 
that will be removed from the current quota
+051   * @param readSize the read size that 
will be removed from the current quota
+052   */
+053  void grabQuota(long writeReqs, long 
writeSize, long readReqs, long readSize);
+054
+055  /**
+056   * Removes or add back some write 
amount to the quota.
+057   * (called at the end of an operation 
in case the estimate quota was off)
+058   */
+059  void consumeWrite(long size);
+060
+061  /**
+062   * Removes or add back some read amount 
to the quota.
+063   * (called at the end of an operation 
in case the estimate quota was off)
+064   */
+065  void consumeRead(long size);
+066
+067  /** @return true if the limiter is a 
noop */
+068  boolean isBypass();
+069
+070/** @return the number of bytes 
available to read to avoid exceeding the quota */
+071  long getReadAvailable();
+072
+073  /** @return the number of bytes 
available to write to avoid exceeding the quota */
+074  long getWriteAvailable();
+075}
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8bc57a1a/devapidocs/src-html/org/apache/hadoop/hbase/quotas/TimeBasedLimiter.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/quotas/TimeBasedLimiter.html 
b/devapidocs/src-html/org/apache/hadoop/hbase/quotas/TimeBasedLimiter.html
index 16a2371..7e6dad2 100644
--- a/devapidocs/src-html/org/apache/hadoop/hbase/quotas/TimeBasedLimiter.html
+++ b/devapidocs/src-html/org/apache/hadoop/hbase/quotas/TimeBasedLimiter.html
@@ -118,92 +118,95 @@
 110  }
 111
 112  @Override
-113  

[06/22] hbase-site git commit: Published site at 46cb5dfa226892fd2580f26ce9ce77225bd7e67c.

2018-04-21 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8bc57a1a/devapidocs/src-html/org/apache/hadoop/hbase/thrift/ThriftServerRunner.ResultScannerWrapper.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/thrift/ThriftServerRunner.ResultScannerWrapper.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/thrift/ThriftServerRunner.ResultScannerWrapper.html
index d6f672a..fd7ab7a 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/thrift/ThriftServerRunner.ResultScannerWrapper.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/thrift/ThriftServerRunner.ResultScannerWrapper.html
@@ -87,1949 +87,1951 @@
 079import 
org.apache.hadoop.hbase.filter.ParseFilter;
 080import 
org.apache.hadoop.hbase.filter.PrefixFilter;
 081import 
org.apache.hadoop.hbase.filter.WhileMatchFilter;
-082import 
org.apache.hadoop.hbase.log.HBaseMarkers;
-083import 
org.apache.hadoop.hbase.security.SaslUtil;
-084import 
org.apache.hadoop.hbase.security.SaslUtil.QualityOfProtection;
-085import 
org.apache.hadoop.hbase.security.SecurityUtil;
-086import 
org.apache.hadoop.hbase.security.UserProvider;
-087import 
org.apache.hadoop.hbase.thrift.generated.AlreadyExists;
-088import 
org.apache.hadoop.hbase.thrift.generated.BatchMutation;
-089import 
org.apache.hadoop.hbase.thrift.generated.ColumnDescriptor;
-090import 
org.apache.hadoop.hbase.thrift.generated.Hbase;
-091import 
org.apache.hadoop.hbase.thrift.generated.IOError;
-092import 
org.apache.hadoop.hbase.thrift.generated.IllegalArgument;
-093import 
org.apache.hadoop.hbase.thrift.generated.Mutation;
-094import 
org.apache.hadoop.hbase.thrift.generated.TAppend;
-095import 
org.apache.hadoop.hbase.thrift.generated.TCell;
-096import 
org.apache.hadoop.hbase.thrift.generated.TIncrement;
-097import 
org.apache.hadoop.hbase.thrift.generated.TRegionInfo;
-098import 
org.apache.hadoop.hbase.thrift.generated.TRowResult;
-099import 
org.apache.hadoop.hbase.thrift.generated.TScan;
-100import 
org.apache.hadoop.hbase.util.Bytes;
-101import 
org.apache.hadoop.hbase.util.ConnectionCache;
-102import 
org.apache.hadoop.hbase.util.DNS;
-103import 
org.apache.hadoop.hbase.util.JvmPauseMonitor;
-104import 
org.apache.hadoop.hbase.util.Strings;
-105import 
org.apache.hadoop.security.SaslRpcServer.SaslGssCallbackHandler;
-106import 
org.apache.hadoop.security.UserGroupInformation;
-107import 
org.apache.hadoop.security.authorize.ProxyUsers;
-108import org.apache.thrift.TException;
-109import org.apache.thrift.TProcessor;
-110import 
org.apache.thrift.protocol.TBinaryProtocol;
-111import 
org.apache.thrift.protocol.TCompactProtocol;
-112import 
org.apache.thrift.protocol.TProtocol;
-113import 
org.apache.thrift.protocol.TProtocolFactory;
-114import 
org.apache.thrift.server.THsHaServer;
-115import 
org.apache.thrift.server.TNonblockingServer;
-116import 
org.apache.thrift.server.TServer;
-117import 
org.apache.thrift.server.TServlet;
-118import 
org.apache.thrift.server.TThreadedSelectorServer;
-119import 
org.apache.thrift.transport.TFramedTransport;
-120import 
org.apache.thrift.transport.TNonblockingServerSocket;
-121import 
org.apache.thrift.transport.TNonblockingServerTransport;
-122import 
org.apache.thrift.transport.TSaslServerTransport;
-123import 
org.apache.thrift.transport.TServerSocket;
-124import 
org.apache.thrift.transport.TServerTransport;
-125import 
org.apache.thrift.transport.TTransportFactory;
-126import 
org.apache.yetus.audience.InterfaceAudience;
-127import 
org.eclipse.jetty.http.HttpVersion;
-128import 
org.eclipse.jetty.server.HttpConfiguration;
-129import 
org.eclipse.jetty.server.HttpConnectionFactory;
-130import 
org.eclipse.jetty.server.SecureRequestCustomizer;
-131import org.eclipse.jetty.server.Server;
-132import 
org.eclipse.jetty.server.ServerConnector;
-133import 
org.eclipse.jetty.server.SslConnectionFactory;
-134import 
org.eclipse.jetty.servlet.ServletContextHandler;
-135import 
org.eclipse.jetty.servlet.ServletHolder;
-136import 
org.eclipse.jetty.util.ssl.SslContextFactory;
-137import 
org.eclipse.jetty.util.thread.QueuedThreadPool;
-138import org.slf4j.Logger;
-139import org.slf4j.LoggerFactory;
-140import 
org.apache.hbase.thirdparty.com.google.common.base.Joiner;
-141import 
org.apache.hbase.thirdparty.com.google.common.base.Throwables;
-142import 
org.apache.hbase.thirdparty.com.google.common.util.concurrent.ThreadFactoryBuilder;
-143import 
org.apache.hbase.thirdparty.org.apache.commons.cli.CommandLine;
-144import 
org.apache.hbase.thirdparty.org.apache.commons.cli.Option;
-145import 
org.apache.hbase.thirdparty.org.apache.commons.cli.OptionGroup;
-146
-147/**
-148 * ThriftServerRunner - this class starts 
up a Thrift server which implements
-149 * the Hbase API specified in the 
Hbase.thrift IDL file.
-150 */
-151@InterfaceAudience.Private
-152public class ThriftServerRunner 
implements Runnable {
-153
-154  private static final Logger LOG = 

[16/22] hbase-site git commit: Published site at 46cb5dfa226892fd2580f26ce9ce77225bd7e67c.

2018-04-21 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8bc57a1a/devapidocs/org/apache/hadoop/hbase/ipc/RpcScheduler.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/ipc/RpcScheduler.html 
b/devapidocs/org/apache/hadoop/hbase/ipc/RpcScheduler.html
index af861b5..1f83c35 100644
--- a/devapidocs/org/apache/hadoop/hbase/ipc/RpcScheduler.html
+++ b/devapidocs/org/apache/hadoop/hbase/ipc/RpcScheduler.html
@@ -18,7 +18,7 @@
 catch(err) {
 }
 //-->
-var methods = 
{"i0":6,"i1":6,"i2":6,"i3":6,"i4":6,"i5":6,"i6":6,"i7":6,"i8":6,"i9":6,"i10":6,"i11":6,"i12":6,"i13":6,"i14":6,"i15":6,"i16":6};
+var methods = 
{"i0":6,"i1":6,"i2":6,"i3":6,"i4":6,"i5":6,"i6":6,"i7":6,"i8":6,"i9":6,"i10":6,"i11":6,"i12":6,"i13":6,"i14":6,"i15":6,"i16":6,"i17":6,"i18":6,"i19":6};
 var tabs = {65535:["t0","All Methods"],2:["t2","Instance 
Methods"],4:["t3","Abstract Methods"]};
 var altColor = "altColor";
 var rowColor = "rowColor";
@@ -205,48 +205,66 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 abstract int
+getActiveGeneralRpcHandlerCount()
+Retrieves the number of active general handler.
+
+
+
+abstract int
+getActivePriorityRpcHandlerCount()
+Retrieves the number of active priority handler.
+
+
+
+abstract int
 getActiveReadRpcHandlerCount()
 Retrieves the number of active write rpc handler when use 
RWQueueRpcExecutor.
 
 
-
+
+abstract int
+getActiveReplicationRpcHandlerCount()
+Retrieves the number of active replication handler.
+
+
+
 abstract int
 getActiveRpcHandlerCount()
-Retrieves the number of active handler.
+Retrieves the total number of active handler.
 
 
-
+
 abstract int
 getActiveScanRpcHandlerCount()
 Retrieves the number of active write rpc handler when use 
RWQueueRpcExecutor.
 
 
-
+
 abstract int
 getActiveWriteRpcHandlerCount()
 Retrieves the number of active write rpc handler when use 
RWQueueRpcExecutor.
 
 
-
+
 abstract CallQueueInfo
 getCallQueueInfo()
 Get call queue information
 
 
-
+
 abstract int
 getGeneralQueueLength()
 Retrieves length of the general queue for metrics.
 
 
-
+
 abstract long
 getNumGeneralCallsDropped()
 If CoDel-based RPC executors are used, retrieves the number 
of Calls that were dropped
  from general queue because RPC executor is under high load; returns 0 
otherwise.
 
 
-
+
 abstract long
 getNumLifoModeSwitches()
 If CoDel-based RPC executors are used, retrieves the number 
of Calls that were
@@ -254,49 +272,49 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
  in the period of overloade we serve last requests first); returns 0 
otherwise.
 
 
-
+
 abstract int
 getPriorityQueueLength()
 Retrieves length of the priority queue for metrics.
 
 
-
+
 abstract int
 getReadQueueLength()
 Retrieves length of the read queue for metrics when use 
RWQueueRpcExecutor.
 
 
-
+
 abstract int
 getReplicationQueueLength()
 Retrieves length of the replication queue for metrics.
 
 
-
+
 abstract int
 getScanQueueLength()
 Retrieves length of the scan queue for metrics when use 
RWQueueRpcExecutor.
 
 
-
+
 abstract int
 getWriteQueueLength()
 Retrieves length of the write queue for metrics when use 
RWQueueRpcExecutor.
 
 
-
+
 abstract void
 init(RpcScheduler.Contextcontext)
 Does some quick initialization.
 
 
-
+
 abstract void
 start()
 Prepares for request serving.
 
 
-
+
 abstract void
 stop()
 Stops serving new requests.
@@ -477,7 +495,37 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 getActiveRpcHandlerCount
 public abstractintgetActiveRpcHandlerCount()
-Retrieves the number of active handler.
+Retrieves the total number of active handler.
+
+
+
+
+
+
+
+getActiveGeneralRpcHandlerCount
+public abstractintgetActiveGeneralRpcHandlerCount()
+Retrieves the number of active general handler.
+
+
+
+
+
+
+
+getActivePriorityRpcHandlerCount
+public abstractintgetActivePriorityRpcHandlerCount()
+Retrieves the number of active priority handler.
+
+
+
+
+
+
+
+getActiveReplicationRpcHandlerCount
+public abstractintgetActiveReplicationRpcHandlerCount()
+Retrieves the number of active replication handler.
 
 
 
@@ -486,7 +534,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 getNumGeneralCallsDropped
-public abstractlonggetNumGeneralCallsDropped()
+public abstractlonggetNumGeneralCallsDropped()
 If CoDel-based RPC executors are used, retrieves the number 
of Calls that were dropped
  from general queue because RPC executor is under high load; returns 0 
otherwise.
 
@@ -497,7 +545,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 getNumLifoModeSwitches
-public abstractlonggetNumLifoModeSwitches()
+public abstractlonggetNumLifoModeSwitches()
 If CoDel-based RPC executors are used, retrieves the number 
of Calls that were
  picked from the tail of the queue (indicating adaptive LIFO mode, when
  in the period of overloade we serve last requests first); returns 

[07/22] hbase-site git commit: Published site at 46cb5dfa226892fd2580f26ce9ce77225bd7e67c.

2018-04-21 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8bc57a1a/devapidocs/src-html/org/apache/hadoop/hbase/thrift/ThriftServerRunner.ImplType.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/thrift/ThriftServerRunner.ImplType.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/thrift/ThriftServerRunner.ImplType.html
index d6f672a..fd7ab7a 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/thrift/ThriftServerRunner.ImplType.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/thrift/ThriftServerRunner.ImplType.html
@@ -87,1949 +87,1951 @@
 079import 
org.apache.hadoop.hbase.filter.ParseFilter;
 080import 
org.apache.hadoop.hbase.filter.PrefixFilter;
 081import 
org.apache.hadoop.hbase.filter.WhileMatchFilter;
-082import 
org.apache.hadoop.hbase.log.HBaseMarkers;
-083import 
org.apache.hadoop.hbase.security.SaslUtil;
-084import 
org.apache.hadoop.hbase.security.SaslUtil.QualityOfProtection;
-085import 
org.apache.hadoop.hbase.security.SecurityUtil;
-086import 
org.apache.hadoop.hbase.security.UserProvider;
-087import 
org.apache.hadoop.hbase.thrift.generated.AlreadyExists;
-088import 
org.apache.hadoop.hbase.thrift.generated.BatchMutation;
-089import 
org.apache.hadoop.hbase.thrift.generated.ColumnDescriptor;
-090import 
org.apache.hadoop.hbase.thrift.generated.Hbase;
-091import 
org.apache.hadoop.hbase.thrift.generated.IOError;
-092import 
org.apache.hadoop.hbase.thrift.generated.IllegalArgument;
-093import 
org.apache.hadoop.hbase.thrift.generated.Mutation;
-094import 
org.apache.hadoop.hbase.thrift.generated.TAppend;
-095import 
org.apache.hadoop.hbase.thrift.generated.TCell;
-096import 
org.apache.hadoop.hbase.thrift.generated.TIncrement;
-097import 
org.apache.hadoop.hbase.thrift.generated.TRegionInfo;
-098import 
org.apache.hadoop.hbase.thrift.generated.TRowResult;
-099import 
org.apache.hadoop.hbase.thrift.generated.TScan;
-100import 
org.apache.hadoop.hbase.util.Bytes;
-101import 
org.apache.hadoop.hbase.util.ConnectionCache;
-102import 
org.apache.hadoop.hbase.util.DNS;
-103import 
org.apache.hadoop.hbase.util.JvmPauseMonitor;
-104import 
org.apache.hadoop.hbase.util.Strings;
-105import 
org.apache.hadoop.security.SaslRpcServer.SaslGssCallbackHandler;
-106import 
org.apache.hadoop.security.UserGroupInformation;
-107import 
org.apache.hadoop.security.authorize.ProxyUsers;
-108import org.apache.thrift.TException;
-109import org.apache.thrift.TProcessor;
-110import 
org.apache.thrift.protocol.TBinaryProtocol;
-111import 
org.apache.thrift.protocol.TCompactProtocol;
-112import 
org.apache.thrift.protocol.TProtocol;
-113import 
org.apache.thrift.protocol.TProtocolFactory;
-114import 
org.apache.thrift.server.THsHaServer;
-115import 
org.apache.thrift.server.TNonblockingServer;
-116import 
org.apache.thrift.server.TServer;
-117import 
org.apache.thrift.server.TServlet;
-118import 
org.apache.thrift.server.TThreadedSelectorServer;
-119import 
org.apache.thrift.transport.TFramedTransport;
-120import 
org.apache.thrift.transport.TNonblockingServerSocket;
-121import 
org.apache.thrift.transport.TNonblockingServerTransport;
-122import 
org.apache.thrift.transport.TSaslServerTransport;
-123import 
org.apache.thrift.transport.TServerSocket;
-124import 
org.apache.thrift.transport.TServerTransport;
-125import 
org.apache.thrift.transport.TTransportFactory;
-126import 
org.apache.yetus.audience.InterfaceAudience;
-127import 
org.eclipse.jetty.http.HttpVersion;
-128import 
org.eclipse.jetty.server.HttpConfiguration;
-129import 
org.eclipse.jetty.server.HttpConnectionFactory;
-130import 
org.eclipse.jetty.server.SecureRequestCustomizer;
-131import org.eclipse.jetty.server.Server;
-132import 
org.eclipse.jetty.server.ServerConnector;
-133import 
org.eclipse.jetty.server.SslConnectionFactory;
-134import 
org.eclipse.jetty.servlet.ServletContextHandler;
-135import 
org.eclipse.jetty.servlet.ServletHolder;
-136import 
org.eclipse.jetty.util.ssl.SslContextFactory;
-137import 
org.eclipse.jetty.util.thread.QueuedThreadPool;
-138import org.slf4j.Logger;
-139import org.slf4j.LoggerFactory;
-140import 
org.apache.hbase.thirdparty.com.google.common.base.Joiner;
-141import 
org.apache.hbase.thirdparty.com.google.common.base.Throwables;
-142import 
org.apache.hbase.thirdparty.com.google.common.util.concurrent.ThreadFactoryBuilder;
-143import 
org.apache.hbase.thirdparty.org.apache.commons.cli.CommandLine;
-144import 
org.apache.hbase.thirdparty.org.apache.commons.cli.Option;
-145import 
org.apache.hbase.thirdparty.org.apache.commons.cli.OptionGroup;
-146
-147/**
-148 * ThriftServerRunner - this class starts 
up a Thrift server which implements
-149 * the Hbase API specified in the 
Hbase.thrift IDL file.
-150 */
-151@InterfaceAudience.Private
-152public class ThriftServerRunner 
implements Runnable {
-153
-154  private static final Logger LOG = 
LoggerFactory.getLogger(ThriftServerRunner.class);
-155
-156  private static final int 

[13/22] hbase-site git commit: Published site at 46cb5dfa226892fd2580f26ce9ce77225bd7e67c.

2018-04-21 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8bc57a1a/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.IOErrorWithCause.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.IOErrorWithCause.html
 
b/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.IOErrorWithCause.html
index 23b678a..c2d61b1 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.IOErrorWithCause.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.IOErrorWithCause.html
@@ -137,7 +137,7 @@ var activeTableTab = "activeTableTab";
 
 
 
-public static class ThriftServerRunner.IOErrorWithCause
+public static class ThriftServerRunner.IOErrorWithCause
 extends org.apache.hadoop.hbase.thrift.generated.IOError
 
 See Also:
@@ -273,7 +273,7 @@ extends 
org.apache.hadoop.hbase.thrift.generated.IOError
 
 
 cause
-privatehttps://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true;
 title="class or interface in java.lang">Throwable cause
+privatehttps://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true;
 title="class or interface in java.lang">Throwable cause
 
 
 
@@ -290,7 +290,7 @@ extends 
org.apache.hadoop.hbase.thrift.generated.IOError
 
 
 IOErrorWithCause
-publicIOErrorWithCause(https://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true;
 title="class or interface in java.lang">Throwablecause)
+publicIOErrorWithCause(https://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true;
 title="class or interface in java.lang">Throwablecause)
 
 
 
@@ -307,7 +307,7 @@ extends 
org.apache.hadoop.hbase.thrift.generated.IOError
 
 
 getCause
-publichttps://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true;
 title="class or interface in java.lang">ThrowablegetCause()
+publichttps://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true;
 title="class or interface in java.lang">ThrowablegetCause()
 
 Overrides:
 https://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true#getCause--;
 title="class or interface in java.lang">getCausein 
classhttps://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true;
 title="class or interface in java.lang">Throwable
@@ -320,7 +320,7 @@ extends 
org.apache.hadoop.hbase.thrift.generated.IOError
 
 
 equals
-publicbooleanequals(https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
 title="class or interface in java.lang">Objectother)
+publicbooleanequals(https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
 title="class or interface in java.lang">Objectother)
 
 Overrides:
 equalsin 
classorg.apache.hadoop.hbase.thrift.generated.IOError
@@ -333,7 +333,7 @@ extends 
org.apache.hadoop.hbase.thrift.generated.IOError
 
 
 hashCode
-publicinthashCode()
+publicinthashCode()
 
 Overrides:
 hashCodein 
classorg.apache.hadoop.hbase.thrift.generated.IOError

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8bc57a1a/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.ImplType.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.ImplType.html 
b/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.ImplType.html
index 82d106c..0392c62 100644
--- a/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.ImplType.html
+++ b/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.ImplType.html
@@ -122,7 +122,7 @@ var activeTableTab = "activeTableTab";
 
 
 
-static enum ThriftServerRunner.ImplType
+static enum ThriftServerRunner.ImplType
 extends https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true;
 title="class or interface in java.lang">EnumThriftServerRunner.ImplType
 An enum of server implementation selections
 
@@ -281,7 +281,7 @@ the order they are declared.
 
 
 HS_HA
-public static finalThriftServerRunner.ImplType 
HS_HA
+public static finalThriftServerRunner.ImplType 
HS_HA
 
 
 
@@ -290,7 +290,7 @@ the order they are declared.
 
 
 NONBLOCKING
-public static finalThriftServerRunner.ImplType 
NONBLOCKING
+public static finalThriftServerRunner.ImplType 
NONBLOCKING
 
 
 
@@ -299,7 +299,7 @@ the order they are declared.
 
 
 THREAD_POOL
-public static finalThriftServerRunner.ImplType 
THREAD_POOL
+public static finalThriftServerRunner.ImplType 
THREAD_POOL
 
 
 
@@ -308,7 +308,7 @@ the order they are declared.
 
 
 THREADED_SELECTOR
-public static finalThriftServerRunner.ImplType 
THREADED_SELECTOR
+public static finalThriftServerRunner.ImplType 
THREADED_SELECTOR
 
 
 
@@ -325,7 +325,7 @@ the order they are declared.
 
 
 DEFAULT
-public static finalThriftServerRunner.ImplType 
DEFAULT
+public static finalThriftServerRunner.ImplType 
DEFAULT
 
 
 
@@ -334,7 +334,7 

[08/22] hbase-site git commit: Published site at 46cb5dfa226892fd2580f26ce9ce77225bd7e67c.

2018-04-21 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8bc57a1a/devapidocs/src-html/org/apache/hadoop/hbase/thrift/ThriftServerRunner.IOErrorWithCause.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/thrift/ThriftServerRunner.IOErrorWithCause.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/thrift/ThriftServerRunner.IOErrorWithCause.html
index d6f672a..fd7ab7a 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/thrift/ThriftServerRunner.IOErrorWithCause.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/thrift/ThriftServerRunner.IOErrorWithCause.html
@@ -87,1949 +87,1951 @@
 079import 
org.apache.hadoop.hbase.filter.ParseFilter;
 080import 
org.apache.hadoop.hbase.filter.PrefixFilter;
 081import 
org.apache.hadoop.hbase.filter.WhileMatchFilter;
-082import 
org.apache.hadoop.hbase.log.HBaseMarkers;
-083import 
org.apache.hadoop.hbase.security.SaslUtil;
-084import 
org.apache.hadoop.hbase.security.SaslUtil.QualityOfProtection;
-085import 
org.apache.hadoop.hbase.security.SecurityUtil;
-086import 
org.apache.hadoop.hbase.security.UserProvider;
-087import 
org.apache.hadoop.hbase.thrift.generated.AlreadyExists;
-088import 
org.apache.hadoop.hbase.thrift.generated.BatchMutation;
-089import 
org.apache.hadoop.hbase.thrift.generated.ColumnDescriptor;
-090import 
org.apache.hadoop.hbase.thrift.generated.Hbase;
-091import 
org.apache.hadoop.hbase.thrift.generated.IOError;
-092import 
org.apache.hadoop.hbase.thrift.generated.IllegalArgument;
-093import 
org.apache.hadoop.hbase.thrift.generated.Mutation;
-094import 
org.apache.hadoop.hbase.thrift.generated.TAppend;
-095import 
org.apache.hadoop.hbase.thrift.generated.TCell;
-096import 
org.apache.hadoop.hbase.thrift.generated.TIncrement;
-097import 
org.apache.hadoop.hbase.thrift.generated.TRegionInfo;
-098import 
org.apache.hadoop.hbase.thrift.generated.TRowResult;
-099import 
org.apache.hadoop.hbase.thrift.generated.TScan;
-100import 
org.apache.hadoop.hbase.util.Bytes;
-101import 
org.apache.hadoop.hbase.util.ConnectionCache;
-102import 
org.apache.hadoop.hbase.util.DNS;
-103import 
org.apache.hadoop.hbase.util.JvmPauseMonitor;
-104import 
org.apache.hadoop.hbase.util.Strings;
-105import 
org.apache.hadoop.security.SaslRpcServer.SaslGssCallbackHandler;
-106import 
org.apache.hadoop.security.UserGroupInformation;
-107import 
org.apache.hadoop.security.authorize.ProxyUsers;
-108import org.apache.thrift.TException;
-109import org.apache.thrift.TProcessor;
-110import 
org.apache.thrift.protocol.TBinaryProtocol;
-111import 
org.apache.thrift.protocol.TCompactProtocol;
-112import 
org.apache.thrift.protocol.TProtocol;
-113import 
org.apache.thrift.protocol.TProtocolFactory;
-114import 
org.apache.thrift.server.THsHaServer;
-115import 
org.apache.thrift.server.TNonblockingServer;
-116import 
org.apache.thrift.server.TServer;
-117import 
org.apache.thrift.server.TServlet;
-118import 
org.apache.thrift.server.TThreadedSelectorServer;
-119import 
org.apache.thrift.transport.TFramedTransport;
-120import 
org.apache.thrift.transport.TNonblockingServerSocket;
-121import 
org.apache.thrift.transport.TNonblockingServerTransport;
-122import 
org.apache.thrift.transport.TSaslServerTransport;
-123import 
org.apache.thrift.transport.TServerSocket;
-124import 
org.apache.thrift.transport.TServerTransport;
-125import 
org.apache.thrift.transport.TTransportFactory;
-126import 
org.apache.yetus.audience.InterfaceAudience;
-127import 
org.eclipse.jetty.http.HttpVersion;
-128import 
org.eclipse.jetty.server.HttpConfiguration;
-129import 
org.eclipse.jetty.server.HttpConnectionFactory;
-130import 
org.eclipse.jetty.server.SecureRequestCustomizer;
-131import org.eclipse.jetty.server.Server;
-132import 
org.eclipse.jetty.server.ServerConnector;
-133import 
org.eclipse.jetty.server.SslConnectionFactory;
-134import 
org.eclipse.jetty.servlet.ServletContextHandler;
-135import 
org.eclipse.jetty.servlet.ServletHolder;
-136import 
org.eclipse.jetty.util.ssl.SslContextFactory;
-137import 
org.eclipse.jetty.util.thread.QueuedThreadPool;
-138import org.slf4j.Logger;
-139import org.slf4j.LoggerFactory;
-140import 
org.apache.hbase.thirdparty.com.google.common.base.Joiner;
-141import 
org.apache.hbase.thirdparty.com.google.common.base.Throwables;
-142import 
org.apache.hbase.thirdparty.com.google.common.util.concurrent.ThreadFactoryBuilder;
-143import 
org.apache.hbase.thirdparty.org.apache.commons.cli.CommandLine;
-144import 
org.apache.hbase.thirdparty.org.apache.commons.cli.Option;
-145import 
org.apache.hbase.thirdparty.org.apache.commons.cli.OptionGroup;
-146
-147/**
-148 * ThriftServerRunner - this class starts 
up a Thrift server which implements
-149 * the Hbase API specified in the 
Hbase.thrift IDL file.
-150 */
-151@InterfaceAudience.Private
-152public class ThriftServerRunner 
implements Runnable {
-153
-154  private static final Logger LOG = 

[02/22] hbase-site git commit: Published site at 46cb5dfa226892fd2580f26ce9ce77225bd7e67c.

2018-04-21 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8bc57a1a/testdevapidocs/src-html/org/apache/hadoop/hbase/http/TestHttpServer.HtmlContentServlet.html
--
diff --git 
a/testdevapidocs/src-html/org/apache/hadoop/hbase/http/TestHttpServer.HtmlContentServlet.html
 
b/testdevapidocs/src-html/org/apache/hadoop/hbase/http/TestHttpServer.HtmlContentServlet.html
index 5681a20..cbcaf55 100644
--- 
a/testdevapidocs/src-html/org/apache/hadoop/hbase/http/TestHttpServer.HtmlContentServlet.html
+++ 
b/testdevapidocs/src-html/org/apache/hadoop/hbase/http/TestHttpServer.HtmlContentServlet.html
@@ -613,21 +613,30 @@
 605myServer.stop();
 606  }
 607
-608
-609
-610  @Test
-611  public void testNoCacheHeader() throws 
Exception {
-612URL url = new URL(baseUrl, 
"/echo?a=bc=d");
-613HttpURLConnection conn = 
(HttpURLConnection) url.openConnection();
-614
assertEquals(HttpURLConnection.HTTP_OK, conn.getResponseCode());
-615assertEquals("no-cache", 
conn.getHeaderField("Cache-Control"));
-616assertEquals("no-cache", 
conn.getHeaderField("Pragma"));
-617
assertNotNull(conn.getHeaderField("Expires"));
-618
assertNotNull(conn.getHeaderField("Date"));
-619
assertEquals(conn.getHeaderField("Expires"), conn.getHeaderField("Date"));
-620assertEquals("DENY", 
conn.getHeaderField("X-Frame-Options"));
-621  }
-622}
+608  @Test
+609  public void testNoCacheHeader() throws 
Exception {
+610URL url = new URL(baseUrl, 
"/echo?a=bc=d");
+611HttpURLConnection conn = 
(HttpURLConnection) url.openConnection();
+612
assertEquals(HttpURLConnection.HTTP_OK, conn.getResponseCode());
+613assertEquals("no-cache", 
conn.getHeaderField("Cache-Control"));
+614assertEquals("no-cache", 
conn.getHeaderField("Pragma"));
+615
assertNotNull(conn.getHeaderField("Expires"));
+616
assertNotNull(conn.getHeaderField("Date"));
+617
assertEquals(conn.getHeaderField("Expires"), conn.getHeaderField("Date"));
+618assertEquals("DENY", 
conn.getHeaderField("X-Frame-Options"));
+619  }
+620
+621  @Test
+622  public void testHttpMethods() throws 
Exception {
+623// HTTP TRACE method should be 
disabled for security
+624// See 
https://www.owasp.org/index.php/Cross_Site_Tracing
+625URL url = new URL(baseUrl, 
"/echo?a=b");
+626HttpURLConnection conn = 
(HttpURLConnection) url.openConnection();
+627conn.setRequestMethod("TRACE");
+628conn.connect();
+629
assertEquals(HttpURLConnection.HTTP_FORBIDDEN, conn.getResponseCode());
+630  }
+631}
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8bc57a1a/testdevapidocs/src-html/org/apache/hadoop/hbase/http/TestHttpServer.LongHeaderServlet.html
--
diff --git 
a/testdevapidocs/src-html/org/apache/hadoop/hbase/http/TestHttpServer.LongHeaderServlet.html
 
b/testdevapidocs/src-html/org/apache/hadoop/hbase/http/TestHttpServer.LongHeaderServlet.html
index 5681a20..cbcaf55 100644
--- 
a/testdevapidocs/src-html/org/apache/hadoop/hbase/http/TestHttpServer.LongHeaderServlet.html
+++ 
b/testdevapidocs/src-html/org/apache/hadoop/hbase/http/TestHttpServer.LongHeaderServlet.html
@@ -613,21 +613,30 @@
 605myServer.stop();
 606  }
 607
-608
-609
-610  @Test
-611  public void testNoCacheHeader() throws 
Exception {
-612URL url = new URL(baseUrl, 
"/echo?a=bc=d");
-613HttpURLConnection conn = 
(HttpURLConnection) url.openConnection();
-614
assertEquals(HttpURLConnection.HTTP_OK, conn.getResponseCode());
-615assertEquals("no-cache", 
conn.getHeaderField("Cache-Control"));
-616assertEquals("no-cache", 
conn.getHeaderField("Pragma"));
-617
assertNotNull(conn.getHeaderField("Expires"));
-618
assertNotNull(conn.getHeaderField("Date"));
-619
assertEquals(conn.getHeaderField("Expires"), conn.getHeaderField("Date"));
-620assertEquals("DENY", 
conn.getHeaderField("X-Frame-Options"));
-621  }
-622}
+608  @Test
+609  public void testNoCacheHeader() throws 
Exception {
+610URL url = new URL(baseUrl, 
"/echo?a=bc=d");
+611HttpURLConnection conn = 
(HttpURLConnection) url.openConnection();
+612
assertEquals(HttpURLConnection.HTTP_OK, conn.getResponseCode());
+613assertEquals("no-cache", 
conn.getHeaderField("Cache-Control"));
+614assertEquals("no-cache", 
conn.getHeaderField("Pragma"));
+615
assertNotNull(conn.getHeaderField("Expires"));
+616
assertNotNull(conn.getHeaderField("Date"));
+617
assertEquals(conn.getHeaderField("Expires"), conn.getHeaderField("Date"));
+618assertEquals("DENY", 
conn.getHeaderField("X-Frame-Options"));
+619  }
+620
+621  @Test
+622  public void testHttpMethods() throws 
Exception {
+623// HTTP TRACE method should be 
disabled for security
+624// See 
https://www.owasp.org/index.php/Cross_Site_Tracing
+625URL url = new URL(baseUrl, 
"/echo?a=b");
+626HttpURLConnection conn = 
(HttpURLConnection) 

[14/22] hbase-site git commit: Published site at 46cb5dfa226892fd2580f26ce9ce77225bd7e67c.

2018-04-21 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8bc57a1a/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.HBaseHandler.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.HBaseHandler.html
 
b/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.HBaseHandler.html
index aeebaec..f338c1a 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.HBaseHandler.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.HBaseHandler.html
@@ -117,7 +117,7 @@ var activeTableTab = "activeTableTab";
 
 
 
-public static class ThriftServerRunner.HBaseHandler
+public static class ThriftServerRunner.HBaseHandler
 extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
 title="class or interface in java.lang">Object
 implements org.apache.hadoop.hbase.thrift.generated.Hbase.Iface
 The HBaseHandler is a glue object that connects Thrift RPC 
calls to the
@@ -721,7 +721,7 @@ implements 
org.apache.hadoop.hbase.thrift.generated.Hbase.Iface
 
 
 conf
-protectedorg.apache.hadoop.conf.Configuration conf
+protectedorg.apache.hadoop.conf.Configuration conf
 
 
 
@@ -730,7 +730,7 @@ implements 
org.apache.hadoop.hbase.thrift.generated.Hbase.Iface
 
 
 LOG
-protected static finalorg.slf4j.Logger LOG
+protected static finalorg.slf4j.Logger LOG
 
 
 
@@ -739,7 +739,7 @@ implements 
org.apache.hadoop.hbase.thrift.generated.Hbase.Iface
 
 
 nextScannerId
-protectedint nextScannerId
+protectedint nextScannerId
 
 
 
@@ -748,7 +748,7 @@ implements 
org.apache.hadoop.hbase.thrift.generated.Hbase.Iface
 
 
 scannerMap
-protectedhttps://docs.oracle.com/javase/8/docs/api/java/util/HashMap.html?is-external=true;
 title="class or interface in java.util">HashMaphttps://docs.oracle.com/javase/8/docs/api/java/lang/Integer.html?is-external=true;
 title="class or interface in java.lang">Integer,ThriftServerRunner.ResultScannerWrapper 
scannerMap
+protectedhttps://docs.oracle.com/javase/8/docs/api/java/util/HashMap.html?is-external=true;
 title="class or interface in java.util">HashMaphttps://docs.oracle.com/javase/8/docs/api/java/lang/Integer.html?is-external=true;
 title="class or interface in java.lang">Integer,ThriftServerRunner.ResultScannerWrapper 
scannerMap
 
 
 
@@ -757,7 +757,7 @@ implements 
org.apache.hadoop.hbase.thrift.generated.Hbase.Iface
 
 
 metrics
-privateThriftMetrics metrics
+privateThriftMetrics metrics
 
 
 
@@ -766,7 +766,7 @@ implements 
org.apache.hadoop.hbase.thrift.generated.Hbase.Iface
 
 
 connectionCache
-private finalConnectionCache connectionCache
+private finalConnectionCache connectionCache
 
 
 
@@ -775,7 +775,7 @@ implements 
org.apache.hadoop.hbase.thrift.generated.Hbase.Iface
 
 
 coalescer
-IncrementCoalescer coalescer
+IncrementCoalescer coalescer
 
 
 
@@ -784,7 +784,7 @@ implements 
org.apache.hadoop.hbase.thrift.generated.Hbase.Iface
 
 
 CLEANUP_INTERVAL
-static finalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String CLEANUP_INTERVAL
+static finalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String CLEANUP_INTERVAL
 
 See Also:
 Constant
 Field Values
@@ -797,7 +797,7 @@ implements 
org.apache.hadoop.hbase.thrift.generated.Hbase.Iface
 
 
 MAX_IDLETIME
-static finalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String MAX_IDLETIME
+static finalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String MAX_IDLETIME
 
 See Also:
 Constant
 Field Values
@@ -818,7 +818,7 @@ implements 
org.apache.hadoop.hbase.thrift.generated.Hbase.Iface
 
 
 HBaseHandler
-protectedHBaseHandler(org.apache.hadoop.conf.Configurationc,
+protectedHBaseHandler(org.apache.hadoop.conf.Configurationc,
UserProvideruserProvider)
 throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
 title="class or interface in java.io">IOException
 
@@ -841,7 +841,7 @@ implements 
org.apache.hadoop.hbase.thrift.generated.Hbase.Iface
 
 
 getAllColumns
-byte[][]getAllColumns(Tabletable)
+byte[][]getAllColumns(Tabletable)
 throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
 title="class or interface in java.io">IOException
 Returns a list of all the column families for a given 
Table.
 
@@ -858,7 +858,7 @@ implements 
org.apache.hadoop.hbase.thrift.generated.Hbase.Iface
 
 
 getTable
-publicTablegetTable(byte[]tableName)
+publicTablegetTable(byte[]tableName)
throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
 title="class or interface in java.io">IOException
 Creates and 

[05/22] hbase-site git commit: Published site at 46cb5dfa226892fd2580f26ce9ce77225bd7e67c.

2018-04-21 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8bc57a1a/devapidocs/src-html/org/apache/hadoop/hbase/thrift/ThriftServerRunner.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/thrift/ThriftServerRunner.html 
b/devapidocs/src-html/org/apache/hadoop/hbase/thrift/ThriftServerRunner.html
index d6f672a..fd7ab7a 100644
--- a/devapidocs/src-html/org/apache/hadoop/hbase/thrift/ThriftServerRunner.html
+++ b/devapidocs/src-html/org/apache/hadoop/hbase/thrift/ThriftServerRunner.html
@@ -87,1949 +87,1951 @@
 079import 
org.apache.hadoop.hbase.filter.ParseFilter;
 080import 
org.apache.hadoop.hbase.filter.PrefixFilter;
 081import 
org.apache.hadoop.hbase.filter.WhileMatchFilter;
-082import 
org.apache.hadoop.hbase.log.HBaseMarkers;
-083import 
org.apache.hadoop.hbase.security.SaslUtil;
-084import 
org.apache.hadoop.hbase.security.SaslUtil.QualityOfProtection;
-085import 
org.apache.hadoop.hbase.security.SecurityUtil;
-086import 
org.apache.hadoop.hbase.security.UserProvider;
-087import 
org.apache.hadoop.hbase.thrift.generated.AlreadyExists;
-088import 
org.apache.hadoop.hbase.thrift.generated.BatchMutation;
-089import 
org.apache.hadoop.hbase.thrift.generated.ColumnDescriptor;
-090import 
org.apache.hadoop.hbase.thrift.generated.Hbase;
-091import 
org.apache.hadoop.hbase.thrift.generated.IOError;
-092import 
org.apache.hadoop.hbase.thrift.generated.IllegalArgument;
-093import 
org.apache.hadoop.hbase.thrift.generated.Mutation;
-094import 
org.apache.hadoop.hbase.thrift.generated.TAppend;
-095import 
org.apache.hadoop.hbase.thrift.generated.TCell;
-096import 
org.apache.hadoop.hbase.thrift.generated.TIncrement;
-097import 
org.apache.hadoop.hbase.thrift.generated.TRegionInfo;
-098import 
org.apache.hadoop.hbase.thrift.generated.TRowResult;
-099import 
org.apache.hadoop.hbase.thrift.generated.TScan;
-100import 
org.apache.hadoop.hbase.util.Bytes;
-101import 
org.apache.hadoop.hbase.util.ConnectionCache;
-102import 
org.apache.hadoop.hbase.util.DNS;
-103import 
org.apache.hadoop.hbase.util.JvmPauseMonitor;
-104import 
org.apache.hadoop.hbase.util.Strings;
-105import 
org.apache.hadoop.security.SaslRpcServer.SaslGssCallbackHandler;
-106import 
org.apache.hadoop.security.UserGroupInformation;
-107import 
org.apache.hadoop.security.authorize.ProxyUsers;
-108import org.apache.thrift.TException;
-109import org.apache.thrift.TProcessor;
-110import 
org.apache.thrift.protocol.TBinaryProtocol;
-111import 
org.apache.thrift.protocol.TCompactProtocol;
-112import 
org.apache.thrift.protocol.TProtocol;
-113import 
org.apache.thrift.protocol.TProtocolFactory;
-114import 
org.apache.thrift.server.THsHaServer;
-115import 
org.apache.thrift.server.TNonblockingServer;
-116import 
org.apache.thrift.server.TServer;
-117import 
org.apache.thrift.server.TServlet;
-118import 
org.apache.thrift.server.TThreadedSelectorServer;
-119import 
org.apache.thrift.transport.TFramedTransport;
-120import 
org.apache.thrift.transport.TNonblockingServerSocket;
-121import 
org.apache.thrift.transport.TNonblockingServerTransport;
-122import 
org.apache.thrift.transport.TSaslServerTransport;
-123import 
org.apache.thrift.transport.TServerSocket;
-124import 
org.apache.thrift.transport.TServerTransport;
-125import 
org.apache.thrift.transport.TTransportFactory;
-126import 
org.apache.yetus.audience.InterfaceAudience;
-127import 
org.eclipse.jetty.http.HttpVersion;
-128import 
org.eclipse.jetty.server.HttpConfiguration;
-129import 
org.eclipse.jetty.server.HttpConnectionFactory;
-130import 
org.eclipse.jetty.server.SecureRequestCustomizer;
-131import org.eclipse.jetty.server.Server;
-132import 
org.eclipse.jetty.server.ServerConnector;
-133import 
org.eclipse.jetty.server.SslConnectionFactory;
-134import 
org.eclipse.jetty.servlet.ServletContextHandler;
-135import 
org.eclipse.jetty.servlet.ServletHolder;
-136import 
org.eclipse.jetty.util.ssl.SslContextFactory;
-137import 
org.eclipse.jetty.util.thread.QueuedThreadPool;
-138import org.slf4j.Logger;
-139import org.slf4j.LoggerFactory;
-140import 
org.apache.hbase.thirdparty.com.google.common.base.Joiner;
-141import 
org.apache.hbase.thirdparty.com.google.common.base.Throwables;
-142import 
org.apache.hbase.thirdparty.com.google.common.util.concurrent.ThreadFactoryBuilder;
-143import 
org.apache.hbase.thirdparty.org.apache.commons.cli.CommandLine;
-144import 
org.apache.hbase.thirdparty.org.apache.commons.cli.Option;
-145import 
org.apache.hbase.thirdparty.org.apache.commons.cli.OptionGroup;
-146
-147/**
-148 * ThriftServerRunner - this class starts 
up a Thrift server which implements
-149 * the Hbase API specified in the 
Hbase.thrift IDL file.
-150 */
-151@InterfaceAudience.Private
-152public class ThriftServerRunner 
implements Runnable {
-153
-154  private static final Logger LOG = 
LoggerFactory.getLogger(ThriftServerRunner.class);
-155
-156  private static final int 
DEFAULT_HTTP_MAX_HEADER_SIZE = 64 * 1024; // 

[10/22] hbase-site git commit: Published site at 46cb5dfa226892fd2580f26ce9ce77225bd7e67c.

2018-04-21 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8bc57a1a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/StoreFileInfo.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/StoreFileInfo.html 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/StoreFileInfo.html
index c467ef4..d4b94dd 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/StoreFileInfo.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/StoreFileInfo.html
@@ -165,427 +165,442 @@
 157
 158  /**
 159   * Create a Store File Info from an 
HFileLink
-160   * @param conf the {@link 
Configuration} to use
-161   * @param fs The current file system to 
use.
+160   * @param conf The {@link 
Configuration} to use
+161   * @param fs The current file system to 
use
 162   * @param fileStatus The {@link 
FileStatus} of the file
 163   */
 164  public StoreFileInfo(final 
Configuration conf, final FileSystem fs, final FileStatus fileStatus,
-165  final HFileLink link)
-166  throws IOException {
-167this.fs = fs;
-168this.conf = conf;
-169// initialPath can be null only if we 
get a link.
-170this.initialPath = (fileStatus == 
null) ? null : fileStatus.getPath();
-171  // HFileLink
-172this.reference = null;
-173this.link = link;
-174  }
-175
-176  /**
-177   * Create a Store File Info from an 
HFileLink
-178   * @param conf
-179   * @param fs
-180   * @param fileStatus
-181   * @param reference
-182   * @throws IOException
-183   */
-184  public StoreFileInfo(final 
Configuration conf, final FileSystem fs, final FileStatus fileStatus,
-185  final Reference reference)
-186  throws IOException {
-187this.fs = fs;
-188this.conf = conf;
-189this.initialPath = 
fileStatus.getPath();
-190this.createdTimestamp = 
fileStatus.getModificationTime();
-191this.reference = reference;
-192this.link = null;
-193  }
-194
-195  /**
-196   * Sets the region coprocessor env.
-197   * @param coprocessorHost
-198   */
-199  public void 
setRegionCoprocessorHost(RegionCoprocessorHost coprocessorHost) {
-200this.coprocessorHost = 
coprocessorHost;
-201  }
-202
-203  /*
-204   * @return the Reference object 
associated to this StoreFileInfo.
-205   * null if the StoreFile is not 
a reference.
-206   */
-207  public Reference getReference() {
-208return this.reference;
-209  }
-210
-211  /** @return True if the store file is a 
Reference */
-212  public boolean isReference() {
-213return this.reference != null;
-214  }
-215
-216  /** @return True if the store file is a 
top Reference */
-217  public boolean isTopReference() {
-218return this.reference != null 
 Reference.isTopFileRegion(this.reference.getFileRegion());
-219  }
-220
-221  /** @return True if the store file is a 
link */
-222  public boolean isLink() {
-223return this.link != null  
this.reference == null;
+165  final HFileLink link) {
+166this.fs = fs;
+167this.conf = conf;
+168// initialPath can be null only if we 
get a link.
+169this.initialPath = (fileStatus == 
null) ? null : fileStatus.getPath();
+170  // HFileLink
+171this.reference = null;
+172this.link = link;
+173  }
+174
+175  /**
+176   * Create a Store File Info from an 
HFileLink
+177   * @param conf The {@link 
Configuration} to use
+178   * @param fs The current file system to 
use
+179   * @param fileStatus The {@link 
FileStatus} of the file
+180   * @param reference The reference 
instance
+181   */
+182  public StoreFileInfo(final 
Configuration conf, final FileSystem fs, final FileStatus fileStatus,
+183  final Reference reference) {
+184this.fs = fs;
+185this.conf = conf;
+186this.initialPath = 
fileStatus.getPath();
+187this.createdTimestamp = 
fileStatus.getModificationTime();
+188this.reference = reference;
+189this.link = null;
+190  }
+191
+192  /**
+193   * Create a Store File Info from an 
HFileLink and a Reference
+194   * @param conf The {@link 
Configuration} to use
+195   * @param fs The current file system to 
use
+196   * @param fileStatus The {@link 
FileStatus} of the file
+197   * @param reference The reference 
instance
+198   * @param link The link instance
+199   */
+200  public StoreFileInfo(final 
Configuration conf, final FileSystem fs, final FileStatus fileStatus,
+201  final Reference reference, final 
HFileLink link) {
+202this.fs = fs;
+203this.conf = conf;
+204this.initialPath = 
fileStatus.getPath();
+205this.createdTimestamp = 
fileStatus.getModificationTime();
+206this.reference = reference;
+207this.link = link;
+208  }
+209
+210  /**
+211   * Sets the region coprocessor env.
+212   * @param coprocessorHost
+213   */
+214  public void 
setRegionCoprocessorHost(RegionCoprocessorHost coprocessorHost) {
+215this.coprocessorHost = 
coprocessorHost;
+216  }
+217
+218  /*
+219   * 

[03/22] hbase-site git commit: Published site at 46cb5dfa226892fd2580f26ce9ce77225bd7e67c.

2018-04-21 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8bc57a1a/testdevapidocs/org/apache/hadoop/hbase/ipc/MetricsHBaseServerWrapperStub.html
--
diff --git 
a/testdevapidocs/org/apache/hadoop/hbase/ipc/MetricsHBaseServerWrapperStub.html 
b/testdevapidocs/org/apache/hadoop/hbase/ipc/MetricsHBaseServerWrapperStub.html
index ba8e14f..75e6b86 100644
--- 
a/testdevapidocs/org/apache/hadoop/hbase/ipc/MetricsHBaseServerWrapperStub.html
+++ 
b/testdevapidocs/org/apache/hadoop/hbase/ipc/MetricsHBaseServerWrapperStub.html
@@ -18,7 +18,7 @@
 catch(err) {
 }
 //-->
-var methods = 
{"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10};
+var methods = 
{"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10};
 var tabs = {65535:["t0","All Methods"],2:["t2","Instance 
Methods"],8:["t4","Concrete Methods"]};
 var altColor = "altColor";
 var rowColor = "rowColor";
@@ -153,57 +153,69 @@ implements 
org.apache.hadoop.hbase.ipc.MetricsHBaseServerWrapper
 
 
 int
-getActiveReadRpcHandlerCount()
+getActiveGeneralRpcHandlerCount()
 
 
 int
-getActiveRpcHandlerCount()
+getActivePriorityRpcHandlerCount()
 
 
 int
-getActiveScanRpcHandlerCount()
+getActiveReadRpcHandlerCount()
 
 
 int
-getActiveWriteRpcHandlerCount()
+getActiveReplicationRpcHandlerCount()
 
 
 int
-getGeneralQueueLength()
+getActiveRpcHandlerCount()
 
 
+int
+getActiveScanRpcHandlerCount()
+
+
+int
+getActiveWriteRpcHandlerCount()
+
+
+int
+getGeneralQueueLength()
+
+
 long
 getNumGeneralCallsDropped()
 
-
+
 long
 getNumLifoModeSwitches()
 
-
+
 int
 getNumOpenConnections()
 
-
+
 int
 getPriorityQueueLength()
 
-
+
 int
 getReadQueueLength()
 
-
+
 int
 getReplicationQueueLength()
 
-
+
 int
 getScanQueueLength()
 
-
+
 long
 getTotalQueueSize()
 
-
+
 int
 getWriteQueueLength()
 
@@ -324,13 +336,52 @@ implements 
org.apache.hadoop.hbase.ipc.MetricsHBaseServerWrapper
 
 
 
+
+
+
+
+
+getActiveGeneralRpcHandlerCount
+publicintgetActiveGeneralRpcHandlerCount()
+
+Specified by:
+getActiveGeneralRpcHandlerCountin 
interfaceorg.apache.hadoop.hbase.ipc.MetricsHBaseServerWrapper
+
+
+
+
+
+
+
+
+getActivePriorityRpcHandlerCount
+publicintgetActivePriorityRpcHandlerCount()
+
+Specified by:
+getActivePriorityRpcHandlerCountin 
interfaceorg.apache.hadoop.hbase.ipc.MetricsHBaseServerWrapper
+
+
+
+
+
+
+
+
+getActiveReplicationRpcHandlerCount
+publicintgetActiveReplicationRpcHandlerCount()
+
+Specified by:
+getActiveReplicationRpcHandlerCountin 
interfaceorg.apache.hadoop.hbase.ipc.MetricsHBaseServerWrapper
+
+
+
 
 
 
 
 
 getNumGeneralCallsDropped
-publiclonggetNumGeneralCallsDropped()
+publiclonggetNumGeneralCallsDropped()
 
 Specified by:
 getNumGeneralCallsDroppedin 
interfaceorg.apache.hadoop.hbase.ipc.MetricsHBaseServerWrapper
@@ -343,7 +394,7 @@ implements 
org.apache.hadoop.hbase.ipc.MetricsHBaseServerWrapper
 
 
 getNumLifoModeSwitches
-publiclonggetNumLifoModeSwitches()
+publiclonggetNumLifoModeSwitches()
 
 Specified by:
 getNumLifoModeSwitchesin 
interfaceorg.apache.hadoop.hbase.ipc.MetricsHBaseServerWrapper
@@ -356,7 +407,7 @@ implements 
org.apache.hadoop.hbase.ipc.MetricsHBaseServerWrapper
 
 
 getWriteQueueLength
-publicintgetWriteQueueLength()
+publicintgetWriteQueueLength()
 
 Specified by:
 getWriteQueueLengthin 
interfaceorg.apache.hadoop.hbase.ipc.MetricsHBaseServerWrapper
@@ -369,7 +420,7 @@ implements 
org.apache.hadoop.hbase.ipc.MetricsHBaseServerWrapper
 
 
 getReadQueueLength
-publicintgetReadQueueLength()
+publicintgetReadQueueLength()
 
 Specified by:
 getReadQueueLengthin 
interfaceorg.apache.hadoop.hbase.ipc.MetricsHBaseServerWrapper
@@ -382,7 +433,7 @@ implements 
org.apache.hadoop.hbase.ipc.MetricsHBaseServerWrapper
 
 
 getScanQueueLength
-publicintgetScanQueueLength()
+publicintgetScanQueueLength()
 
 Specified by:
 getScanQueueLengthin 
interfaceorg.apache.hadoop.hbase.ipc.MetricsHBaseServerWrapper
@@ -395,7 +446,7 @@ implements 
org.apache.hadoop.hbase.ipc.MetricsHBaseServerWrapper
 
 
 getActiveWriteRpcHandlerCount
-publicintgetActiveWriteRpcHandlerCount()
+publicintgetActiveWriteRpcHandlerCount()
 
 Specified by:
 getActiveWriteRpcHandlerCountin 
interfaceorg.apache.hadoop.hbase.ipc.MetricsHBaseServerWrapper
@@ -408,7 +459,7 @@ implements 
org.apache.hadoop.hbase.ipc.MetricsHBaseServerWrapper
 
 
 getActiveReadRpcHandlerCount
-publicintgetActiveReadRpcHandlerCount()
+publicintgetActiveReadRpcHandlerCount()
 
 Specified by:
 getActiveReadRpcHandlerCountin 
interfaceorg.apache.hadoop.hbase.ipc.MetricsHBaseServerWrapper
@@ -421,7 +472,7 @@ implements 
org.apache.hadoop.hbase.ipc.MetricsHBaseServerWrapper
 
 
 getActiveScanRpcHandlerCount
-publicintgetActiveScanRpcHandlerCount()
+publicintgetActiveScanRpcHandlerCount()
 
 Specified by:
 getActiveScanRpcHandlerCountin 

[18/22] hbase-site git commit: Published site at 46cb5dfa226892fd2580f26ce9ce77225bd7e67c.

2018-04-21 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8bc57a1a/checkstyle.rss
--
diff --git a/checkstyle.rss b/checkstyle.rss
index 2da7c60..e2f8b27 100644
--- a/checkstyle.rss
+++ b/checkstyle.rss
@@ -26,7 +26,7 @@ under the License.
 2007 - 2018 The Apache Software Foundation
 
   File: 3604,
- Errors: 15864,
+ Errors: 15859,
  Warnings: 0,
  Infos: 0
   
@@ -3065,7 +3065,7 @@ under the License.
   0
 
 
-  23
+  18
 
   
   

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8bc57a1a/coc.html
--
diff --git a/coc.html b/coc.html
index 0144c87..d9b9654 100644
--- a/coc.html
+++ b/coc.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase  
   Code of Conduct Policy
@@ -375,7 +375,7 @@ email to mailto:priv...@hbase.apache.org;>the priv
 https://www.apache.org/;>The Apache Software 
Foundation.
 All rights reserved.  
 
-  Last Published: 
2018-04-20
+  Last Published: 
2018-04-21
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8bc57a1a/dependencies.html
--
diff --git a/dependencies.html b/dependencies.html
index 461e974..d32dc6e 100644
--- a/dependencies.html
+++ b/dependencies.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase  Project Dependencies
 
@@ -440,7 +440,7 @@
 https://www.apache.org/;>The Apache Software 
Foundation.
 All rights reserved.  
 
-  Last Published: 
2018-04-20
+  Last Published: 
2018-04-21
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8bc57a1a/dependency-convergence.html
--
diff --git a/dependency-convergence.html b/dependency-convergence.html
index 76564c8..64f6cd2 100644
--- a/dependency-convergence.html
+++ b/dependency-convergence.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase  Reactor Dependency Convergence
 
@@ -1105,7 +1105,7 @@
 https://www.apache.org/;>The Apache Software 
Foundation.
 All rights reserved.  
 
-  Last Published: 
2018-04-20
+  Last Published: 
2018-04-21
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8bc57a1a/dependency-info.html
--
diff --git a/dependency-info.html b/dependency-info.html
index 6d9a8c7..e6384eb 100644
--- a/dependency-info.html
+++ b/dependency-info.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase  Dependency Information
 
@@ -313,7 +313,7 @@
 https://www.apache.org/;>The Apache Software 
Foundation.
 All rights reserved.  
 
-  Last Published: 
2018-04-20
+  Last Published: 
2018-04-21
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8bc57a1a/dependency-management.html
--
diff --git a/dependency-management.html b/dependency-management.html
index fa99e3e..87dab1b 100644
--- a/dependency-management.html
+++ b/dependency-management.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase  Project Dependency Management
 
@@ -969,7 +969,7 @@
 https://www.apache.org/;>The Apache Software 
Foundation.
 All rights reserved.  
 
-  Last Published: 
2018-04-20
+  Last Published: 
2018-04-21
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8bc57a1a/devapidocs/constant-values.html
--
diff --git a/devapidocs/constant-values.html b/devapidocs/constant-values.html
index 60dc3eb..ee70906 100644
--- a/devapidocs/constant-values.html
+++ b/devapidocs/constant-values.html
@@ -3768,21 +3768,21 @@
 
 publicstaticfinalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String
 date
-"Fri Apr 20 14:39:14 UTC 2018"
+"Sat Apr 21 14:39:09 UTC 2018"
 
 
 
 
 publicstaticfinalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String
 revision

[22/22] hbase-site git commit: Published site at 46cb5dfa226892fd2580f26ce9ce77225bd7e67c.

2018-04-21 Thread git-site-role
Published site at 46cb5dfa226892fd2580f26ce9ce77225bd7e67c.


Project: http://git-wip-us.apache.org/repos/asf/hbase-site/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase-site/commit/8bc57a1a
Tree: http://git-wip-us.apache.org/repos/asf/hbase-site/tree/8bc57a1a
Diff: http://git-wip-us.apache.org/repos/asf/hbase-site/diff/8bc57a1a

Branch: refs/heads/asf-site
Commit: 8bc57a1a277bc0ef1998aa33e1043ce08e5df096
Parents: 84a296a
Author: jenkins 
Authored: Sat Apr 21 14:47:17 2018 +
Committer: jenkins 
Committed: Sat Apr 21 14:47:17 2018 +

--
 acid-semantics.html | 4 +-
 apache_hbase_reference_guide.pdf|  9510 ++---
 book.html   | 2 +-
 bulk-loads.html | 4 +-
 checkstyle-aggregate.html   | 11892 -
 checkstyle.rss  | 4 +-
 coc.html| 4 +-
 dependencies.html   | 4 +-
 dependency-convergence.html | 4 +-
 dependency-info.html| 4 +-
 dependency-management.html  | 4 +-
 devapidocs/constant-values.html |50 +-
 devapidocs/index-all.html   |66 +-
 .../hadoop/hbase/backup/package-tree.html   | 4 +-
 .../hadoop/hbase/client/package-tree.html   |22 +-
 .../hadoop/hbase/executor/package-tree.html | 2 +-
 .../hadoop/hbase/filter/package-tree.html   |10 +-
 .../hadoop/hbase/io/class-use/HFileLink.html| 9 +
 .../hadoop/hbase/io/class-use/Reference.html| 9 +
 .../hadoop/hbase/io/hfile/package-tree.html | 6 +-
 .../hadoop/hbase/ipc/FifoRpcScheduler.html  |   117 +-
 .../hbase/ipc/MetricsHBaseServerSource.html |   146 +-
 .../hbase/ipc/MetricsHBaseServerSourceImpl.html | 2 +-
 .../hbase/ipc/MetricsHBaseServerWrapper.html|83 +-
 .../ipc/MetricsHBaseServerWrapperImpl.html  |97 +-
 .../apache/hadoop/hbase/ipc/RpcScheduler.html   |   100 +-
 .../hadoop/hbase/ipc/SimpleRpcScheduler.html|   119 +-
 .../apache/hadoop/hbase/ipc/package-tree.html   | 2 +-
 .../hadoop/hbase/mapreduce/package-tree.html| 4 +-
 .../hbase/master/balancer/package-tree.html | 2 +-
 .../hadoop/hbase/master/package-tree.html   | 4 +-
 .../hbase/master/procedure/package-tree.html| 2 +-
 .../org/apache/hadoop/hbase/package-tree.html   |14 +-
 .../hadoop/hbase/procedure2/package-tree.html   | 4 +-
 .../hadoop/hbase/quotas/NoopQuotaLimiter.html   |34 +-
 .../hadoop/hbase/quotas/QuotaLimiter.html   |36 +-
 .../hadoop/hbase/quotas/TimeBasedLimiter.html   |54 +-
 .../class-use/RpcThrottlingException.html   |14 +-
 .../hadoop/hbase/quotas/package-tree.html   | 8 +-
 .../hbase/regionserver/StoreFileInfo.html   |   113 +-
 .../hadoop/hbase/regionserver/package-tree.html |14 +-
 .../regionserver/querymatcher/package-tree.html | 2 +-
 .../hbase/regionserver/wal/package-tree.html| 2 +-
 .../replication/regionserver/package-tree.html  | 2 +-
 .../hbase/security/access/package-tree.html | 2 +-
 .../hadoop/hbase/security/package-tree.html | 2 +-
 .../thrift/ThriftServerRunner.HBaseHandler.html |   142 +-
 .../ThriftServerRunner.IOErrorWithCause.html|12 +-
 .../thrift/ThriftServerRunner.ImplType.html |38 +-
 ...ThriftServerRunner.ResultScannerWrapper.html |12 +-
 .../hadoop/hbase/thrift/ThriftServerRunner.html |   108 +-
 .../hadoop/hbase/thrift/package-tree.html   | 2 +-
 .../hbase/util/ServerRegionReplicaUtil.html |12 +-
 .../apache/hadoop/hbase/util/package-tree.html  |10 +-
 .../apache/hadoop/hbase/wal/package-tree.html   | 2 +-
 .../org/apache/hadoop/hbase/Version.html| 6 +-
 .../ipc/FifoRpcScheduler.FifoCallRunner.html|96 +-
 .../hadoop/hbase/ipc/FifoRpcScheduler.html  |96 +-
 .../hbase/ipc/MetricsHBaseServerSource.html |64 +-
 .../hbase/ipc/MetricsHBaseServerSourceImpl.html |50 +-
 .../hbase/ipc/MetricsHBaseServerWrapper.html|24 +-
 .../ipc/MetricsHBaseServerWrapperImpl.html  |58 +-
 .../hadoop/hbase/ipc/RpcScheduler.Context.html  |65 +-
 .../apache/hadoop/hbase/ipc/RpcScheduler.html   |65 +-
 .../hadoop/hbase/ipc/SimpleRpcScheduler.html|   158 +-
 .../hbase/quotas/DefaultOperationQuota.html | 4 +-
 .../hadoop/hbase/quotas/NoopQuotaLimiter.html   | 6 +-
 .../hadoop/hbase/quotas/QuotaLimiter.html   |80 +-
 .../hadoop/hbase/quotas/TimeBasedLimiter.html   |   173 +-
 .../hbase/regionserver/StoreFileInfo.html   |   831 +-
 .../thrift/ThriftServerRunner.HBaseHandler.html |  3880 +++---
 .../ThriftServerRunner.IOErrorWithCause.html 

[04/22] hbase-site git commit: Published site at 46cb5dfa226892fd2580f26ce9ce77225bd7e67c.

2018-04-21 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8bc57a1a/devapidocs/src-html/org/apache/hadoop/hbase/util/ServerRegionReplicaUtil.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/util/ServerRegionReplicaUtil.html 
b/devapidocs/src-html/org/apache/hadoop/hbase/util/ServerRegionReplicaUtil.html
index 6bdffeb..3dbb30a 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/util/ServerRegionReplicaUtil.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/util/ServerRegionReplicaUtil.html
@@ -136,76 +136,89 @@
 128}
 129
 130// else create a store file link. The 
link file does not exists on filesystem though.
-131HFileLink link = 
HFileLink.build(conf, regionInfoForFs.getTable(),
-132
regionInfoForFs.getEncodedName(), familyName, path.getName());
-133
-134if (StoreFileInfo.isReference(path)) 
{
-135  Reference reference = 
Reference.read(fs, path);
-136  return new StoreFileInfo(conf, fs, 
link.getFileStatus(fs), reference);
-137}
-138
-139return new StoreFileInfo(conf, fs, 
link.getFileStatus(fs), link);
-140  }
-141
-142  /**
-143   * Create replication peer for 
replicating to region replicas if needed.
-144   * @param conf configuration to use
-145   * @throws IOException
-146   */
-147  public static void 
setupRegionReplicaReplication(Configuration conf) throws IOException {
-148if 
(!isRegionReplicaReplicationEnabled(conf)) {
-149  return;
-150}
-151Admin admin = 
ConnectionFactory.createConnection(conf).getAdmin();
-152ReplicationPeerConfig peerConfig = 
null;
-153try {
-154  peerConfig = 
admin.getReplicationPeerConfig(REGION_REPLICA_REPLICATION_PEER);
-155} catch 
(ReplicationPeerNotFoundException e) {
-156  LOG.warn("Region replica 
replication peer id=" + REGION_REPLICA_REPLICATION_PEER
-157  + " not exist", e);
-158}
-159try {
-160  if (peerConfig == null) {
-161LOG.info("Region replica 
replication peer id=" + REGION_REPLICA_REPLICATION_PEER
-162+ " not exist. 
Creating...");
-163peerConfig = new 
ReplicationPeerConfig();
-164
peerConfig.setClusterKey(ZKConfig.getZooKeeperClusterKey(conf));
-165
peerConfig.setReplicationEndpointImpl(RegionReplicaReplicationEndpoint.class.getName());
-166
admin.addReplicationPeer(REGION_REPLICA_REPLICATION_PEER, peerConfig);
-167  }
-168} finally {
-169  admin.close();
-170}
-171  }
-172
-173  public static boolean 
isRegionReplicaReplicationEnabled(Configuration conf) {
-174return 
conf.getBoolean(REGION_REPLICA_REPLICATION_CONF_KEY,
-175  
DEFAULT_REGION_REPLICA_REPLICATION);
-176  }
-177
-178  public static boolean 
isRegionReplicaWaitForPrimaryFlushEnabled(Configuration conf) {
-179return 
conf.getBoolean(REGION_REPLICA_WAIT_FOR_PRIMARY_FLUSH_CONF_KEY,
-180  
DEFAULT_REGION_REPLICA_WAIT_FOR_PRIMARY_FLUSH);
-181  }
-182
-183  public static boolean 
isRegionReplicaStoreFileRefreshEnabled(Configuration conf) {
-184return 
conf.getBoolean(REGION_REPLICA_STORE_FILE_REFRESH,
-185  
DEFAULT_REGION_REPLICA_STORE_FILE_REFRESH);
-186  }
-187
-188  public static double 
getRegionReplicaStoreFileRefreshMultiplier(Configuration conf) {
-189return 
conf.getDouble(REGION_REPLICA_STORE_FILE_REFRESH_MEMSTORE_MULTIPLIER,
-190  
DEFAULT_REGION_REPLICA_STORE_FILE_REFRESH_MEMSTORE_MULTIPLIER);
-191  }
-192
-193  /**
-194   * Return the peer id used for 
replicating to secondary region replicas
-195   */
-196  public static String 
getReplicationPeerId() {
-197return 
REGION_REPLICA_REPLICATION_PEER;
-198  }
-199
-200}
+131if (HFileLink.isHFileLink(path) || 
StoreFileInfo.isHFile(path)) {
+132  HFileLink link = HFileLink
+133  .build(conf, 
regionInfoForFs.getTable(), regionInfoForFs.getEncodedName(), familyName,
+134  path.getName());
+135  return new StoreFileInfo(conf, fs, 
link.getFileStatus(fs), link);
+136} else if 
(StoreFileInfo.isReference(path)) {
+137  Reference reference = 
Reference.read(fs, path);
+138  Path referencePath = 
StoreFileInfo.getReferredToFile(path);
+139  if 
(HFileLink.isHFileLink(referencePath)) {
+140// HFileLink Reference
+141HFileLink link = 
HFileLink.buildFromHFileLinkPattern(conf, referencePath);
+142return new StoreFileInfo(conf, 
fs, link.getFileStatus(fs), reference, link);
+143  } else {
+144// Reference
+145HFileLink link = HFileLink
+146.build(conf, 
regionInfoForFs.getTable(), regionInfoForFs.getEncodedName(), familyName,
+147path.getName());
+148return new StoreFileInfo(conf, 
fs, link.getFileStatus(fs), reference);
+149  }
+150} else {
+151  throw new IOException("path=" + 
path + " doesn't look like a valid StoreFile");
+152}
+153  }
+154
+155  /**
+156   * Create replication peer for 
replicating