hbase git commit: HBASE-20459 Majority of scan CPU time in HBase-1 spent in size estimation
Repository: hbase Updated Branches: refs/heads/branch-2.0 51069e6e8 -> 9d0a2e850 HBASE-20459 Majority of scan CPU time in HBase-1 spent in size estimation Project: http://git-wip-us.apache.org/repos/asf/hbase/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/9d0a2e85 Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/9d0a2e85 Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/9d0a2e85 Branch: refs/heads/branch-2.0 Commit: 9d0a2e850ea0f685435009894adc8f6250342169 Parents: 51069e6 Author: Lars HofhanslAuthored: Sat Apr 21 08:09:18 2018 -0700 Committer: Michael Stack Committed: Sat Apr 21 20:06:51 2018 -0700 -- .../src/main/java/org/apache/hadoop/hbase/util/ClassSize.java| 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hbase/blob/9d0a2e85/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ClassSize.java -- diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ClassSize.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ClassSize.java index 82f6e60..6f88c00 100644 --- a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ClassSize.java +++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ClassSize.java @@ -172,7 +172,7 @@ public class ClassSize { } long sizeOfByteArray(int len) { - return align(arrayHeaderSize() + len); + return align(ARRAY + len); } } @@ -217,7 +217,7 @@ public class ClassSize { @Override @SuppressWarnings("static-access") long sizeOfByteArray(int len) { - return align(arrayHeaderSize() + len * UnsafeAccess.theUnsafe.ARRAY_BYTE_INDEX_SCALE); + return align(ARRAY + len * UnsafeAccess.theUnsafe.ARRAY_BYTE_INDEX_SCALE); } }
hbase git commit: HBASE-20459 Majority of scan CPU time in HBase-1 spent in size estimation
Repository: hbase Updated Branches: refs/heads/branch-2 cb1aaa683 -> 42d2445cf HBASE-20459 Majority of scan CPU time in HBase-1 spent in size estimation Project: http://git-wip-us.apache.org/repos/asf/hbase/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/42d2445c Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/42d2445c Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/42d2445c Branch: refs/heads/branch-2 Commit: 42d2445cf527f717d6f6af17ac56c77dc8d44d05 Parents: cb1aaa6 Author: Lars HofhanslAuthored: Sat Apr 21 08:09:18 2018 -0700 Committer: Michael Stack Committed: Sat Apr 21 20:07:22 2018 -0700 -- .../src/main/java/org/apache/hadoop/hbase/util/ClassSize.java| 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hbase/blob/42d2445c/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ClassSize.java -- diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ClassSize.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ClassSize.java index 82f6e60..6f88c00 100644 --- a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ClassSize.java +++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ClassSize.java @@ -172,7 +172,7 @@ public class ClassSize { } long sizeOfByteArray(int len) { - return align(arrayHeaderSize() + len); + return align(ARRAY + len); } } @@ -217,7 +217,7 @@ public class ClassSize { @Override @SuppressWarnings("static-access") long sizeOfByteArray(int len) { - return align(arrayHeaderSize() + len * UnsafeAccess.theUnsafe.ARRAY_BYTE_INDEX_SCALE); + return align(ARRAY + len * UnsafeAccess.theUnsafe.ARRAY_BYTE_INDEX_SCALE); } }
[hbase] Git Push Summary
Repository: hbase Updated Branches: refs/heads/HBASE-20388 [deleted] 6e4efa3b1
hbase git commit: HBASE-20388 nightly tests running on a feature branch should only comment on that feature branch's jira
Repository: hbase Updated Branches: refs/heads/master 00fca5a4c -> f4a39043e HBASE-20388 nightly tests running on a feature branch should only comment on that feature branch's jira Signed-off-by: Mike DrobProject: http://git-wip-us.apache.org/repos/asf/hbase/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/f4a39043 Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/f4a39043 Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/f4a39043 Branch: refs/heads/master Commit: f4a39043e2fb3da69e378b4e1cbcb4bd7e2a6662 Parents: 00fca5a Author: Sean Busbey Authored: Thu Apr 12 21:10:53 2018 -0500 Committer: Sean Busbey Committed: Sat Apr 21 13:50:40 2018 -0500 -- dev-support/Jenkinsfile | 35 +++ 1 file changed, 23 insertions(+), 12 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hbase/blob/f4a39043/dev-support/Jenkinsfile -- diff --git a/dev-support/Jenkinsfile b/dev-support/Jenkinsfile index 6e320ca..821d20e 100644 --- a/dev-support/Jenkinsfile +++ b/dev-support/Jenkinsfile @@ -528,8 +528,14 @@ curl -L -o personality.sh "${env.PROJECT_PERSONALITY}" echo "[INFO] Comment:" echo comment echo "" - echo "[INFO] There are ${currentBuild.changeSets.size()} change sets." - getJirasToComment(currentBuild).each { currentIssue -> + echo "[DEBUG] checking to see if feature branch" + def jiras = getJirasToComment(env.BRANCH_NAME, []) + if (jiras.isEmpty()) { + echo "[DEBUG] non-feature branch, checking change messages for jira keys." + echo "[INFO] There are ${currentBuild.changeSets.size()} change sets." + jiras = getJirasToCommentFromChangesets(currentBuild) + } + jiras.each { currentIssue -> jiraComment issueKey: currentIssue, body: comment } } catch (Exception exception) { @@ -542,7 +548,7 @@ curl -L -o personality.sh "${env.PROJECT_PERSONALITY}" } import org.jenkinsci.plugins.workflow.support.steps.build.RunWrapper @NonCPS -List getJirasToComment(RunWrapper thisBuild) { +List getJirasToCommentFromChangesets(RunWrapper thisBuild) { def seenJiras = [] thisBuild.changeSets.each { cs -> cs.getItems().each { change -> @@ -552,16 +558,21 @@ List getJirasToComment(RunWrapper thisBuild) { echo " ${change.commitId}" echo " ${change.author}" echo "" - msg.eachMatch("HBASE-[0-9]+") { currentIssue -> -echo "[DEBUG] found jira key: ${currentIssue}" -if (currentIssue in seenJiras) { - echo "[DEBUG] already commented on ${currentIssue}." -} else { - echo "[INFO] commenting on ${currentIssue}." - seenJiras << currentIssue -} - } + seenJiras = getJirasToComment(msg, seenJiras) } } return seenJiras } +@NonCPS +List getJirasToComment(CharSequence source, List seen) { + source.eachMatch("HBASE-[0-9]+") { currentIssue -> +echo "[DEBUG] found jira key: ${currentIssue}" +if (currentIssue in seen) { + echo "[DEBUG] already commented on ${currentIssue}." +} else { + echo "[INFO] commenting on ${currentIssue}." + seen << currentIssue +} + } + return seen +}
hbase git commit: HBASE-20459 Majority of scan CPU time in HBase-1 spent in size estimation.
Repository: hbase Updated Branches: refs/heads/master 46cb5dfa2 -> 00fca5a4c HBASE-20459 Majority of scan CPU time in HBase-1 spent in size estimation. Project: http://git-wip-us.apache.org/repos/asf/hbase/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/00fca5a4 Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/00fca5a4 Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/00fca5a4 Branch: refs/heads/master Commit: 00fca5a4c9bdc73b4e0c962af0cbfcc7db2029fc Parents: 46cb5df Author: Lars HofhanslAuthored: Sat Apr 21 10:12:24 2018 -0700 Committer: Lars Hofhansl Committed: Sat Apr 21 10:12:24 2018 -0700 -- .../src/main/java/org/apache/hadoop/hbase/util/ClassSize.java| 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hbase/blob/00fca5a4/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ClassSize.java -- diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ClassSize.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ClassSize.java index 82f6e60..6f88c00 100644 --- a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ClassSize.java +++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ClassSize.java @@ -172,7 +172,7 @@ public class ClassSize { } long sizeOfByteArray(int len) { - return align(arrayHeaderSize() + len); + return align(ARRAY + len); } } @@ -217,7 +217,7 @@ public class ClassSize { @Override @SuppressWarnings("static-access") long sizeOfByteArray(int len) { - return align(arrayHeaderSize() + len * UnsafeAccess.theUnsafe.ARRAY_BYTE_INDEX_SCALE); + return align(ARRAY + len * UnsafeAccess.theUnsafe.ARRAY_BYTE_INDEX_SCALE); } }
hbase git commit: HBASE-20459 Majority of scan CPU time in HBase-1 spent in size estimation.
Repository: hbase Updated Branches: refs/heads/branch-1.4 1be417a99 -> e30023071 HBASE-20459 Majority of scan CPU time in HBase-1 spent in size estimation. Project: http://git-wip-us.apache.org/repos/asf/hbase/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/e3002307 Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/e3002307 Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/e3002307 Branch: refs/heads/branch-1.4 Commit: e30023071a012d8332925b65a508392cbf738df7 Parents: 1be417a Author: Lars HofhanslAuthored: Sat Apr 21 10:04:15 2018 -0700 Committer: Lars Hofhansl Committed: Sat Apr 21 10:04:15 2018 -0700 -- .../src/main/java/org/apache/hadoop/hbase/util/ClassSize.java| 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hbase/blob/e3002307/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ClassSize.java -- diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ClassSize.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ClassSize.java index edb5b2d..5e6627f 100644 --- a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ClassSize.java +++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ClassSize.java @@ -141,7 +141,7 @@ public class ClassSize { } long sizeOf(byte[] b, int len) { - return align(arrayHeaderSize() + len); + return align(ARRAY + len); } } @@ -186,7 +186,7 @@ public class ClassSize { @Override @SuppressWarnings("static-access") long sizeOf(byte[] b, int len) { - return align(arrayHeaderSize() + len * UnsafeAccess.theUnsafe.ARRAY_BYTE_INDEX_SCALE); + return align(ARRAY + len * UnsafeAccess.theUnsafe.ARRAY_BYTE_INDEX_SCALE); } }
hbase git commit: HBASE-20459 Majority of scan CPU time in HBase-1 spent in size estimation.
Repository: hbase Updated Branches: refs/heads/branch-1 e60189035 -> 6685b62ab HBASE-20459 Majority of scan CPU time in HBase-1 spent in size estimation. Project: http://git-wip-us.apache.org/repos/asf/hbase/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/6685b62a Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/6685b62a Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/6685b62a Branch: refs/heads/branch-1 Commit: 6685b62abe44227ee98b4e88386c12a02d718ce3 Parents: e601890 Author: Lars HofhanslAuthored: Sat Apr 21 10:01:14 2018 -0700 Committer: Lars Hofhansl Committed: Sat Apr 21 10:02:07 2018 -0700 -- .../src/main/java/org/apache/hadoop/hbase/util/ClassSize.java| 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hbase/blob/6685b62a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ClassSize.java -- diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ClassSize.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ClassSize.java index edb5b2d..5e6627f 100644 --- a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ClassSize.java +++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ClassSize.java @@ -141,7 +141,7 @@ public class ClassSize { } long sizeOf(byte[] b, int len) { - return align(arrayHeaderSize() + len); + return align(ARRAY + len); } } @@ -186,7 +186,7 @@ public class ClassSize { @Override @SuppressWarnings("static-access") long sizeOf(byte[] b, int len) { - return align(arrayHeaderSize() + len * UnsafeAccess.theUnsafe.ARRAY_BYTE_INDEX_SCALE); + return align(ARRAY + len * UnsafeAccess.theUnsafe.ARRAY_BYTE_INDEX_SCALE); } }
hbase-site git commit: INFRA-10751 Empty commit
Repository: hbase-site Updated Branches: refs/heads/asf-site 8bc57a1a2 -> 1a85017e7 INFRA-10751 Empty commit Project: http://git-wip-us.apache.org/repos/asf/hbase-site/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase-site/commit/1a85017e Tree: http://git-wip-us.apache.org/repos/asf/hbase-site/tree/1a85017e Diff: http://git-wip-us.apache.org/repos/asf/hbase-site/diff/1a85017e Branch: refs/heads/asf-site Commit: 1a85017e791e22693b6121ddacba1cd02a42fa03 Parents: 8bc57a1 Author: jenkinsAuthored: Sat Apr 21 14:47:37 2018 + Committer: jenkins Committed: Sat Apr 21 14:47:37 2018 + -- --
[21/22] hbase-site git commit: Published site at 46cb5dfa226892fd2580f26ce9ce77225bd7e67c.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8bc57a1a/apache_hbase_reference_guide.pdf -- diff --git a/apache_hbase_reference_guide.pdf b/apache_hbase_reference_guide.pdf index 7ac23b6..396a172 100644 --- a/apache_hbase_reference_guide.pdf +++ b/apache_hbase_reference_guide.pdf @@ -5,8 +5,8 @@ /Author (Apache HBase Team) /Creator (Asciidoctor PDF 1.5.0.alpha.15, based on Prawn 2.2.2) /Producer (Apache HBase Team) -/ModDate (D:20180420144439+00'00') -/CreationDate (D:20180420144439+00'00') +/ModDate (D:20180421142950+00'00') +/CreationDate (D:20180421144439+00'00') >> endobj 2 0 obj @@ -32028,7 +32028,7 @@ endobj endobj 157 0 obj << /Limits [(__anchor-top) (adding.new.node)] -/Names [(__anchor-top) 25 0 R (__indexterm-7613324) 3545 0 R (__indexterm-7615728) 3547 0 R (__indexterm-7617276) 3549 0 R (__indexterm-7619796) 3552 0 R (acid) 936 0 R (acl) 3356 0 R (add-metric-name-and-function-to-hadoop-compat-interface) 3645 0 R (add-the-implementation-to-both-hadoop-1-and-hadoop-2-compat-modules) 3647 0 R (add.metrics) 3643 0 R (adding-a-new-chapter-to-the-hbase-reference-guide) 3887 0 R (adding.new.node) 3102 0 R] +/Names [(__anchor-top) 25 0 R (__indexterm-2002) 3545 0 R (__indexterm-2004) 3547 0 R (__indexterm-2006) 3549 0 R (__indexterm-2008) 3552 0 R (acid) 936 0 R (acl) 3356 0 R (add-metric-name-and-function-to-hadoop-compat-interface) 3645 0 R (add-the-implementation-to-both-hadoop-1-and-hadoop-2-compat-modules) 3647 0 R (add.metrics) 3643 0 R (adding-a-new-chapter-to-the-hbase-reference-guide) 3887 0 R (adding.new.node) 3102 0 R] >> endobj 158 0 obj @@ -800821,4762 +800821,4762 @@ xref 519751 0 n 520162 0 n 520209 0 n -520717 0 n -521058 0 n -521207 0 n -521427 0 n -521474 0 n -536662 0 n -537081 0 n -537127 0 n -537252 0 n -537378 0 n -537424 0 n -537548 0 n -537594 0 n -537640 0 n -548886 0 n -549337 0 n -549521 0 n -549753 0 n -549983 0 n -550029 0 n -550157 0 n -550291 0 n -550338 0 n -550484 0 n -550609 0 n -569270 0 n -569655 0 n -569785 0 n -569929 0 n -574618 0 n -574977 0 n -585005 0 n -585395 0 n -585441 0 n -585572 0 n -596438 0 n -596841 0 n -596887 0 n -596933 0 n -597075 0 n -597121 0 n -597167 0 n -597214 0 n -597261 0 n -605049 0 n -605408 0 n -605454 0 n -605501 0 n -605548 0 n -606152 0 n -606199 0 n -606246 0 n -606292 0 n -615083 0 n -615442 0 n -615488 0 n -615535 0 n -615582 0 n -615629 0 n -615676 0 n -615722 0 n -623349 0 n -623708 0 n -623756 0 n -624442 0 n -624490 0 n -624538 0 n -624586 0 n -624634 0 n -624681 0 n -635763 0 n -636122 0 n -636169 0 n -636216 0 n -636262 0 n -644211 0 n -644570 0 n -644617 0 n -644664 0 n -644711 0 n -644758 0 n -644805 0 n -645578 0 n -645624 0 n -654093 0 n -654452 0 n -654499 0 n -654546 0 n -654593 0 n -654640 0 n -654687 0 n -654733 0 n -663824 0 n -664209 0 n -664256 0 n -664489 0 n -664721 0 n -664768 0 n -664815 0 n -664862 0 n -664907 0 n -672944 0 n -673345 0 n -673393 0 n -673639 0 n -673884 0 n -673932 0 n -674178 0 n -674423 0 n -674471 0 n -674519 0 n -675052 0 n -675100 0 n -675148 0 n -684161 0 n -684533 0 n -684581 0 n -684629 0 n -684677 0 n -684725 0 n -684773 0 n -694088 0 n -694473 0 n -694521 0 n -694569 0 n -694617 0 n -694665 0 n -695180 0 n -695377 0 n -695572 0 n -695620 0 n -703955 0 n -704314 0 n -704361 0 n -704408 0 n -704455 0 n -704502 0 n -704549 0 n -704595 0 n -711508 0 n -711867 0 n -711914 0 n -711961 0 n -712008 0 n -712055 0 n -712102 0 n -712149 0 n -721699 0 n -722058 0 n -722104 0 n -722151 0 n -722198 0 n -722245 0 n
[12/22] hbase-site git commit: Published site at 46cb5dfa226892fd2580f26ce9ce77225bd7e67c.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8bc57a1a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/FifoRpcScheduler.FifoCallRunner.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/FifoRpcScheduler.FifoCallRunner.html b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/FifoRpcScheduler.FifoCallRunner.html index 9014fce..231ba93 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/FifoRpcScheduler.FifoCallRunner.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/FifoRpcScheduler.FifoCallRunner.html @@ -142,78 +142,92 @@ 134 } 135 136 @Override -137 public long getNumGeneralCallsDropped() { -138return 0; +137 public int getActiveGeneralRpcHandlerCount() { +138return getActiveRpcHandlerCount(); 139 } 140 141 @Override -142 public long getNumLifoModeSwitches() { +142 public int getActivePriorityRpcHandlerCount() { 143return 0; 144 } 145 146 @Override -147 public int getWriteQueueLength() { +147 public int getActiveReplicationRpcHandlerCount() { 148return 0; 149 } 150 151 @Override -152 public int getReadQueueLength() { +152 public long getNumGeneralCallsDropped() { 153return 0; 154 } 155 156 @Override -157 public int getScanQueueLength() { +157 public long getNumLifoModeSwitches() { 158return 0; 159 } 160 161 @Override -162 public int getActiveWriteRpcHandlerCount() { +162 public int getWriteQueueLength() { 163return 0; 164 } 165 166 @Override -167 public int getActiveReadRpcHandlerCount() { +167 public int getReadQueueLength() { 168return 0; 169 } 170 171 @Override -172 public int getActiveScanRpcHandlerCount() { +172 public int getScanQueueLength() { 173return 0; 174 } 175 176 @Override -177 public CallQueueInfo getCallQueueInfo() { -178String queueName = "Fifo Queue"; -179 -180HashMapString, Long methodCount = new HashMap(); -181HashMapString, Long methodSize = new HashMap(); -182 -183CallQueueInfo callQueueInfo = new CallQueueInfo(); -184 callQueueInfo.setCallMethodCount(queueName, methodCount); -185 callQueueInfo.setCallMethodSize(queueName, methodSize); -186 -187 -188for (Runnable r:executor.getQueue()) { -189 FifoCallRunner mcr = (FifoCallRunner) r; -190 RpcCall rpcCall = mcr.getCallRunner().getRpcCall(); -191 -192 String method; -193 -194 if (null==rpcCall.getMethod() || -195 StringUtil.isNullOrEmpty(method = rpcCall.getMethod().getName())) { -196method = "Unknown"; -197 } -198 -199 long size = rpcCall.getSize(); -200 -201 methodCount.put(method, 1 + methodCount.getOrDefault(method, 0L)); -202 methodSize.put(method, size + methodSize.getOrDefault(method, 0L)); -203} -204 -205return callQueueInfo; -206 } -207 -208} +177 public int getActiveWriteRpcHandlerCount() { +178return 0; +179 } +180 +181 @Override +182 public int getActiveReadRpcHandlerCount() { +183return 0; +184 } +185 +186 @Override +187 public int getActiveScanRpcHandlerCount() { +188return 0; +189 } +190 +191 @Override +192 public CallQueueInfo getCallQueueInfo() { +193String queueName = "Fifo Queue"; +194 +195HashMapString, Long methodCount = new HashMap(); +196HashMapString, Long methodSize = new HashMap(); +197 +198CallQueueInfo callQueueInfo = new CallQueueInfo(); +199 callQueueInfo.setCallMethodCount(queueName, methodCount); +200 callQueueInfo.setCallMethodSize(queueName, methodSize); +201 +202 +203for (Runnable r:executor.getQueue()) { +204 FifoCallRunner mcr = (FifoCallRunner) r; +205 RpcCall rpcCall = mcr.getCallRunner().getRpcCall(); +206 +207 String method; +208 +209 if (null==rpcCall.getMethod() || +210 StringUtil.isNullOrEmpty(method = rpcCall.getMethod().getName())) { +211method = "Unknown"; +212 } +213 +214 long size = rpcCall.getSize(); +215 +216 methodCount.put(method, 1 + methodCount.getOrDefault(method, 0L)); +217 methodSize.put(method, size + methodSize.getOrDefault(method, 0L)); +218} +219 +220return callQueueInfo; +221 } +222} http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8bc57a1a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/FifoRpcScheduler.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/FifoRpcScheduler.html b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/FifoRpcScheduler.html index 9014fce..231ba93 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/FifoRpcScheduler.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/FifoRpcScheduler.html @@ -142,78 +142,92 @@ 134 } 135 136 @Override -137 public long getNumGeneralCallsDropped() { -138return 0; +137 public int getActiveGeneralRpcHandlerCount() { +138return
[20/22] hbase-site git commit: Published site at 46cb5dfa226892fd2580f26ce9ce77225bd7e67c.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8bc57a1a/book.html -- diff --git a/book.html b/book.html index 8de5bcb..d82af18 100644 --- a/book.html +++ b/book.html @@ -37824,7 +37824,7 @@ The server will return cellblocks compressed using this same compressor as long Version 3.0.0-SNAPSHOT -Last updated 2018-04-20 14:29:52 UTC +Last updated 2018-04-21 14:29:50 UTC http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8bc57a1a/bulk-loads.html -- diff --git a/bulk-loads.html b/bulk-loads.html index 2c97d08..111fe28 100644 --- a/bulk-loads.html +++ b/bulk-loads.html @@ -7,7 +7,7 @@ - + Apache HBase Bulk Loads in Apache HBase (TM) @@ -306,7 +306,7 @@ under the License. --> https://www.apache.org/;>The Apache Software Foundation. All rights reserved. - Last Published: 2018-04-20 + Last Published: 2018-04-21
[01/22] hbase-site git commit: Published site at 46cb5dfa226892fd2580f26ce9ce77225bd7e67c.
Repository: hbase-site Updated Branches: refs/heads/asf-site 84a296a43 -> 8bc57a1a2 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8bc57a1a/testdevapidocs/src-html/org/apache/hadoop/hbase/thrift/TestThriftHttpServer.html -- diff --git a/testdevapidocs/src-html/org/apache/hadoop/hbase/thrift/TestThriftHttpServer.html b/testdevapidocs/src-html/org/apache/hadoop/hbase/thrift/TestThriftHttpServer.html index 2d059c0..0e3c33c 100644 --- a/testdevapidocs/src-html/org/apache/hadoop/hbase/thrift/TestThriftHttpServer.html +++ b/testdevapidocs/src-html/org/apache/hadoop/hbase/thrift/TestThriftHttpServer.html @@ -29,214 +29,227 @@ 021import static org.junit.Assert.assertNull; 022import static org.junit.Assert.fail; 023 -024import java.util.ArrayList; -025import java.util.List; -026import org.apache.hadoop.conf.Configuration; -027import org.apache.hadoop.hbase.HBaseClassTestRule; -028import org.apache.hadoop.hbase.HBaseTestingUtility; -029import org.apache.hadoop.hbase.HConstants; -030import org.apache.hadoop.hbase.testclassification.ClientTests; -031import org.apache.hadoop.hbase.testclassification.LargeTests; -032import org.apache.hadoop.hbase.thrift.generated.Hbase; -033import org.apache.hadoop.hbase.util.EnvironmentEdgeManager; -034import org.apache.hadoop.hbase.util.EnvironmentEdgeManagerTestHelper; -035import org.apache.hadoop.hbase.util.IncrementingEnvironmentEdge; -036import org.apache.thrift.protocol.TBinaryProtocol; -037import org.apache.thrift.protocol.TProtocol; -038import org.apache.thrift.transport.THttpClient; -039import org.apache.thrift.transport.TTransportException; -040import org.junit.AfterClass; -041import org.junit.BeforeClass; -042import org.junit.ClassRule; -043import org.junit.Rule; -044import org.junit.Test; -045import org.junit.experimental.categories.Category; -046import org.junit.rules.ExpectedException; -047import org.slf4j.Logger; -048import org.slf4j.LoggerFactory; -049 -050import org.apache.hbase.thirdparty.com.google.common.base.Joiner; -051 -052/** -053 * Start the HBase Thrift HTTP server on a random port through the command-line -054 * interface and talk to it from client side. -055 */ -056@Category({ClientTests.class, LargeTests.class}) -057 -058public class TestThriftHttpServer { -059 -060 @ClassRule -061 public static final HBaseClassTestRule CLASS_RULE = -062 HBaseClassTestRule.forClass(TestThriftHttpServer.class); -063 -064 private static final Logger LOG = -065 LoggerFactory.getLogger(TestThriftHttpServer.class); +024import java.net.HttpURLConnection; +025import java.net.URL; +026import java.util.ArrayList; +027import java.util.List; +028import org.apache.hadoop.conf.Configuration; +029import org.apache.hadoop.hbase.HBaseClassTestRule; +030import org.apache.hadoop.hbase.HBaseTestingUtility; +031import org.apache.hadoop.hbase.HConstants; +032import org.apache.hadoop.hbase.testclassification.ClientTests; +033import org.apache.hadoop.hbase.testclassification.LargeTests; +034import org.apache.hadoop.hbase.thrift.generated.Hbase; +035import org.apache.hadoop.hbase.util.EnvironmentEdgeManager; +036import org.apache.hadoop.hbase.util.EnvironmentEdgeManagerTestHelper; +037import org.apache.hadoop.hbase.util.IncrementingEnvironmentEdge; +038import org.apache.thrift.protocol.TBinaryProtocol; +039import org.apache.thrift.protocol.TProtocol; +040import org.apache.thrift.transport.THttpClient; +041import org.apache.thrift.transport.TTransportException; +042import org.junit.AfterClass; +043import org.junit.Assert; +044import org.junit.BeforeClass; +045import org.junit.ClassRule; +046import org.junit.Rule; +047import org.junit.Test; +048import org.junit.experimental.categories.Category; +049import org.junit.rules.ExpectedException; +050import org.slf4j.Logger; +051import org.slf4j.LoggerFactory; +052 +053import org.apache.hbase.thirdparty.com.google.common.base.Joiner; +054 +055/** +056 * Start the HBase Thrift HTTP server on a random port through the command-line +057 * interface and talk to it from client side. +058 */ +059@Category({ClientTests.class, LargeTests.class}) +060 +061public class TestThriftHttpServer { +062 +063 @ClassRule +064 public static final HBaseClassTestRule CLASS_RULE = +065 HBaseClassTestRule.forClass(TestThriftHttpServer.class); 066 -067 private static final HBaseTestingUtility TEST_UTIL = -068 new HBaseTestingUtility(); +067 private static final Logger LOG = +068 LoggerFactory.getLogger(TestThriftHttpServer.class); 069 -070 private Thread httpServerThread; -071 private volatile Exception httpServerException; +070 private static final HBaseTestingUtility TEST_UTIL = +071 new HBaseTestingUtility(); 072 -073 private Exception clientSideException; -074 -075 private ThriftServer thriftServer; -076 private int port; +073 private Thread httpServerThread; +074 private volatile Exception
[17/22] hbase-site git commit: Published site at 46cb5dfa226892fd2580f26ce9ce77225bd7e67c.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8bc57a1a/devapidocs/org/apache/hadoop/hbase/ipc/MetricsHBaseServerSource.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/ipc/MetricsHBaseServerSource.html b/devapidocs/org/apache/hadoop/hbase/ipc/MetricsHBaseServerSource.html index 5db80d5..b3fddf3 100644 --- a/devapidocs/org/apache/hadoop/hbase/ipc/MetricsHBaseServerSource.html +++ b/devapidocs/org/apache/hadoop/hbase/ipc/MetricsHBaseServerSource.html @@ -180,6 +180,14 @@ extends static https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String +NUM_ACTIVE_GENERAL_HANDLER_DESC + + +static https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String +NUM_ACTIVE_GENERAL_HANDLER_NAME + + +static https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String NUM_ACTIVE_HANDLER_DESC @@ -188,6 +196,14 @@ extends static https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String +NUM_ACTIVE_PRIORITY_HANDLER_DESC + + +static https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String +NUM_ACTIVE_PRIORITY_HANDLER_NAME + + +static https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String NUM_ACTIVE_READ_HANDLER_DESC @@ -196,6 +212,14 @@ extends static https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String +NUM_ACTIVE_REPLICATION_HANDLER_DESC + + +static https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String +NUM_ACTIVE_REPLICATION_HANDLER_NAME + + +static https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String NUM_ACTIVE_SCAN_HANDLER_DESC @@ -990,13 +1014,91 @@ extends + + + + +NUM_ACTIVE_GENERAL_HANDLER_NAME +static finalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String NUM_ACTIVE_GENERAL_HANDLER_NAME + +See Also: +Constant Field Values + + + + + + + + +NUM_ACTIVE_GENERAL_HANDLER_DESC +static finalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String NUM_ACTIVE_GENERAL_HANDLER_DESC + +See Also: +Constant Field Values + + + + + + + + +NUM_ACTIVE_PRIORITY_HANDLER_NAME +static finalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String NUM_ACTIVE_PRIORITY_HANDLER_NAME + +See Also: +Constant Field Values + + + + + + + + +NUM_ACTIVE_PRIORITY_HANDLER_DESC +static finalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String NUM_ACTIVE_PRIORITY_HANDLER_DESC + +See Also: +Constant Field Values + + + + + + + + +NUM_ACTIVE_REPLICATION_HANDLER_NAME +static finalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String NUM_ACTIVE_REPLICATION_HANDLER_NAME + +See Also: +Constant Field Values + + + + + + + + +NUM_ACTIVE_REPLICATION_HANDLER_DESC +static finalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String NUM_ACTIVE_REPLICATION_HANDLER_DESC + +See Also: +Constant Field Values + + + NUM_ACTIVE_WRITE_HANDLER_NAME -static finalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String NUM_ACTIVE_WRITE_HANDLER_NAME +static finalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String NUM_ACTIVE_WRITE_HANDLER_NAME See Also: Constant Field Values @@ -1009,7 +,7 @@ extends NUM_ACTIVE_WRITE_HANDLER_DESC -static finalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String NUM_ACTIVE_WRITE_HANDLER_DESC +static finalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String NUM_ACTIVE_WRITE_HANDLER_DESC See Also: Constant Field Values @@ -1022,7 +1124,7 @@ extends NUM_ACTIVE_READ_HANDLER_NAME -static finalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String NUM_ACTIVE_READ_HANDLER_NAME +static
[19/22] hbase-site git commit: Published site at 46cb5dfa226892fd2580f26ce9ce77225bd7e67c.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8bc57a1a/checkstyle-aggregate.html -- diff --git a/checkstyle-aggregate.html b/checkstyle-aggregate.html index 67a62c1..88750a0 100644 --- a/checkstyle-aggregate.html +++ b/checkstyle-aggregate.html @@ -7,7 +7,7 @@ - + Apache HBase Checkstyle Results @@ -284,7 +284,7 @@ 3604 0 0 -15864 +15859 Files @@ -6947,7 +6947,7 @@ org/apache/hadoop/hbase/regionserver/StoreFileInfo.java 0 0 -23 +18 org/apache/hadoop/hbase/regionserver/StoreFileManager.java 0 @@ -10332,7 +10332,7 @@ http://checkstyle.sourceforge.net/config_javadoc.html#NonEmptyAtclauseDescription;>NonEmptyAtclauseDescription -3831 +3826 Error misc @@ -52880,7 +52880,7 @@ coding InnerAssignment Inner assignments should be avoided. -195 +210 org/apache/hadoop/hbase/ipc/HBaseRpcController.java @@ -54608,37 +54608,37 @@ sizes LineLength Line is longer than 100 characters (found 101). -90 +93 Error indentation Indentation 'array initialization' child have incorrect indentation level 8, expected level should be one of the following: 6, 35. -153 +156 Error indentation Indentation 'array initialization' child have incorrect indentation level 8, expected level should be one of the following: 6, 35. -154 +157 Error indentation Indentation 'array initialization' child have incorrect indentation level 8, expected level should be one of the following: 6, 35. -157 +160 Error indentation Indentation 'array initialization' child have incorrect indentation level 8, expected level should be one of the following: 6, 35. -158 +161 Error indentation Indentation 'array initialization' child have incorrect indentation level 8, expected level should be one of the following: 6, 35. -159 +162 org/apache/hadoop/hbase/ipc/TestSimpleRpcScheduler.java @@ -78137,37 +78137,37 @@ blocks NeedBraces 'if' construct must use '{}'s. -189 +192 Error blocks NeedBraces 'if' construct must use '{}'s. -190 +193 Error blocks NeedBraces 'if' construct must use '{}'s. -191 +194 Error blocks NeedBraces 'if' construct must use '{}'s. -192 +195 Error blocks NeedBraces 'if' construct must use '{}'s. -193 +196 Error blocks NeedBraces 'if' construct must use '{}'s. -194 +197 org/apache/hadoop/hbase/quotas/UserQuotaState.java @@ -85844,137 +85844,107 @@ javadoc NonEmptyAtclauseDescription At-clause should have a non-empty description. -178 - -Error -javadoc -NonEmptyAtclauseDescription -At-clause should have a non-empty description. -179 - -Error -javadoc -NonEmptyAtclauseDescription -At-clause should have a non-empty description. -180 - -Error -javadoc -NonEmptyAtclauseDescription -At-clause should have a non-empty description. -181 - -Error -javadoc -NonEmptyAtclauseDescription -At-clause should have a non-empty description. -182 +212 Error -javadoc -NonEmptyAtclauseDescription -At-clause should have a non-empty description. -197 +blocks +NeedBraces +'if' construct must use '{}'s. +486 Error blocks NeedBraces 'if' construct must use '{}'s. -471 +500 Error blocks NeedBraces 'if' construct must use '{}'s. -485 +549 Error blocks NeedBraces 'if' construct must use '{}'s. -534 +550 Error blocks NeedBraces 'if' construct must use '{}'s. -535 +552 Error blocks NeedBraces 'if' construct must use '{}'s. -537 +555 Error blocks NeedBraces 'if' construct must use '{}'s. -540 +556 Error blocks NeedBraces 'if' construct must use '{}'s. -541 +557 Error blocks NeedBraces 'if' construct must use '{}'s. -542 +560 Error blocks NeedBraces 'if' construct must use '{}'s. -545 +561 Error blocks NeedBraces 'if' construct must use '{}'s. -546 +562 Error blocks NeedBraces 'if' construct must use '{}'s. -547 +565 Error blocks NeedBraces 'if' construct must use '{}'s. -550 +566 Error blocks NeedBraces 'if' construct must use '{}'s. -551 - -Error -blocks -NeedBraces -'if' construct must use '{}'s. -552 +567 org/apache/hadoop/hbase/regionserver/StoreFileManager.java - + Severity Category Rule Message Line - + Error javadoc NonEmptyAtclauseDescription At-clause should have a non-empty description. 68 - + Error javadoc NonEmptyAtclauseDescription @@ -85983,31 +85953,31 @@ org/apache/hadoop/hbase/regionserver/StoreFileReader.java - + Severity Category Rule Message Line - + Error imports ImportOrder Wrong order for 'org.apache.hadoop.hbase.KeyValue' import. 39 - + Error blocks NeedBraces 'if' construct must use '{}'s. 469 - + Error blocks NeedBraces 'if' construct must use '{}'s. 489 - + Error blocks NeedBraces @@ -86016,67 +85986,67 @@ org/apache/hadoop/hbase/regionserver/StoreFileScanner.java - + Severity Category Rule Message Line - + Error imports ImportOrder Wrong order for 'org.apache.hadoop.hbase.client.Scan'
[09/22] hbase-site git commit: Published site at 46cb5dfa226892fd2580f26ce9ce77225bd7e67c.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8bc57a1a/devapidocs/src-html/org/apache/hadoop/hbase/thrift/ThriftServerRunner.HBaseHandler.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/thrift/ThriftServerRunner.HBaseHandler.html b/devapidocs/src-html/org/apache/hadoop/hbase/thrift/ThriftServerRunner.HBaseHandler.html index d6f672a..fd7ab7a 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/thrift/ThriftServerRunner.HBaseHandler.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/thrift/ThriftServerRunner.HBaseHandler.html @@ -87,1949 +87,1951 @@ 079import org.apache.hadoop.hbase.filter.ParseFilter; 080import org.apache.hadoop.hbase.filter.PrefixFilter; 081import org.apache.hadoop.hbase.filter.WhileMatchFilter; -082import org.apache.hadoop.hbase.log.HBaseMarkers; -083import org.apache.hadoop.hbase.security.SaslUtil; -084import org.apache.hadoop.hbase.security.SaslUtil.QualityOfProtection; -085import org.apache.hadoop.hbase.security.SecurityUtil; -086import org.apache.hadoop.hbase.security.UserProvider; -087import org.apache.hadoop.hbase.thrift.generated.AlreadyExists; -088import org.apache.hadoop.hbase.thrift.generated.BatchMutation; -089import org.apache.hadoop.hbase.thrift.generated.ColumnDescriptor; -090import org.apache.hadoop.hbase.thrift.generated.Hbase; -091import org.apache.hadoop.hbase.thrift.generated.IOError; -092import org.apache.hadoop.hbase.thrift.generated.IllegalArgument; -093import org.apache.hadoop.hbase.thrift.generated.Mutation; -094import org.apache.hadoop.hbase.thrift.generated.TAppend; -095import org.apache.hadoop.hbase.thrift.generated.TCell; -096import org.apache.hadoop.hbase.thrift.generated.TIncrement; -097import org.apache.hadoop.hbase.thrift.generated.TRegionInfo; -098import org.apache.hadoop.hbase.thrift.generated.TRowResult; -099import org.apache.hadoop.hbase.thrift.generated.TScan; -100import org.apache.hadoop.hbase.util.Bytes; -101import org.apache.hadoop.hbase.util.ConnectionCache; -102import org.apache.hadoop.hbase.util.DNS; -103import org.apache.hadoop.hbase.util.JvmPauseMonitor; -104import org.apache.hadoop.hbase.util.Strings; -105import org.apache.hadoop.security.SaslRpcServer.SaslGssCallbackHandler; -106import org.apache.hadoop.security.UserGroupInformation; -107import org.apache.hadoop.security.authorize.ProxyUsers; -108import org.apache.thrift.TException; -109import org.apache.thrift.TProcessor; -110import org.apache.thrift.protocol.TBinaryProtocol; -111import org.apache.thrift.protocol.TCompactProtocol; -112import org.apache.thrift.protocol.TProtocol; -113import org.apache.thrift.protocol.TProtocolFactory; -114import org.apache.thrift.server.THsHaServer; -115import org.apache.thrift.server.TNonblockingServer; -116import org.apache.thrift.server.TServer; -117import org.apache.thrift.server.TServlet; -118import org.apache.thrift.server.TThreadedSelectorServer; -119import org.apache.thrift.transport.TFramedTransport; -120import org.apache.thrift.transport.TNonblockingServerSocket; -121import org.apache.thrift.transport.TNonblockingServerTransport; -122import org.apache.thrift.transport.TSaslServerTransport; -123import org.apache.thrift.transport.TServerSocket; -124import org.apache.thrift.transport.TServerTransport; -125import org.apache.thrift.transport.TTransportFactory; -126import org.apache.yetus.audience.InterfaceAudience; -127import org.eclipse.jetty.http.HttpVersion; -128import org.eclipse.jetty.server.HttpConfiguration; -129import org.eclipse.jetty.server.HttpConnectionFactory; -130import org.eclipse.jetty.server.SecureRequestCustomizer; -131import org.eclipse.jetty.server.Server; -132import org.eclipse.jetty.server.ServerConnector; -133import org.eclipse.jetty.server.SslConnectionFactory; -134import org.eclipse.jetty.servlet.ServletContextHandler; -135import org.eclipse.jetty.servlet.ServletHolder; -136import org.eclipse.jetty.util.ssl.SslContextFactory; -137import org.eclipse.jetty.util.thread.QueuedThreadPool; -138import org.slf4j.Logger; -139import org.slf4j.LoggerFactory; -140import org.apache.hbase.thirdparty.com.google.common.base.Joiner; -141import org.apache.hbase.thirdparty.com.google.common.base.Throwables; -142import org.apache.hbase.thirdparty.com.google.common.util.concurrent.ThreadFactoryBuilder; -143import org.apache.hbase.thirdparty.org.apache.commons.cli.CommandLine; -144import org.apache.hbase.thirdparty.org.apache.commons.cli.Option; -145import org.apache.hbase.thirdparty.org.apache.commons.cli.OptionGroup; -146 -147/** -148 * ThriftServerRunner - this class starts up a Thrift server which implements -149 * the Hbase API specified in the Hbase.thrift IDL file. -150 */ -151@InterfaceAudience.Private -152public class ThriftServerRunner implements Runnable { -153 -154 private static final Logger LOG = LoggerFactory.getLogger(ThriftServerRunner.class); -155 -156
[15/22] hbase-site git commit: Published site at 46cb5dfa226892fd2580f26ce9ce77225bd7e67c.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8bc57a1a/devapidocs/org/apache/hadoop/hbase/quotas/class-use/RpcThrottlingException.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/quotas/class-use/RpcThrottlingException.html b/devapidocs/org/apache/hadoop/hbase/quotas/class-use/RpcThrottlingException.html index f7bd99d..f5a70f3 100644 --- a/devapidocs/org/apache/hadoop/hbase/quotas/class-use/RpcThrottlingException.html +++ b/devapidocs/org/apache/hadoop/hbase/quotas/class-use/RpcThrottlingException.html @@ -124,19 +124,25 @@ void -TimeBasedLimiter.checkQuota(longwriteSize, - longreadSize) +TimeBasedLimiter.checkQuota(longwriteReqs, + longestimateWriteSize, + longreadReqs, + longestimateReadSize) void -QuotaLimiter.checkQuota(longestimateWriteSize, +QuotaLimiter.checkQuota(longwriteReqs, + longestimateWriteSize, + longreadReqs, longestimateReadSize) Checks if it is possible to execute the specified operation. void -NoopQuotaLimiter.checkQuota(longestimateWriteSize, +NoopQuotaLimiter.checkQuota(longwriteReqs, + longestimateWriteSize, + longreadReqs, longestimateReadSize) http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8bc57a1a/devapidocs/org/apache/hadoop/hbase/quotas/package-tree.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/quotas/package-tree.html b/devapidocs/org/apache/hadoop/hbase/quotas/package-tree.html index 4e25e3e..152e2b7 100644 --- a/devapidocs/org/apache/hadoop/hbase/quotas/package-tree.html +++ b/devapidocs/org/apache/hadoop/hbase/quotas/package-tree.html @@ -229,13 +229,13 @@ java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true; title="class or interface in java.lang">EnumE (implements java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true; title="class or interface in java.lang">ComparableT, java.io.https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true; title="class or interface in java.io">Serializable) -org.apache.hadoop.hbase.quotas.SpaceViolationPolicy -org.apache.hadoop.hbase.quotas.OperationQuota.OperationType +org.apache.hadoop.hbase.quotas.QuotaType org.apache.hadoop.hbase.quotas.ThrottleType -org.apache.hadoop.hbase.quotas.QuotaScope org.apache.hadoop.hbase.quotas.RpcThrottlingException.Type +org.apache.hadoop.hbase.quotas.OperationQuota.OperationType +org.apache.hadoop.hbase.quotas.SpaceViolationPolicy org.apache.hadoop.hbase.quotas.ThrottlingException.Type -org.apache.hadoop.hbase.quotas.QuotaType +org.apache.hadoop.hbase.quotas.QuotaScope http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8bc57a1a/devapidocs/org/apache/hadoop/hbase/regionserver/StoreFileInfo.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/StoreFileInfo.html b/devapidocs/org/apache/hadoop/hbase/regionserver/StoreFileInfo.html index 1158ff7..df2be42 100644 --- a/devapidocs/org/apache/hadoop/hbase/regionserver/StoreFileInfo.html +++ b/devapidocs/org/apache/hadoop/hbase/regionserver/StoreFileInfo.html @@ -237,6 +237,15 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html +StoreFileInfo(org.apache.hadoop.conf.Configurationconf, + org.apache.hadoop.fs.FileSystemfs, + org.apache.hadoop.fs.FileStatusfileStatus, + Referencereference, + HFileLinklink) +Create a Store File Info from an HFileLink and a Reference + + + StoreFileInfo(org.apache.hadoop.conf.Configurationconf, org.apache.hadoop.fs.FileSystemfs, org.apache.hadoop.fs.PathinitialPath) @@ -622,39 +631,55 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html publicStoreFileInfo(org.apache.hadoop.conf.Configurationconf, org.apache.hadoop.fs.FileSystemfs, org.apache.hadoop.fs.FileStatusfileStatus, - HFileLinklink) - throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException + HFileLinklink) Create a Store File Info from an HFileLink Parameters: -conf - the Configuration to use -fs - The current file system to use. +conf - The Configuration to use +fs - The current file system to use fileStatus - The FileStatus of the file -Throws: -https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException - + StoreFileInfo -publicStoreFileInfo(org.apache.hadoop.conf.Configurationconf, +publicStoreFileInfo(org.apache.hadoop.conf.Configurationconf,
[11/22] hbase-site git commit: Published site at 46cb5dfa226892fd2580f26ce9ce77225bd7e67c.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8bc57a1a/devapidocs/src-html/org/apache/hadoop/hbase/quotas/QuotaLimiter.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/quotas/QuotaLimiter.html b/devapidocs/src-html/org/apache/hadoop/hbase/quotas/QuotaLimiter.html index 80bd305..ee8178a 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/quotas/QuotaLimiter.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/quotas/QuotaLimiter.html @@ -39,44 +39,48 @@ 031 /** 032 * Checks if it is possible to execute the specified operation. 033 * -034 * @param estimateWriteSize the write size that will be checked against the available quota -035 * @param estimateReadSize the read size that will be checked against the available quota -036 * @throws RpcThrottlingException thrown if not enough avialable resources to perform operation. -037 */ -038 void checkQuota(long estimateWriteSize, long estimateReadSize) -039throws RpcThrottlingException; -040 -041 /** -042 * Removes the specified write and read amount from the quota. -043 * At this point the write and read amount will be an estimate, -044 * that will be later adjusted with a consumeWrite()/consumeRead() call. -045 * -046 * @param writeSize the write size that will be removed from the current quota -047 * @param readSize the read size that will be removed from the current quota -048 */ -049 void grabQuota(long writeSize, long readSize); -050 -051 /** -052 * Removes or add back some write amount to the quota. -053 * (called at the end of an operation in case the estimate quota was off) -054 */ -055 void consumeWrite(long size); -056 -057 /** -058 * Removes or add back some read amount to the quota. -059 * (called at the end of an operation in case the estimate quota was off) -060 */ -061 void consumeRead(long size); -062 -063 /** @return true if the limiter is a noop */ -064 boolean isBypass(); -065 -066/** @return the number of bytes available to read to avoid exceeding the quota */ -067 long getReadAvailable(); -068 -069 /** @return the number of bytes available to write to avoid exceeding the quota */ -070 long getWriteAvailable(); -071} +034 * @param writeReqs the write requests that will be checked against the available quota +035 * @param estimateWriteSize the write size that will be checked against the available quota +036 * @param readReqs the read requests that will be checked against the available quota +037 * @param estimateReadSize the read size that will be checked against the available quota +038 * @throws RpcThrottlingException thrown if not enough available resources to perform operation. +039 */ +040 void checkQuota(long writeReqs, long estimateWriteSize, long readReqs, long estimateReadSize) +041 throws RpcThrottlingException; +042 +043 /** +044 * Removes the specified write and read amount from the quota. +045 * At this point the write and read amount will be an estimate, +046 * that will be later adjusted with a consumeWrite()/consumeRead() call. +047 * +048 * @param writeReqs the write requests that will be removed from the current quota +049 * @param writeSize the write size that will be removed from the current quota +050 * @param readReqs the read requests that will be removed from the current quota +051 * @param readSize the read size that will be removed from the current quota +052 */ +053 void grabQuota(long writeReqs, long writeSize, long readReqs, long readSize); +054 +055 /** +056 * Removes or add back some write amount to the quota. +057 * (called at the end of an operation in case the estimate quota was off) +058 */ +059 void consumeWrite(long size); +060 +061 /** +062 * Removes or add back some read amount to the quota. +063 * (called at the end of an operation in case the estimate quota was off) +064 */ +065 void consumeRead(long size); +066 +067 /** @return true if the limiter is a noop */ +068 boolean isBypass(); +069 +070/** @return the number of bytes available to read to avoid exceeding the quota */ +071 long getReadAvailable(); +072 +073 /** @return the number of bytes available to write to avoid exceeding the quota */ +074 long getWriteAvailable(); +075} http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8bc57a1a/devapidocs/src-html/org/apache/hadoop/hbase/quotas/TimeBasedLimiter.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/quotas/TimeBasedLimiter.html b/devapidocs/src-html/org/apache/hadoop/hbase/quotas/TimeBasedLimiter.html index 16a2371..7e6dad2 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/quotas/TimeBasedLimiter.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/quotas/TimeBasedLimiter.html @@ -118,92 +118,95 @@ 110 } 111 112 @Override -113
[06/22] hbase-site git commit: Published site at 46cb5dfa226892fd2580f26ce9ce77225bd7e67c.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8bc57a1a/devapidocs/src-html/org/apache/hadoop/hbase/thrift/ThriftServerRunner.ResultScannerWrapper.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/thrift/ThriftServerRunner.ResultScannerWrapper.html b/devapidocs/src-html/org/apache/hadoop/hbase/thrift/ThriftServerRunner.ResultScannerWrapper.html index d6f672a..fd7ab7a 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/thrift/ThriftServerRunner.ResultScannerWrapper.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/thrift/ThriftServerRunner.ResultScannerWrapper.html @@ -87,1949 +87,1951 @@ 079import org.apache.hadoop.hbase.filter.ParseFilter; 080import org.apache.hadoop.hbase.filter.PrefixFilter; 081import org.apache.hadoop.hbase.filter.WhileMatchFilter; -082import org.apache.hadoop.hbase.log.HBaseMarkers; -083import org.apache.hadoop.hbase.security.SaslUtil; -084import org.apache.hadoop.hbase.security.SaslUtil.QualityOfProtection; -085import org.apache.hadoop.hbase.security.SecurityUtil; -086import org.apache.hadoop.hbase.security.UserProvider; -087import org.apache.hadoop.hbase.thrift.generated.AlreadyExists; -088import org.apache.hadoop.hbase.thrift.generated.BatchMutation; -089import org.apache.hadoop.hbase.thrift.generated.ColumnDescriptor; -090import org.apache.hadoop.hbase.thrift.generated.Hbase; -091import org.apache.hadoop.hbase.thrift.generated.IOError; -092import org.apache.hadoop.hbase.thrift.generated.IllegalArgument; -093import org.apache.hadoop.hbase.thrift.generated.Mutation; -094import org.apache.hadoop.hbase.thrift.generated.TAppend; -095import org.apache.hadoop.hbase.thrift.generated.TCell; -096import org.apache.hadoop.hbase.thrift.generated.TIncrement; -097import org.apache.hadoop.hbase.thrift.generated.TRegionInfo; -098import org.apache.hadoop.hbase.thrift.generated.TRowResult; -099import org.apache.hadoop.hbase.thrift.generated.TScan; -100import org.apache.hadoop.hbase.util.Bytes; -101import org.apache.hadoop.hbase.util.ConnectionCache; -102import org.apache.hadoop.hbase.util.DNS; -103import org.apache.hadoop.hbase.util.JvmPauseMonitor; -104import org.apache.hadoop.hbase.util.Strings; -105import org.apache.hadoop.security.SaslRpcServer.SaslGssCallbackHandler; -106import org.apache.hadoop.security.UserGroupInformation; -107import org.apache.hadoop.security.authorize.ProxyUsers; -108import org.apache.thrift.TException; -109import org.apache.thrift.TProcessor; -110import org.apache.thrift.protocol.TBinaryProtocol; -111import org.apache.thrift.protocol.TCompactProtocol; -112import org.apache.thrift.protocol.TProtocol; -113import org.apache.thrift.protocol.TProtocolFactory; -114import org.apache.thrift.server.THsHaServer; -115import org.apache.thrift.server.TNonblockingServer; -116import org.apache.thrift.server.TServer; -117import org.apache.thrift.server.TServlet; -118import org.apache.thrift.server.TThreadedSelectorServer; -119import org.apache.thrift.transport.TFramedTransport; -120import org.apache.thrift.transport.TNonblockingServerSocket; -121import org.apache.thrift.transport.TNonblockingServerTransport; -122import org.apache.thrift.transport.TSaslServerTransport; -123import org.apache.thrift.transport.TServerSocket; -124import org.apache.thrift.transport.TServerTransport; -125import org.apache.thrift.transport.TTransportFactory; -126import org.apache.yetus.audience.InterfaceAudience; -127import org.eclipse.jetty.http.HttpVersion; -128import org.eclipse.jetty.server.HttpConfiguration; -129import org.eclipse.jetty.server.HttpConnectionFactory; -130import org.eclipse.jetty.server.SecureRequestCustomizer; -131import org.eclipse.jetty.server.Server; -132import org.eclipse.jetty.server.ServerConnector; -133import org.eclipse.jetty.server.SslConnectionFactory; -134import org.eclipse.jetty.servlet.ServletContextHandler; -135import org.eclipse.jetty.servlet.ServletHolder; -136import org.eclipse.jetty.util.ssl.SslContextFactory; -137import org.eclipse.jetty.util.thread.QueuedThreadPool; -138import org.slf4j.Logger; -139import org.slf4j.LoggerFactory; -140import org.apache.hbase.thirdparty.com.google.common.base.Joiner; -141import org.apache.hbase.thirdparty.com.google.common.base.Throwables; -142import org.apache.hbase.thirdparty.com.google.common.util.concurrent.ThreadFactoryBuilder; -143import org.apache.hbase.thirdparty.org.apache.commons.cli.CommandLine; -144import org.apache.hbase.thirdparty.org.apache.commons.cli.Option; -145import org.apache.hbase.thirdparty.org.apache.commons.cli.OptionGroup; -146 -147/** -148 * ThriftServerRunner - this class starts up a Thrift server which implements -149 * the Hbase API specified in the Hbase.thrift IDL file. -150 */ -151@InterfaceAudience.Private -152public class ThriftServerRunner implements Runnable { -153 -154 private static final Logger LOG =
[16/22] hbase-site git commit: Published site at 46cb5dfa226892fd2580f26ce9ce77225bd7e67c.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8bc57a1a/devapidocs/org/apache/hadoop/hbase/ipc/RpcScheduler.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/ipc/RpcScheduler.html b/devapidocs/org/apache/hadoop/hbase/ipc/RpcScheduler.html index af861b5..1f83c35 100644 --- a/devapidocs/org/apache/hadoop/hbase/ipc/RpcScheduler.html +++ b/devapidocs/org/apache/hadoop/hbase/ipc/RpcScheduler.html @@ -18,7 +18,7 @@ catch(err) { } //--> -var methods = {"i0":6,"i1":6,"i2":6,"i3":6,"i4":6,"i5":6,"i6":6,"i7":6,"i8":6,"i9":6,"i10":6,"i11":6,"i12":6,"i13":6,"i14":6,"i15":6,"i16":6}; +var methods = {"i0":6,"i1":6,"i2":6,"i3":6,"i4":6,"i5":6,"i6":6,"i7":6,"i8":6,"i9":6,"i10":6,"i11":6,"i12":6,"i13":6,"i14":6,"i15":6,"i16":6,"i17":6,"i18":6,"i19":6}; var tabs = {65535:["t0","All Methods"],2:["t2","Instance Methods"],4:["t3","Abstract Methods"]}; var altColor = "altColor"; var rowColor = "rowColor"; @@ -205,48 +205,66 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html abstract int +getActiveGeneralRpcHandlerCount() +Retrieves the number of active general handler. + + + +abstract int +getActivePriorityRpcHandlerCount() +Retrieves the number of active priority handler. + + + +abstract int getActiveReadRpcHandlerCount() Retrieves the number of active write rpc handler when use RWQueueRpcExecutor. - + +abstract int +getActiveReplicationRpcHandlerCount() +Retrieves the number of active replication handler. + + + abstract int getActiveRpcHandlerCount() -Retrieves the number of active handler. +Retrieves the total number of active handler. - + abstract int getActiveScanRpcHandlerCount() Retrieves the number of active write rpc handler when use RWQueueRpcExecutor. - + abstract int getActiveWriteRpcHandlerCount() Retrieves the number of active write rpc handler when use RWQueueRpcExecutor. - + abstract CallQueueInfo getCallQueueInfo() Get call queue information - + abstract int getGeneralQueueLength() Retrieves length of the general queue for metrics. - + abstract long getNumGeneralCallsDropped() If CoDel-based RPC executors are used, retrieves the number of Calls that were dropped from general queue because RPC executor is under high load; returns 0 otherwise. - + abstract long getNumLifoModeSwitches() If CoDel-based RPC executors are used, retrieves the number of Calls that were @@ -254,49 +272,49 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html in the period of overloade we serve last requests first); returns 0 otherwise. - + abstract int getPriorityQueueLength() Retrieves length of the priority queue for metrics. - + abstract int getReadQueueLength() Retrieves length of the read queue for metrics when use RWQueueRpcExecutor. - + abstract int getReplicationQueueLength() Retrieves length of the replication queue for metrics. - + abstract int getScanQueueLength() Retrieves length of the scan queue for metrics when use RWQueueRpcExecutor. - + abstract int getWriteQueueLength() Retrieves length of the write queue for metrics when use RWQueueRpcExecutor. - + abstract void init(RpcScheduler.Contextcontext) Does some quick initialization. - + abstract void start() Prepares for request serving. - + abstract void stop() Stops serving new requests. @@ -477,7 +495,37 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html getActiveRpcHandlerCount public abstractintgetActiveRpcHandlerCount() -Retrieves the number of active handler. +Retrieves the total number of active handler. + + + + + + + +getActiveGeneralRpcHandlerCount +public abstractintgetActiveGeneralRpcHandlerCount() +Retrieves the number of active general handler. + + + + + + + +getActivePriorityRpcHandlerCount +public abstractintgetActivePriorityRpcHandlerCount() +Retrieves the number of active priority handler. + + + + + + + +getActiveReplicationRpcHandlerCount +public abstractintgetActiveReplicationRpcHandlerCount() +Retrieves the number of active replication handler. @@ -486,7 +534,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html getNumGeneralCallsDropped -public abstractlonggetNumGeneralCallsDropped() +public abstractlonggetNumGeneralCallsDropped() If CoDel-based RPC executors are used, retrieves the number of Calls that were dropped from general queue because RPC executor is under high load; returns 0 otherwise. @@ -497,7 +545,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html getNumLifoModeSwitches -public abstractlonggetNumLifoModeSwitches() +public abstractlonggetNumLifoModeSwitches() If CoDel-based RPC executors are used, retrieves the number of Calls that were picked from the tail of the queue (indicating adaptive LIFO mode, when in the period of overloade we serve last requests first); returns
[07/22] hbase-site git commit: Published site at 46cb5dfa226892fd2580f26ce9ce77225bd7e67c.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8bc57a1a/devapidocs/src-html/org/apache/hadoop/hbase/thrift/ThriftServerRunner.ImplType.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/thrift/ThriftServerRunner.ImplType.html b/devapidocs/src-html/org/apache/hadoop/hbase/thrift/ThriftServerRunner.ImplType.html index d6f672a..fd7ab7a 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/thrift/ThriftServerRunner.ImplType.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/thrift/ThriftServerRunner.ImplType.html @@ -87,1949 +87,1951 @@ 079import org.apache.hadoop.hbase.filter.ParseFilter; 080import org.apache.hadoop.hbase.filter.PrefixFilter; 081import org.apache.hadoop.hbase.filter.WhileMatchFilter; -082import org.apache.hadoop.hbase.log.HBaseMarkers; -083import org.apache.hadoop.hbase.security.SaslUtil; -084import org.apache.hadoop.hbase.security.SaslUtil.QualityOfProtection; -085import org.apache.hadoop.hbase.security.SecurityUtil; -086import org.apache.hadoop.hbase.security.UserProvider; -087import org.apache.hadoop.hbase.thrift.generated.AlreadyExists; -088import org.apache.hadoop.hbase.thrift.generated.BatchMutation; -089import org.apache.hadoop.hbase.thrift.generated.ColumnDescriptor; -090import org.apache.hadoop.hbase.thrift.generated.Hbase; -091import org.apache.hadoop.hbase.thrift.generated.IOError; -092import org.apache.hadoop.hbase.thrift.generated.IllegalArgument; -093import org.apache.hadoop.hbase.thrift.generated.Mutation; -094import org.apache.hadoop.hbase.thrift.generated.TAppend; -095import org.apache.hadoop.hbase.thrift.generated.TCell; -096import org.apache.hadoop.hbase.thrift.generated.TIncrement; -097import org.apache.hadoop.hbase.thrift.generated.TRegionInfo; -098import org.apache.hadoop.hbase.thrift.generated.TRowResult; -099import org.apache.hadoop.hbase.thrift.generated.TScan; -100import org.apache.hadoop.hbase.util.Bytes; -101import org.apache.hadoop.hbase.util.ConnectionCache; -102import org.apache.hadoop.hbase.util.DNS; -103import org.apache.hadoop.hbase.util.JvmPauseMonitor; -104import org.apache.hadoop.hbase.util.Strings; -105import org.apache.hadoop.security.SaslRpcServer.SaslGssCallbackHandler; -106import org.apache.hadoop.security.UserGroupInformation; -107import org.apache.hadoop.security.authorize.ProxyUsers; -108import org.apache.thrift.TException; -109import org.apache.thrift.TProcessor; -110import org.apache.thrift.protocol.TBinaryProtocol; -111import org.apache.thrift.protocol.TCompactProtocol; -112import org.apache.thrift.protocol.TProtocol; -113import org.apache.thrift.protocol.TProtocolFactory; -114import org.apache.thrift.server.THsHaServer; -115import org.apache.thrift.server.TNonblockingServer; -116import org.apache.thrift.server.TServer; -117import org.apache.thrift.server.TServlet; -118import org.apache.thrift.server.TThreadedSelectorServer; -119import org.apache.thrift.transport.TFramedTransport; -120import org.apache.thrift.transport.TNonblockingServerSocket; -121import org.apache.thrift.transport.TNonblockingServerTransport; -122import org.apache.thrift.transport.TSaslServerTransport; -123import org.apache.thrift.transport.TServerSocket; -124import org.apache.thrift.transport.TServerTransport; -125import org.apache.thrift.transport.TTransportFactory; -126import org.apache.yetus.audience.InterfaceAudience; -127import org.eclipse.jetty.http.HttpVersion; -128import org.eclipse.jetty.server.HttpConfiguration; -129import org.eclipse.jetty.server.HttpConnectionFactory; -130import org.eclipse.jetty.server.SecureRequestCustomizer; -131import org.eclipse.jetty.server.Server; -132import org.eclipse.jetty.server.ServerConnector; -133import org.eclipse.jetty.server.SslConnectionFactory; -134import org.eclipse.jetty.servlet.ServletContextHandler; -135import org.eclipse.jetty.servlet.ServletHolder; -136import org.eclipse.jetty.util.ssl.SslContextFactory; -137import org.eclipse.jetty.util.thread.QueuedThreadPool; -138import org.slf4j.Logger; -139import org.slf4j.LoggerFactory; -140import org.apache.hbase.thirdparty.com.google.common.base.Joiner; -141import org.apache.hbase.thirdparty.com.google.common.base.Throwables; -142import org.apache.hbase.thirdparty.com.google.common.util.concurrent.ThreadFactoryBuilder; -143import org.apache.hbase.thirdparty.org.apache.commons.cli.CommandLine; -144import org.apache.hbase.thirdparty.org.apache.commons.cli.Option; -145import org.apache.hbase.thirdparty.org.apache.commons.cli.OptionGroup; -146 -147/** -148 * ThriftServerRunner - this class starts up a Thrift server which implements -149 * the Hbase API specified in the Hbase.thrift IDL file. -150 */ -151@InterfaceAudience.Private -152public class ThriftServerRunner implements Runnable { -153 -154 private static final Logger LOG = LoggerFactory.getLogger(ThriftServerRunner.class); -155 -156 private static final int
[13/22] hbase-site git commit: Published site at 46cb5dfa226892fd2580f26ce9ce77225bd7e67c.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8bc57a1a/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.IOErrorWithCause.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.IOErrorWithCause.html b/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.IOErrorWithCause.html index 23b678a..c2d61b1 100644 --- a/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.IOErrorWithCause.html +++ b/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.IOErrorWithCause.html @@ -137,7 +137,7 @@ var activeTableTab = "activeTableTab"; -public static class ThriftServerRunner.IOErrorWithCause +public static class ThriftServerRunner.IOErrorWithCause extends org.apache.hadoop.hbase.thrift.generated.IOError See Also: @@ -273,7 +273,7 @@ extends org.apache.hadoop.hbase.thrift.generated.IOError cause -privatehttps://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true; title="class or interface in java.lang">Throwable cause +privatehttps://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true; title="class or interface in java.lang">Throwable cause @@ -290,7 +290,7 @@ extends org.apache.hadoop.hbase.thrift.generated.IOError IOErrorWithCause -publicIOErrorWithCause(https://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true; title="class or interface in java.lang">Throwablecause) +publicIOErrorWithCause(https://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true; title="class or interface in java.lang">Throwablecause) @@ -307,7 +307,7 @@ extends org.apache.hadoop.hbase.thrift.generated.IOError getCause -publichttps://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true; title="class or interface in java.lang">ThrowablegetCause() +publichttps://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true; title="class or interface in java.lang">ThrowablegetCause() Overrides: https://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true#getCause--; title="class or interface in java.lang">getCausein classhttps://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true; title="class or interface in java.lang">Throwable @@ -320,7 +320,7 @@ extends org.apache.hadoop.hbase.thrift.generated.IOError equals -publicbooleanequals(https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Objectother) +publicbooleanequals(https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Objectother) Overrides: equalsin classorg.apache.hadoop.hbase.thrift.generated.IOError @@ -333,7 +333,7 @@ extends org.apache.hadoop.hbase.thrift.generated.IOError hashCode -publicinthashCode() +publicinthashCode() Overrides: hashCodein classorg.apache.hadoop.hbase.thrift.generated.IOError http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8bc57a1a/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.ImplType.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.ImplType.html b/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.ImplType.html index 82d106c..0392c62 100644 --- a/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.ImplType.html +++ b/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.ImplType.html @@ -122,7 +122,7 @@ var activeTableTab = "activeTableTab"; -static enum ThriftServerRunner.ImplType +static enum ThriftServerRunner.ImplType extends https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true; title="class or interface in java.lang">EnumThriftServerRunner.ImplType An enum of server implementation selections @@ -281,7 +281,7 @@ the order they are declared. HS_HA -public static finalThriftServerRunner.ImplType HS_HA +public static finalThriftServerRunner.ImplType HS_HA @@ -290,7 +290,7 @@ the order they are declared. NONBLOCKING -public static finalThriftServerRunner.ImplType NONBLOCKING +public static finalThriftServerRunner.ImplType NONBLOCKING @@ -299,7 +299,7 @@ the order they are declared. THREAD_POOL -public static finalThriftServerRunner.ImplType THREAD_POOL +public static finalThriftServerRunner.ImplType THREAD_POOL @@ -308,7 +308,7 @@ the order they are declared. THREADED_SELECTOR -public static finalThriftServerRunner.ImplType THREADED_SELECTOR +public static finalThriftServerRunner.ImplType THREADED_SELECTOR @@ -325,7 +325,7 @@ the order they are declared. DEFAULT -public static finalThriftServerRunner.ImplType DEFAULT +public static finalThriftServerRunner.ImplType DEFAULT @@ -334,7 +334,7
[08/22] hbase-site git commit: Published site at 46cb5dfa226892fd2580f26ce9ce77225bd7e67c.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8bc57a1a/devapidocs/src-html/org/apache/hadoop/hbase/thrift/ThriftServerRunner.IOErrorWithCause.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/thrift/ThriftServerRunner.IOErrorWithCause.html b/devapidocs/src-html/org/apache/hadoop/hbase/thrift/ThriftServerRunner.IOErrorWithCause.html index d6f672a..fd7ab7a 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/thrift/ThriftServerRunner.IOErrorWithCause.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/thrift/ThriftServerRunner.IOErrorWithCause.html @@ -87,1949 +87,1951 @@ 079import org.apache.hadoop.hbase.filter.ParseFilter; 080import org.apache.hadoop.hbase.filter.PrefixFilter; 081import org.apache.hadoop.hbase.filter.WhileMatchFilter; -082import org.apache.hadoop.hbase.log.HBaseMarkers; -083import org.apache.hadoop.hbase.security.SaslUtil; -084import org.apache.hadoop.hbase.security.SaslUtil.QualityOfProtection; -085import org.apache.hadoop.hbase.security.SecurityUtil; -086import org.apache.hadoop.hbase.security.UserProvider; -087import org.apache.hadoop.hbase.thrift.generated.AlreadyExists; -088import org.apache.hadoop.hbase.thrift.generated.BatchMutation; -089import org.apache.hadoop.hbase.thrift.generated.ColumnDescriptor; -090import org.apache.hadoop.hbase.thrift.generated.Hbase; -091import org.apache.hadoop.hbase.thrift.generated.IOError; -092import org.apache.hadoop.hbase.thrift.generated.IllegalArgument; -093import org.apache.hadoop.hbase.thrift.generated.Mutation; -094import org.apache.hadoop.hbase.thrift.generated.TAppend; -095import org.apache.hadoop.hbase.thrift.generated.TCell; -096import org.apache.hadoop.hbase.thrift.generated.TIncrement; -097import org.apache.hadoop.hbase.thrift.generated.TRegionInfo; -098import org.apache.hadoop.hbase.thrift.generated.TRowResult; -099import org.apache.hadoop.hbase.thrift.generated.TScan; -100import org.apache.hadoop.hbase.util.Bytes; -101import org.apache.hadoop.hbase.util.ConnectionCache; -102import org.apache.hadoop.hbase.util.DNS; -103import org.apache.hadoop.hbase.util.JvmPauseMonitor; -104import org.apache.hadoop.hbase.util.Strings; -105import org.apache.hadoop.security.SaslRpcServer.SaslGssCallbackHandler; -106import org.apache.hadoop.security.UserGroupInformation; -107import org.apache.hadoop.security.authorize.ProxyUsers; -108import org.apache.thrift.TException; -109import org.apache.thrift.TProcessor; -110import org.apache.thrift.protocol.TBinaryProtocol; -111import org.apache.thrift.protocol.TCompactProtocol; -112import org.apache.thrift.protocol.TProtocol; -113import org.apache.thrift.protocol.TProtocolFactory; -114import org.apache.thrift.server.THsHaServer; -115import org.apache.thrift.server.TNonblockingServer; -116import org.apache.thrift.server.TServer; -117import org.apache.thrift.server.TServlet; -118import org.apache.thrift.server.TThreadedSelectorServer; -119import org.apache.thrift.transport.TFramedTransport; -120import org.apache.thrift.transport.TNonblockingServerSocket; -121import org.apache.thrift.transport.TNonblockingServerTransport; -122import org.apache.thrift.transport.TSaslServerTransport; -123import org.apache.thrift.transport.TServerSocket; -124import org.apache.thrift.transport.TServerTransport; -125import org.apache.thrift.transport.TTransportFactory; -126import org.apache.yetus.audience.InterfaceAudience; -127import org.eclipse.jetty.http.HttpVersion; -128import org.eclipse.jetty.server.HttpConfiguration; -129import org.eclipse.jetty.server.HttpConnectionFactory; -130import org.eclipse.jetty.server.SecureRequestCustomizer; -131import org.eclipse.jetty.server.Server; -132import org.eclipse.jetty.server.ServerConnector; -133import org.eclipse.jetty.server.SslConnectionFactory; -134import org.eclipse.jetty.servlet.ServletContextHandler; -135import org.eclipse.jetty.servlet.ServletHolder; -136import org.eclipse.jetty.util.ssl.SslContextFactory; -137import org.eclipse.jetty.util.thread.QueuedThreadPool; -138import org.slf4j.Logger; -139import org.slf4j.LoggerFactory; -140import org.apache.hbase.thirdparty.com.google.common.base.Joiner; -141import org.apache.hbase.thirdparty.com.google.common.base.Throwables; -142import org.apache.hbase.thirdparty.com.google.common.util.concurrent.ThreadFactoryBuilder; -143import org.apache.hbase.thirdparty.org.apache.commons.cli.CommandLine; -144import org.apache.hbase.thirdparty.org.apache.commons.cli.Option; -145import org.apache.hbase.thirdparty.org.apache.commons.cli.OptionGroup; -146 -147/** -148 * ThriftServerRunner - this class starts up a Thrift server which implements -149 * the Hbase API specified in the Hbase.thrift IDL file. -150 */ -151@InterfaceAudience.Private -152public class ThriftServerRunner implements Runnable { -153 -154 private static final Logger LOG =
[02/22] hbase-site git commit: Published site at 46cb5dfa226892fd2580f26ce9ce77225bd7e67c.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8bc57a1a/testdevapidocs/src-html/org/apache/hadoop/hbase/http/TestHttpServer.HtmlContentServlet.html -- diff --git a/testdevapidocs/src-html/org/apache/hadoop/hbase/http/TestHttpServer.HtmlContentServlet.html b/testdevapidocs/src-html/org/apache/hadoop/hbase/http/TestHttpServer.HtmlContentServlet.html index 5681a20..cbcaf55 100644 --- a/testdevapidocs/src-html/org/apache/hadoop/hbase/http/TestHttpServer.HtmlContentServlet.html +++ b/testdevapidocs/src-html/org/apache/hadoop/hbase/http/TestHttpServer.HtmlContentServlet.html @@ -613,21 +613,30 @@ 605myServer.stop(); 606 } 607 -608 -609 -610 @Test -611 public void testNoCacheHeader() throws Exception { -612URL url = new URL(baseUrl, "/echo?a=bc=d"); -613HttpURLConnection conn = (HttpURLConnection) url.openConnection(); -614 assertEquals(HttpURLConnection.HTTP_OK, conn.getResponseCode()); -615assertEquals("no-cache", conn.getHeaderField("Cache-Control")); -616assertEquals("no-cache", conn.getHeaderField("Pragma")); -617 assertNotNull(conn.getHeaderField("Expires")); -618 assertNotNull(conn.getHeaderField("Date")); -619 assertEquals(conn.getHeaderField("Expires"), conn.getHeaderField("Date")); -620assertEquals("DENY", conn.getHeaderField("X-Frame-Options")); -621 } -622} +608 @Test +609 public void testNoCacheHeader() throws Exception { +610URL url = new URL(baseUrl, "/echo?a=bc=d"); +611HttpURLConnection conn = (HttpURLConnection) url.openConnection(); +612 assertEquals(HttpURLConnection.HTTP_OK, conn.getResponseCode()); +613assertEquals("no-cache", conn.getHeaderField("Cache-Control")); +614assertEquals("no-cache", conn.getHeaderField("Pragma")); +615 assertNotNull(conn.getHeaderField("Expires")); +616 assertNotNull(conn.getHeaderField("Date")); +617 assertEquals(conn.getHeaderField("Expires"), conn.getHeaderField("Date")); +618assertEquals("DENY", conn.getHeaderField("X-Frame-Options")); +619 } +620 +621 @Test +622 public void testHttpMethods() throws Exception { +623// HTTP TRACE method should be disabled for security +624// See https://www.owasp.org/index.php/Cross_Site_Tracing +625URL url = new URL(baseUrl, "/echo?a=b"); +626HttpURLConnection conn = (HttpURLConnection) url.openConnection(); +627conn.setRequestMethod("TRACE"); +628conn.connect(); +629 assertEquals(HttpURLConnection.HTTP_FORBIDDEN, conn.getResponseCode()); +630 } +631} http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8bc57a1a/testdevapidocs/src-html/org/apache/hadoop/hbase/http/TestHttpServer.LongHeaderServlet.html -- diff --git a/testdevapidocs/src-html/org/apache/hadoop/hbase/http/TestHttpServer.LongHeaderServlet.html b/testdevapidocs/src-html/org/apache/hadoop/hbase/http/TestHttpServer.LongHeaderServlet.html index 5681a20..cbcaf55 100644 --- a/testdevapidocs/src-html/org/apache/hadoop/hbase/http/TestHttpServer.LongHeaderServlet.html +++ b/testdevapidocs/src-html/org/apache/hadoop/hbase/http/TestHttpServer.LongHeaderServlet.html @@ -613,21 +613,30 @@ 605myServer.stop(); 606 } 607 -608 -609 -610 @Test -611 public void testNoCacheHeader() throws Exception { -612URL url = new URL(baseUrl, "/echo?a=bc=d"); -613HttpURLConnection conn = (HttpURLConnection) url.openConnection(); -614 assertEquals(HttpURLConnection.HTTP_OK, conn.getResponseCode()); -615assertEquals("no-cache", conn.getHeaderField("Cache-Control")); -616assertEquals("no-cache", conn.getHeaderField("Pragma")); -617 assertNotNull(conn.getHeaderField("Expires")); -618 assertNotNull(conn.getHeaderField("Date")); -619 assertEquals(conn.getHeaderField("Expires"), conn.getHeaderField("Date")); -620assertEquals("DENY", conn.getHeaderField("X-Frame-Options")); -621 } -622} +608 @Test +609 public void testNoCacheHeader() throws Exception { +610URL url = new URL(baseUrl, "/echo?a=bc=d"); +611HttpURLConnection conn = (HttpURLConnection) url.openConnection(); +612 assertEquals(HttpURLConnection.HTTP_OK, conn.getResponseCode()); +613assertEquals("no-cache", conn.getHeaderField("Cache-Control")); +614assertEquals("no-cache", conn.getHeaderField("Pragma")); +615 assertNotNull(conn.getHeaderField("Expires")); +616 assertNotNull(conn.getHeaderField("Date")); +617 assertEquals(conn.getHeaderField("Expires"), conn.getHeaderField("Date")); +618assertEquals("DENY", conn.getHeaderField("X-Frame-Options")); +619 } +620 +621 @Test +622 public void testHttpMethods() throws Exception { +623// HTTP TRACE method should be disabled for security +624// See https://www.owasp.org/index.php/Cross_Site_Tracing +625URL url = new URL(baseUrl, "/echo?a=b"); +626HttpURLConnection conn = (HttpURLConnection)
[14/22] hbase-site git commit: Published site at 46cb5dfa226892fd2580f26ce9ce77225bd7e67c.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8bc57a1a/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.HBaseHandler.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.HBaseHandler.html b/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.HBaseHandler.html index aeebaec..f338c1a 100644 --- a/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.HBaseHandler.html +++ b/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.HBaseHandler.html @@ -117,7 +117,7 @@ var activeTableTab = "activeTableTab"; -public static class ThriftServerRunner.HBaseHandler +public static class ThriftServerRunner.HBaseHandler extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object implements org.apache.hadoop.hbase.thrift.generated.Hbase.Iface The HBaseHandler is a glue object that connects Thrift RPC calls to the @@ -721,7 +721,7 @@ implements org.apache.hadoop.hbase.thrift.generated.Hbase.Iface conf -protectedorg.apache.hadoop.conf.Configuration conf +protectedorg.apache.hadoop.conf.Configuration conf @@ -730,7 +730,7 @@ implements org.apache.hadoop.hbase.thrift.generated.Hbase.Iface LOG -protected static finalorg.slf4j.Logger LOG +protected static finalorg.slf4j.Logger LOG @@ -739,7 +739,7 @@ implements org.apache.hadoop.hbase.thrift.generated.Hbase.Iface nextScannerId -protectedint nextScannerId +protectedint nextScannerId @@ -748,7 +748,7 @@ implements org.apache.hadoop.hbase.thrift.generated.Hbase.Iface scannerMap -protectedhttps://docs.oracle.com/javase/8/docs/api/java/util/HashMap.html?is-external=true; title="class or interface in java.util">HashMaphttps://docs.oracle.com/javase/8/docs/api/java/lang/Integer.html?is-external=true; title="class or interface in java.lang">Integer,ThriftServerRunner.ResultScannerWrapper scannerMap +protectedhttps://docs.oracle.com/javase/8/docs/api/java/util/HashMap.html?is-external=true; title="class or interface in java.util">HashMaphttps://docs.oracle.com/javase/8/docs/api/java/lang/Integer.html?is-external=true; title="class or interface in java.lang">Integer,ThriftServerRunner.ResultScannerWrapper scannerMap @@ -757,7 +757,7 @@ implements org.apache.hadoop.hbase.thrift.generated.Hbase.Iface metrics -privateThriftMetrics metrics +privateThriftMetrics metrics @@ -766,7 +766,7 @@ implements org.apache.hadoop.hbase.thrift.generated.Hbase.Iface connectionCache -private finalConnectionCache connectionCache +private finalConnectionCache connectionCache @@ -775,7 +775,7 @@ implements org.apache.hadoop.hbase.thrift.generated.Hbase.Iface coalescer -IncrementCoalescer coalescer +IncrementCoalescer coalescer @@ -784,7 +784,7 @@ implements org.apache.hadoop.hbase.thrift.generated.Hbase.Iface CLEANUP_INTERVAL -static finalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String CLEANUP_INTERVAL +static finalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String CLEANUP_INTERVAL See Also: Constant Field Values @@ -797,7 +797,7 @@ implements org.apache.hadoop.hbase.thrift.generated.Hbase.Iface MAX_IDLETIME -static finalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String MAX_IDLETIME +static finalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String MAX_IDLETIME See Also: Constant Field Values @@ -818,7 +818,7 @@ implements org.apache.hadoop.hbase.thrift.generated.Hbase.Iface HBaseHandler -protectedHBaseHandler(org.apache.hadoop.conf.Configurationc, +protectedHBaseHandler(org.apache.hadoop.conf.Configurationc, UserProvideruserProvider) throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException @@ -841,7 +841,7 @@ implements org.apache.hadoop.hbase.thrift.generated.Hbase.Iface getAllColumns -byte[][]getAllColumns(Tabletable) +byte[][]getAllColumns(Tabletable) throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException Returns a list of all the column families for a given Table. @@ -858,7 +858,7 @@ implements org.apache.hadoop.hbase.thrift.generated.Hbase.Iface getTable -publicTablegetTable(byte[]tableName) +publicTablegetTable(byte[]tableName) throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException Creates and
[05/22] hbase-site git commit: Published site at 46cb5dfa226892fd2580f26ce9ce77225bd7e67c.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8bc57a1a/devapidocs/src-html/org/apache/hadoop/hbase/thrift/ThriftServerRunner.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/thrift/ThriftServerRunner.html b/devapidocs/src-html/org/apache/hadoop/hbase/thrift/ThriftServerRunner.html index d6f672a..fd7ab7a 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/thrift/ThriftServerRunner.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/thrift/ThriftServerRunner.html @@ -87,1949 +87,1951 @@ 079import org.apache.hadoop.hbase.filter.ParseFilter; 080import org.apache.hadoop.hbase.filter.PrefixFilter; 081import org.apache.hadoop.hbase.filter.WhileMatchFilter; -082import org.apache.hadoop.hbase.log.HBaseMarkers; -083import org.apache.hadoop.hbase.security.SaslUtil; -084import org.apache.hadoop.hbase.security.SaslUtil.QualityOfProtection; -085import org.apache.hadoop.hbase.security.SecurityUtil; -086import org.apache.hadoop.hbase.security.UserProvider; -087import org.apache.hadoop.hbase.thrift.generated.AlreadyExists; -088import org.apache.hadoop.hbase.thrift.generated.BatchMutation; -089import org.apache.hadoop.hbase.thrift.generated.ColumnDescriptor; -090import org.apache.hadoop.hbase.thrift.generated.Hbase; -091import org.apache.hadoop.hbase.thrift.generated.IOError; -092import org.apache.hadoop.hbase.thrift.generated.IllegalArgument; -093import org.apache.hadoop.hbase.thrift.generated.Mutation; -094import org.apache.hadoop.hbase.thrift.generated.TAppend; -095import org.apache.hadoop.hbase.thrift.generated.TCell; -096import org.apache.hadoop.hbase.thrift.generated.TIncrement; -097import org.apache.hadoop.hbase.thrift.generated.TRegionInfo; -098import org.apache.hadoop.hbase.thrift.generated.TRowResult; -099import org.apache.hadoop.hbase.thrift.generated.TScan; -100import org.apache.hadoop.hbase.util.Bytes; -101import org.apache.hadoop.hbase.util.ConnectionCache; -102import org.apache.hadoop.hbase.util.DNS; -103import org.apache.hadoop.hbase.util.JvmPauseMonitor; -104import org.apache.hadoop.hbase.util.Strings; -105import org.apache.hadoop.security.SaslRpcServer.SaslGssCallbackHandler; -106import org.apache.hadoop.security.UserGroupInformation; -107import org.apache.hadoop.security.authorize.ProxyUsers; -108import org.apache.thrift.TException; -109import org.apache.thrift.TProcessor; -110import org.apache.thrift.protocol.TBinaryProtocol; -111import org.apache.thrift.protocol.TCompactProtocol; -112import org.apache.thrift.protocol.TProtocol; -113import org.apache.thrift.protocol.TProtocolFactory; -114import org.apache.thrift.server.THsHaServer; -115import org.apache.thrift.server.TNonblockingServer; -116import org.apache.thrift.server.TServer; -117import org.apache.thrift.server.TServlet; -118import org.apache.thrift.server.TThreadedSelectorServer; -119import org.apache.thrift.transport.TFramedTransport; -120import org.apache.thrift.transport.TNonblockingServerSocket; -121import org.apache.thrift.transport.TNonblockingServerTransport; -122import org.apache.thrift.transport.TSaslServerTransport; -123import org.apache.thrift.transport.TServerSocket; -124import org.apache.thrift.transport.TServerTransport; -125import org.apache.thrift.transport.TTransportFactory; -126import org.apache.yetus.audience.InterfaceAudience; -127import org.eclipse.jetty.http.HttpVersion; -128import org.eclipse.jetty.server.HttpConfiguration; -129import org.eclipse.jetty.server.HttpConnectionFactory; -130import org.eclipse.jetty.server.SecureRequestCustomizer; -131import org.eclipse.jetty.server.Server; -132import org.eclipse.jetty.server.ServerConnector; -133import org.eclipse.jetty.server.SslConnectionFactory; -134import org.eclipse.jetty.servlet.ServletContextHandler; -135import org.eclipse.jetty.servlet.ServletHolder; -136import org.eclipse.jetty.util.ssl.SslContextFactory; -137import org.eclipse.jetty.util.thread.QueuedThreadPool; -138import org.slf4j.Logger; -139import org.slf4j.LoggerFactory; -140import org.apache.hbase.thirdparty.com.google.common.base.Joiner; -141import org.apache.hbase.thirdparty.com.google.common.base.Throwables; -142import org.apache.hbase.thirdparty.com.google.common.util.concurrent.ThreadFactoryBuilder; -143import org.apache.hbase.thirdparty.org.apache.commons.cli.CommandLine; -144import org.apache.hbase.thirdparty.org.apache.commons.cli.Option; -145import org.apache.hbase.thirdparty.org.apache.commons.cli.OptionGroup; -146 -147/** -148 * ThriftServerRunner - this class starts up a Thrift server which implements -149 * the Hbase API specified in the Hbase.thrift IDL file. -150 */ -151@InterfaceAudience.Private -152public class ThriftServerRunner implements Runnable { -153 -154 private static final Logger LOG = LoggerFactory.getLogger(ThriftServerRunner.class); -155 -156 private static final int DEFAULT_HTTP_MAX_HEADER_SIZE = 64 * 1024; //
[10/22] hbase-site git commit: Published site at 46cb5dfa226892fd2580f26ce9ce77225bd7e67c.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8bc57a1a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/StoreFileInfo.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/StoreFileInfo.html b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/StoreFileInfo.html index c467ef4..d4b94dd 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/StoreFileInfo.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/StoreFileInfo.html @@ -165,427 +165,442 @@ 157 158 /** 159 * Create a Store File Info from an HFileLink -160 * @param conf the {@link Configuration} to use -161 * @param fs The current file system to use. +160 * @param conf The {@link Configuration} to use +161 * @param fs The current file system to use 162 * @param fileStatus The {@link FileStatus} of the file 163 */ 164 public StoreFileInfo(final Configuration conf, final FileSystem fs, final FileStatus fileStatus, -165 final HFileLink link) -166 throws IOException { -167this.fs = fs; -168this.conf = conf; -169// initialPath can be null only if we get a link. -170this.initialPath = (fileStatus == null) ? null : fileStatus.getPath(); -171 // HFileLink -172this.reference = null; -173this.link = link; -174 } -175 -176 /** -177 * Create a Store File Info from an HFileLink -178 * @param conf -179 * @param fs -180 * @param fileStatus -181 * @param reference -182 * @throws IOException -183 */ -184 public StoreFileInfo(final Configuration conf, final FileSystem fs, final FileStatus fileStatus, -185 final Reference reference) -186 throws IOException { -187this.fs = fs; -188this.conf = conf; -189this.initialPath = fileStatus.getPath(); -190this.createdTimestamp = fileStatus.getModificationTime(); -191this.reference = reference; -192this.link = null; -193 } -194 -195 /** -196 * Sets the region coprocessor env. -197 * @param coprocessorHost -198 */ -199 public void setRegionCoprocessorHost(RegionCoprocessorHost coprocessorHost) { -200this.coprocessorHost = coprocessorHost; -201 } -202 -203 /* -204 * @return the Reference object associated to this StoreFileInfo. -205 * null if the StoreFile is not a reference. -206 */ -207 public Reference getReference() { -208return this.reference; -209 } -210 -211 /** @return True if the store file is a Reference */ -212 public boolean isReference() { -213return this.reference != null; -214 } -215 -216 /** @return True if the store file is a top Reference */ -217 public boolean isTopReference() { -218return this.reference != null Reference.isTopFileRegion(this.reference.getFileRegion()); -219 } -220 -221 /** @return True if the store file is a link */ -222 public boolean isLink() { -223return this.link != null this.reference == null; +165 final HFileLink link) { +166this.fs = fs; +167this.conf = conf; +168// initialPath can be null only if we get a link. +169this.initialPath = (fileStatus == null) ? null : fileStatus.getPath(); +170 // HFileLink +171this.reference = null; +172this.link = link; +173 } +174 +175 /** +176 * Create a Store File Info from an HFileLink +177 * @param conf The {@link Configuration} to use +178 * @param fs The current file system to use +179 * @param fileStatus The {@link FileStatus} of the file +180 * @param reference The reference instance +181 */ +182 public StoreFileInfo(final Configuration conf, final FileSystem fs, final FileStatus fileStatus, +183 final Reference reference) { +184this.fs = fs; +185this.conf = conf; +186this.initialPath = fileStatus.getPath(); +187this.createdTimestamp = fileStatus.getModificationTime(); +188this.reference = reference; +189this.link = null; +190 } +191 +192 /** +193 * Create a Store File Info from an HFileLink and a Reference +194 * @param conf The {@link Configuration} to use +195 * @param fs The current file system to use +196 * @param fileStatus The {@link FileStatus} of the file +197 * @param reference The reference instance +198 * @param link The link instance +199 */ +200 public StoreFileInfo(final Configuration conf, final FileSystem fs, final FileStatus fileStatus, +201 final Reference reference, final HFileLink link) { +202this.fs = fs; +203this.conf = conf; +204this.initialPath = fileStatus.getPath(); +205this.createdTimestamp = fileStatus.getModificationTime(); +206this.reference = reference; +207this.link = link; +208 } +209 +210 /** +211 * Sets the region coprocessor env. +212 * @param coprocessorHost +213 */ +214 public void setRegionCoprocessorHost(RegionCoprocessorHost coprocessorHost) { +215this.coprocessorHost = coprocessorHost; +216 } +217 +218 /* +219 *
[03/22] hbase-site git commit: Published site at 46cb5dfa226892fd2580f26ce9ce77225bd7e67c.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8bc57a1a/testdevapidocs/org/apache/hadoop/hbase/ipc/MetricsHBaseServerWrapperStub.html -- diff --git a/testdevapidocs/org/apache/hadoop/hbase/ipc/MetricsHBaseServerWrapperStub.html b/testdevapidocs/org/apache/hadoop/hbase/ipc/MetricsHBaseServerWrapperStub.html index ba8e14f..75e6b86 100644 --- a/testdevapidocs/org/apache/hadoop/hbase/ipc/MetricsHBaseServerWrapperStub.html +++ b/testdevapidocs/org/apache/hadoop/hbase/ipc/MetricsHBaseServerWrapperStub.html @@ -18,7 +18,7 @@ catch(err) { } //--> -var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10}; +var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10}; var tabs = {65535:["t0","All Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]}; var altColor = "altColor"; var rowColor = "rowColor"; @@ -153,57 +153,69 @@ implements org.apache.hadoop.hbase.ipc.MetricsHBaseServerWrapper int -getActiveReadRpcHandlerCount() +getActiveGeneralRpcHandlerCount() int -getActiveRpcHandlerCount() +getActivePriorityRpcHandlerCount() int -getActiveScanRpcHandlerCount() +getActiveReadRpcHandlerCount() int -getActiveWriteRpcHandlerCount() +getActiveReplicationRpcHandlerCount() int -getGeneralQueueLength() +getActiveRpcHandlerCount() +int +getActiveScanRpcHandlerCount() + + +int +getActiveWriteRpcHandlerCount() + + +int +getGeneralQueueLength() + + long getNumGeneralCallsDropped() - + long getNumLifoModeSwitches() - + int getNumOpenConnections() - + int getPriorityQueueLength() - + int getReadQueueLength() - + int getReplicationQueueLength() - + int getScanQueueLength() - + long getTotalQueueSize() - + int getWriteQueueLength() @@ -324,13 +336,52 @@ implements org.apache.hadoop.hbase.ipc.MetricsHBaseServerWrapper + + + + + +getActiveGeneralRpcHandlerCount +publicintgetActiveGeneralRpcHandlerCount() + +Specified by: +getActiveGeneralRpcHandlerCountin interfaceorg.apache.hadoop.hbase.ipc.MetricsHBaseServerWrapper + + + + + + + + +getActivePriorityRpcHandlerCount +publicintgetActivePriorityRpcHandlerCount() + +Specified by: +getActivePriorityRpcHandlerCountin interfaceorg.apache.hadoop.hbase.ipc.MetricsHBaseServerWrapper + + + + + + + + +getActiveReplicationRpcHandlerCount +publicintgetActiveReplicationRpcHandlerCount() + +Specified by: +getActiveReplicationRpcHandlerCountin interfaceorg.apache.hadoop.hbase.ipc.MetricsHBaseServerWrapper + + + getNumGeneralCallsDropped -publiclonggetNumGeneralCallsDropped() +publiclonggetNumGeneralCallsDropped() Specified by: getNumGeneralCallsDroppedin interfaceorg.apache.hadoop.hbase.ipc.MetricsHBaseServerWrapper @@ -343,7 +394,7 @@ implements org.apache.hadoop.hbase.ipc.MetricsHBaseServerWrapper getNumLifoModeSwitches -publiclonggetNumLifoModeSwitches() +publiclonggetNumLifoModeSwitches() Specified by: getNumLifoModeSwitchesin interfaceorg.apache.hadoop.hbase.ipc.MetricsHBaseServerWrapper @@ -356,7 +407,7 @@ implements org.apache.hadoop.hbase.ipc.MetricsHBaseServerWrapper getWriteQueueLength -publicintgetWriteQueueLength() +publicintgetWriteQueueLength() Specified by: getWriteQueueLengthin interfaceorg.apache.hadoop.hbase.ipc.MetricsHBaseServerWrapper @@ -369,7 +420,7 @@ implements org.apache.hadoop.hbase.ipc.MetricsHBaseServerWrapper getReadQueueLength -publicintgetReadQueueLength() +publicintgetReadQueueLength() Specified by: getReadQueueLengthin interfaceorg.apache.hadoop.hbase.ipc.MetricsHBaseServerWrapper @@ -382,7 +433,7 @@ implements org.apache.hadoop.hbase.ipc.MetricsHBaseServerWrapper getScanQueueLength -publicintgetScanQueueLength() +publicintgetScanQueueLength() Specified by: getScanQueueLengthin interfaceorg.apache.hadoop.hbase.ipc.MetricsHBaseServerWrapper @@ -395,7 +446,7 @@ implements org.apache.hadoop.hbase.ipc.MetricsHBaseServerWrapper getActiveWriteRpcHandlerCount -publicintgetActiveWriteRpcHandlerCount() +publicintgetActiveWriteRpcHandlerCount() Specified by: getActiveWriteRpcHandlerCountin interfaceorg.apache.hadoop.hbase.ipc.MetricsHBaseServerWrapper @@ -408,7 +459,7 @@ implements org.apache.hadoop.hbase.ipc.MetricsHBaseServerWrapper getActiveReadRpcHandlerCount -publicintgetActiveReadRpcHandlerCount() +publicintgetActiveReadRpcHandlerCount() Specified by: getActiveReadRpcHandlerCountin interfaceorg.apache.hadoop.hbase.ipc.MetricsHBaseServerWrapper @@ -421,7 +472,7 @@ implements org.apache.hadoop.hbase.ipc.MetricsHBaseServerWrapper getActiveScanRpcHandlerCount -publicintgetActiveScanRpcHandlerCount() +publicintgetActiveScanRpcHandlerCount() Specified by: getActiveScanRpcHandlerCountin
[18/22] hbase-site git commit: Published site at 46cb5dfa226892fd2580f26ce9ce77225bd7e67c.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8bc57a1a/checkstyle.rss -- diff --git a/checkstyle.rss b/checkstyle.rss index 2da7c60..e2f8b27 100644 --- a/checkstyle.rss +++ b/checkstyle.rss @@ -26,7 +26,7 @@ under the License. 2007 - 2018 The Apache Software Foundation File: 3604, - Errors: 15864, + Errors: 15859, Warnings: 0, Infos: 0 @@ -3065,7 +3065,7 @@ under the License. 0 - 23 + 18 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8bc57a1a/coc.html -- diff --git a/coc.html b/coc.html index 0144c87..d9b9654 100644 --- a/coc.html +++ b/coc.html @@ -7,7 +7,7 @@ - + Apache HBase Code of Conduct Policy @@ -375,7 +375,7 @@ email to mailto:priv...@hbase.apache.org;>the priv https://www.apache.org/;>The Apache Software Foundation. All rights reserved. - Last Published: 2018-04-20 + Last Published: 2018-04-21 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8bc57a1a/dependencies.html -- diff --git a/dependencies.html b/dependencies.html index 461e974..d32dc6e 100644 --- a/dependencies.html +++ b/dependencies.html @@ -7,7 +7,7 @@ - + Apache HBase Project Dependencies @@ -440,7 +440,7 @@ https://www.apache.org/;>The Apache Software Foundation. All rights reserved. - Last Published: 2018-04-20 + Last Published: 2018-04-21 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8bc57a1a/dependency-convergence.html -- diff --git a/dependency-convergence.html b/dependency-convergence.html index 76564c8..64f6cd2 100644 --- a/dependency-convergence.html +++ b/dependency-convergence.html @@ -7,7 +7,7 @@ - + Apache HBase Reactor Dependency Convergence @@ -1105,7 +1105,7 @@ https://www.apache.org/;>The Apache Software Foundation. All rights reserved. - Last Published: 2018-04-20 + Last Published: 2018-04-21 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8bc57a1a/dependency-info.html -- diff --git a/dependency-info.html b/dependency-info.html index 6d9a8c7..e6384eb 100644 --- a/dependency-info.html +++ b/dependency-info.html @@ -7,7 +7,7 @@ - + Apache HBase Dependency Information @@ -313,7 +313,7 @@ https://www.apache.org/;>The Apache Software Foundation. All rights reserved. - Last Published: 2018-04-20 + Last Published: 2018-04-21 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8bc57a1a/dependency-management.html -- diff --git a/dependency-management.html b/dependency-management.html index fa99e3e..87dab1b 100644 --- a/dependency-management.html +++ b/dependency-management.html @@ -7,7 +7,7 @@ - + Apache HBase Project Dependency Management @@ -969,7 +969,7 @@ https://www.apache.org/;>The Apache Software Foundation. All rights reserved. - Last Published: 2018-04-20 + Last Published: 2018-04-21 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8bc57a1a/devapidocs/constant-values.html -- diff --git a/devapidocs/constant-values.html b/devapidocs/constant-values.html index 60dc3eb..ee70906 100644 --- a/devapidocs/constant-values.html +++ b/devapidocs/constant-values.html @@ -3768,21 +3768,21 @@ publicstaticfinalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String date -"Fri Apr 20 14:39:14 UTC 2018" +"Sat Apr 21 14:39:09 UTC 2018" publicstaticfinalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String revision
[22/22] hbase-site git commit: Published site at 46cb5dfa226892fd2580f26ce9ce77225bd7e67c.
Published site at 46cb5dfa226892fd2580f26ce9ce77225bd7e67c. Project: http://git-wip-us.apache.org/repos/asf/hbase-site/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase-site/commit/8bc57a1a Tree: http://git-wip-us.apache.org/repos/asf/hbase-site/tree/8bc57a1a Diff: http://git-wip-us.apache.org/repos/asf/hbase-site/diff/8bc57a1a Branch: refs/heads/asf-site Commit: 8bc57a1a277bc0ef1998aa33e1043ce08e5df096 Parents: 84a296a Author: jenkinsAuthored: Sat Apr 21 14:47:17 2018 + Committer: jenkins Committed: Sat Apr 21 14:47:17 2018 + -- acid-semantics.html | 4 +- apache_hbase_reference_guide.pdf| 9510 ++--- book.html | 2 +- bulk-loads.html | 4 +- checkstyle-aggregate.html | 11892 - checkstyle.rss | 4 +- coc.html| 4 +- dependencies.html | 4 +- dependency-convergence.html | 4 +- dependency-info.html| 4 +- dependency-management.html | 4 +- devapidocs/constant-values.html |50 +- devapidocs/index-all.html |66 +- .../hadoop/hbase/backup/package-tree.html | 4 +- .../hadoop/hbase/client/package-tree.html |22 +- .../hadoop/hbase/executor/package-tree.html | 2 +- .../hadoop/hbase/filter/package-tree.html |10 +- .../hadoop/hbase/io/class-use/HFileLink.html| 9 + .../hadoop/hbase/io/class-use/Reference.html| 9 + .../hadoop/hbase/io/hfile/package-tree.html | 6 +- .../hadoop/hbase/ipc/FifoRpcScheduler.html | 117 +- .../hbase/ipc/MetricsHBaseServerSource.html | 146 +- .../hbase/ipc/MetricsHBaseServerSourceImpl.html | 2 +- .../hbase/ipc/MetricsHBaseServerWrapper.html|83 +- .../ipc/MetricsHBaseServerWrapperImpl.html |97 +- .../apache/hadoop/hbase/ipc/RpcScheduler.html | 100 +- .../hadoop/hbase/ipc/SimpleRpcScheduler.html| 119 +- .../apache/hadoop/hbase/ipc/package-tree.html | 2 +- .../hadoop/hbase/mapreduce/package-tree.html| 4 +- .../hbase/master/balancer/package-tree.html | 2 +- .../hadoop/hbase/master/package-tree.html | 4 +- .../hbase/master/procedure/package-tree.html| 2 +- .../org/apache/hadoop/hbase/package-tree.html |14 +- .../hadoop/hbase/procedure2/package-tree.html | 4 +- .../hadoop/hbase/quotas/NoopQuotaLimiter.html |34 +- .../hadoop/hbase/quotas/QuotaLimiter.html |36 +- .../hadoop/hbase/quotas/TimeBasedLimiter.html |54 +- .../class-use/RpcThrottlingException.html |14 +- .../hadoop/hbase/quotas/package-tree.html | 8 +- .../hbase/regionserver/StoreFileInfo.html | 113 +- .../hadoop/hbase/regionserver/package-tree.html |14 +- .../regionserver/querymatcher/package-tree.html | 2 +- .../hbase/regionserver/wal/package-tree.html| 2 +- .../replication/regionserver/package-tree.html | 2 +- .../hbase/security/access/package-tree.html | 2 +- .../hadoop/hbase/security/package-tree.html | 2 +- .../thrift/ThriftServerRunner.HBaseHandler.html | 142 +- .../ThriftServerRunner.IOErrorWithCause.html|12 +- .../thrift/ThriftServerRunner.ImplType.html |38 +- ...ThriftServerRunner.ResultScannerWrapper.html |12 +- .../hadoop/hbase/thrift/ThriftServerRunner.html | 108 +- .../hadoop/hbase/thrift/package-tree.html | 2 +- .../hbase/util/ServerRegionReplicaUtil.html |12 +- .../apache/hadoop/hbase/util/package-tree.html |10 +- .../apache/hadoop/hbase/wal/package-tree.html | 2 +- .../org/apache/hadoop/hbase/Version.html| 6 +- .../ipc/FifoRpcScheduler.FifoCallRunner.html|96 +- .../hadoop/hbase/ipc/FifoRpcScheduler.html |96 +- .../hbase/ipc/MetricsHBaseServerSource.html |64 +- .../hbase/ipc/MetricsHBaseServerSourceImpl.html |50 +- .../hbase/ipc/MetricsHBaseServerWrapper.html|24 +- .../ipc/MetricsHBaseServerWrapperImpl.html |58 +- .../hadoop/hbase/ipc/RpcScheduler.Context.html |65 +- .../apache/hadoop/hbase/ipc/RpcScheduler.html |65 +- .../hadoop/hbase/ipc/SimpleRpcScheduler.html| 158 +- .../hbase/quotas/DefaultOperationQuota.html | 4 +- .../hadoop/hbase/quotas/NoopQuotaLimiter.html | 6 +- .../hadoop/hbase/quotas/QuotaLimiter.html |80 +- .../hadoop/hbase/quotas/TimeBasedLimiter.html | 173 +- .../hbase/regionserver/StoreFileInfo.html | 831 +- .../thrift/ThriftServerRunner.HBaseHandler.html | 3880 +++--- .../ThriftServerRunner.IOErrorWithCause.html
[04/22] hbase-site git commit: Published site at 46cb5dfa226892fd2580f26ce9ce77225bd7e67c.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8bc57a1a/devapidocs/src-html/org/apache/hadoop/hbase/util/ServerRegionReplicaUtil.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/util/ServerRegionReplicaUtil.html b/devapidocs/src-html/org/apache/hadoop/hbase/util/ServerRegionReplicaUtil.html index 6bdffeb..3dbb30a 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/util/ServerRegionReplicaUtil.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/util/ServerRegionReplicaUtil.html @@ -136,76 +136,89 @@ 128} 129 130// else create a store file link. The link file does not exists on filesystem though. -131HFileLink link = HFileLink.build(conf, regionInfoForFs.getTable(), -132 regionInfoForFs.getEncodedName(), familyName, path.getName()); -133 -134if (StoreFileInfo.isReference(path)) { -135 Reference reference = Reference.read(fs, path); -136 return new StoreFileInfo(conf, fs, link.getFileStatus(fs), reference); -137} -138 -139return new StoreFileInfo(conf, fs, link.getFileStatus(fs), link); -140 } -141 -142 /** -143 * Create replication peer for replicating to region replicas if needed. -144 * @param conf configuration to use -145 * @throws IOException -146 */ -147 public static void setupRegionReplicaReplication(Configuration conf) throws IOException { -148if (!isRegionReplicaReplicationEnabled(conf)) { -149 return; -150} -151Admin admin = ConnectionFactory.createConnection(conf).getAdmin(); -152ReplicationPeerConfig peerConfig = null; -153try { -154 peerConfig = admin.getReplicationPeerConfig(REGION_REPLICA_REPLICATION_PEER); -155} catch (ReplicationPeerNotFoundException e) { -156 LOG.warn("Region replica replication peer id=" + REGION_REPLICA_REPLICATION_PEER -157 + " not exist", e); -158} -159try { -160 if (peerConfig == null) { -161LOG.info("Region replica replication peer id=" + REGION_REPLICA_REPLICATION_PEER -162+ " not exist. Creating..."); -163peerConfig = new ReplicationPeerConfig(); -164 peerConfig.setClusterKey(ZKConfig.getZooKeeperClusterKey(conf)); -165 peerConfig.setReplicationEndpointImpl(RegionReplicaReplicationEndpoint.class.getName()); -166 admin.addReplicationPeer(REGION_REPLICA_REPLICATION_PEER, peerConfig); -167 } -168} finally { -169 admin.close(); -170} -171 } -172 -173 public static boolean isRegionReplicaReplicationEnabled(Configuration conf) { -174return conf.getBoolean(REGION_REPLICA_REPLICATION_CONF_KEY, -175 DEFAULT_REGION_REPLICA_REPLICATION); -176 } -177 -178 public static boolean isRegionReplicaWaitForPrimaryFlushEnabled(Configuration conf) { -179return conf.getBoolean(REGION_REPLICA_WAIT_FOR_PRIMARY_FLUSH_CONF_KEY, -180 DEFAULT_REGION_REPLICA_WAIT_FOR_PRIMARY_FLUSH); -181 } -182 -183 public static boolean isRegionReplicaStoreFileRefreshEnabled(Configuration conf) { -184return conf.getBoolean(REGION_REPLICA_STORE_FILE_REFRESH, -185 DEFAULT_REGION_REPLICA_STORE_FILE_REFRESH); -186 } -187 -188 public static double getRegionReplicaStoreFileRefreshMultiplier(Configuration conf) { -189return conf.getDouble(REGION_REPLICA_STORE_FILE_REFRESH_MEMSTORE_MULTIPLIER, -190 DEFAULT_REGION_REPLICA_STORE_FILE_REFRESH_MEMSTORE_MULTIPLIER); -191 } -192 -193 /** -194 * Return the peer id used for replicating to secondary region replicas -195 */ -196 public static String getReplicationPeerId() { -197return REGION_REPLICA_REPLICATION_PEER; -198 } -199 -200} +131if (HFileLink.isHFileLink(path) || StoreFileInfo.isHFile(path)) { +132 HFileLink link = HFileLink +133 .build(conf, regionInfoForFs.getTable(), regionInfoForFs.getEncodedName(), familyName, +134 path.getName()); +135 return new StoreFileInfo(conf, fs, link.getFileStatus(fs), link); +136} else if (StoreFileInfo.isReference(path)) { +137 Reference reference = Reference.read(fs, path); +138 Path referencePath = StoreFileInfo.getReferredToFile(path); +139 if (HFileLink.isHFileLink(referencePath)) { +140// HFileLink Reference +141HFileLink link = HFileLink.buildFromHFileLinkPattern(conf, referencePath); +142return new StoreFileInfo(conf, fs, link.getFileStatus(fs), reference, link); +143 } else { +144// Reference +145HFileLink link = HFileLink +146.build(conf, regionInfoForFs.getTable(), regionInfoForFs.getEncodedName(), familyName, +147path.getName()); +148return new StoreFileInfo(conf, fs, link.getFileStatus(fs), reference); +149 } +150} else { +151 throw new IOException("path=" + path + " doesn't look like a valid StoreFile"); +152} +153 } +154 +155 /** +156 * Create replication peer for replicating