[05/11] hbase-site git commit: Published site at e963694259aa9242aa709e9b954754abd9a681cd.

2018-07-28 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e10887fe/devapidocs/src-html/org/apache/hadoop/hbase/backup/util/RestoreTool.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/backup/util/RestoreTool.html 
b/devapidocs/src-html/org/apache/hadoop/hbase/backup/util/RestoreTool.html
index bbbf2a1..4153319 100644
--- a/devapidocs/src-html/org/apache/hadoop/hbase/backup/util/RestoreTool.html
+++ b/devapidocs/src-html/org/apache/hadoop/hbase/backup/util/RestoreTool.html
@@ -244,279 +244,280 @@
 236Path tableInfoPath = null;
 237
 238// can't build the path directly as 
the timestamp values are different
-239FileStatus[] snapshots = 
fs.listStatus(tableSnapShotPath);
-240for (FileStatus snapshot : snapshots) 
{
-241  tableInfoPath = 
snapshot.getPath();
-242  // 
SnapshotManifest.DATA_MANIFEST_NAME = "data.manifest";
-243  if 
(tableInfoPath.getName().endsWith("data.manifest")) {
-244break;
-245  }
-246}
-247return tableInfoPath;
-248  }
-249
-250  /**
-251   * Get table descriptor
-252   * @param tableName is the table backed 
up
-253   * @return {@link TableDescriptor} 
saved in backup image of the table
-254   */
-255  TableDescriptor getTableDesc(TableName 
tableName) throws IOException {
-256Path tableInfoPath = 
this.getTableInfoPath(tableName);
-257SnapshotDescription desc = 
SnapshotDescriptionUtils.readSnapshotInfo(fs, tableInfoPath);
-258SnapshotManifest manifest = 
SnapshotManifest.open(conf, fs, tableInfoPath, desc);
-259TableDescriptor tableDescriptor = 
manifest.getTableDescriptor();
-260if 
(!tableDescriptor.getTableName().equals(tableName)) {
-261  LOG.error("couldn't find Table Desc 
for table: " + tableName + " under tableInfoPath: "
-262  + 
tableInfoPath.toString());
-263  
LOG.error("tableDescriptor.getNameAsString() = "
-264  + 
tableDescriptor.getTableName().getNameAsString());
-265  throw new 
FileNotFoundException("couldn't find Table Desc for table: " + tableName
-266  + " under tableInfoPath: " + 
tableInfoPath.toString());
-267}
-268return tableDescriptor;
-269  }
-270
-271  private TableDescriptor 
getTableDescriptor(FileSystem fileSys, TableName tableName,
-272  String lastIncrBackupId) throws 
IOException {
-273if (lastIncrBackupId != null) {
-274  String target =
-275  
BackupUtils.getTableBackupDir(backupRootPath.toString(),
-276lastIncrBackupId, 
tableName);
-277  return 
FSTableDescriptors.getTableDescriptorFromFs(fileSys, new Path(target));
-278}
-279return null;
-280  }
-281
-282  private void 
createAndRestoreTable(Connection conn, TableName tableName, TableName 
newTableName,
-283  Path tableBackupPath, boolean 
truncateIfExists, String lastIncrBackupId) throws IOException {
-284if (newTableName == null) {
-285  newTableName = tableName;
-286}
-287FileSystem fileSys = 
tableBackupPath.getFileSystem(this.conf);
-288
-289// get table descriptor first
-290TableDescriptor tableDescriptor = 
getTableDescriptor(fileSys, tableName, lastIncrBackupId);
-291if (tableDescriptor != null) {
-292  LOG.debug("Retrieved descriptor: " 
+ tableDescriptor + " thru " + lastIncrBackupId);
-293}
-294
-295if (tableDescriptor == null) {
-296  Path tableSnapshotPath = 
getTableSnapshotPath(backupRootPath, tableName, backupId);
-297  if 
(fileSys.exists(tableSnapshotPath)) {
-298// snapshot path exist means the 
backup path is in HDFS
-299// check whether snapshot dir 
already recorded for target table
-300if (snapshotMap.get(tableName) != 
null) {
-301  SnapshotDescription desc =
-302  
SnapshotDescriptionUtils.readSnapshotInfo(fileSys, tableSnapshotPath);
-303  SnapshotManifest manifest = 
SnapshotManifest.open(conf, fileSys, tableSnapshotPath, desc);
-304  tableDescriptor = 
manifest.getTableDescriptor();
-305} else {
-306  tableDescriptor = 
getTableDesc(tableName);
-307  snapshotMap.put(tableName, 
getTableInfoPath(tableName));
-308}
-309if (tableDescriptor == null) {
-310  LOG.debug("Found no table 
descriptor in the snapshot dir, previous schema would be lost");
-311}
-312  } else {
-313throw new IOException("Table 
snapshot directory: " +
-314tableSnapshotPath + " does 
not exist.");
-315  }
-316}
-317
-318Path tableArchivePath = 
getTableArchivePath(tableName);
-319if (tableArchivePath == null) {
-320  if (tableDescriptor != null) {
-321// find table descriptor but no 
archive dir means the table is empty, create table and exit
-322if (LOG.isDebugEnabled()) {
-323  LOG.debug("find table 
descriptor but no archive dir for table " + tableName
-324  + ", will only create 
table");
-325}
-326   

[02/11] hbase-site git commit: Published site at e963694259aa9242aa709e9b954754abd9a681cd.

2018-07-28 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e10887fe/devapidocs/src-html/org/apache/hadoop/hbase/util/CommonFSUtils.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/util/CommonFSUtils.html 
b/devapidocs/src-html/org/apache/hadoop/hbase/util/CommonFSUtils.html
index c6bfdca..cae3d07 100644
--- a/devapidocs/src-html/org/apache/hadoop/hbase/util/CommonFSUtils.html
+++ b/devapidocs/src-html/org/apache/hadoop/hbase/util/CommonFSUtils.html
@@ -829,90 +829,217 @@
 821conf.setIfUnset(dfsKey, 
Integer.toString(hbaseSize));
 822  }
 823
-824  // Holder singleton idiom. JVM spec 
ensures this will be run at most once per Classloader, and
-825  // not until we attempt to reference 
it.
-826  private static class StreamCapabilities 
{
-827public static final boolean 
PRESENT;
-828public static final Class 
CLASS;
-829public static final Method METHOD;
-830static {
-831  boolean tmp = false;
-832  Class clazz = null;
-833  Method method = null;
-834  try {
-835clazz = 
Class.forName("org.apache.hadoop.fs.StreamCapabilities");
-836method = 
clazz.getMethod("hasCapability", String.class);
-837tmp = true;
-838  } 
catch(ClassNotFoundException|NoSuchMethodException|SecurityException exception) 
{
-839LOG.warn("Your Hadoop 
installation does not include the StreamCapabilities class from " +
-840 "HDFS-11644, so we will 
skip checking if any FSDataOutputStreams actually " +
-841 "support hflush/hsync. 
If you are running on top of HDFS this probably just " +
-842 "means you have an older 
version and this can be ignored. If you are running on " +
-843 "top of an alternate 
FileSystem implementation you should manually verify that " +
-844 "hflush and hsync are 
implemented; otherwise you risk data loss and hard to " +
-845 "diagnose errors when 
our assumptions are violated.");
-846LOG.debug("The first request to 
check for StreamCapabilities came from this stacktrace.",
-847exception);
-848  } finally {
-849PRESENT = tmp;
-850CLASS = clazz;
-851METHOD = method;
-852  }
-853}
-854  }
-855
-856  /**
-857   * If our FileSystem version includes 
the StreamCapabilities class, check if
-858   * the given stream has a particular 
capability.
-859   * @param stream capabilities are 
per-stream instance, so check this one specifically. must not be
-860   *null
-861   * @param capability what to look for, 
per Hadoop Common's FileSystem docs
-862   * @return true if there are no 
StreamCapabilities. false if there are, but this stream doesn't
-863   * implement it. return result 
of asking the stream otherwise.
-864   */
-865  public static boolean 
hasCapability(FSDataOutputStream stream, String capability) {
-866// be consistent whether or not 
StreamCapabilities is present
-867if (stream == null) {
-868  throw new 
NullPointerException("stream parameter must not be null.");
-869}
-870// If o.a.h.fs.StreamCapabilities 
doesn't exist, assume everyone does everything
-871// otherwise old versions of Hadoop 
will break.
-872boolean result = true;
-873if (StreamCapabilities.PRESENT) {
-874  // if StreamCapabilities is 
present, but the stream doesn't implement it
-875  // or we run into a problem 
invoking the method,
-876  // we treat that as equivalent to 
not declaring anything
-877  result = false;
-878  if 
(StreamCapabilities.CLASS.isAssignableFrom(stream.getClass())) {
+824  private static class DfsBuilderUtility 
{
+825static Class dfsClass = 
null;
+826static Method createMethod;
+827static Method overwriteMethod;
+828static Method bufferSizeMethod;
+829static Method blockSizeMethod;
+830static Method recursiveMethod;
+831static Method replicateMethod;
+832static Method replicationMethod;
+833static Method buildMethod;
+834static boolean allMethodsPresent = 
false;
+835
+836static {
+837  String dfsName = 
"org.apache.hadoop.hdfs.DistributedFileSystem";
+838  String builderName = dfsName + 
"$HdfsDataOutputStreamBuilder";
+839  Class builderClass = 
null;
+840
+841  try {
+842dfsClass = 
Class.forName(dfsName);
+843  } catch (ClassNotFoundException e) 
{
+844LOG.debug("{} not available, will 
not use builder API for file creation.", dfsName);
+845  }
+846  try {
+847builderClass = 
Class.forName(builderName);
+848  } catch (ClassNotFoundException e) 
{
+849LOG.debug("{} not available, will 
not use builder API for file creation.", builderName);
+850  }
+851
+852  if (dfsClass != null && 
builderClass != null) {
+853try {
+854  createMethod = 
dfsClass.getMethod("createFile", Path.class);
+855  over

[09/11] hbase-site git commit: Published site at e963694259aa9242aa709e9b954754abd9a681cd.

2018-07-28 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e10887fe/checkstyle.rss
--
diff --git a/checkstyle.rss b/checkstyle.rss
index 579f337..2ac991b 100644
--- a/checkstyle.rss
+++ b/checkstyle.rss
@@ -25,8 +25,8 @@ under the License.
 en-us
 ©2007 - 2018 The Apache Software Foundation
 
-  File: 3696,
- Errors: 15632,
+  File: 3697,
+ Errors: 15630,
  Warnings: 0,
  Infos: 0
   
@@ -5669,7 +5669,7 @@ under the License.
   0
 
 
-  2
+  0
 
   
   
@@ -45532,6 +45532,20 @@ under the License.
   
   
 
+  http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.regionserver.wal.TestHBaseWalOnEC.java";>org/apache/hadoop/hbase/regionserver/wal/TestHBaseWalOnEC.java
+
+
+  0
+
+
+  0
+
+
+  0
+
+  
+  
+
   http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.thrift2.generated.TResult.java";>org/apache/hadoop/hbase/thrift2/generated/TResult.java
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e10887fe/coc.html
--
diff --git a/coc.html b/coc.html
index c44ea99..1c93089 100644
--- a/coc.html
+++ b/coc.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase – 
   Code of Conduct Policy
@@ -375,7 +375,7 @@ email to mailto:priv...@hbase.apache.org";>the priv
 https://www.apache.org/";>The Apache Software 
Foundation.
 All rights reserved.  
 
-  Last Published: 
2018-07-27
+  Last Published: 
2018-07-28
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e10887fe/dependencies.html
--
diff --git a/dependencies.html b/dependencies.html
index b8acd13..dd4686f 100644
--- a/dependencies.html
+++ b/dependencies.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase – Project Dependencies
 
@@ -440,7 +440,7 @@
 https://www.apache.org/";>The Apache Software 
Foundation.
 All rights reserved.  
 
-  Last Published: 
2018-07-27
+  Last Published: 
2018-07-28
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e10887fe/dependency-convergence.html
--
diff --git a/dependency-convergence.html b/dependency-convergence.html
index 6887e4d..cda2363 100644
--- a/dependency-convergence.html
+++ b/dependency-convergence.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase – Reactor Dependency Convergence
 
@@ -905,7 +905,7 @@
 https://www.apache.org/";>The Apache Software 
Foundation.
 All rights reserved.  
 
-  Last Published: 
2018-07-27
+  Last Published: 
2018-07-28
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e10887fe/dependency-info.html
--
diff --git a/dependency-info.html b/dependency-info.html
index 860411e..cff8007 100644
--- a/dependency-info.html
+++ b/dependency-info.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase – Dependency Information
 
@@ -313,7 +313,7 @@
 https://www.apache.org/";>The Apache Software 
Foundation.
 All rights reserved.  
 
-  Last Published: 
2018-07-27
+  Last Published: 
2018-07-28
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e10887fe/dependency-management.html
--
diff --git a/dependency-management.html b/dependency-management.html
index cdc598a..29db2b6 100644
--- a/dependency-management.html
+++ b/dependency-management.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase – Project Dependency Management
 
@@ -1005,7 +1005,7 @@
 https://www.apache.org/";>The Apache Software 
Foundation.
 All rights reserved.  
 
-  Last Published: 
2018-07-27
+  Last Published: 
2018-07-28
 
 
 

http://

[10/11] hbase-site git commit: Published site at e963694259aa9242aa709e9b954754abd9a681cd.

2018-07-28 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e10887fe/checkstyle-aggregate.html
--
diff --git a/checkstyle-aggregate.html b/checkstyle-aggregate.html
index fe45785..80490cd 100644
--- a/checkstyle-aggregate.html
+++ b/checkstyle-aggregate.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase – Checkstyle Results
 
@@ -281,10 +281,10 @@
  Warnings
  Errors
 
-3696
+3697
 0
 0
-15632
+15630
 
 Files
 
@@ -7584,2216 +7584,2211 @@
 0
 22
 
-org/apache/hadoop/hbase/regionserver/wal/ProtobufLogWriter.java
-0
-0
-2
-
 org/apache/hadoop/hbase/regionserver/wal/ReaderBase.java
 0
 0
 1
-
+
 org/apache/hadoop/hbase/regionserver/wal/SecureProtobufLogReader.java
 0
 0
 2
-
+
 org/apache/hadoop/hbase/regionserver/wal/SecureProtobufLogWriter.java
 0
 0
 1
-
+
 org/apache/hadoop/hbase/regionserver/wal/SecureWALCellCodec.java
 0
 0
 2
-
+
 org/apache/hadoop/hbase/regionserver/wal/SequenceIdAccounting.java
 0
 0
 13
-
+
 org/apache/hadoop/hbase/regionserver/wal/TestLogRolling.java
 0
 0
 3
-
+
 org/apache/hadoop/hbase/regionserver/wal/TestLogRollingNoCluster.java
 0
 0
 4
-
+
 org/apache/hadoop/hbase/regionserver/wal/TestWALConfiguration.java
 0
 0
 3
-
+
 org/apache/hadoop/hbase/regionserver/wal/WALActionsListener.java
 0
 0
 2
-
+
 org/apache/hadoop/hbase/regionserver/wal/WALCellCodec.java
 0
 0
 6
-
+
 org/apache/hadoop/hbase/regionserver/wal/WALCoprocessorHost.java
 0
 0
 1
-
+
 org/apache/hadoop/hbase/regionserver/wal/WALUtil.java
 0
 0
 5
-
+
 org/apache/hadoop/hbase/replication/BaseReplicationEndpoint.java
 0
 0
 3
-
+
 org/apache/hadoop/hbase/replication/BulkLoadCellFilter.java
 0
 0
 1
-
+
 org/apache/hadoop/hbase/replication/ClusterMarkingEntryFilter.java
 0
 0
 3
-
+
 org/apache/hadoop/hbase/replication/HBaseReplicationEndpoint.java
 0
 0
 3
-
+
 org/apache/hadoop/hbase/replication/ReplicationEndpoint.java
 0
 0
 4
-
+
 org/apache/hadoop/hbase/replication/ReplicationException.java
 0
 0
 1
-
+
 org/apache/hadoop/hbase/replication/SystemTableWALEntryFilter.java
 0
 0
 1
-
+
 org/apache/hadoop/hbase/replication/TestMasterReplication.java
 0
 0
 12
-
+
 org/apache/hadoop/hbase/replication/TestMultiSlaveReplication.java
 0
 0
 7
-
+
 org/apache/hadoop/hbase/replication/TestPerTableCFReplication.java
 0
 0
 3
-
+
 org/apache/hadoop/hbase/replication/TestReplicationDisableInactivePeer.java
 0
 0
 1
-
+
 org/apache/hadoop/hbase/replication/TestReplicationEndpoint.java
 0
 0
 7
-
+
 org/apache/hadoop/hbase/replication/TestReplicationKillMasterRSCompressed.java
 0
 0
 1
-
+
 org/apache/hadoop/hbase/replication/TestReplicationWithTags.java
 0
 0
 1
-
+
 org/apache/hadoop/hbase/replication/TestVerifyReplication.java
 0
 0
 9
-
+
 org/apache/hadoop/hbase/replication/WALCellFilter.java
 0
 0
 2
-
+
 org/apache/hadoop/hbase/replication/WALEntryFilter.java
 0
 0
 2
-
+
 org/apache/hadoop/hbase/replication/ZKReplicationQueueStorage.java
 0
 0
 3
-
+
 org/apache/hadoop/hbase/replication/master/ReplicationLogCleaner.java
 0
 0
 1
-
+
 org/apache/hadoop/hbase/replication/regionserver/DumpReplicationQueues.java
 0
 0
 6
-
+
 org/apache/hadoop/hbase/replication/regionserver/HFileReplicator.java
 0
 0
 3
-
+
 org/apache/hadoop/hbase/replication/regionserver/MetricsSink.java
 0
 0
 2
-
+
 org/apache/hadoop/hbase/replication/regionserver/MetricsSource.java
 0
 0
 2
-
+
 org/apache/hadoop/hbase/replication/regionserver/RegionReplicaReplicationEndpoint.java
 0
 0
 3
-
+
 org/apache/hadoop/hbase/replication/regionserver/Replication.java
 0
 0
 2
-
+
 org/apache/hadoop/hbase/replication/regionserver/ReplicationLoad.java
 0
 0
 6
-
+
 org/apache/hadoop/hbase/replication/regionserver/ReplicationObserver.java
 0
 0
 4
-
+
 org/apache/hadoop/hbase/replication/regionserver/ReplicationSink.java
 0
 0
 11
-
+
 org/apache/hadoop/hbase/replication/regionserver/ReplicationSourceFactory.java
 0
 0
 2
-
+
 org/apache/hadoop/hbase/replication/regionserver/ReplicationSourceInterface.java
 0
 0
 1
-
+
 org/apache/hadoop/hbase/replication/regionserver/ReplicationSourceManager.java
 0
 0
 3
-
+
 org/apache/hadoop/hbase/replication/regionserver/ReplicationThrottler.java
 0
 0
 2
-
+
 org/apache/hadoop/hbase/replication/regionserver/TestGlobalReplicationThrottler.java
 0
 0
 3
-
+
 org/apache/hadoop/hbase/replication/regionserver/TestRegionReplicaReplicationEndpointNoMaster.java
 0
 0
 1
-
+
 org/apache/hadoop/hbase/replication/regionserver/TestReplicationSink.java
 0
 0
 11
-
+
 org/apache/hadoop/hbase/replication/regionserver/TestReplicationSource.java
 0
 0
 10
-
+
 org/apache/hadoop/hbase/replication/regionserver/TestReplicationSourceManager.java
 0
 0
 3
-
+
 org/apache/hadoop/hbase/replication/regionserver/TestWALEntrySinkFilter.java
 0
 0
 16
-
+
 org/apache/hadoop/hbase/replication/regionserver/TestWALEntryStream.java
 0
 0
 1
-
+
 org/apache/hadoop/hbase/replication/regionserver/WALEntrySinkFilter.java
 0
 0
 2
-
+
 org/apache/hadoop/hbase/replication

[08/11] hbase-site git commit: Published site at e963694259aa9242aa709e9b954754abd9a681cd.

2018-07-28 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e10887fe/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html 
b/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html
index 482efa8..4eb2488 100644
--- a/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html
+++ b/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html
@@ -704,20 +704,20 @@
 
 java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true";
 title="class or interface in java.lang">Enum (implements java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true";
 title="class or interface in java.lang">Comparable, java.io.https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true";
 title="class or interface in java.io">Serializable)
 
-org.apache.hadoop.hbase.regionserver.ScannerContext.NextState
-org.apache.hadoop.hbase.regionserver.CompactingMemStore.IndexType
-org.apache.hadoop.hbase.regionserver.DefaultHeapMemoryTuner.StepDirection
-org.apache.hadoop.hbase.regionserver.ChunkCreator.ChunkType
-org.apache.hadoop.hbase.regionserver.HRegion.FlushResult.Result
 org.apache.hadoop.hbase.regionserver.SplitLogWorker.TaskExecutor.Status
 org.apache.hadoop.hbase.regionserver.TimeRangeTracker.Type
 org.apache.hadoop.hbase.regionserver.ScanType
-org.apache.hadoop.hbase.regionserver.MetricsRegionServerSourceFactoryImpl.FactoryStorage
+org.apache.hadoop.hbase.regionserver.ScannerContext.LimitScope
 org.apache.hadoop.hbase.regionserver.Region.Operation
+org.apache.hadoop.hbase.regionserver.ChunkCreator.ChunkType
+org.apache.hadoop.hbase.regionserver.CompactingMemStore.IndexType
+org.apache.hadoop.hbase.regionserver.ScannerContext.NextState
+org.apache.hadoop.hbase.regionserver.HRegion.FlushResult.Result
+org.apache.hadoop.hbase.regionserver.MetricsRegionServerSourceFactoryImpl.FactoryStorage
+org.apache.hadoop.hbase.regionserver.BloomType
 org.apache.hadoop.hbase.regionserver.MemStoreCompactionStrategy.Action
 org.apache.hadoop.hbase.regionserver.FlushType
-org.apache.hadoop.hbase.regionserver.BloomType
-org.apache.hadoop.hbase.regionserver.ScannerContext.LimitScope
+org.apache.hadoop.hbase.regionserver.DefaultHeapMemoryTuner.StepDirection
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e10887fe/devapidocs/org/apache/hadoop/hbase/regionserver/querymatcher/package-tree.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/regionserver/querymatcher/package-tree.html
 
b/devapidocs/org/apache/hadoop/hbase/regionserver/querymatcher/package-tree.html
index 858ccf6..b377318 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/regionserver/querymatcher/package-tree.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/regionserver/querymatcher/package-tree.html
@@ -130,9 +130,9 @@
 
 java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true";
 title="class or interface in java.lang">Enum (implements java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true";
 title="class or interface in java.lang">Comparable, java.io.https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true";
 title="class or interface in java.io">Serializable)
 
-org.apache.hadoop.hbase.regionserver.querymatcher.DeleteTracker.DeleteResult
 org.apache.hadoop.hbase.regionserver.querymatcher.StripeCompactionScanQueryMatcher.DropDeletesInOutput
 org.apache.hadoop.hbase.regionserver.querymatcher.ScanQueryMatcher.MatchCode
+org.apache.hadoop.hbase.regionserver.querymatcher.DeleteTracker.DeleteResult
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e10887fe/devapidocs/org/apache/hadoop/hbase/regionserver/wal/ProtobufLogWriter.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/regionserver/wal/ProtobufLogWriter.html 
b/devapidocs/org/apache/hadoop/hbase/regionserver/wal/ProtobufLogWriter.html
index 221b016..1e59005 100644
--- a/devapidocs/org/apache/hadoop/hbase/regionserver/wal/ProtobufLogWriter.html
+++ b/devapidocs/org/apache/hadoop/hbase/regionserver/wal/ProtobufLogWriter.html
@@ -359,7 +359,7 @@ implements 
 
 sync
-public void sync(boolean forceSync)
+public void sync(boolean forceSync)
   throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true";
 title="class or interface in java.io">IOException
 
 Specified by:
@@ -375,7 +375,7 @@ implements 
 
 getStream
-public org.apache.hadoop.fs.FSDataOutputStream getStream()
+public org.apache.hadoop.fs.FSDataOutputStream getStream()
 
 
 
@@ -384,7 +384,7 @@ implements 
 
 initOutput
-protected void initOutput(org.apache.hadoop.fs.FileSystem fs,
+protected void in

[04/11] hbase-site git commit: Published site at e963694259aa9242aa709e9b954754abd9a681cd.

2018-07-28 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e10887fe/devapidocs/src-html/org/apache/hadoop/hbase/util/CommonFSUtils.DfsBuilderUtility.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/util/CommonFSUtils.DfsBuilderUtility.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/util/CommonFSUtils.DfsBuilderUtility.html
new file mode 100644
index 000..cae3d07
--- /dev/null
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/util/CommonFSUtils.DfsBuilderUtility.html
@@ -0,0 +1,1106 @@
+http://www.w3.org/TR/html4/loose.dtd";>
+
+
+Source code
+
+
+
+
+001/**
+002 *
+003 * Licensed to the Apache Software 
Foundation (ASF) under one
+004 * or more contributor license 
agreements.  See the NOTICE file
+005 * distributed with this work for 
additional information
+006 * regarding copyright ownership.  The 
ASF licenses this file
+007 * to you under the Apache License, 
Version 2.0 (the
+008 * "License"); you may not use this file 
except in compliance
+009 * with the License.  You may obtain a 
copy of the License at
+010 *
+011 * 
http://www.apache.org/licenses/LICENSE-2.0
+012 *
+013 * Unless required by applicable law or 
agreed to in writing, software
+014 * distributed under the License is 
distributed on an "AS IS" BASIS,
+015 * WITHOUT WARRANTIES OR CONDITIONS OF 
ANY KIND, either express or implied.
+016 * See the License for the specific 
language governing permissions and
+017 * limitations under the License.
+018 */
+019package org.apache.hadoop.hbase.util;
+020
+021import java.io.FileNotFoundException;
+022import java.io.IOException;
+023import 
java.lang.reflect.InvocationTargetException;
+024import java.lang.reflect.Method;
+025import java.net.URI;
+026import java.net.URISyntaxException;
+027import java.util.List;
+028import java.util.Locale;
+029import java.util.Map;
+030import 
java.util.concurrent.ConcurrentHashMap;
+031
+032import 
org.apache.hadoop.HadoopIllegalArgumentException;
+033import 
org.apache.hadoop.conf.Configuration;
+034import 
org.apache.hadoop.fs.FSDataOutputStream;
+035import org.apache.hadoop.fs.FileStatus;
+036import org.apache.hadoop.fs.FileSystem;
+037import 
org.apache.hadoop.fs.LocatedFileStatus;
+038import org.apache.hadoop.fs.Path;
+039import org.apache.hadoop.fs.PathFilter;
+040import 
org.apache.hadoop.fs.RemoteIterator;
+041import 
org.apache.hadoop.fs.permission.FsPermission;
+042import 
org.apache.hadoop.hbase.HConstants;
+043import 
org.apache.hadoop.hbase.TableName;
+044import 
org.apache.hadoop.ipc.RemoteException;
+045import 
org.apache.yetus.audience.InterfaceAudience;
+046import org.slf4j.Logger;
+047import org.slf4j.LoggerFactory;
+048
+049import 
org.apache.hbase.thirdparty.com.google.common.annotations.VisibleForTesting;
+050import 
org.apache.hbase.thirdparty.com.google.common.collect.Lists;
+051
+052/**
+053 * Utility methods for interacting with 
the underlying file system.
+054 * 

+055 * Note that {@link #setStoragePolicy(FileSystem, Path, String)} is tested in TestFSUtils and +056 * pre-commit will run the hbase-server tests if there's code change in this class. See +057 * HBASE-20838; for more details. +058 */ +059@InterfaceAudience.Private +060public abstract class CommonFSUtils { +061 private static final Logger LOG = LoggerFactory.getLogger(CommonFSUtils.class); +062 +063 /** Parameter name for HBase WAL directory */ +064 public static final String HBASE_WAL_DIR = "hbase.wal.dir"; +065 +066 /** Parameter to disable stream capability enforcement checks */ +067 public static final String UNSAFE_STREAM_CAPABILITY_ENFORCE = "hbase.unsafe.stream.capability.enforce"; +068 +069 /** Full access permissions (starting point for a umask) */ +070 public static final String FULL_RWX_PERMISSIONS = "777"; +071 +072 protected CommonFSUtils() { +073super(); +074 } +075 +076 /** +077 * Compare of path component. Does not consider schema; i.e. if schemas +078 * different but path starts with rootPath, +079 * then the function returns true +080 * @param rootPath value to check for +081 * @param path subject to check +082 * @return True if path starts with rootPath +083 */ +084 public static boolean isStartingWithPath(final Path rootPath, final String path) { +085String uriRootPath = rootPath.toUri().getPath(); +086String tailUriPath = (new Path(path)).toUri().getPath(); +087return tailUriPath.startsWith(uriRootPath); +088 } +089 +090 /** +091 * Compare path component of the Path URI; e.g. if hdfs://a/b/c and /a/b/c, it will compare the +092 * '/a/b/c' part. Does not consider schema; i.e. if schemas different but path or subpath matches, +093 * the two will equate. +094 * @param pathToSearch Path we will be trying to match against. +095 * @param pathTail what to match +096 * @retur


[03/11] hbase-site git commit: Published site at e963694259aa9242aa709e9b954754abd9a681cd.

2018-07-28 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e10887fe/devapidocs/src-html/org/apache/hadoop/hbase/util/CommonFSUtils.StreamCapabilities.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/util/CommonFSUtils.StreamCapabilities.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/util/CommonFSUtils.StreamCapabilities.html
index c6bfdca..cae3d07 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/util/CommonFSUtils.StreamCapabilities.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/util/CommonFSUtils.StreamCapabilities.html
@@ -829,90 +829,217 @@
 821conf.setIfUnset(dfsKey, 
Integer.toString(hbaseSize));
 822  }
 823
-824  // Holder singleton idiom. JVM spec 
ensures this will be run at most once per Classloader, and
-825  // not until we attempt to reference 
it.
-826  private static class StreamCapabilities 
{
-827public static final boolean 
PRESENT;
-828public static final Class 
CLASS;
-829public static final Method METHOD;
-830static {
-831  boolean tmp = false;
-832  Class clazz = null;
-833  Method method = null;
-834  try {
-835clazz = 
Class.forName("org.apache.hadoop.fs.StreamCapabilities");
-836method = 
clazz.getMethod("hasCapability", String.class);
-837tmp = true;
-838  } 
catch(ClassNotFoundException|NoSuchMethodException|SecurityException exception) 
{
-839LOG.warn("Your Hadoop 
installation does not include the StreamCapabilities class from " +
-840 "HDFS-11644, so we will 
skip checking if any FSDataOutputStreams actually " +
-841 "support hflush/hsync. 
If you are running on top of HDFS this probably just " +
-842 "means you have an older 
version and this can be ignored. If you are running on " +
-843 "top of an alternate 
FileSystem implementation you should manually verify that " +
-844 "hflush and hsync are 
implemented; otherwise you risk data loss and hard to " +
-845 "diagnose errors when 
our assumptions are violated.");
-846LOG.debug("The first request to 
check for StreamCapabilities came from this stacktrace.",
-847exception);
-848  } finally {
-849PRESENT = tmp;
-850CLASS = clazz;
-851METHOD = method;
-852  }
-853}
-854  }
-855
-856  /**
-857   * If our FileSystem version includes 
the StreamCapabilities class, check if
-858   * the given stream has a particular 
capability.
-859   * @param stream capabilities are 
per-stream instance, so check this one specifically. must not be
-860   *null
-861   * @param capability what to look for, 
per Hadoop Common's FileSystem docs
-862   * @return true if there are no 
StreamCapabilities. false if there are, but this stream doesn't
-863   * implement it. return result 
of asking the stream otherwise.
-864   */
-865  public static boolean 
hasCapability(FSDataOutputStream stream, String capability) {
-866// be consistent whether or not 
StreamCapabilities is present
-867if (stream == null) {
-868  throw new 
NullPointerException("stream parameter must not be null.");
-869}
-870// If o.a.h.fs.StreamCapabilities 
doesn't exist, assume everyone does everything
-871// otherwise old versions of Hadoop 
will break.
-872boolean result = true;
-873if (StreamCapabilities.PRESENT) {
-874  // if StreamCapabilities is 
present, but the stream doesn't implement it
-875  // or we run into a problem 
invoking the method,
-876  // we treat that as equivalent to 
not declaring anything
-877  result = false;
-878  if 
(StreamCapabilities.CLASS.isAssignableFrom(stream.getClass())) {
+824  private static class DfsBuilderUtility 
{
+825static Class dfsClass = 
null;
+826static Method createMethod;
+827static Method overwriteMethod;
+828static Method bufferSizeMethod;
+829static Method blockSizeMethod;
+830static Method recursiveMethod;
+831static Method replicateMethod;
+832static Method replicationMethod;
+833static Method buildMethod;
+834static boolean allMethodsPresent = 
false;
+835
+836static {
+837  String dfsName = 
"org.apache.hadoop.hdfs.DistributedFileSystem";
+838  String builderName = dfsName + 
"$HdfsDataOutputStreamBuilder";
+839  Class builderClass = 
null;
+840
+841  try {
+842dfsClass = 
Class.forName(dfsName);
+843  } catch (ClassNotFoundException e) 
{
+844LOG.debug("{} not available, will 
not use builder API for file creation.", dfsName);
+845  }
+846  try {
+847builderClass = 
Class.forName(builderName);
+848  } catch (ClassNotFoundException e) 
{
+849LOG.debug("{} not available, will 
not use builder API for file creation.", builderName);
+850  }
+851
+852  if (dfsClass != null && 
builderClass != null) {
+853tr

hbase-site git commit: INFRA-10751 Empty commit

2018-07-28 Thread git-site-role
Repository: hbase-site
Updated Branches:
  refs/heads/asf-site e10887fef -> 09f27cd5b


INFRA-10751 Empty commit


Project: http://git-wip-us.apache.org/repos/asf/hbase-site/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase-site/commit/09f27cd5
Tree: http://git-wip-us.apache.org/repos/asf/hbase-site/tree/09f27cd5
Diff: http://git-wip-us.apache.org/repos/asf/hbase-site/diff/09f27cd5

Branch: refs/heads/asf-site
Commit: 09f27cd5b21b1ce13ee07d6cc6b64086f95cc26c
Parents: e10887f
Author: jenkins 
Authored: Sat Jul 28 14:49:08 2018 +
Committer: jenkins 
Committed: Sat Jul 28 14:49:08 2018 +

--

--




[11/11] hbase-site git commit: Published site at e963694259aa9242aa709e9b954754abd9a681cd.

2018-07-28 Thread git-site-role
Published site at e963694259aa9242aa709e9b954754abd9a681cd.


Project: http://git-wip-us.apache.org/repos/asf/hbase-site/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase-site/commit/e10887fe
Tree: http://git-wip-us.apache.org/repos/asf/hbase-site/tree/e10887fe
Diff: http://git-wip-us.apache.org/repos/asf/hbase-site/diff/e10887fe

Branch: refs/heads/asf-site
Commit: e10887fef6c65851f2fcba83f242b0dd360f4878
Parents: e0dc0c6
Author: jenkins 
Authored: Sat Jul 28 14:48:45 2018 +
Committer: jenkins 
Committed: Sat Jul 28 14:48:45 2018 +

--
 acid-semantics.html | 4 +-
 apache_hbase_reference_guide.pdf| 4 +-
 book.html   | 2 +-
 bulk-loads.html | 4 +-
 checkstyle-aggregate.html   | 26878 -
 checkstyle.rss  |20 +-
 coc.html| 4 +-
 dependencies.html   | 4 +-
 dependency-convergence.html | 4 +-
 dependency-info.html| 4 +-
 dependency-management.html  | 4 +-
 devapidocs/allclasses-frame.html| 1 +
 devapidocs/allclasses-noframe.html  | 1 +
 devapidocs/constant-values.html | 6 +-
 devapidocs/index-all.html   |44 +
 .../hadoop/hbase/backup/package-tree.html   | 4 +-
 .../hadoop/hbase/backup/util/RestoreTool.html   |12 +-
 .../hadoop/hbase/client/package-tree.html   |22 +-
 .../hadoop/hbase/coprocessor/package-tree.html  | 2 +-
 .../hadoop/hbase/executor/package-tree.html | 2 +-
 .../hadoop/hbase/filter/package-tree.html   |10 +-
 .../hadoop/hbase/io/hfile/package-tree.html | 6 +-
 .../apache/hadoop/hbase/ipc/package-tree.html   | 2 +-
 .../hadoop/hbase/mapreduce/package-tree.html| 4 +-
 .../hbase/master/balancer/package-tree.html | 2 +-
 .../hadoop/hbase/master/package-tree.html   | 4 +-
 .../hbase/master/procedure/package-tree.html| 4 +-
 .../org/apache/hadoop/hbase/package-tree.html   |18 +-
 .../hadoop/hbase/procedure2/package-tree.html   | 4 +-
 .../hadoop/hbase/quotas/package-tree.html   | 8 +-
 .../hadoop/hbase/regionserver/package-tree.html |16 +-
 .../regionserver/querymatcher/package-tree.html | 2 +-
 .../regionserver/wal/ProtobufLogWriter.html |12 +-
 .../hbase/regionserver/wal/package-tree.html| 2 +-
 .../replication/regionserver/package-tree.html  | 2 +-
 .../hadoop/hbase/rest/model/package-tree.html   | 2 +-
 .../hbase/security/access/package-tree.html | 2 +-
 .../hadoop/hbase/security/package-tree.html | 2 +-
 .../hadoop/hbase/thrift/package-tree.html   | 2 +-
 .../util/CommonFSUtils.DfsBuilderUtility.html   |   480 +
 .../util/CommonFSUtils.StreamCapabilities.html  |14 +-
 ...nFSUtils.StreamLacksCapabilityException.html | 6 +-
 .../apache/hadoop/hbase/util/CommonFSUtils.html |   154 +-
 .../apache/hadoop/hbase/util/FSHDFSUtils.html   | 2 +-
 .../org/apache/hadoop/hbase/util/FSUtils.html   | 2 +-
 .../CommonFSUtils.DfsBuilderUtility.html|   125 +
 .../apache/hadoop/hbase/util/package-frame.html | 1 +
 .../hadoop/hbase/util/package-summary.html  |   274 +-
 .../apache/hadoop/hbase/util/package-tree.html  | 7 +-
 .../apache/hadoop/hbase/wal/package-tree.html   | 2 +-
 devapidocs/overview-tree.html   | 1 +
 .../org/apache/hadoop/hbase/Version.html| 6 +-
 .../hadoop/hbase/backup/util/RestoreTool.html   |   547 +-
 .../hbase/io/asyncfs/AsyncFSOutputHelper.html   |35 +-
 .../wal/WALProcedureStore.LeaseRecovery.html| 2 +-
 .../store/wal/WALProcedureStore.PushType.html   | 2 +-
 .../wal/WALProcedureStore.SyncMetrics.html  | 2 +-
 .../procedure2/store/wal/WALProcedureStore.html | 2 +-
 .../regionserver/wal/ProtobufLogWriter.html |   124 +-
 .../util/CommonFSUtils.DfsBuilderUtility.html   |  1106 +
 .../util/CommonFSUtils.StreamCapabilities.html  |   293 +-
 ...nFSUtils.StreamLacksCapabilityException.html |   293 +-
 .../apache/hadoop/hbase/util/CommonFSUtils.html |   293 +-
 downloads.html  | 4 +-
 export_control.html | 4 +-
 index.html  | 4 +-
 integration.html| 4 +-
 issue-tracking.html | 4 +-
 license.html| 4 +-
 mail-lists.html | 4 +-
 metrics.html| 4 +-
 old_news.html   | 4 +-
 plugin-management.html  

[06/11] hbase-site git commit: Published site at e963694259aa9242aa709e9b954754abd9a681cd.

2018-07-28 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e10887fe/devapidocs/org/apache/hadoop/hbase/util/package-summary.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/util/package-summary.html 
b/devapidocs/org/apache/hadoop/hbase/util/package-summary.html
index 7deac38..051c486 100644
--- a/devapidocs/org/apache/hadoop/hbase/util/package-summary.html
+++ b/devapidocs/org/apache/hadoop/hbase/util/package-summary.html
@@ -479,358 +479,362 @@
 
 
 
-CommonFSUtils.StreamCapabilities
+CommonFSUtils.DfsBuilderUtility
  
 
 
+CommonFSUtils.StreamCapabilities
+ 
+
+
 CompressionTest
 
 Compression validation test.
 
 
-
+
 ConcatenatedLists
 
 A collection class that contains multiple sub-lists, which 
allows us to not copy lists.
 
 
-
+
 ConfigurationUtil
 
 Utilities for storing more complex collection types in
  Configuration instances.
 
 
-
+
 ConnectionCache
 
 A utility to store user specific HConnections in 
memory.
 
 
-
+
 CoprocessorClassLoader
 
 ClassLoader used to load classes for Coprocessor 
instances.
 
 
-
+
 Counter
 Deprecated
 use https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/atomic/LongAdder.html?is-external=true";
 title="class or interface in 
java.util.concurrent.atomic">LongAdder instead.
 
 
-
+
 Counter.Cell
  
 
-
+
 Counter.Container
  
 
-
+
 DefaultEnvironmentEdge
 
 Default implementation of an environment edge.
 
 
-
+
 DirectMemoryUtils
 
 Utilities for interacting with and monitoring 
DirectByteBuffer allocations.
 
 
-
+
 DNS
 
 Wrapper around Hadoop's DNS class to hide reflection.
 
 
-
+
 DynamicClassLoader
 
 This is a class loader that can load classes dynamically 
from new
  jar files under a configured folder.
 
 
-
+
 EncryptionTest
  
 
-
+
 EnvironmentEdgeManager
 
 Manages a singleton instance of the environment edge.
 
 
-
+
 ExceptionUtil
 
 This class handles the different interruption classes.
 
 
-
+
 ForeignExceptionUtil
 
 Helper to convert Exceptions and StackTraces from/to 
protobuf.
 
 
-
+
 FSHDFSUtils
 
 Implementation for hdfs
 
 
-
+
 FSRegionScanner
 
 Thread that walks over the filesystem, and computes the 
mappings
  Region -> BestHost and Region -> Map
 
 
-
+
 FSTableDescriptors
 
 Implementation of TableDescriptors 
that reads descriptors from the
  passed filesystem.
 
 
-
+
 FSUtils
 
 Utility methods for interacting with the underlying file 
system.
 
 
-
+
 FSUtils.BlackListDirFilter
 
 Directory filter that doesn't include any of the 
directories in the specified blacklist
 
 
-
+
 FSUtils.DirFilter
 
 A PathFilter that only allows 
directories.
 
 
-
+
 FSUtils.FamilyDirFilter
 
 Filter for all dirs that are legal column family 
names.
 
 
-
+
 FSUtils.FileFilter
 
 A PathFilter that returns only regular 
files.
 
 
-
+
 FSUtils.HFileFilter
 
 Filter for HFiles that excludes reference files.
 
 
-
+
 FSUtils.HFileLinkFilter
 
 Filter for HFileLinks (StoreFiles and HFiles not 
included).
 
 
-
+
 FSUtils.ReferenceFileFilter
  
 
-
+
 FSUtils.RegionDirFilter
 
 Filter for all dirs that don't start with '.'
 
 
-
+
 FSUtils.UserTableDirFilter
 
 A PathFilter that returns usertable 
directories.
 
 
-
+
 FSVisitor
 
 Utility methods for interacting with the hbase.root file 
system.
 
 
-
+
 GetJavaProperty
 
 A generic way for querying Java properties.
 
 
-
+
 Hash
 
 This class represents a common API for hashing 
functions.
 
 
-
+
 HashedBytes
 
 This class encapsulates a byte array and overrides hashCode 
and equals so
  that it's identity is based on the data rather than the array instance.
 
 
-
+
 HashKey
 
 Used to calculate the hash Hash algorithms for 
Bloomfilters.
 
 
-
+
 HasThread
 
 Abstract class which contains a Thread and delegates the 
common Thread
  methods to that instance.
 
 
-
+
 HBaseConfTool
 
 Tool that prints out a configuration.
 
 
-
+
 HBaseFsck
 
 HBaseFsck (hbck) is a tool for checking and repairing 
region consistency and
  table integrity problems in a corrupted HBase.
 
 
-
+
 HBaseFsck.HBaseFsckTool
 
 This is a Tool wrapper that gathers -Dxxx=yyy configuration 
settings from the command line.
 
 
-
+
 HBaseFsck.HbckInfo
 
 Maintain information about a particular region.
 
 
-
+
 HBaseFsck.HdfsEntry
 
 Stores the regioninfo entries from HDFS
 
 
-
+
 HBaseFsck.MetaEntry
 
 Stores the regioninfo entries scanned from META
 
 
-
+
 HBaseFsck.OnlineEntry
 
 Stores the regioninfo retrieved from Online region 
servers.
 
 
-
+
 HBaseFsck.PrintingErrorReporter
  
 
-
+
 HBaseFsck.RegionBoundariesInformation
  
 
-
+
 HBaseFsck.WorkItemHdfsRegionInfo
 
 Contact hdfs and get all information about specified table 
directory into
  regioninfo list.
 
 
-
+
 HBaseFsck.WorkItemOverlapMerge
  
 
-
+
 HBaseFsck.WorkItemRegion
 
 Contact a region server and get all information from 
it
 
 
-
+
 HBaseFsckRepair
 
 This class contains helper methods that repair parts of 
hbase's files

[07/11] hbase-site git commit: Published site at e963694259aa9242aa709e9b954754abd9a681cd.

2018-07-28 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e10887fe/devapidocs/org/apache/hadoop/hbase/util/CommonFSUtils.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/util/CommonFSUtils.html 
b/devapidocs/org/apache/hadoop/hbase/util/CommonFSUtils.html
index 40667fc..6705fc8 100644
--- a/devapidocs/org/apache/hadoop/hbase/util/CommonFSUtils.html
+++ b/devapidocs/org/apache/hadoop/hbase/util/CommonFSUtils.html
@@ -18,7 +18,7 @@
 catch(err) {
 }
 //-->
-var methods = 
{"i0":9,"i1":9,"i2":9,"i3":9,"i4":9,"i5":9,"i6":9,"i7":9,"i8":9,"i9":9,"i10":9,"i11":9,"i12":9,"i13":9,"i14":9,"i15":9,"i16":9,"i17":9,"i18":9,"i19":9,"i20":9,"i21":9,"i22":9,"i23":9,"i24":9,"i25":9,"i26":9,"i27":9,"i28":9,"i29":9,"i30":9,"i31":9,"i32":9,"i33":9,"i34":9,"i35":9,"i36":9,"i37":9,"i38":9,"i39":9};
+var methods = 
{"i0":9,"i1":9,"i2":9,"i3":9,"i4":9,"i5":9,"i6":9,"i7":9,"i8":9,"i9":9,"i10":9,"i11":9,"i12":9,"i13":9,"i14":9,"i15":9,"i16":9,"i17":9,"i18":9,"i19":9,"i20":9,"i21":9,"i22":9,"i23":9,"i24":9,"i25":9,"i26":9,"i27":9,"i28":9,"i29":9,"i30":9,"i31":9,"i32":9,"i33":9,"i34":9,"i35":9,"i36":9,"i37":9,"i38":9,"i39":9,"i40":9,"i41":9};
 var tabs = {65535:["t0","All Methods"],1:["t1","Static 
Methods"],8:["t4","Concrete Methods"]};
 var altColor = "altColor";
 var rowColor = "rowColor";
@@ -50,7 +50,7 @@ var activeTableTab = "activeTableTab";
 
 
 Prev Class
-Next Class
+Next Class
 
 
 Frames
@@ -141,9 +141,13 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 private static class 
-CommonFSUtils.StreamCapabilities 
+CommonFSUtils.DfsBuilderUtility 
 
 
+private static class 
+CommonFSUtils.StreamCapabilities 
+
+
 static class 
 CommonFSUtils.StreamLacksCapabilityException
 Helper exception for those cases where the place where we 
need to check a stream capability
@@ -241,6 +245,28 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 
+static 
org.apache.hadoop.fs.FSDataOutputStream
+createForWal(org.apache.hadoop.fs.FileSystem fs,
+org.apache.hadoop.fs.Path path,
+boolean overwritable)
+Attempt to use builder API via reflection to create a file 
with the given parameters and
+ replication enabled.
+
+
+
+static 
org.apache.hadoop.fs.FSDataOutputStream
+createForWal(org.apache.hadoop.fs.FileSystem fs,
+org.apache.hadoop.fs.Path path,
+boolean overwritable,
+int bufferSize,
+short replication,
+long blockSize,
+boolean isRecursive)
+Attempt to use builder API via reflection to create a file 
with the given parameters and
+ replication enabled.
+
+
+
 static boolean
 delete(org.apache.hadoop.fs.FileSystem fs,
   org.apache.hadoop.fs.Path path,
@@ -248,18 +274,18 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 Calls fs.delete() and returns the value returned by the 
fs.delete()
 
 
-
+
 static boolean
 deleteDirectory(org.apache.hadoop.fs.FileSystem fs,
org.apache.hadoop.fs.Path dir)
 Delete if exists.
 
 
-
+
 static org.apache.hadoop.fs.FileSystem
 getCurrentFileSystem(org.apache.hadoop.conf.Configuration conf) 
 
-
+
 static long
 getDefaultBlockSize(org.apache.hadoop.fs.FileSystem fs,
org.apache.hadoop.fs.Path path)
@@ -267,18 +293,18 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
  be split into to minimize i/o time.
 
 
-
+
 static int
 getDefaultBufferSize(org.apache.hadoop.fs.FileSystem fs)
 Returns the default buffer size to use during writes.
 
 
-
+
 static short
 getDefaultReplication(org.apache.hadoop.fs.FileSystem fs,
  org.apache.hadoop.fs.Path path) 
 
-
+
 static 
org.apache.hadoop.fs.permission.FsPermission
 getFilePermissions(org.apache.hadoop.fs.FileSystem fs,
   org.apache.hadoop.conf.Configuration conf,
@@ -287,7 +313,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
  enabled.
 
 
-
+
 static org.apache.hadoop.fs.Path
 getNamespaceDir(org.apache.hadoop.fs.Path rootdir,
https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String namespace)
@@ -295,21 +321,21 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
  the namespace directory under path rootdir
 
 
-
+
 static https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String
 getPath(org.apache.hadoop.fs.Path p)
 Return the 'path' component of a Path.
 
 
-
+
 static org.apache.hadoop.fs.Path
 getRootDir(org.apache.hadoop.conf.Configuration c) 
 
-
+
 static org.apache.hadoop.fs.FileSystem
 getRootDirFileSystem(org.apache.hadoop.conf.Configuration c) 
 
-
+
 static org.apache.hadoop.fs.Path
 getTableDir(org.apache.hadoop.fs.Path rootdir,
TableName tableName)
@@ -317,7 +343,7 @@ ext

[01/11] hbase-site git commit: Published site at e963694259aa9242aa709e9b954754abd9a681cd.

2018-07-28 Thread git-site-role
Repository: hbase-site
Updated Branches:
  refs/heads/asf-site e0dc0c667 -> e10887fef


http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e10887fe/testdevapidocs/org/apache/hadoop/hbase/class-use/HBaseTestingUtility.html
--
diff --git 
a/testdevapidocs/org/apache/hadoop/hbase/class-use/HBaseTestingUtility.html 
b/testdevapidocs/org/apache/hadoop/hbase/class-use/HBaseTestingUtility.html
index 38cc11c..1165b4f 100644
--- a/testdevapidocs/org/apache/hadoop/hbase/class-use/HBaseTestingUtility.html
+++ b/testdevapidocs/org/apache/hadoop/hbase/class-use/HBaseTestingUtility.html
@@ -3263,6 +3263,10 @@
 
 
 private static HBaseTestingUtility
+TestHBaseWalOnEC.util 
+
+
+private static HBaseTestingUtility
 TestCombinedAsyncWriter.UTIL 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e10887fe/testdevapidocs/org/apache/hadoop/hbase/io/hfile/package-tree.html
--
diff --git a/testdevapidocs/org/apache/hadoop/hbase/io/hfile/package-tree.html 
b/testdevapidocs/org/apache/hadoop/hbase/io/hfile/package-tree.html
index ec56445..4837d80 100644
--- a/testdevapidocs/org/apache/hadoop/hbase/io/hfile/package-tree.html
+++ b/testdevapidocs/org/apache/hadoop/hbase/io/hfile/package-tree.html
@@ -158,8 +158,8 @@
 
 java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true";
 title="class or interface in java.lang">Enum (implements java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true";
 title="class or interface in java.lang">Comparable, java.io.https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true";
 title="class or interface in java.io">Serializable)
 
-org.apache.hadoop.hbase.io.hfile.TestCacheOnWrite.CacheOnWriteType
 org.apache.hadoop.hbase.io.hfile.TagUsage
+org.apache.hadoop.hbase.io.hfile.TestCacheOnWrite.CacheOnWriteType
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e10887fe/testdevapidocs/org/apache/hadoop/hbase/package-tree.html
--
diff --git a/testdevapidocs/org/apache/hadoop/hbase/package-tree.html 
b/testdevapidocs/org/apache/hadoop/hbase/package-tree.html
index 3d372b1..55dede9 100644
--- a/testdevapidocs/org/apache/hadoop/hbase/package-tree.html
+++ b/testdevapidocs/org/apache/hadoop/hbase/package-tree.html
@@ -576,15 +576,15 @@
 
 java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true";
 title="class or interface in java.lang">Enum (implements java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true";
 title="class or interface in java.lang">Comparable, java.io.https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true";
 title="class or interface in java.io">Serializable)
 
+org.apache.hadoop.hbase.RESTApiClusterManager.Service
 org.apache.hadoop.hbase.RESTApiClusterManager.RoleCommand
 org.apache.hadoop.hbase.IntegrationTestRegionReplicaPerf.Stat
-org.apache.hadoop.hbase.ClusterManager.ServiceType
-org.apache.hadoop.hbase.PerformanceEvaluation.Counter
 org.apache.hadoop.hbase.ScanPerformanceEvaluation.ScanCounter
-org.apache.hadoop.hbase.RESTApiClusterManager.Service
-org.apache.hadoop.hbase.ResourceChecker.Phase
+org.apache.hadoop.hbase.ClusterManager.ServiceType
 org.apache.hadoop.hbase.HBaseClusterManager.CommandProvider.Operation
 org.apache.hadoop.hbase.IntegrationTestDDLMasterFailover.ACTION
+org.apache.hadoop.hbase.PerformanceEvaluation.Counter
+org.apache.hadoop.hbase.ResourceChecker.Phase
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e10887fe/testdevapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html
--
diff --git 
a/testdevapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html 
b/testdevapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html
index 47095d2..9ff71c8 100644
--- a/testdevapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html
+++ b/testdevapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html
@@ -667,11 +667,11 @@
 
 java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true";
 title="class or interface in java.lang">Enum (implements java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true";
 title="class or interface in java.lang">Comparable, java.io.https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true";
 title="class or interface in java.io">Serializable)
 
+org.apache.hadoop.hbase.regionserver.TestRegionServerReadRequestMetrics.Metric
 org.apache.hadoop.hbase.regionserver.TestMultiLogThreshold.ActionType
+org.apache.hadoop.hbase.regionserver.DataBlockEncodingTool.Manipulation
 org.apache.hadoop.hbase.regionserver.TestAto