[05/11] hbase-site git commit: Published site at e963694259aa9242aa709e9b954754abd9a681cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e10887fe/devapidocs/src-html/org/apache/hadoop/hbase/backup/util/RestoreTool.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/backup/util/RestoreTool.html b/devapidocs/src-html/org/apache/hadoop/hbase/backup/util/RestoreTool.html index bbbf2a1..4153319 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/backup/util/RestoreTool.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/backup/util/RestoreTool.html @@ -244,279 +244,280 @@ 236Path tableInfoPath = null; 237 238// can't build the path directly as the timestamp values are different -239FileStatus[] snapshots = fs.listStatus(tableSnapShotPath); -240for (FileStatus snapshot : snapshots) { -241 tableInfoPath = snapshot.getPath(); -242 // SnapshotManifest.DATA_MANIFEST_NAME = "data.manifest"; -243 if (tableInfoPath.getName().endsWith("data.manifest")) { -244break; -245 } -246} -247return tableInfoPath; -248 } -249 -250 /** -251 * Get table descriptor -252 * @param tableName is the table backed up -253 * @return {@link TableDescriptor} saved in backup image of the table -254 */ -255 TableDescriptor getTableDesc(TableName tableName) throws IOException { -256Path tableInfoPath = this.getTableInfoPath(tableName); -257SnapshotDescription desc = SnapshotDescriptionUtils.readSnapshotInfo(fs, tableInfoPath); -258SnapshotManifest manifest = SnapshotManifest.open(conf, fs, tableInfoPath, desc); -259TableDescriptor tableDescriptor = manifest.getTableDescriptor(); -260if (!tableDescriptor.getTableName().equals(tableName)) { -261 LOG.error("couldn't find Table Desc for table: " + tableName + " under tableInfoPath: " -262 + tableInfoPath.toString()); -263 LOG.error("tableDescriptor.getNameAsString() = " -264 + tableDescriptor.getTableName().getNameAsString()); -265 throw new FileNotFoundException("couldn't find Table Desc for table: " + tableName -266 + " under tableInfoPath: " + tableInfoPath.toString()); -267} -268return tableDescriptor; -269 } -270 -271 private TableDescriptor getTableDescriptor(FileSystem fileSys, TableName tableName, -272 String lastIncrBackupId) throws IOException { -273if (lastIncrBackupId != null) { -274 String target = -275 BackupUtils.getTableBackupDir(backupRootPath.toString(), -276lastIncrBackupId, tableName); -277 return FSTableDescriptors.getTableDescriptorFromFs(fileSys, new Path(target)); -278} -279return null; -280 } -281 -282 private void createAndRestoreTable(Connection conn, TableName tableName, TableName newTableName, -283 Path tableBackupPath, boolean truncateIfExists, String lastIncrBackupId) throws IOException { -284if (newTableName == null) { -285 newTableName = tableName; -286} -287FileSystem fileSys = tableBackupPath.getFileSystem(this.conf); -288 -289// get table descriptor first -290TableDescriptor tableDescriptor = getTableDescriptor(fileSys, tableName, lastIncrBackupId); -291if (tableDescriptor != null) { -292 LOG.debug("Retrieved descriptor: " + tableDescriptor + " thru " + lastIncrBackupId); -293} -294 -295if (tableDescriptor == null) { -296 Path tableSnapshotPath = getTableSnapshotPath(backupRootPath, tableName, backupId); -297 if (fileSys.exists(tableSnapshotPath)) { -298// snapshot path exist means the backup path is in HDFS -299// check whether snapshot dir already recorded for target table -300if (snapshotMap.get(tableName) != null) { -301 SnapshotDescription desc = -302 SnapshotDescriptionUtils.readSnapshotInfo(fileSys, tableSnapshotPath); -303 SnapshotManifest manifest = SnapshotManifest.open(conf, fileSys, tableSnapshotPath, desc); -304 tableDescriptor = manifest.getTableDescriptor(); -305} else { -306 tableDescriptor = getTableDesc(tableName); -307 snapshotMap.put(tableName, getTableInfoPath(tableName)); -308} -309if (tableDescriptor == null) { -310 LOG.debug("Found no table descriptor in the snapshot dir, previous schema would be lost"); -311} -312 } else { -313throw new IOException("Table snapshot directory: " + -314tableSnapshotPath + " does not exist."); -315 } -316} -317 -318Path tableArchivePath = getTableArchivePath(tableName); -319if (tableArchivePath == null) { -320 if (tableDescriptor != null) { -321// find table descriptor but no archive dir means the table is empty, create table and exit -322if (LOG.isDebugEnabled()) { -323 LOG.debug("find table descriptor but no archive dir for table " + tableName -324 + ", will only create table"); -325} -326
[02/11] hbase-site git commit: Published site at e963694259aa9242aa709e9b954754abd9a681cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e10887fe/devapidocs/src-html/org/apache/hadoop/hbase/util/CommonFSUtils.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/util/CommonFSUtils.html b/devapidocs/src-html/org/apache/hadoop/hbase/util/CommonFSUtils.html index c6bfdca..cae3d07 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/util/CommonFSUtils.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/util/CommonFSUtils.html @@ -829,90 +829,217 @@ 821conf.setIfUnset(dfsKey, Integer.toString(hbaseSize)); 822 } 823 -824 // Holder singleton idiom. JVM spec ensures this will be run at most once per Classloader, and -825 // not until we attempt to reference it. -826 private static class StreamCapabilities { -827public static final boolean PRESENT; -828public static final Class> CLASS; -829public static final Method METHOD; -830static { -831 boolean tmp = false; -832 Class> clazz = null; -833 Method method = null; -834 try { -835clazz = Class.forName("org.apache.hadoop.fs.StreamCapabilities"); -836method = clazz.getMethod("hasCapability", String.class); -837tmp = true; -838 } catch(ClassNotFoundException|NoSuchMethodException|SecurityException exception) { -839LOG.warn("Your Hadoop installation does not include the StreamCapabilities class from " + -840 "HDFS-11644, so we will skip checking if any FSDataOutputStreams actually " + -841 "support hflush/hsync. If you are running on top of HDFS this probably just " + -842 "means you have an older version and this can be ignored. If you are running on " + -843 "top of an alternate FileSystem implementation you should manually verify that " + -844 "hflush and hsync are implemented; otherwise you risk data loss and hard to " + -845 "diagnose errors when our assumptions are violated."); -846LOG.debug("The first request to check for StreamCapabilities came from this stacktrace.", -847exception); -848 } finally { -849PRESENT = tmp; -850CLASS = clazz; -851METHOD = method; -852 } -853} -854 } -855 -856 /** -857 * If our FileSystem version includes the StreamCapabilities class, check if -858 * the given stream has a particular capability. -859 * @param stream capabilities are per-stream instance, so check this one specifically. must not be -860 *null -861 * @param capability what to look for, per Hadoop Common's FileSystem docs -862 * @return true if there are no StreamCapabilities. false if there are, but this stream doesn't -863 * implement it. return result of asking the stream otherwise. -864 */ -865 public static boolean hasCapability(FSDataOutputStream stream, String capability) { -866// be consistent whether or not StreamCapabilities is present -867if (stream == null) { -868 throw new NullPointerException("stream parameter must not be null."); -869} -870// If o.a.h.fs.StreamCapabilities doesn't exist, assume everyone does everything -871// otherwise old versions of Hadoop will break. -872boolean result = true; -873if (StreamCapabilities.PRESENT) { -874 // if StreamCapabilities is present, but the stream doesn't implement it -875 // or we run into a problem invoking the method, -876 // we treat that as equivalent to not declaring anything -877 result = false; -878 if (StreamCapabilities.CLASS.isAssignableFrom(stream.getClass())) { +824 private static class DfsBuilderUtility { +825static Class> dfsClass = null; +826static Method createMethod; +827static Method overwriteMethod; +828static Method bufferSizeMethod; +829static Method blockSizeMethod; +830static Method recursiveMethod; +831static Method replicateMethod; +832static Method replicationMethod; +833static Method buildMethod; +834static boolean allMethodsPresent = false; +835 +836static { +837 String dfsName = "org.apache.hadoop.hdfs.DistributedFileSystem"; +838 String builderName = dfsName + "$HdfsDataOutputStreamBuilder"; +839 Class> builderClass = null; +840 +841 try { +842dfsClass = Class.forName(dfsName); +843 } catch (ClassNotFoundException e) { +844LOG.debug("{} not available, will not use builder API for file creation.", dfsName); +845 } +846 try { +847builderClass = Class.forName(builderName); +848 } catch (ClassNotFoundException e) { +849LOG.debug("{} not available, will not use builder API for file creation.", builderName); +850 } +851 +852 if (dfsClass != null && builderClass != null) { +853try { +854 createMethod = dfsClass.getMethod("createFile", Path.class); +855 over
[09/11] hbase-site git commit: Published site at e963694259aa9242aa709e9b954754abd9a681cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e10887fe/checkstyle.rss -- diff --git a/checkstyle.rss b/checkstyle.rss index 579f337..2ac991b 100644 --- a/checkstyle.rss +++ b/checkstyle.rss @@ -25,8 +25,8 @@ under the License. en-us ©2007 - 2018 The Apache Software Foundation - File: 3696, - Errors: 15632, + File: 3697, + Errors: 15630, Warnings: 0, Infos: 0 @@ -5669,7 +5669,7 @@ under the License. 0 - 2 + 0 @@ -45532,6 +45532,20 @@ under the License. + http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.regionserver.wal.TestHBaseWalOnEC.java";>org/apache/hadoop/hbase/regionserver/wal/TestHBaseWalOnEC.java + + + 0 + + + 0 + + + 0 + + + + http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.thrift2.generated.TResult.java";>org/apache/hadoop/hbase/thrift2/generated/TResult.java http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e10887fe/coc.html -- diff --git a/coc.html b/coc.html index c44ea99..1c93089 100644 --- a/coc.html +++ b/coc.html @@ -7,7 +7,7 @@ - + Apache HBase – Code of Conduct Policy @@ -375,7 +375,7 @@ email to mailto:priv...@hbase.apache.org";>the priv https://www.apache.org/";>The Apache Software Foundation. All rights reserved. - Last Published: 2018-07-27 + Last Published: 2018-07-28 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e10887fe/dependencies.html -- diff --git a/dependencies.html b/dependencies.html index b8acd13..dd4686f 100644 --- a/dependencies.html +++ b/dependencies.html @@ -7,7 +7,7 @@ - + Apache HBase – Project Dependencies @@ -440,7 +440,7 @@ https://www.apache.org/";>The Apache Software Foundation. All rights reserved. - Last Published: 2018-07-27 + Last Published: 2018-07-28 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e10887fe/dependency-convergence.html -- diff --git a/dependency-convergence.html b/dependency-convergence.html index 6887e4d..cda2363 100644 --- a/dependency-convergence.html +++ b/dependency-convergence.html @@ -7,7 +7,7 @@ - + Apache HBase – Reactor Dependency Convergence @@ -905,7 +905,7 @@ https://www.apache.org/";>The Apache Software Foundation. All rights reserved. - Last Published: 2018-07-27 + Last Published: 2018-07-28 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e10887fe/dependency-info.html -- diff --git a/dependency-info.html b/dependency-info.html index 860411e..cff8007 100644 --- a/dependency-info.html +++ b/dependency-info.html @@ -7,7 +7,7 @@ - + Apache HBase – Dependency Information @@ -313,7 +313,7 @@ https://www.apache.org/";>The Apache Software Foundation. All rights reserved. - Last Published: 2018-07-27 + Last Published: 2018-07-28 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e10887fe/dependency-management.html -- diff --git a/dependency-management.html b/dependency-management.html index cdc598a..29db2b6 100644 --- a/dependency-management.html +++ b/dependency-management.html @@ -7,7 +7,7 @@ - + Apache HBase – Project Dependency Management @@ -1005,7 +1005,7 @@ https://www.apache.org/";>The Apache Software Foundation. All rights reserved. - Last Published: 2018-07-27 + Last Published: 2018-07-28 http://
[10/11] hbase-site git commit: Published site at e963694259aa9242aa709e9b954754abd9a681cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e10887fe/checkstyle-aggregate.html -- diff --git a/checkstyle-aggregate.html b/checkstyle-aggregate.html index fe45785..80490cd 100644 --- a/checkstyle-aggregate.html +++ b/checkstyle-aggregate.html @@ -7,7 +7,7 @@ - + Apache HBase – Checkstyle Results @@ -281,10 +281,10 @@ Warnings Errors -3696 +3697 0 0 -15632 +15630 Files @@ -7584,2216 +7584,2211 @@ 0 22 -org/apache/hadoop/hbase/regionserver/wal/ProtobufLogWriter.java -0 -0 -2 - org/apache/hadoop/hbase/regionserver/wal/ReaderBase.java 0 0 1 - + org/apache/hadoop/hbase/regionserver/wal/SecureProtobufLogReader.java 0 0 2 - + org/apache/hadoop/hbase/regionserver/wal/SecureProtobufLogWriter.java 0 0 1 - + org/apache/hadoop/hbase/regionserver/wal/SecureWALCellCodec.java 0 0 2 - + org/apache/hadoop/hbase/regionserver/wal/SequenceIdAccounting.java 0 0 13 - + org/apache/hadoop/hbase/regionserver/wal/TestLogRolling.java 0 0 3 - + org/apache/hadoop/hbase/regionserver/wal/TestLogRollingNoCluster.java 0 0 4 - + org/apache/hadoop/hbase/regionserver/wal/TestWALConfiguration.java 0 0 3 - + org/apache/hadoop/hbase/regionserver/wal/WALActionsListener.java 0 0 2 - + org/apache/hadoop/hbase/regionserver/wal/WALCellCodec.java 0 0 6 - + org/apache/hadoop/hbase/regionserver/wal/WALCoprocessorHost.java 0 0 1 - + org/apache/hadoop/hbase/regionserver/wal/WALUtil.java 0 0 5 - + org/apache/hadoop/hbase/replication/BaseReplicationEndpoint.java 0 0 3 - + org/apache/hadoop/hbase/replication/BulkLoadCellFilter.java 0 0 1 - + org/apache/hadoop/hbase/replication/ClusterMarkingEntryFilter.java 0 0 3 - + org/apache/hadoop/hbase/replication/HBaseReplicationEndpoint.java 0 0 3 - + org/apache/hadoop/hbase/replication/ReplicationEndpoint.java 0 0 4 - + org/apache/hadoop/hbase/replication/ReplicationException.java 0 0 1 - + org/apache/hadoop/hbase/replication/SystemTableWALEntryFilter.java 0 0 1 - + org/apache/hadoop/hbase/replication/TestMasterReplication.java 0 0 12 - + org/apache/hadoop/hbase/replication/TestMultiSlaveReplication.java 0 0 7 - + org/apache/hadoop/hbase/replication/TestPerTableCFReplication.java 0 0 3 - + org/apache/hadoop/hbase/replication/TestReplicationDisableInactivePeer.java 0 0 1 - + org/apache/hadoop/hbase/replication/TestReplicationEndpoint.java 0 0 7 - + org/apache/hadoop/hbase/replication/TestReplicationKillMasterRSCompressed.java 0 0 1 - + org/apache/hadoop/hbase/replication/TestReplicationWithTags.java 0 0 1 - + org/apache/hadoop/hbase/replication/TestVerifyReplication.java 0 0 9 - + org/apache/hadoop/hbase/replication/WALCellFilter.java 0 0 2 - + org/apache/hadoop/hbase/replication/WALEntryFilter.java 0 0 2 - + org/apache/hadoop/hbase/replication/ZKReplicationQueueStorage.java 0 0 3 - + org/apache/hadoop/hbase/replication/master/ReplicationLogCleaner.java 0 0 1 - + org/apache/hadoop/hbase/replication/regionserver/DumpReplicationQueues.java 0 0 6 - + org/apache/hadoop/hbase/replication/regionserver/HFileReplicator.java 0 0 3 - + org/apache/hadoop/hbase/replication/regionserver/MetricsSink.java 0 0 2 - + org/apache/hadoop/hbase/replication/regionserver/MetricsSource.java 0 0 2 - + org/apache/hadoop/hbase/replication/regionserver/RegionReplicaReplicationEndpoint.java 0 0 3 - + org/apache/hadoop/hbase/replication/regionserver/Replication.java 0 0 2 - + org/apache/hadoop/hbase/replication/regionserver/ReplicationLoad.java 0 0 6 - + org/apache/hadoop/hbase/replication/regionserver/ReplicationObserver.java 0 0 4 - + org/apache/hadoop/hbase/replication/regionserver/ReplicationSink.java 0 0 11 - + org/apache/hadoop/hbase/replication/regionserver/ReplicationSourceFactory.java 0 0 2 - + org/apache/hadoop/hbase/replication/regionserver/ReplicationSourceInterface.java 0 0 1 - + org/apache/hadoop/hbase/replication/regionserver/ReplicationSourceManager.java 0 0 3 - + org/apache/hadoop/hbase/replication/regionserver/ReplicationThrottler.java 0 0 2 - + org/apache/hadoop/hbase/replication/regionserver/TestGlobalReplicationThrottler.java 0 0 3 - + org/apache/hadoop/hbase/replication/regionserver/TestRegionReplicaReplicationEndpointNoMaster.java 0 0 1 - + org/apache/hadoop/hbase/replication/regionserver/TestReplicationSink.java 0 0 11 - + org/apache/hadoop/hbase/replication/regionserver/TestReplicationSource.java 0 0 10 - + org/apache/hadoop/hbase/replication/regionserver/TestReplicationSourceManager.java 0 0 3 - + org/apache/hadoop/hbase/replication/regionserver/TestWALEntrySinkFilter.java 0 0 16 - + org/apache/hadoop/hbase/replication/regionserver/TestWALEntryStream.java 0 0 1 - + org/apache/hadoop/hbase/replication/regionserver/WALEntrySinkFilter.java 0 0 2 - + org/apache/hadoop/hbase/replication
[08/11] hbase-site git commit: Published site at e963694259aa9242aa709e9b954754abd9a681cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e10887fe/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html b/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html index 482efa8..4eb2488 100644 --- a/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html +++ b/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html @@ -704,20 +704,20 @@ java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true"; title="class or interface in java.lang">Enum(implements java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true"; title="class or interface in java.lang">Comparable , java.io.https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true"; title="class or interface in java.io">Serializable) -org.apache.hadoop.hbase.regionserver.ScannerContext.NextState -org.apache.hadoop.hbase.regionserver.CompactingMemStore.IndexType -org.apache.hadoop.hbase.regionserver.DefaultHeapMemoryTuner.StepDirection -org.apache.hadoop.hbase.regionserver.ChunkCreator.ChunkType -org.apache.hadoop.hbase.regionserver.HRegion.FlushResult.Result org.apache.hadoop.hbase.regionserver.SplitLogWorker.TaskExecutor.Status org.apache.hadoop.hbase.regionserver.TimeRangeTracker.Type org.apache.hadoop.hbase.regionserver.ScanType -org.apache.hadoop.hbase.regionserver.MetricsRegionServerSourceFactoryImpl.FactoryStorage +org.apache.hadoop.hbase.regionserver.ScannerContext.LimitScope org.apache.hadoop.hbase.regionserver.Region.Operation +org.apache.hadoop.hbase.regionserver.ChunkCreator.ChunkType +org.apache.hadoop.hbase.regionserver.CompactingMemStore.IndexType +org.apache.hadoop.hbase.regionserver.ScannerContext.NextState +org.apache.hadoop.hbase.regionserver.HRegion.FlushResult.Result +org.apache.hadoop.hbase.regionserver.MetricsRegionServerSourceFactoryImpl.FactoryStorage +org.apache.hadoop.hbase.regionserver.BloomType org.apache.hadoop.hbase.regionserver.MemStoreCompactionStrategy.Action org.apache.hadoop.hbase.regionserver.FlushType -org.apache.hadoop.hbase.regionserver.BloomType -org.apache.hadoop.hbase.regionserver.ScannerContext.LimitScope +org.apache.hadoop.hbase.regionserver.DefaultHeapMemoryTuner.StepDirection http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e10887fe/devapidocs/org/apache/hadoop/hbase/regionserver/querymatcher/package-tree.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/querymatcher/package-tree.html b/devapidocs/org/apache/hadoop/hbase/regionserver/querymatcher/package-tree.html index 858ccf6..b377318 100644 --- a/devapidocs/org/apache/hadoop/hbase/regionserver/querymatcher/package-tree.html +++ b/devapidocs/org/apache/hadoop/hbase/regionserver/querymatcher/package-tree.html @@ -130,9 +130,9 @@ java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true"; title="class or interface in java.lang">Enum (implements java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true"; title="class or interface in java.lang">Comparable , java.io.https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true"; title="class or interface in java.io">Serializable) -org.apache.hadoop.hbase.regionserver.querymatcher.DeleteTracker.DeleteResult org.apache.hadoop.hbase.regionserver.querymatcher.StripeCompactionScanQueryMatcher.DropDeletesInOutput org.apache.hadoop.hbase.regionserver.querymatcher.ScanQueryMatcher.MatchCode +org.apache.hadoop.hbase.regionserver.querymatcher.DeleteTracker.DeleteResult http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e10887fe/devapidocs/org/apache/hadoop/hbase/regionserver/wal/ProtobufLogWriter.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/wal/ProtobufLogWriter.html b/devapidocs/org/apache/hadoop/hbase/regionserver/wal/ProtobufLogWriter.html index 221b016..1e59005 100644 --- a/devapidocs/org/apache/hadoop/hbase/regionserver/wal/ProtobufLogWriter.html +++ b/devapidocs/org/apache/hadoop/hbase/regionserver/wal/ProtobufLogWriter.html @@ -359,7 +359,7 @@ implements sync -public void sync(boolean forceSync) +public void sync(boolean forceSync) throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true"; title="class or interface in java.io">IOException Specified by: @@ -375,7 +375,7 @@ implements getStream -public org.apache.hadoop.fs.FSDataOutputStream getStream() +public org.apache.hadoop.fs.FSDataOutputStream getStream() @@ -384,7 +384,7 @@ implements initOutput -protected void initOutput(org.apache.hadoop.fs.FileSystem fs, +protected void in
[04/11] hbase-site git commit: Published site at e963694259aa9242aa709e9b954754abd9a681cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e10887fe/devapidocs/src-html/org/apache/hadoop/hbase/util/CommonFSUtils.DfsBuilderUtility.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/util/CommonFSUtils.DfsBuilderUtility.html b/devapidocs/src-html/org/apache/hadoop/hbase/util/CommonFSUtils.DfsBuilderUtility.html new file mode 100644 index 000..cae3d07 --- /dev/null +++ b/devapidocs/src-html/org/apache/hadoop/hbase/util/CommonFSUtils.DfsBuilderUtility.html @@ -0,0 +1,1106 @@ +http://www.w3.org/TR/html4/loose.dtd";> + + +Source code + + + + +001/** +002 * +003 * Licensed to the Apache Software Foundation (ASF) under one +004 * or more contributor license agreements. See the NOTICE file +005 * distributed with this work for additional information +006 * regarding copyright ownership. The ASF licenses this file +007 * to you under the Apache License, Version 2.0 (the +008 * "License"); you may not use this file except in compliance +009 * with the License. You may obtain a copy of the License at +010 * +011 * http://www.apache.org/licenses/LICENSE-2.0 +012 * +013 * Unless required by applicable law or agreed to in writing, software +014 * distributed under the License is distributed on an "AS IS" BASIS, +015 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +016 * See the License for the specific language governing permissions and +017 * limitations under the License. +018 */ +019package org.apache.hadoop.hbase.util; +020 +021import java.io.FileNotFoundException; +022import java.io.IOException; +023import java.lang.reflect.InvocationTargetException; +024import java.lang.reflect.Method; +025import java.net.URI; +026import java.net.URISyntaxException; +027import java.util.List; +028import java.util.Locale; +029import java.util.Map; +030import java.util.concurrent.ConcurrentHashMap; +031 +032import org.apache.hadoop.HadoopIllegalArgumentException; +033import org.apache.hadoop.conf.Configuration; +034import org.apache.hadoop.fs.FSDataOutputStream; +035import org.apache.hadoop.fs.FileStatus; +036import org.apache.hadoop.fs.FileSystem; +037import org.apache.hadoop.fs.LocatedFileStatus; +038import org.apache.hadoop.fs.Path; +039import org.apache.hadoop.fs.PathFilter; +040import org.apache.hadoop.fs.RemoteIterator; +041import org.apache.hadoop.fs.permission.FsPermission; +042import org.apache.hadoop.hbase.HConstants; +043import org.apache.hadoop.hbase.TableName; +044import org.apache.hadoop.ipc.RemoteException; +045import org.apache.yetus.audience.InterfaceAudience; +046import org.slf4j.Logger; +047import org.slf4j.LoggerFactory; +048 +049import org.apache.hbase.thirdparty.com.google.common.annotations.VisibleForTesting; +050import org.apache.hbase.thirdparty.com.google.common.collect.Lists; +051 +052/** +053 * Utility methods for interacting with the underlying file system. +054 * +055 * Note that {@link #setStoragePolicy(FileSystem, Path, String)} is tested in TestFSUtils and +056 * pre-commit will run the hbase-server tests if there's code change in this class. See +057 * HBASE-20838; for more details. +058 */ +059@InterfaceAudience.Private +060public abstract class CommonFSUtils { +061 private static final Logger LOG = LoggerFactory.getLogger(CommonFSUtils.class); +062 +063 /** Parameter name for HBase WAL directory */ +064 public static final String HBASE_WAL_DIR = "hbase.wal.dir"; +065 +066 /** Parameter to disable stream capability enforcement checks */ +067 public static final String UNSAFE_STREAM_CAPABILITY_ENFORCE = "hbase.unsafe.stream.capability.enforce"; +068 +069 /** Full access permissions (starting point for a umask) */ +070 public static final String FULL_RWX_PERMISSIONS = "777"; +071 +072 protected CommonFSUtils() { +073super(); +074 } +075 +076 /** +077 * Compare of path component. Does not consider schema; i.e. if schemas +078 * different butpath
starts withrootPath
, +079 * then the function returns true +080 * @param rootPath value to check for +081 * @param path subject to check +082 * @return True ifpath
starts withrootPath
+083 */ +084 public static boolean isStartingWithPath(final Path rootPath, final String path) { +085String uriRootPath = rootPath.toUri().getPath(); +086String tailUriPath = (new Path(path)).toUri().getPath(); +087return tailUriPath.startsWith(uriRootPath); +088 } +089 +090 /** +091 * Compare path component of the Path URI; e.g. if hdfs://a/b/c and /a/b/c, it will compare the +092 * '/a/b/c' part. Does not consider schema; i.e. if schemas different but path or subpath matches, +093 * the two will equate. +094 * @param pathToSearch Path we will be trying to match against. +095 * @param pathTail what to match +096 * @retur
[03/11] hbase-site git commit: Published site at e963694259aa9242aa709e9b954754abd9a681cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e10887fe/devapidocs/src-html/org/apache/hadoop/hbase/util/CommonFSUtils.StreamCapabilities.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/util/CommonFSUtils.StreamCapabilities.html b/devapidocs/src-html/org/apache/hadoop/hbase/util/CommonFSUtils.StreamCapabilities.html index c6bfdca..cae3d07 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/util/CommonFSUtils.StreamCapabilities.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/util/CommonFSUtils.StreamCapabilities.html @@ -829,90 +829,217 @@ 821conf.setIfUnset(dfsKey, Integer.toString(hbaseSize)); 822 } 823 -824 // Holder singleton idiom. JVM spec ensures this will be run at most once per Classloader, and -825 // not until we attempt to reference it. -826 private static class StreamCapabilities { -827public static final boolean PRESENT; -828public static final Class> CLASS; -829public static final Method METHOD; -830static { -831 boolean tmp = false; -832 Class> clazz = null; -833 Method method = null; -834 try { -835clazz = Class.forName("org.apache.hadoop.fs.StreamCapabilities"); -836method = clazz.getMethod("hasCapability", String.class); -837tmp = true; -838 } catch(ClassNotFoundException|NoSuchMethodException|SecurityException exception) { -839LOG.warn("Your Hadoop installation does not include the StreamCapabilities class from " + -840 "HDFS-11644, so we will skip checking if any FSDataOutputStreams actually " + -841 "support hflush/hsync. If you are running on top of HDFS this probably just " + -842 "means you have an older version and this can be ignored. If you are running on " + -843 "top of an alternate FileSystem implementation you should manually verify that " + -844 "hflush and hsync are implemented; otherwise you risk data loss and hard to " + -845 "diagnose errors when our assumptions are violated."); -846LOG.debug("The first request to check for StreamCapabilities came from this stacktrace.", -847exception); -848 } finally { -849PRESENT = tmp; -850CLASS = clazz; -851METHOD = method; -852 } -853} -854 } -855 -856 /** -857 * If our FileSystem version includes the StreamCapabilities class, check if -858 * the given stream has a particular capability. -859 * @param stream capabilities are per-stream instance, so check this one specifically. must not be -860 *null -861 * @param capability what to look for, per Hadoop Common's FileSystem docs -862 * @return true if there are no StreamCapabilities. false if there are, but this stream doesn't -863 * implement it. return result of asking the stream otherwise. -864 */ -865 public static boolean hasCapability(FSDataOutputStream stream, String capability) { -866// be consistent whether or not StreamCapabilities is present -867if (stream == null) { -868 throw new NullPointerException("stream parameter must not be null."); -869} -870// If o.a.h.fs.StreamCapabilities doesn't exist, assume everyone does everything -871// otherwise old versions of Hadoop will break. -872boolean result = true; -873if (StreamCapabilities.PRESENT) { -874 // if StreamCapabilities is present, but the stream doesn't implement it -875 // or we run into a problem invoking the method, -876 // we treat that as equivalent to not declaring anything -877 result = false; -878 if (StreamCapabilities.CLASS.isAssignableFrom(stream.getClass())) { +824 private static class DfsBuilderUtility { +825static Class> dfsClass = null; +826static Method createMethod; +827static Method overwriteMethod; +828static Method bufferSizeMethod; +829static Method blockSizeMethod; +830static Method recursiveMethod; +831static Method replicateMethod; +832static Method replicationMethod; +833static Method buildMethod; +834static boolean allMethodsPresent = false; +835 +836static { +837 String dfsName = "org.apache.hadoop.hdfs.DistributedFileSystem"; +838 String builderName = dfsName + "$HdfsDataOutputStreamBuilder"; +839 Class> builderClass = null; +840 +841 try { +842dfsClass = Class.forName(dfsName); +843 } catch (ClassNotFoundException e) { +844LOG.debug("{} not available, will not use builder API for file creation.", dfsName); +845 } +846 try { +847builderClass = Class.forName(builderName); +848 } catch (ClassNotFoundException e) { +849LOG.debug("{} not available, will not use builder API for file creation.", builderName); +850 } +851 +852 if (dfsClass != null && builderClass != null) { +853tr
hbase-site git commit: INFRA-10751 Empty commit
Repository: hbase-site Updated Branches: refs/heads/asf-site e10887fef -> 09f27cd5b INFRA-10751 Empty commit Project: http://git-wip-us.apache.org/repos/asf/hbase-site/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase-site/commit/09f27cd5 Tree: http://git-wip-us.apache.org/repos/asf/hbase-site/tree/09f27cd5 Diff: http://git-wip-us.apache.org/repos/asf/hbase-site/diff/09f27cd5 Branch: refs/heads/asf-site Commit: 09f27cd5b21b1ce13ee07d6cc6b64086f95cc26c Parents: e10887f Author: jenkins Authored: Sat Jul 28 14:49:08 2018 + Committer: jenkins Committed: Sat Jul 28 14:49:08 2018 + -- --
[11/11] hbase-site git commit: Published site at e963694259aa9242aa709e9b954754abd9a681cd.
Published site at e963694259aa9242aa709e9b954754abd9a681cd. Project: http://git-wip-us.apache.org/repos/asf/hbase-site/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase-site/commit/e10887fe Tree: http://git-wip-us.apache.org/repos/asf/hbase-site/tree/e10887fe Diff: http://git-wip-us.apache.org/repos/asf/hbase-site/diff/e10887fe Branch: refs/heads/asf-site Commit: e10887fef6c65851f2fcba83f242b0dd360f4878 Parents: e0dc0c6 Author: jenkins Authored: Sat Jul 28 14:48:45 2018 + Committer: jenkins Committed: Sat Jul 28 14:48:45 2018 + -- acid-semantics.html | 4 +- apache_hbase_reference_guide.pdf| 4 +- book.html | 2 +- bulk-loads.html | 4 +- checkstyle-aggregate.html | 26878 - checkstyle.rss |20 +- coc.html| 4 +- dependencies.html | 4 +- dependency-convergence.html | 4 +- dependency-info.html| 4 +- dependency-management.html | 4 +- devapidocs/allclasses-frame.html| 1 + devapidocs/allclasses-noframe.html | 1 + devapidocs/constant-values.html | 6 +- devapidocs/index-all.html |44 + .../hadoop/hbase/backup/package-tree.html | 4 +- .../hadoop/hbase/backup/util/RestoreTool.html |12 +- .../hadoop/hbase/client/package-tree.html |22 +- .../hadoop/hbase/coprocessor/package-tree.html | 2 +- .../hadoop/hbase/executor/package-tree.html | 2 +- .../hadoop/hbase/filter/package-tree.html |10 +- .../hadoop/hbase/io/hfile/package-tree.html | 6 +- .../apache/hadoop/hbase/ipc/package-tree.html | 2 +- .../hadoop/hbase/mapreduce/package-tree.html| 4 +- .../hbase/master/balancer/package-tree.html | 2 +- .../hadoop/hbase/master/package-tree.html | 4 +- .../hbase/master/procedure/package-tree.html| 4 +- .../org/apache/hadoop/hbase/package-tree.html |18 +- .../hadoop/hbase/procedure2/package-tree.html | 4 +- .../hadoop/hbase/quotas/package-tree.html | 8 +- .../hadoop/hbase/regionserver/package-tree.html |16 +- .../regionserver/querymatcher/package-tree.html | 2 +- .../regionserver/wal/ProtobufLogWriter.html |12 +- .../hbase/regionserver/wal/package-tree.html| 2 +- .../replication/regionserver/package-tree.html | 2 +- .../hadoop/hbase/rest/model/package-tree.html | 2 +- .../hbase/security/access/package-tree.html | 2 +- .../hadoop/hbase/security/package-tree.html | 2 +- .../hadoop/hbase/thrift/package-tree.html | 2 +- .../util/CommonFSUtils.DfsBuilderUtility.html | 480 + .../util/CommonFSUtils.StreamCapabilities.html |14 +- ...nFSUtils.StreamLacksCapabilityException.html | 6 +- .../apache/hadoop/hbase/util/CommonFSUtils.html | 154 +- .../apache/hadoop/hbase/util/FSHDFSUtils.html | 2 +- .../org/apache/hadoop/hbase/util/FSUtils.html | 2 +- .../CommonFSUtils.DfsBuilderUtility.html| 125 + .../apache/hadoop/hbase/util/package-frame.html | 1 + .../hadoop/hbase/util/package-summary.html | 274 +- .../apache/hadoop/hbase/util/package-tree.html | 7 +- .../apache/hadoop/hbase/wal/package-tree.html | 2 +- devapidocs/overview-tree.html | 1 + .../org/apache/hadoop/hbase/Version.html| 6 +- .../hadoop/hbase/backup/util/RestoreTool.html | 547 +- .../hbase/io/asyncfs/AsyncFSOutputHelper.html |35 +- .../wal/WALProcedureStore.LeaseRecovery.html| 2 +- .../store/wal/WALProcedureStore.PushType.html | 2 +- .../wal/WALProcedureStore.SyncMetrics.html | 2 +- .../procedure2/store/wal/WALProcedureStore.html | 2 +- .../regionserver/wal/ProtobufLogWriter.html | 124 +- .../util/CommonFSUtils.DfsBuilderUtility.html | 1106 + .../util/CommonFSUtils.StreamCapabilities.html | 293 +- ...nFSUtils.StreamLacksCapabilityException.html | 293 +- .../apache/hadoop/hbase/util/CommonFSUtils.html | 293 +- downloads.html | 4 +- export_control.html | 4 +- index.html | 4 +- integration.html| 4 +- issue-tracking.html | 4 +- license.html| 4 +- mail-lists.html | 4 +- metrics.html| 4 +- old_news.html | 4 +- plugin-management.html
[06/11] hbase-site git commit: Published site at e963694259aa9242aa709e9b954754abd9a681cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e10887fe/devapidocs/org/apache/hadoop/hbase/util/package-summary.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/util/package-summary.html b/devapidocs/org/apache/hadoop/hbase/util/package-summary.html index 7deac38..051c486 100644 --- a/devapidocs/org/apache/hadoop/hbase/util/package-summary.html +++ b/devapidocs/org/apache/hadoop/hbase/util/package-summary.html @@ -479,358 +479,362 @@ -CommonFSUtils.StreamCapabilities +CommonFSUtils.DfsBuilderUtility +CommonFSUtils.StreamCapabilities + + + CompressionTest Compression validation test. - + ConcatenatedListsA collection class that contains multiple sub-lists, which allows us to not copy lists. - + ConfigurationUtil Utilities for storing more complex collection types in Configuration instances. - + ConnectionCache A utility to store user specific HConnections in memory. - + CoprocessorClassLoader ClassLoader used to load classes for Coprocessor instances. - + Counter Deprecated use https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/atomic/LongAdder.html?is-external=true"; title="class or interface in java.util.concurrent.atomic">LongAdder instead. - + Counter.Cell - + Counter.Container - + DefaultEnvironmentEdge Default implementation of an environment edge. - + DirectMemoryUtils Utilities for interacting with and monitoring DirectByteBuffer allocations. - + DNS Wrapper around Hadoop's DNS class to hide reflection. - + DynamicClassLoader This is a class loader that can load classes dynamically from new jar files under a configured folder. - + EncryptionTest - + EnvironmentEdgeManager Manages a singleton instance of the environment edge. - + ExceptionUtil This class handles the different interruption classes. - + ForeignExceptionUtil Helper to convert Exceptions and StackTraces from/to protobuf. - + FSHDFSUtils Implementation for hdfs - + FSRegionScanner Thread that walks over the filesystem, and computes the mappings Region -> BestHost and Region -> Map - + FSTableDescriptors Implementation of TableDescriptors that reads descriptors from the passed filesystem. - + FSUtils Utility methods for interacting with the underlying file system. - + FSUtils.BlackListDirFilter Directory filter that doesn't include any of the directories in the specified blacklist - + FSUtils.DirFilter A PathFilter that only allows directories. - + FSUtils.FamilyDirFilter Filter for all dirs that are legal column family names. - + FSUtils.FileFilter A PathFilter that returns only regular files. - + FSUtils.HFileFilter Filter for HFiles that excludes reference files. - + FSUtils.HFileLinkFilter Filter for HFileLinks (StoreFiles and HFiles not included). - + FSUtils.ReferenceFileFilter - + FSUtils.RegionDirFilter Filter for all dirs that don't start with '.' - + FSUtils.UserTableDirFilter A PathFilter that returns usertable directories. - + FSVisitor Utility methods for interacting with the hbase.root file system. - + GetJavaProperty A generic way for querying Java properties. - + Hash This class represents a common API for hashing functions. - + HashedBytes This class encapsulates a byte array and overrides hashCode and equals so that it's identity is based on the data rather than the array instance. - + HashKey Used to calculate the hash Hash algorithms for Bloomfilters. - + HasThread Abstract class which contains a Thread and delegates the common Thread methods to that instance. - + HBaseConfTool Tool that prints out a configuration. - + HBaseFsck HBaseFsck (hbck) is a tool for checking and repairing region consistency and table integrity problems in a corrupted HBase. - + HBaseFsck.HBaseFsckTool This is a Tool wrapper that gathers -Dxxx=yyy configuration settings from the command line. - + HBaseFsck.HbckInfo Maintain information about a particular region. - + HBaseFsck.HdfsEntry Stores the regioninfo entries from HDFS - + HBaseFsck.MetaEntry Stores the regioninfo entries scanned from META - + HBaseFsck.OnlineEntry Stores the regioninfo retrieved from Online region servers. - + HBaseFsck.PrintingErrorReporter - + HBaseFsck.RegionBoundariesInformation - + HBaseFsck.WorkItemHdfsRegionInfo Contact hdfs and get all information about specified table directory into regioninfo list. - + HBaseFsck.WorkItemOverlapMerge - + HBaseFsck.WorkItemRegion Contact a region server and get all information from it - + HBaseFsckRepair This class contains helper methods that repair parts of hbase's files
[07/11] hbase-site git commit: Published site at e963694259aa9242aa709e9b954754abd9a681cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e10887fe/devapidocs/org/apache/hadoop/hbase/util/CommonFSUtils.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/util/CommonFSUtils.html b/devapidocs/org/apache/hadoop/hbase/util/CommonFSUtils.html index 40667fc..6705fc8 100644 --- a/devapidocs/org/apache/hadoop/hbase/util/CommonFSUtils.html +++ b/devapidocs/org/apache/hadoop/hbase/util/CommonFSUtils.html @@ -18,7 +18,7 @@ catch(err) { } //--> -var methods = {"i0":9,"i1":9,"i2":9,"i3":9,"i4":9,"i5":9,"i6":9,"i7":9,"i8":9,"i9":9,"i10":9,"i11":9,"i12":9,"i13":9,"i14":9,"i15":9,"i16":9,"i17":9,"i18":9,"i19":9,"i20":9,"i21":9,"i22":9,"i23":9,"i24":9,"i25":9,"i26":9,"i27":9,"i28":9,"i29":9,"i30":9,"i31":9,"i32":9,"i33":9,"i34":9,"i35":9,"i36":9,"i37":9,"i38":9,"i39":9}; +var methods = {"i0":9,"i1":9,"i2":9,"i3":9,"i4":9,"i5":9,"i6":9,"i7":9,"i8":9,"i9":9,"i10":9,"i11":9,"i12":9,"i13":9,"i14":9,"i15":9,"i16":9,"i17":9,"i18":9,"i19":9,"i20":9,"i21":9,"i22":9,"i23":9,"i24":9,"i25":9,"i26":9,"i27":9,"i28":9,"i29":9,"i30":9,"i31":9,"i32":9,"i33":9,"i34":9,"i35":9,"i36":9,"i37":9,"i38":9,"i39":9,"i40":9,"i41":9}; var tabs = {65535:["t0","All Methods"],1:["t1","Static Methods"],8:["t4","Concrete Methods"]}; var altColor = "altColor"; var rowColor = "rowColor"; @@ -50,7 +50,7 @@ var activeTableTab = "activeTableTab"; Prev Class -Next Class +Next Class Frames @@ -141,9 +141,13 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html private static class -CommonFSUtils.StreamCapabilities +CommonFSUtils.DfsBuilderUtility +private static class +CommonFSUtils.StreamCapabilities + + static class CommonFSUtils.StreamLacksCapabilityException Helper exception for those cases where the place where we need to check a stream capability @@ -241,6 +245,28 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html +static org.apache.hadoop.fs.FSDataOutputStream +createForWal(org.apache.hadoop.fs.FileSystem fs, +org.apache.hadoop.fs.Path path, +boolean overwritable) +Attempt to use builder API via reflection to create a file with the given parameters and + replication enabled. + + + +static org.apache.hadoop.fs.FSDataOutputStream +createForWal(org.apache.hadoop.fs.FileSystem fs, +org.apache.hadoop.fs.Path path, +boolean overwritable, +int bufferSize, +short replication, +long blockSize, +boolean isRecursive) +Attempt to use builder API via reflection to create a file with the given parameters and + replication enabled. + + + static boolean delete(org.apache.hadoop.fs.FileSystem fs, org.apache.hadoop.fs.Path path, @@ -248,18 +274,18 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html Calls fs.delete() and returns the value returned by the fs.delete() - + static boolean deleteDirectory(org.apache.hadoop.fs.FileSystem fs, org.apache.hadoop.fs.Path dir) Delete if exists. - + static org.apache.hadoop.fs.FileSystem getCurrentFileSystem(org.apache.hadoop.conf.Configuration conf) - + static long getDefaultBlockSize(org.apache.hadoop.fs.FileSystem fs, org.apache.hadoop.fs.Path path) @@ -267,18 +293,18 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html be split into to minimize i/o time. - + static int getDefaultBufferSize(org.apache.hadoop.fs.FileSystem fs) Returns the default buffer size to use during writes. - + static short getDefaultReplication(org.apache.hadoop.fs.FileSystem fs, org.apache.hadoop.fs.Path path) - + static org.apache.hadoop.fs.permission.FsPermission getFilePermissions(org.apache.hadoop.fs.FileSystem fs, org.apache.hadoop.conf.Configuration conf, @@ -287,7 +313,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html enabled. - + static org.apache.hadoop.fs.Path getNamespaceDir(org.apache.hadoop.fs.Path rootdir, https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true"; title="class or interface in java.lang">String namespace) @@ -295,21 +321,21 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html the namespace directory under path rootdir - + static https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true"; title="class or interface in java.lang">String getPath(org.apache.hadoop.fs.Path p) Return the 'path' component of a Path. - + static org.apache.hadoop.fs.Path getRootDir(org.apache.hadoop.conf.Configuration c) - + static org.apache.hadoop.fs.FileSystem getRootDirFileSystem(org.apache.hadoop.conf.Configuration c) - + static org.apache.hadoop.fs.Path getTableDir(org.apache.hadoop.fs.Path rootdir, TableName tableName) @@ -317,7 +343,7 @@ ext
[01/11] hbase-site git commit: Published site at e963694259aa9242aa709e9b954754abd9a681cd.
Repository: hbase-site Updated Branches: refs/heads/asf-site e0dc0c667 -> e10887fef http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e10887fe/testdevapidocs/org/apache/hadoop/hbase/class-use/HBaseTestingUtility.html -- diff --git a/testdevapidocs/org/apache/hadoop/hbase/class-use/HBaseTestingUtility.html b/testdevapidocs/org/apache/hadoop/hbase/class-use/HBaseTestingUtility.html index 38cc11c..1165b4f 100644 --- a/testdevapidocs/org/apache/hadoop/hbase/class-use/HBaseTestingUtility.html +++ b/testdevapidocs/org/apache/hadoop/hbase/class-use/HBaseTestingUtility.html @@ -3263,6 +3263,10 @@ private static HBaseTestingUtility +TestHBaseWalOnEC.util + + +private static HBaseTestingUtility TestCombinedAsyncWriter.UTIL http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e10887fe/testdevapidocs/org/apache/hadoop/hbase/io/hfile/package-tree.html -- diff --git a/testdevapidocs/org/apache/hadoop/hbase/io/hfile/package-tree.html b/testdevapidocs/org/apache/hadoop/hbase/io/hfile/package-tree.html index ec56445..4837d80 100644 --- a/testdevapidocs/org/apache/hadoop/hbase/io/hfile/package-tree.html +++ b/testdevapidocs/org/apache/hadoop/hbase/io/hfile/package-tree.html @@ -158,8 +158,8 @@ java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true"; title="class or interface in java.lang">Enum(implements java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true"; title="class or interface in java.lang">Comparable , java.io.https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true"; title="class or interface in java.io">Serializable) -org.apache.hadoop.hbase.io.hfile.TestCacheOnWrite.CacheOnWriteType org.apache.hadoop.hbase.io.hfile.TagUsage +org.apache.hadoop.hbase.io.hfile.TestCacheOnWrite.CacheOnWriteType http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e10887fe/testdevapidocs/org/apache/hadoop/hbase/package-tree.html -- diff --git a/testdevapidocs/org/apache/hadoop/hbase/package-tree.html b/testdevapidocs/org/apache/hadoop/hbase/package-tree.html index 3d372b1..55dede9 100644 --- a/testdevapidocs/org/apache/hadoop/hbase/package-tree.html +++ b/testdevapidocs/org/apache/hadoop/hbase/package-tree.html @@ -576,15 +576,15 @@ java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true"; title="class or interface in java.lang">Enum (implements java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true"; title="class or interface in java.lang">Comparable , java.io.https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true"; title="class or interface in java.io">Serializable) +org.apache.hadoop.hbase.RESTApiClusterManager.Service org.apache.hadoop.hbase.RESTApiClusterManager.RoleCommand org.apache.hadoop.hbase.IntegrationTestRegionReplicaPerf.Stat -org.apache.hadoop.hbase.ClusterManager.ServiceType -org.apache.hadoop.hbase.PerformanceEvaluation.Counter org.apache.hadoop.hbase.ScanPerformanceEvaluation.ScanCounter -org.apache.hadoop.hbase.RESTApiClusterManager.Service -org.apache.hadoop.hbase.ResourceChecker.Phase +org.apache.hadoop.hbase.ClusterManager.ServiceType org.apache.hadoop.hbase.HBaseClusterManager.CommandProvider.Operation org.apache.hadoop.hbase.IntegrationTestDDLMasterFailover.ACTION +org.apache.hadoop.hbase.PerformanceEvaluation.Counter +org.apache.hadoop.hbase.ResourceChecker.Phase http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e10887fe/testdevapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html -- diff --git a/testdevapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html b/testdevapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html index 47095d2..9ff71c8 100644 --- a/testdevapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html +++ b/testdevapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html @@ -667,11 +667,11 @@ java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true"; title="class or interface in java.lang">Enum (implements java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true"; title="class or interface in java.lang">Comparable , java.io.https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true"; title="class or interface in java.io">Serializable) +org.apache.hadoop.hbase.regionserver.TestRegionServerReadRequestMetrics.Metric org.apache.hadoop.hbase.regionserver.TestMultiLogThreshold.ActionType +org.apache.hadoop.hbase.regionserver.DataBlockEncodingTool.Manipulation org.apache.hadoop.hbase.regionserver.TestAto