[21/35] hbase-site git commit: Published site at 8cc56bd18c40ba9a7131336e97c74f8d97d8b2be.

2018-10-17 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/713132a3/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/compactions/Compactor.InternalScannerFactory.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/compactions/Compactor.InternalScannerFactory.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/compactions/Compactor.InternalScannerFactory.html
index c2698b3..9142fe1 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/compactions/Compactor.InternalScannerFactory.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/compactions/Compactor.InternalScannerFactory.html
@@ -207,307 +207,309 @@
 199  }
 200  tmp = 
fileInfo.get(TIMERANGE_KEY);
 201  fd.latestPutTs = tmp == null ? 
HConstants.LATEST_TIMESTAMP: TimeRangeTracker.parseFrom(tmp).getMax();
-202  LOG.debug("Compacting {}, 
keycount={}, bloomtype={}, size={}, encoding={}, seqNum={}{}",
-203  (file.getPath() == null? null: 
file.getPath().getName()),
-204  keyCount,
-205  
r.getBloomFilterType().toString(),
-206  
TraditionalBinaryPrefix.long2String(r.length(), "", 1),
-207  
r.getHFileReader().getDataBlockEncoding(),
-208  seqNum,
-209  (allFiles? ", earliestPutTs=" + 
earliestPutTs: ""));
-210}
-211return fd;
-212  }
-213
-214  /**
-215   * Creates file scanners for 
compaction.
-216   * @param filesToCompact Files.
-217   * @return Scanners.
-218   */
-219  private ListStoreFileScanner 
createFileScanners(CollectionHStoreFile filesToCompact,
-220  long smallestReadPoint, boolean 
useDropBehind) throws IOException {
-221return 
StoreFileScanner.getScannersForCompaction(filesToCompact, useDropBehind,
-222  smallestReadPoint);
-223  }
-224
-225  private long getSmallestReadPoint() {
-226return 
store.getSmallestReadPoint();
-227  }
-228
-229  protected interface 
InternalScannerFactory {
+202  LOG.debug("Compacting {}, 
keycount={}, bloomtype={}, size={}, "
+203  + "encoding={}, 
compression={}, seqNum={}{}",
+204  (file.getPath() == null? null: 
file.getPath().getName()),
+205  keyCount,
+206  
r.getBloomFilterType().toString(),
+207  
TraditionalBinaryPrefix.long2String(r.length(), "", 1),
+208  
r.getHFileReader().getDataBlockEncoding(),
+209  compactionCompression,
+210  seqNum,
+211  (allFiles? ", earliestPutTs=" + 
earliestPutTs: ""));
+212}
+213return fd;
+214  }
+215
+216  /**
+217   * Creates file scanners for 
compaction.
+218   * @param filesToCompact Files.
+219   * @return Scanners.
+220   */
+221  private ListStoreFileScanner 
createFileScanners(CollectionHStoreFile filesToCompact,
+222  long smallestReadPoint, boolean 
useDropBehind) throws IOException {
+223return 
StoreFileScanner.getScannersForCompaction(filesToCompact, useDropBehind,
+224  smallestReadPoint);
+225  }
+226
+227  private long getSmallestReadPoint() {
+228return 
store.getSmallestReadPoint();
+229  }
 230
-231ScanType 
getScanType(CompactionRequestImpl request);
+231  protected interface 
InternalScannerFactory {
 232
-233InternalScanner 
createScanner(ScanInfo scanInfo, ListStoreFileScanner scanners, 
ScanType scanType,
-234FileDetails fd, long 
smallestReadPoint) throws IOException;
-235  }
-236
-237  protected final InternalScannerFactory 
defaultScannerFactory = new InternalScannerFactory() {
+233ScanType 
getScanType(CompactionRequestImpl request);
+234
+235InternalScanner 
createScanner(ScanInfo scanInfo, ListStoreFileScanner scanners, 
ScanType scanType,
+236FileDetails fd, long 
smallestReadPoint) throws IOException;
+237  }
 238
-239@Override
-240public ScanType 
getScanType(CompactionRequestImpl request) {
-241  return request.isAllFiles() ? 
COMPACT_DROP_DELETES : COMPACT_RETAIN_DELETES;
-242}
-243
-244@Override
-245public InternalScanner 
createScanner(ScanInfo scanInfo, ListStoreFileScanner scanners,
-246ScanType scanType, FileDetails 
fd, long smallestReadPoint) throws IOException {
-247  return 
Compactor.this.createScanner(store, scanInfo, scanners, scanType, 
smallestReadPoint,
-248fd.earliestPutTs);
-249}
-250  };
-251
-252  /**
-253   * Creates a writer for a new file in a 
temporary directory.
-254   * @param fd The file details.
-255   * @return Writer for a new StoreFile 
in the tmp dir.
-256   * @throws IOException if creation 
failed
-257   */
-258  protected final StoreFileWriter 
createTmpWriter(FileDetails fd, boolean shouldDropBehind)
-259  throws IOException {
-260// When all MVCC readpoints are 0, 
don't write them.
-261// See HBASE-8166, HBASE-12600, and 
HBASE-13389.
-262return 
store.createWriterInTmp(fd.maxKeyCount, this.compactionCompression, true,
-263fd.maxMVCCReadpoint  0, 

[21/35] hbase-site git commit: Published site at 42d5447cfbc593becfc26684e03f482eb3a0fc49.

2018-04-17 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a610f23a/testdevapidocs/org/apache/hadoop/hbase/replication/TestReplicationSmallTests.html
--
diff --git 
a/testdevapidocs/org/apache/hadoop/hbase/replication/TestReplicationSmallTests.html
 
b/testdevapidocs/org/apache/hadoop/hbase/replication/TestReplicationSmallTests.html
index 7dbb5cc..79759cc 100644
--- 
a/testdevapidocs/org/apache/hadoop/hbase/replication/TestReplicationSmallTests.html
+++ 
b/testdevapidocs/org/apache/hadoop/hbase/replication/TestReplicationSmallTests.html
@@ -18,8 +18,8 @@
 catch(err) {
 }
 //-->
-var methods = 
{"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10};
-var tabs = {65535:["t0","All Methods"],2:["t2","Instance 
Methods"],8:["t4","Concrete Methods"]};
+var methods = 
{"i0":10,"i1":9,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10};
+var tabs = {65535:["t0","All Methods"],1:["t1","Static 
Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]};
 var altColor = "altColor";
 var rowColor = "rowColor";
 var tableTab = "tableTab";
@@ -114,7 +114,7 @@ var activeTableTab = "activeTableTab";
 
 
 
-public class TestReplicationSmallTests
+public class TestReplicationSmallTests
 extends TestReplicationBase
 
 
@@ -146,13 +146,17 @@ extends private static https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String
 PEER_ID
 
+
+boolean
+serialPeer
+
 
 
 
 
 
 Fields inherited from classorg.apache.hadoop.hbase.replication.TestReplicationBase
-admin,
 CONF_WITH_LOCALFS,
 conf1,
 conf2,
 famName,
 hbaseAdmin,
 htable1,
 htable2,
 NB_RETRIES,
 NB_ROWS_IN_BATCH,
 NB_ROWS_IN_BIG_BATCH,
 noRepfamName,
 row,
 scopes,
 seperateOldWALs,
 SLEEP_TIME,
 tableName,
 utility1, utility2,
 zkw1,
 zkw2
+admin,
 CONF_WITH_LOCALFS,
 conf1,
 conf2,
 famName,
 hbaseAdmin,
 htable1,
 htable2,
 NB_RETRIES,
 NB_ROWS_IN_BATCH,
 NB_ROWS_IN_BIG_BATCH,
 noRepfamName,
 PEER_ID2,
 row,
 scopes,
 SLEEP_TIME,
 tableName,
 utility1, utility2,
 zkw1,
 zkw2
 
 
 
@@ -180,60 +184,68 @@ extends 
-All MethodsInstance MethodsConcrete Methods
+All MethodsStatic MethodsInstance MethodsConcrete Methods
 
 Modifier and Type
 Method and Description
 
 
+protected boolean
+isSerialPeer()
+
+
+static https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">Listhttps://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true;
 title="class or interface in java.lang">Boolean
+parameters()
+
+
 void
 setUp()
 
-
+
 void
 testAddAndRemoveClusters()
 Integration test for TestReplicationAdmin, removes and 
re-add a peer cluster
 
 
-
+
 void
 testDeleteTypes()
 Verify that version and column delete marker types are 
replicated correctly.
 
 
-
+
 void
 testDisableEnable()
 Test disable/enable replication, trying to insert, make 
sure nothing's replicated, enable it,
  the insert should be replicated
 
 
-
+
 void
 testLoading()
 Do a more intense version testSmallBatch, one that will 
trigger wal rolling and other
  non-trivial code paths
 
 
-
+
 void
 testReplicationInReplay()
 Test for HBase-15259 WALEdits under replay will also be 
replicated
 
 
-
+
 void
 testSimplePutDelete()
 Add a row, check it's replicated, delete it, check's 
gone
 
 
-
+
 void
 testSmallBatch()
 Try a small batch upload using the write buffer, check it's 
replicated
 
 
-
+
 void
 testVerifyListReplicatedTable()
 Test for HBASE-8663
@@ -245,7 +257,7 @@ extends TestReplicationBase
-cleanUp,
 loadData,
 params,
 runSimplePutDeleteTest,
 runSmallBatchTest,
 setUpBeforeClass,
 tearDownAfterClass,
 w
 aitForReplication
+cleanUp,
 loadData,
 runSimplePutDeleteTest,
 runSmallBatchTest,
 setUpBase,
 setUpBeforeClass,
 tearDownAfterClass,
 tearDown
 Base, waitForReplication
 
 
 
@@ -274,7 +286,7 @@ extends 
 
 CLASS_RULE
-public static finalHBaseClassTestRule CLASS_RULE
+public static finalHBaseClassTestRule CLASS_RULE
 
 
 
@@ -283,22 +295,31 @@ extends 
 
 LOG
-private static finalorg.slf4j.Logger LOG
+private static finalorg.slf4j.Logger LOG
 
 
 
 
 
-
+
 
 PEER_ID
-private static finalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String PEER_ID
+private static finalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String PEER_ID
 
 See Also:
 Constant
 Field Values
 
 
 
+
+
+
+
+
+serialPeer
+publicboolean serialPeer
+
+
 
 
 
@@ -313,7 +334,7 @@ extends 
 
 TestReplicationSmallTests
-publicTestReplicationSmallTests()
+publicTestReplicationSmallTests()
 
 
 
@@ -324,13 +345,35 @@ extends 
+
+
+
+
+isSerialPeer
+protectedbooleanisSerialPeer()
+
+Overrides:
+isSerialPeerin
 classTestReplicationBase
+
+
+
+
+
+
+
+
+parameters
+public 

[21/35] hbase-site git commit: Published site at .

2018-02-05 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/04d647a7/devapidocs/src-html/org/apache/hadoop/hbase/util/HBaseFsck.FileLockCallable.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/util/HBaseFsck.FileLockCallable.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/util/HBaseFsck.FileLockCallable.html
index bf1a2cc..89317aa 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/util/HBaseFsck.FileLockCallable.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/util/HBaseFsck.FileLockCallable.html
@@ -1813,3325 +1813,3330 @@
 1805  private void loadTableStates()
 1806  throws IOException {
 1807tableStates = 
MetaTableAccessor.getTableStates(connection);
-1808  }
-1809
-1810  /**
-1811   * Check if the specified region's 
table is disabled.
-1812   * @param tableName table to check 
status of
-1813   */
-1814  private boolean 
isTableDisabled(TableName tableName) {
-1815return 
tableStates.containsKey(tableName)
-1816 
tableStates.get(tableName)
-1817
.inStates(TableState.State.DISABLED, TableState.State.DISABLING);
-1818  }
-1819
-1820  /**
-1821   * Scan HDFS for all regions, 
recording their information into
-1822   * regionInfoMap
-1823   */
-1824  public void loadHdfsRegionDirs() 
throws IOException, InterruptedException {
-1825Path rootDir = 
FSUtils.getRootDir(getConf());
-1826FileSystem fs = 
rootDir.getFileSystem(getConf());
-1827
-1828// list all tables from HDFS
-1829ListFileStatus tableDirs = 
Lists.newArrayList();
-1830
-1831boolean foundVersionFile = 
fs.exists(new Path(rootDir, HConstants.VERSION_FILE_NAME));
+1808// Add hbase:meta so this tool keeps 
working. In hbase2, meta is always enabled though it
+1809// has no entry in the table states. 
HBCK doesn't work right w/ hbase2 but just do this in
+1810// meantime.
+1811
this.tableStates.put(TableName.META_TABLE_NAME,
+1812new 
TableState(TableName.META_TABLE_NAME, TableState.State.ENABLED));
+1813  }
+1814
+1815  /**
+1816   * Check if the specified region's 
table is disabled.
+1817   * @param tableName table to check 
status of
+1818   */
+1819  private boolean 
isTableDisabled(TableName tableName) {
+1820return 
tableStates.containsKey(tableName)
+1821 
tableStates.get(tableName)
+1822
.inStates(TableState.State.DISABLED, TableState.State.DISABLING);
+1823  }
+1824
+1825  /**
+1826   * Scan HDFS for all regions, 
recording their information into
+1827   * regionInfoMap
+1828   */
+1829  public void loadHdfsRegionDirs() 
throws IOException, InterruptedException {
+1830Path rootDir = 
FSUtils.getRootDir(getConf());
+1831FileSystem fs = 
rootDir.getFileSystem(getConf());
 1832
-1833ListPath paths = 
FSUtils.getTableDirs(fs, rootDir);
-1834for (Path path : paths) {
-1835  TableName tableName = 
FSUtils.getTableName(path);
-1836   if ((!checkMetaOnly 
-1837   isTableIncluded(tableName)) 
||
-1838   
tableName.equals(TableName.META_TABLE_NAME)) {
-1839 
tableDirs.add(fs.getFileStatus(path));
-1840   }
-1841}
-1842
-1843// verify that version file exists
-1844if (!foundVersionFile) {
-1845  
errors.reportError(ERROR_CODE.NO_VERSION_FILE,
-1846  "Version file does not exist 
in root dir " + rootDir);
-1847  if (shouldFixVersionFile()) {
-1848LOG.info("Trying to create a new 
" + HConstants.VERSION_FILE_NAME
-1849+ " file.");
-1850setShouldRerun();
-1851FSUtils.setVersion(fs, rootDir, 
getConf().getInt(
-1852
HConstants.THREAD_WAKE_FREQUENCY, 10 * 1000), getConf().getInt(
-1853
HConstants.VERSION_FILE_WRITE_ATTEMPTS,
-1854
HConstants.DEFAULT_VERSION_FILE_WRITE_ATTEMPTS));
-1855  }
-1856}
-1857
-1858// Avoid multithreading at 
table-level because already multithreaded internally at
-1859// region-level.  Additionally 
multithreading at table-level can lead to deadlock
-1860// if there are many tables in the 
cluster.  Since there are a limited # of threads
-1861// in the executor's thread pool and 
if we multithread at the table-level by putting
-1862// WorkItemHdfsDir callables into 
the executor, then we will have some threads in the
-1863// executor tied up solely in 
waiting for the tables' region-level calls to complete.
-1864// If there are enough tables then 
there will be no actual threads in the pool left
-1865// for the region-level callables to 
be serviced.
-1866for (FileStatus tableDir : 
tableDirs) {
-1867  LOG.debug("Loading region dirs 
from " +tableDir.getPath());
-1868  WorkItemHdfsDir item = new 
WorkItemHdfsDir(fs, errors, tableDir);
-1869  try {
-1870item.call();
-1871  } catch (ExecutionException e) {
-1872LOG.warn("Could not completely 
load table dir " +
-1873tableDir.getPath(), 
e.getCause());
-1874

[21/35] hbase-site git commit: Published site at 9250bf809155ebe93fd6ae8a0485b22c744fdf70.

2016-11-14 Thread stack
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f17356a7/devapidocs/src-html/org/apache/hadoop/hbase/tmpl/master/MasterStatusTmpl.Intf.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/tmpl/master/MasterStatusTmpl.Intf.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/tmpl/master/MasterStatusTmpl.Intf.html
index 2b27b2b..0a4e021 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/tmpl/master/MasterStatusTmpl.Intf.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/tmpl/master/MasterStatusTmpl.Intf.html
@@ -69,15 +69,15 @@
 061  requiredArguments = {
 062@org.jamon.annotations.Argument(name 
= "master", type = "HMaster")},
 063  optionalArguments = {
-064@org.jamon.annotations.Argument(name 
= "frags", type = "MapString,Integer"),
-065@org.jamon.annotations.Argument(name 
= "metaLocation", type = "ServerName"),
+064@org.jamon.annotations.Argument(name 
= "filter", type = "String"),
+065@org.jamon.annotations.Argument(name 
= "deadServers", type = "SetServerName"),
 066@org.jamon.annotations.Argument(name 
= "catalogJanitorEnabled", type = "boolean"),
-067@org.jamon.annotations.Argument(name 
= "deadServers", type = "SetServerName"),
-068@org.jamon.annotations.Argument(name 
= "assignmentManager", type = "AssignmentManager"),
-069@org.jamon.annotations.Argument(name 
= "servers", type = "ListServerName"),
-070@org.jamon.annotations.Argument(name 
= "format", type = "String"),
-071@org.jamon.annotations.Argument(name 
= "filter", type = "String"),
-072@org.jamon.annotations.Argument(name 
= "serverManager", type = "ServerManager")})
+067@org.jamon.annotations.Argument(name 
= "format", type = "String"),
+068@org.jamon.annotations.Argument(name 
= "serverManager", type = "ServerManager"),
+069@org.jamon.annotations.Argument(name 
= "frags", type = "MapString,Integer"),
+070@org.jamon.annotations.Argument(name 
= "metaLocation", type = "ServerName"),
+071@org.jamon.annotations.Argument(name 
= "assignmentManager", type = "AssignmentManager"),
+072@org.jamon.annotations.Argument(name 
= "servers", type = "ListServerName")})
 073public class MasterStatusTmpl
 074  extends 
org.jamon.AbstractTemplateProxy
 075{
@@ -118,40 +118,40 @@
 110  return m_master;
 111}
 112private HMaster m_master;
-113// 21, 1
-114public void 
setFrags(MapString,Integer frags)
+113// 26, 1
+114public void setFilter(String 
filter)
 115{
-116  // 21, 1
-117  m_frags = frags;
-118  m_frags__IsNotDefault = true;
+116  // 26, 1
+117  m_filter = filter;
+118  m_filter__IsNotDefault = true;
 119}
-120public MapString,Integer 
getFrags()
+120public String getFilter()
 121{
-122  return m_frags;
+122  return m_filter;
 123}
-124private MapString,Integer 
m_frags;
-125public boolean 
getFrags__IsNotDefault()
+124private String m_filter;
+125public boolean 
getFilter__IsNotDefault()
 126{
-127  return m_frags__IsNotDefault;
+127  return m_filter__IsNotDefault;
 128}
-129private boolean 
m_frags__IsNotDefault;
-130// 22, 1
-131public void 
setMetaLocation(ServerName metaLocation)
+129private boolean 
m_filter__IsNotDefault;
+130// 24, 1
+131public void 
setDeadServers(SetServerName deadServers)
 132{
-133  // 22, 1
-134  m_metaLocation = metaLocation;
-135  m_metaLocation__IsNotDefault = 
true;
+133  // 24, 1
+134  m_deadServers = deadServers;
+135  m_deadServers__IsNotDefault = 
true;
 136}
-137public ServerName getMetaLocation()
+137public SetServerName 
getDeadServers()
 138{
-139  return m_metaLocation;
+139  return m_deadServers;
 140}
-141private ServerName m_metaLocation;
-142public boolean 
getMetaLocation__IsNotDefault()
+141private SetServerName 
m_deadServers;
+142public boolean 
getDeadServers__IsNotDefault()
 143{
-144  return 
m_metaLocation__IsNotDefault;
+144  return 
m_deadServers__IsNotDefault;
 145}
-146private boolean 
m_metaLocation__IsNotDefault;
+146private boolean 
m_deadServers__IsNotDefault;
 147// 25, 1
 148public void 
setCatalogJanitorEnabled(boolean catalogJanitorEnabled)
 149{
@@ -169,108 +169,108 @@
 161  return 
m_catalogJanitorEnabled__IsNotDefault;
 162}
 163private boolean 
m_catalogJanitorEnabled__IsNotDefault;
-164// 24, 1
-165public void 
setDeadServers(SetServerName deadServers)
+164// 27, 1
+165public void setFormat(String 
format)
 166{
-167  // 24, 1
-168  m_deadServers = deadServers;
-169  m_deadServers__IsNotDefault = 
true;
+167  // 27, 1
+168  m_format = format;
+169  m_format__IsNotDefault = true;
 170}
-171public SetServerName 
getDeadServers()
+171public String getFormat()
 172{
-173  return m_deadServers;