[21/35] hbase-site git commit: Published site at 8cc56bd18c40ba9a7131336e97c74f8d97d8b2be.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/713132a3/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/compactions/Compactor.InternalScannerFactory.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/compactions/Compactor.InternalScannerFactory.html b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/compactions/Compactor.InternalScannerFactory.html index c2698b3..9142fe1 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/compactions/Compactor.InternalScannerFactory.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/compactions/Compactor.InternalScannerFactory.html @@ -207,307 +207,309 @@ 199 } 200 tmp = fileInfo.get(TIMERANGE_KEY); 201 fd.latestPutTs = tmp == null ? HConstants.LATEST_TIMESTAMP: TimeRangeTracker.parseFrom(tmp).getMax(); -202 LOG.debug("Compacting {}, keycount={}, bloomtype={}, size={}, encoding={}, seqNum={}{}", -203 (file.getPath() == null? null: file.getPath().getName()), -204 keyCount, -205 r.getBloomFilterType().toString(), -206 TraditionalBinaryPrefix.long2String(r.length(), "", 1), -207 r.getHFileReader().getDataBlockEncoding(), -208 seqNum, -209 (allFiles? ", earliestPutTs=" + earliestPutTs: "")); -210} -211return fd; -212 } -213 -214 /** -215 * Creates file scanners for compaction. -216 * @param filesToCompact Files. -217 * @return Scanners. -218 */ -219 private ListStoreFileScanner createFileScanners(CollectionHStoreFile filesToCompact, -220 long smallestReadPoint, boolean useDropBehind) throws IOException { -221return StoreFileScanner.getScannersForCompaction(filesToCompact, useDropBehind, -222 smallestReadPoint); -223 } -224 -225 private long getSmallestReadPoint() { -226return store.getSmallestReadPoint(); -227 } -228 -229 protected interface InternalScannerFactory { +202 LOG.debug("Compacting {}, keycount={}, bloomtype={}, size={}, " +203 + "encoding={}, compression={}, seqNum={}{}", +204 (file.getPath() == null? null: file.getPath().getName()), +205 keyCount, +206 r.getBloomFilterType().toString(), +207 TraditionalBinaryPrefix.long2String(r.length(), "", 1), +208 r.getHFileReader().getDataBlockEncoding(), +209 compactionCompression, +210 seqNum, +211 (allFiles? ", earliestPutTs=" + earliestPutTs: "")); +212} +213return fd; +214 } +215 +216 /** +217 * Creates file scanners for compaction. +218 * @param filesToCompact Files. +219 * @return Scanners. +220 */ +221 private ListStoreFileScanner createFileScanners(CollectionHStoreFile filesToCompact, +222 long smallestReadPoint, boolean useDropBehind) throws IOException { +223return StoreFileScanner.getScannersForCompaction(filesToCompact, useDropBehind, +224 smallestReadPoint); +225 } +226 +227 private long getSmallestReadPoint() { +228return store.getSmallestReadPoint(); +229 } 230 -231ScanType getScanType(CompactionRequestImpl request); +231 protected interface InternalScannerFactory { 232 -233InternalScanner createScanner(ScanInfo scanInfo, ListStoreFileScanner scanners, ScanType scanType, -234FileDetails fd, long smallestReadPoint) throws IOException; -235 } -236 -237 protected final InternalScannerFactory defaultScannerFactory = new InternalScannerFactory() { +233ScanType getScanType(CompactionRequestImpl request); +234 +235InternalScanner createScanner(ScanInfo scanInfo, ListStoreFileScanner scanners, ScanType scanType, +236FileDetails fd, long smallestReadPoint) throws IOException; +237 } 238 -239@Override -240public ScanType getScanType(CompactionRequestImpl request) { -241 return request.isAllFiles() ? COMPACT_DROP_DELETES : COMPACT_RETAIN_DELETES; -242} -243 -244@Override -245public InternalScanner createScanner(ScanInfo scanInfo, ListStoreFileScanner scanners, -246ScanType scanType, FileDetails fd, long smallestReadPoint) throws IOException { -247 return Compactor.this.createScanner(store, scanInfo, scanners, scanType, smallestReadPoint, -248fd.earliestPutTs); -249} -250 }; -251 -252 /** -253 * Creates a writer for a new file in a temporary directory. -254 * @param fd The file details. -255 * @return Writer for a new StoreFile in the tmp dir. -256 * @throws IOException if creation failed -257 */ -258 protected final StoreFileWriter createTmpWriter(FileDetails fd, boolean shouldDropBehind) -259 throws IOException { -260// When all MVCC readpoints are 0, don't write them. -261// See HBASE-8166, HBASE-12600, and HBASE-13389. -262return store.createWriterInTmp(fd.maxKeyCount, this.compactionCompression, true, -263fd.maxMVCCReadpoint 0,
[21/35] hbase-site git commit: Published site at 42d5447cfbc593becfc26684e03f482eb3a0fc49.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a610f23a/testdevapidocs/org/apache/hadoop/hbase/replication/TestReplicationSmallTests.html -- diff --git a/testdevapidocs/org/apache/hadoop/hbase/replication/TestReplicationSmallTests.html b/testdevapidocs/org/apache/hadoop/hbase/replication/TestReplicationSmallTests.html index 7dbb5cc..79759cc 100644 --- a/testdevapidocs/org/apache/hadoop/hbase/replication/TestReplicationSmallTests.html +++ b/testdevapidocs/org/apache/hadoop/hbase/replication/TestReplicationSmallTests.html @@ -18,8 +18,8 @@ catch(err) { } //--> -var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10}; -var tabs = {65535:["t0","All Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]}; +var methods = {"i0":10,"i1":9,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10}; +var tabs = {65535:["t0","All Methods"],1:["t1","Static Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]}; var altColor = "altColor"; var rowColor = "rowColor"; var tableTab = "tableTab"; @@ -114,7 +114,7 @@ var activeTableTab = "activeTableTab"; -public class TestReplicationSmallTests +public class TestReplicationSmallTests extends TestReplicationBase @@ -146,13 +146,17 @@ extends private static https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String PEER_ID + +boolean +serialPeer + Fields inherited from classorg.apache.hadoop.hbase.replication.TestReplicationBase -admin, CONF_WITH_LOCALFS, conf1, conf2, famName, hbaseAdmin, htable1, htable2, NB_RETRIES, NB_ROWS_IN_BATCH, NB_ROWS_IN_BIG_BATCH, noRepfamName, row, scopes, seperateOldWALs, SLEEP_TIME, tableName, utility1, utility2, zkw1, zkw2 +admin, CONF_WITH_LOCALFS, conf1, conf2, famName, hbaseAdmin, htable1, htable2, NB_RETRIES, NB_ROWS_IN_BATCH, NB_ROWS_IN_BIG_BATCH, noRepfamName, PEER_ID2, row, scopes, SLEEP_TIME, tableName, utility1, utility2, zkw1, zkw2 @@ -180,60 +184,68 @@ extends -All MethodsInstance MethodsConcrete Methods +All MethodsStatic MethodsInstance MethodsConcrete Methods Modifier and Type Method and Description +protected boolean +isSerialPeer() + + +static https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">Listhttps://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true; title="class or interface in java.lang">Boolean +parameters() + + void setUp() - + void testAddAndRemoveClusters() Integration test for TestReplicationAdmin, removes and re-add a peer cluster - + void testDeleteTypes() Verify that version and column delete marker types are replicated correctly. - + void testDisableEnable() Test disable/enable replication, trying to insert, make sure nothing's replicated, enable it, the insert should be replicated - + void testLoading() Do a more intense version testSmallBatch, one that will trigger wal rolling and other non-trivial code paths - + void testReplicationInReplay() Test for HBase-15259 WALEdits under replay will also be replicated - + void testSimplePutDelete() Add a row, check it's replicated, delete it, check's gone - + void testSmallBatch() Try a small batch upload using the write buffer, check it's replicated - + void testVerifyListReplicatedTable() Test for HBASE-8663 @@ -245,7 +257,7 @@ extends TestReplicationBase -cleanUp, loadData, params, runSimplePutDeleteTest, runSmallBatchTest, setUpBeforeClass, tearDownAfterClass, w aitForReplication +cleanUp, loadData, runSimplePutDeleteTest, runSmallBatchTest, setUpBase, setUpBeforeClass, tearDownAfterClass, tearDown Base, waitForReplication @@ -274,7 +286,7 @@ extends CLASS_RULE -public static finalHBaseClassTestRule CLASS_RULE +public static finalHBaseClassTestRule CLASS_RULE @@ -283,22 +295,31 @@ extends LOG -private static finalorg.slf4j.Logger LOG +private static finalorg.slf4j.Logger LOG - + PEER_ID -private static finalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String PEER_ID +private static finalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String PEER_ID See Also: Constant Field Values + + + + + +serialPeer +publicboolean serialPeer + + @@ -313,7 +334,7 @@ extends TestReplicationSmallTests -publicTestReplicationSmallTests() +publicTestReplicationSmallTests() @@ -324,13 +345,35 @@ extends + + + + +isSerialPeer +protectedbooleanisSerialPeer() + +Overrides: +isSerialPeerin classTestReplicationBase + + + + + + + + +parameters +public
[21/35] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/04d647a7/devapidocs/src-html/org/apache/hadoop/hbase/util/HBaseFsck.FileLockCallable.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/util/HBaseFsck.FileLockCallable.html b/devapidocs/src-html/org/apache/hadoop/hbase/util/HBaseFsck.FileLockCallable.html index bf1a2cc..89317aa 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/util/HBaseFsck.FileLockCallable.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/util/HBaseFsck.FileLockCallable.html @@ -1813,3325 +1813,3330 @@ 1805 private void loadTableStates() 1806 throws IOException { 1807tableStates = MetaTableAccessor.getTableStates(connection); -1808 } -1809 -1810 /** -1811 * Check if the specified region's table is disabled. -1812 * @param tableName table to check status of -1813 */ -1814 private boolean isTableDisabled(TableName tableName) { -1815return tableStates.containsKey(tableName) -1816 tableStates.get(tableName) -1817 .inStates(TableState.State.DISABLED, TableState.State.DISABLING); -1818 } -1819 -1820 /** -1821 * Scan HDFS for all regions, recording their information into -1822 * regionInfoMap -1823 */ -1824 public void loadHdfsRegionDirs() throws IOException, InterruptedException { -1825Path rootDir = FSUtils.getRootDir(getConf()); -1826FileSystem fs = rootDir.getFileSystem(getConf()); -1827 -1828// list all tables from HDFS -1829ListFileStatus tableDirs = Lists.newArrayList(); -1830 -1831boolean foundVersionFile = fs.exists(new Path(rootDir, HConstants.VERSION_FILE_NAME)); +1808// Add hbase:meta so this tool keeps working. In hbase2, meta is always enabled though it +1809// has no entry in the table states. HBCK doesn't work right w/ hbase2 but just do this in +1810// meantime. +1811 this.tableStates.put(TableName.META_TABLE_NAME, +1812new TableState(TableName.META_TABLE_NAME, TableState.State.ENABLED)); +1813 } +1814 +1815 /** +1816 * Check if the specified region's table is disabled. +1817 * @param tableName table to check status of +1818 */ +1819 private boolean isTableDisabled(TableName tableName) { +1820return tableStates.containsKey(tableName) +1821 tableStates.get(tableName) +1822 .inStates(TableState.State.DISABLED, TableState.State.DISABLING); +1823 } +1824 +1825 /** +1826 * Scan HDFS for all regions, recording their information into +1827 * regionInfoMap +1828 */ +1829 public void loadHdfsRegionDirs() throws IOException, InterruptedException { +1830Path rootDir = FSUtils.getRootDir(getConf()); +1831FileSystem fs = rootDir.getFileSystem(getConf()); 1832 -1833ListPath paths = FSUtils.getTableDirs(fs, rootDir); -1834for (Path path : paths) { -1835 TableName tableName = FSUtils.getTableName(path); -1836 if ((!checkMetaOnly -1837 isTableIncluded(tableName)) || -1838 tableName.equals(TableName.META_TABLE_NAME)) { -1839 tableDirs.add(fs.getFileStatus(path)); -1840 } -1841} -1842 -1843// verify that version file exists -1844if (!foundVersionFile) { -1845 errors.reportError(ERROR_CODE.NO_VERSION_FILE, -1846 "Version file does not exist in root dir " + rootDir); -1847 if (shouldFixVersionFile()) { -1848LOG.info("Trying to create a new " + HConstants.VERSION_FILE_NAME -1849+ " file."); -1850setShouldRerun(); -1851FSUtils.setVersion(fs, rootDir, getConf().getInt( -1852 HConstants.THREAD_WAKE_FREQUENCY, 10 * 1000), getConf().getInt( -1853 HConstants.VERSION_FILE_WRITE_ATTEMPTS, -1854 HConstants.DEFAULT_VERSION_FILE_WRITE_ATTEMPTS)); -1855 } -1856} -1857 -1858// Avoid multithreading at table-level because already multithreaded internally at -1859// region-level. Additionally multithreading at table-level can lead to deadlock -1860// if there are many tables in the cluster. Since there are a limited # of threads -1861// in the executor's thread pool and if we multithread at the table-level by putting -1862// WorkItemHdfsDir callables into the executor, then we will have some threads in the -1863// executor tied up solely in waiting for the tables' region-level calls to complete. -1864// If there are enough tables then there will be no actual threads in the pool left -1865// for the region-level callables to be serviced. -1866for (FileStatus tableDir : tableDirs) { -1867 LOG.debug("Loading region dirs from " +tableDir.getPath()); -1868 WorkItemHdfsDir item = new WorkItemHdfsDir(fs, errors, tableDir); -1869 try { -1870item.call(); -1871 } catch (ExecutionException e) { -1872LOG.warn("Could not completely load table dir " + -1873tableDir.getPath(), e.getCause()); -1874
[21/35] hbase-site git commit: Published site at 9250bf809155ebe93fd6ae8a0485b22c744fdf70.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f17356a7/devapidocs/src-html/org/apache/hadoop/hbase/tmpl/master/MasterStatusTmpl.Intf.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/tmpl/master/MasterStatusTmpl.Intf.html b/devapidocs/src-html/org/apache/hadoop/hbase/tmpl/master/MasterStatusTmpl.Intf.html index 2b27b2b..0a4e021 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/tmpl/master/MasterStatusTmpl.Intf.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/tmpl/master/MasterStatusTmpl.Intf.html @@ -69,15 +69,15 @@ 061 requiredArguments = { 062@org.jamon.annotations.Argument(name = "master", type = "HMaster")}, 063 optionalArguments = { -064@org.jamon.annotations.Argument(name = "frags", type = "MapString,Integer"), -065@org.jamon.annotations.Argument(name = "metaLocation", type = "ServerName"), +064@org.jamon.annotations.Argument(name = "filter", type = "String"), +065@org.jamon.annotations.Argument(name = "deadServers", type = "SetServerName"), 066@org.jamon.annotations.Argument(name = "catalogJanitorEnabled", type = "boolean"), -067@org.jamon.annotations.Argument(name = "deadServers", type = "SetServerName"), -068@org.jamon.annotations.Argument(name = "assignmentManager", type = "AssignmentManager"), -069@org.jamon.annotations.Argument(name = "servers", type = "ListServerName"), -070@org.jamon.annotations.Argument(name = "format", type = "String"), -071@org.jamon.annotations.Argument(name = "filter", type = "String"), -072@org.jamon.annotations.Argument(name = "serverManager", type = "ServerManager")}) +067@org.jamon.annotations.Argument(name = "format", type = "String"), +068@org.jamon.annotations.Argument(name = "serverManager", type = "ServerManager"), +069@org.jamon.annotations.Argument(name = "frags", type = "MapString,Integer"), +070@org.jamon.annotations.Argument(name = "metaLocation", type = "ServerName"), +071@org.jamon.annotations.Argument(name = "assignmentManager", type = "AssignmentManager"), +072@org.jamon.annotations.Argument(name = "servers", type = "ListServerName")}) 073public class MasterStatusTmpl 074 extends org.jamon.AbstractTemplateProxy 075{ @@ -118,40 +118,40 @@ 110 return m_master; 111} 112private HMaster m_master; -113// 21, 1 -114public void setFrags(MapString,Integer frags) +113// 26, 1 +114public void setFilter(String filter) 115{ -116 // 21, 1 -117 m_frags = frags; -118 m_frags__IsNotDefault = true; +116 // 26, 1 +117 m_filter = filter; +118 m_filter__IsNotDefault = true; 119} -120public MapString,Integer getFrags() +120public String getFilter() 121{ -122 return m_frags; +122 return m_filter; 123} -124private MapString,Integer m_frags; -125public boolean getFrags__IsNotDefault() +124private String m_filter; +125public boolean getFilter__IsNotDefault() 126{ -127 return m_frags__IsNotDefault; +127 return m_filter__IsNotDefault; 128} -129private boolean m_frags__IsNotDefault; -130// 22, 1 -131public void setMetaLocation(ServerName metaLocation) +129private boolean m_filter__IsNotDefault; +130// 24, 1 +131public void setDeadServers(SetServerName deadServers) 132{ -133 // 22, 1 -134 m_metaLocation = metaLocation; -135 m_metaLocation__IsNotDefault = true; +133 // 24, 1 +134 m_deadServers = deadServers; +135 m_deadServers__IsNotDefault = true; 136} -137public ServerName getMetaLocation() +137public SetServerName getDeadServers() 138{ -139 return m_metaLocation; +139 return m_deadServers; 140} -141private ServerName m_metaLocation; -142public boolean getMetaLocation__IsNotDefault() +141private SetServerName m_deadServers; +142public boolean getDeadServers__IsNotDefault() 143{ -144 return m_metaLocation__IsNotDefault; +144 return m_deadServers__IsNotDefault; 145} -146private boolean m_metaLocation__IsNotDefault; +146private boolean m_deadServers__IsNotDefault; 147// 25, 1 148public void setCatalogJanitorEnabled(boolean catalogJanitorEnabled) 149{ @@ -169,108 +169,108 @@ 161 return m_catalogJanitorEnabled__IsNotDefault; 162} 163private boolean m_catalogJanitorEnabled__IsNotDefault; -164// 24, 1 -165public void setDeadServers(SetServerName deadServers) +164// 27, 1 +165public void setFormat(String format) 166{ -167 // 24, 1 -168 m_deadServers = deadServers; -169 m_deadServers__IsNotDefault = true; +167 // 27, 1 +168 m_format = format; +169 m_format__IsNotDefault = true; 170} -171public SetServerName getDeadServers() +171public String getFormat() 172{ -173 return m_deadServers;