[04/22] hbase-site git commit: Published site at ac5bb8155b618194fe9cf1131f0e72c99b7b534c.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a7f7dba1/downloads.html -- diff --git a/downloads.html b/downloads.html index 8979bea..b86726d 100644 --- a/downloads.html +++ b/downloads.html @@ -7,7 +7,7 @@ - + Apache HBase – Apache HBase Downloads @@ -366,7 +366,7 @@ under the License. --> https://www.apache.org/";>The Apache Software Foundation. All rights reserved. - Last Published: 2018-06-15 + Last Published: 2018-06-16 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a7f7dba1/export_control.html -- diff --git a/export_control.html b/export_control.html index 452535e..6a16070 100644 --- a/export_control.html +++ b/export_control.html @@ -7,7 +7,7 @@ - + Apache HBase – Export Control @@ -331,7 +331,7 @@ for more details. https://www.apache.org/";>The Apache Software Foundation. All rights reserved. - Last Published: 2018-06-15 + Last Published: 2018-06-16 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a7f7dba1/index.html -- diff --git a/index.html b/index.html index 0195bc9..6b39910 100644 --- a/index.html +++ b/index.html @@ -7,7 +7,7 @@ - + Apache HBase – Apache HBase⢠Home @@ -411,7 +411,7 @@ Apache HBase is an open-source, distributed, versioned, non-relational database https://www.apache.org/";>The Apache Software Foundation. All rights reserved. - Last Published: 2018-06-15 + Last Published: 2018-06-16 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a7f7dba1/integration.html -- diff --git a/integration.html b/integration.html index bc63fe6..cf808bf 100644 --- a/integration.html +++ b/integration.html @@ -7,7 +7,7 @@ - + Apache HBase – CI Management @@ -291,7 +291,7 @@ https://www.apache.org/";>The Apache Software Foundation. All rights reserved. - Last Published: 2018-06-15 + Last Published: 2018-06-16 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a7f7dba1/issue-tracking.html -- diff --git a/issue-tracking.html b/issue-tracking.html index 0ce809d..639cb70 100644 --- a/issue-tracking.html +++ b/issue-tracking.html @@ -7,7 +7,7 @@ - + Apache HBase – Issue Management @@ -288,7 +288,7 @@ https://www.apache.org/";>The Apache Software Foundation. All rights reserved. - Last Published: 2018-06-15 + Last Published: 2018-06-16 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a7f7dba1/license.html -- diff --git a/license.html b/license.html index 0485a2b..37e3114 100644 --- a/license.html +++ b/license.html @@ -7,7 +7,7 @@ - + Apache HBase – Project Licenses @@ -491,7 +491,7 @@ https://www.apache.org/";>The Apache Software Foundation. All rights reserved. - Last Published: 2018-06-15 + Last Published: 2018-06-16 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a7f7dba1/mail-lists.html -- diff --git a/mail-lists.html b/mail-lists.html index 5549fad..75ac1c2 100644 --- a/mail-lists.html +++ b/mail-lists.html @@ -7,7 +7,7 @@ - + Apache HBase – Project Mailing Lists @@ -341,7 +341,7 @@ https://www.apache.org/";>The Apache Software Foundation. All rights reserved. - Last Published: 2018-06-15 + Last Published: 2018-06-16 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a7f7dba1/metrics.html -- diff --git a/metrics.html b/metrics.html index c180dce..457149d 100644 --- a/metrics.html +++ b/m
[04/22] hbase-site git commit: Published site at 46cb5dfa226892fd2580f26ce9ce77225bd7e67c.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8bc57a1a/devapidocs/src-html/org/apache/hadoop/hbase/util/ServerRegionReplicaUtil.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/util/ServerRegionReplicaUtil.html b/devapidocs/src-html/org/apache/hadoop/hbase/util/ServerRegionReplicaUtil.html index 6bdffeb..3dbb30a 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/util/ServerRegionReplicaUtil.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/util/ServerRegionReplicaUtil.html @@ -136,76 +136,89 @@ 128} 129 130// else create a store file link. The link file does not exists on filesystem though. -131HFileLink link = HFileLink.build(conf, regionInfoForFs.getTable(), -132 regionInfoForFs.getEncodedName(), familyName, path.getName()); -133 -134if (StoreFileInfo.isReference(path)) { -135 Reference reference = Reference.read(fs, path); -136 return new StoreFileInfo(conf, fs, link.getFileStatus(fs), reference); -137} -138 -139return new StoreFileInfo(conf, fs, link.getFileStatus(fs), link); -140 } -141 -142 /** -143 * Create replication peer for replicating to region replicas if needed. -144 * @param conf configuration to use -145 * @throws IOException -146 */ -147 public static void setupRegionReplicaReplication(Configuration conf) throws IOException { -148if (!isRegionReplicaReplicationEnabled(conf)) { -149 return; -150} -151Admin admin = ConnectionFactory.createConnection(conf).getAdmin(); -152ReplicationPeerConfig peerConfig = null; -153try { -154 peerConfig = admin.getReplicationPeerConfig(REGION_REPLICA_REPLICATION_PEER); -155} catch (ReplicationPeerNotFoundException e) { -156 LOG.warn("Region replica replication peer id=" + REGION_REPLICA_REPLICATION_PEER -157 + " not exist", e); -158} -159try { -160 if (peerConfig == null) { -161LOG.info("Region replica replication peer id=" + REGION_REPLICA_REPLICATION_PEER -162+ " not exist. Creating..."); -163peerConfig = new ReplicationPeerConfig(); -164 peerConfig.setClusterKey(ZKConfig.getZooKeeperClusterKey(conf)); -165 peerConfig.setReplicationEndpointImpl(RegionReplicaReplicationEndpoint.class.getName()); -166 admin.addReplicationPeer(REGION_REPLICA_REPLICATION_PEER, peerConfig); -167 } -168} finally { -169 admin.close(); -170} -171 } -172 -173 public static boolean isRegionReplicaReplicationEnabled(Configuration conf) { -174return conf.getBoolean(REGION_REPLICA_REPLICATION_CONF_KEY, -175 DEFAULT_REGION_REPLICA_REPLICATION); -176 } -177 -178 public static boolean isRegionReplicaWaitForPrimaryFlushEnabled(Configuration conf) { -179return conf.getBoolean(REGION_REPLICA_WAIT_FOR_PRIMARY_FLUSH_CONF_KEY, -180 DEFAULT_REGION_REPLICA_WAIT_FOR_PRIMARY_FLUSH); -181 } -182 -183 public static boolean isRegionReplicaStoreFileRefreshEnabled(Configuration conf) { -184return conf.getBoolean(REGION_REPLICA_STORE_FILE_REFRESH, -185 DEFAULT_REGION_REPLICA_STORE_FILE_REFRESH); -186 } -187 -188 public static double getRegionReplicaStoreFileRefreshMultiplier(Configuration conf) { -189return conf.getDouble(REGION_REPLICA_STORE_FILE_REFRESH_MEMSTORE_MULTIPLIER, -190 DEFAULT_REGION_REPLICA_STORE_FILE_REFRESH_MEMSTORE_MULTIPLIER); -191 } -192 -193 /** -194 * Return the peer id used for replicating to secondary region replicas -195 */ -196 public static String getReplicationPeerId() { -197return REGION_REPLICA_REPLICATION_PEER; -198 } -199 -200} +131if (HFileLink.isHFileLink(path) || StoreFileInfo.isHFile(path)) { +132 HFileLink link = HFileLink +133 .build(conf, regionInfoForFs.getTable(), regionInfoForFs.getEncodedName(), familyName, +134 path.getName()); +135 return new StoreFileInfo(conf, fs, link.getFileStatus(fs), link); +136} else if (StoreFileInfo.isReference(path)) { +137 Reference reference = Reference.read(fs, path); +138 Path referencePath = StoreFileInfo.getReferredToFile(path); +139 if (HFileLink.isHFileLink(referencePath)) { +140// HFileLink Reference +141HFileLink link = HFileLink.buildFromHFileLinkPattern(conf, referencePath); +142return new StoreFileInfo(conf, fs, link.getFileStatus(fs), reference, link); +143 } else { +144// Reference +145HFileLink link = HFileLink +146.build(conf, regionInfoForFs.getTable(), regionInfoForFs.getEncodedName(), familyName, +147path.getName()); +148return new StoreFileInfo(conf, fs, link.getFileStatus(fs), reference); +149 } +150} else { +151 throw new IOException("path=" + path + " doesn't look like a valid StoreFile"); +152} +153 } +154 +155 /** +156 * Create replication peer for replicating to
[04/22] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/34a201e5/testdevapidocs/src-html/org/apache/hadoop/hbase/IntegrationTestAcidGuarantees.html -- diff --git a/testdevapidocs/src-html/org/apache/hadoop/hbase/IntegrationTestAcidGuarantees.html b/testdevapidocs/src-html/org/apache/hadoop/hbase/IntegrationTestAcidGuarantees.html index ee99b78..59c0ced 100644 --- a/testdevapidocs/src-html/org/apache/hadoop/hbase/IntegrationTestAcidGuarantees.html +++ b/testdevapidocs/src-html/org/apache/hadoop/hbase/IntegrationTestAcidGuarantees.html @@ -25,108 +25,104 @@ 017 */ 018package org.apache.hadoop.hbase; 019 -020import org.apache.hadoop.hbase.shaded.com.google.common.collect.Sets; -021import org.apache.hadoop.conf.Configuration; -022import org.apache.hadoop.hbase.regionserver.CompactingMemStore; -023import org.apache.hadoop.hbase.regionserver.ConstantSizeRegionSplitPolicy; -024import org.apache.hadoop.hbase.regionserver.MemStoreCompactor; -025import org.apache.hadoop.hbase.testclassification.IntegrationTests; -026import org.apache.hadoop.hbase.util.Bytes; -027import org.apache.hadoop.util.ToolRunner; -028import org.junit.Test; -029import org.junit.experimental.categories.Category; -030 -031import java.util.Set; -032 -033/** -034 * This Integration Test verifies acid guarantees across column families by frequently writing -035 * values to rows with multiple column families and concurrently reading entire rows that expect all -036 * column families. -037 * -038 *-039 * Sample usage: -040 *
-041 * hbase org.apache.hadoop.hbase.IntegrationTestAcidGuarantees -Dmillis=1 -DnumWriters=50 -042 * -DnumGetters=2 -DnumScanners=2 -DnumUniqueRows=5 -043 *-044 */ -045@Category(IntegrationTests.class) -046public class IntegrationTestAcidGuarantees extends IntegrationTestBase { -047 private static final int SERVER_COUNT = 1; // number of slaves for the smallest cluster -048 -049 // The unit test version. -050 TestAcidGuarantees tag; -051 -052 @Override -053 public int runTestFromCommandLine() throws Exception { -054Configuration c = getConf(); -055int millis = c.getInt("millis", 5000); -056int numWriters = c.getInt("numWriters", 50); -057int numGetters = c.getInt("numGetters", 2); -058int numScanners = c.getInt("numScanners", 2); -059int numUniqueRows = c.getInt("numUniqueRows", 3); -060boolean useMob = c.getBoolean("useMob",false); -061tag.runTestAtomicity(millis, numWriters, numGetters, numScanners, numUniqueRows, true, useMob); -062return 0; -063 } -064 -065 @Override -066 public void setUpCluster() throws Exception { -067// Set small flush size for minicluster so we exercise reseeking scanners -068util = getTestingUtil(getConf()); -069 util.initializeCluster(SERVER_COUNT); -070conf = getConf(); -071 conf.set(HConstants.HREGION_MEMSTORE_FLUSH_SIZE, String.valueOf(128*1024)); -072// prevent aggressive region split -073 conf.set(HConstants.HBASE_REGION_SPLIT_POLICY_KEY, -074 ConstantSizeRegionSplitPolicy.class.getName()); -075 this.setConf(util.getConfiguration()); -076 -077// replace the HBaseTestingUtility in the unit test with the integration test's -078// IntegrationTestingUtility -079tag = new TestAcidGuarantees(CompactingMemStore.COMPACTING_MEMSTORE_TYPE_DEFAULT); -080tag.setHBaseTestingUtil(util); -081 } -082 -083 @Override -084 public TableName getTablename() { -085return TestAcidGuarantees.TABLE_NAME; -086 } -087 -088 @Override -089 protected SetgetColumnFamilies() { -090return Sets.newHashSet(Bytes.toString(TestAcidGuarantees.FAMILY_A), -091 Bytes.toString(TestAcidGuarantees.FAMILY_B), -092 Bytes.toString(TestAcidGuarantees.FAMILY_C)); -093 } -094 -095 // * Actual integration tests -096 -097 @Test -098 public void testGetAtomicity() throws Exception { -099tag.runTestAtomicity(2, 4, 4, 0, 3); -100 } -101 -102 @Test -103 public void testScanAtomicity() throws Exception { -104tag.runTestAtomicity(2, 3, 0, 2, 3); -105 } -106 -107 @Test -108 public void testMixedAtomicity() throws Exception { -109tag.runTestAtomicity(2, 4, 2, 2, 3); -110 } -111 -112 -113 // Command line hook -114 -115 public static void main(String[] args) throws Exception { -116Configuration conf = HBaseConfiguration.create(); -117 IntegrationTestingUtility.setUseDistributedCluster(conf); -118int ret = ToolRunner.run(conf, new IntegrationTestAcidGuarantees(), args); -119System.exit(ret); -120 } -121} +020import static org.apache.hadoop.hbase.AcidGuaranteesTestTool.FAMILY_A; +021import static org.apache.hadoop.hbase.AcidGuaranteesTestTool.FAMILY_B; +022import static org.apache.hadoop.hbase.AcidGuaranteesTestTool.FAMILY_C; +023import static org.apache.hadoop.hbase.AcidGuaranteesTestTool.TABLE_NAME; +024 +025i
[04/22] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/6f0c8299/devapidocs/src-html/org/apache/hadoop/hbase/zookeeper/MiniZooKeeperCluster.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/zookeeper/MiniZooKeeperCluster.html b/devapidocs/src-html/org/apache/hadoop/hbase/zookeeper/MiniZooKeeperCluster.html index 9d4a644..77768b8 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/zookeeper/MiniZooKeeperCluster.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/zookeeper/MiniZooKeeperCluster.html @@ -233,248 +233,251 @@ 225 } 226 227 ZooKeeperServer server = new ZooKeeperServer(dir, dir, tickTimeToUse); -228 NIOServerCnxnFactory standaloneServerFactory; -229 while (true) { -230try { -231 standaloneServerFactory = new NIOServerCnxnFactory(); -232 standaloneServerFactory.configure( -233new InetSocketAddress(currentClientPort), -234 configuration.getInt(HConstants.ZOOKEEPER_MAX_CLIENT_CNXNS, 1000)); -235} catch (BindException e) { -236 LOG.debug("Failed binding ZK Server to client port: " + -237 currentClientPort, e); -238 // We're told to use some port but it's occupied, fail -239 if (hasValidClientPortInList(i)) { -240return -1; -241 } -242 // This port is already in use, try to use another. -243 tentativePort = selectClientPort(tentativePort); -244 currentClientPort = tentativePort; -245 continue; -246} -247break; -248 } -249 -250 // Start up this ZK server -251 standaloneServerFactory.startup(server); -252 // Runs a 'stat' against the servers. -253 if (!waitForServerUp(currentClientPort, connectionTimeout)) { -254throw new IOException("Waiting for startup of standalone server"); -255 } -256 -257 // We have selected a port as a client port. Update clientPortList if necessary. -258 if (clientPortList.size() <= i) { // it is not in the list, add the port -259 clientPortList.add(currentClientPort); -260 } -261 else if (clientPortList.get(i) <= 0) { // the list has invalid port, update with valid port -262clientPortList.remove(i); -263clientPortList.add(i, currentClientPort); -264 } -265 -266 standaloneServerFactoryList.add(standaloneServerFactory); -267 zooKeeperServers.add(server); -268} -269 -270// set the first one to be active ZK; Others are backups -271activeZKServerIndex = 0; -272started = true; -273int clientPort = clientPortList.get(activeZKServerIndex); -274LOG.info("Started MiniZooKeeperCluster and ran successful 'stat' " + -275"on client port=" + clientPort); -276return clientPort; -277 } -278 -279 private void createDir(File dir) throws IOException { -280try { -281 if (!dir.exists()) { -282dir.mkdirs(); -283 } -284} catch (SecurityException e) { -285 throw new IOException("creating dir: " + dir, e); -286} -287 } -288 -289 /** -290 * @throws IOException -291 */ -292 public void shutdown() throws IOException { -293// shut down all the zk servers -294for (int i = 0; i < standaloneServerFactoryList.size(); i++) { -295 NIOServerCnxnFactory standaloneServerFactory = -296 standaloneServerFactoryList.get(i); -297 int clientPort = clientPortList.get(i); -298 -299 standaloneServerFactory.shutdown(); -300 if (!waitForServerDown(clientPort, connectionTimeout)) { -301throw new IOException("Waiting for shutdown of standalone server"); -302 } -303} -304 standaloneServerFactoryList.clear(); -305 -306for (ZooKeeperServer zkServer: zooKeeperServers) { -307 //explicitly close ZKDatabase since ZookeeperServer does not close them -308 zkServer.getZKDatabase().close(); -309} -310zooKeeperServers.clear(); -311 -312// clear everything -313if (started) { -314 started = false; -315 activeZKServerIndex = 0; -316 clientPortList.clear(); -317 LOG.info("Shutdown MiniZK cluster with all ZK servers"); -318} -319 } -320 -321 /**@return clientPort return clientPort if there is another ZK backup can run -322 * when killing the current active; return -1, if there is no backups. -323 * @throws IOException -324 * @throws InterruptedException -325 */ -326 public int killCurrentActiveZooKeeperServer() throws IOException, -327 InterruptedException { -328if (!started || activeZKServerIndex < 0) { -329 return -1; -330} -331 -332// Shutdown the current active one -333NIOServerCnxnFactory standaloneServerFactory = -334 standaloneServerFactoryList.get(activeZKServerIndex); -335int clientPort = clientPortList.get(activeZKServerIndex); -336 -337
[04/22] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/b9830530/hbase-archetypes/hbase-archetype-builder/dependency-management.html -- diff --git a/hbase-archetypes/hbase-archetype-builder/dependency-management.html b/hbase-archetypes/hbase-archetype-builder/dependency-management.html index 58c52c1..2595788 100644 --- a/hbase-archetypes/hbase-archetype-builder/dependency-management.html +++ b/hbase-archetypes/hbase-archetype-builder/dependency-management.html @@ -1,5 +1,5 @@ http://www.w3.org/TR/xhtml1/DTD/xhtml1-transitional.dtd";> - + http://www.w3.org/1999/xhtml"; xml:lang="en" lang="en"> @@ -10,7 +10,7 @@ @import url("./css/site.css"); - + @@ -27,7 +27,7 @@ -Last Published: 2017-06-16 +Last Published: 2017-06-17 | Version: 3.0.0-SNAPSHOT Apache HBase - Archetype builder http://git-wip-us.apache.org/repos/asf/hbase-site/blob/b9830530/hbase-archetypes/hbase-archetype-builder/index.html -- diff --git a/hbase-archetypes/hbase-archetype-builder/index.html b/hbase-archetypes/hbase-archetype-builder/index.html index 0f1641f..7cc3a28 100644 --- a/hbase-archetypes/hbase-archetype-builder/index.html +++ b/hbase-archetypes/hbase-archetype-builder/index.html @@ -1,5 +1,5 @@ http://www.w3.org/TR/xhtml1/DTD/xhtml1-transitional.dtd";> - + http://www.w3.org/1999/xhtml"; xml:lang="en" lang="en"> @@ -10,7 +10,7 @@ @import url("./css/site.css"); - + @@ -27,7 +27,7 @@ -Last Published: 2017-06-16 +Last Published: 2017-06-17 | Version: 3.0.0-SNAPSHOT Apache HBase - Archetype builder http://git-wip-us.apache.org/repos/asf/hbase-site/blob/b9830530/hbase-archetypes/hbase-archetype-builder/integration.html -- diff --git a/hbase-archetypes/hbase-archetype-builder/integration.html b/hbase-archetypes/hbase-archetype-builder/integration.html index c802ba4..58b1470 100644 --- a/hbase-archetypes/hbase-archetype-builder/integration.html +++ b/hbase-archetypes/hbase-archetype-builder/integration.html @@ -1,5 +1,5 @@ http://www.w3.org/TR/xhtml1/DTD/xhtml1-transitional.dtd";> - + http://www.w3.org/1999/xhtml"; xml:lang="en" lang="en"> @@ -10,7 +10,7 @@ @import url("./css/site.css"); - + @@ -27,7 +27,7 @@ -Last Published: 2017-06-16 +Last Published: 2017-06-17 | Version: 3.0.0-SNAPSHOT Apache HBase - Archetype builder http://git-wip-us.apache.org/repos/asf/hbase-site/blob/b9830530/hbase-archetypes/hbase-archetype-builder/issue-tracking.html -- diff --git a/hbase-archetypes/hbase-archetype-builder/issue-tracking.html b/hbase-archetypes/hbase-archetype-builder/issue-tracking.html index 914435f..47f5b59 100644 --- a/hbase-archetypes/hbase-archetype-builder/issue-tracking.html +++ b/hbase-archetypes/hbase-archetype-builder/issue-tracking.html @@ -1,5 +1,5 @@ http://www.w3.org/TR/xhtml1/DTD/xhtml1-transitional.dtd";> - + http://www.w3.org/1999/xhtml"; xml:lang="en" lang="en"> @@ -10,7 +10,7 @@ @import url("./css/site.css"); - + @@ -27,7 +27,7 @@ -Last Published: 2017-06-16 +Last Published: 2017-06-17 | Version: 3.0.0-SNAPSHOT Apache HBase - Archetype builder http://git-wip-us.apache.org/repos/asf/hbase-site/blob/b9830530/hbase-archetypes/hbase-archetype-builder/license.html -- diff --git a/hbase-archetypes/hbase-archetype-builder/license.html b/hbase-archetypes/hbase-archetype-builder/license.html index 383c399..34c183f 100644 --- a/hbase-archetypes/hbase-archetype-builder/license.html +++ b/hbase-archetypes/hbase-archetype-builder/license.html @@ -1,5 +1,5 @@ http://www.w3.org/TR/xhtml1/DTD/xhtml1-transitional.dtd";> - + http://www.w3.org/1999/xhtml"; xml:lang="en" lang="en"> @@ -10,7 +10,7 @@ @import url("./css/site.css"); - + @@ -27,7 +27,7 @@ -Last Published: 2017-06-16 +Last Published: 2017-06-17 | Version:
[04/22] hbase-site git commit: Published site at da5fb27eabed4a4b4d251be973ee945fb52895bf.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7b1830cf/hbase-archetypes/hbase-shaded-client-project/checkstyle.html -- diff --git a/hbase-archetypes/hbase-shaded-client-project/checkstyle.html b/hbase-archetypes/hbase-shaded-client-project/checkstyle.html index cdc9a65..6fe6ac6 100644 --- a/hbase-archetypes/hbase-shaded-client-project/checkstyle.html +++ b/hbase-archetypes/hbase-shaded-client-project/checkstyle.html @@ -1,5 +1,5 @@ http://www.w3.org/TR/xhtml1/DTD/xhtml1-transitional.dtd";> - + http://www.w3.org/1999/xhtml"; xml:lang="en" lang="en"> @@ -10,7 +10,7 @@ @import url("./css/site.css"); - + @@ -27,7 +27,7 @@ -Last Published: 2017-04-12 +Last Published: 2017-04-13 | Version: 2.0.0-SNAPSHOT Apache HBase - Exemplar for hbase-shaded-client archetype http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7b1830cf/hbase-archetypes/hbase-shaded-client-project/dependencies.html -- diff --git a/hbase-archetypes/hbase-shaded-client-project/dependencies.html b/hbase-archetypes/hbase-shaded-client-project/dependencies.html index 3f1c052..2a519d2 100644 --- a/hbase-archetypes/hbase-shaded-client-project/dependencies.html +++ b/hbase-archetypes/hbase-shaded-client-project/dependencies.html @@ -1,5 +1,5 @@ http://www.w3.org/TR/xhtml1/DTD/xhtml1-transitional.dtd";> - + http://www.w3.org/1999/xhtml"; xml:lang="en" lang="en"> @@ -10,7 +10,7 @@ @import url("./css/site.css"); - + @@ -27,7 +27,7 @@ -Last Published: 2017-04-12 +Last Published: 2017-04-13 | Version: 2.0.0-SNAPSHOT Apache HBase - Exemplar for hbase-shaded-client archetype http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7b1830cf/hbase-archetypes/hbase-shaded-client-project/dependency-convergence.html -- diff --git a/hbase-archetypes/hbase-shaded-client-project/dependency-convergence.html b/hbase-archetypes/hbase-shaded-client-project/dependency-convergence.html index e626fe6..9b24a4d 100644 --- a/hbase-archetypes/hbase-shaded-client-project/dependency-convergence.html +++ b/hbase-archetypes/hbase-shaded-client-project/dependency-convergence.html @@ -1,5 +1,5 @@ http://www.w3.org/TR/xhtml1/DTD/xhtml1-transitional.dtd";> - + http://www.w3.org/1999/xhtml"; xml:lang="en" lang="en"> @@ -10,7 +10,7 @@ @import url("./css/site.css"); - + @@ -27,7 +27,7 @@ -Last Published: 2017-04-12 +Last Published: 2017-04-13 | Version: 2.0.0-SNAPSHOT Apache HBase - Exemplar for hbase-shaded-client archetype http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7b1830cf/hbase-archetypes/hbase-shaded-client-project/dependency-info.html -- diff --git a/hbase-archetypes/hbase-shaded-client-project/dependency-info.html b/hbase-archetypes/hbase-shaded-client-project/dependency-info.html index b8b1760..76b6002 100644 --- a/hbase-archetypes/hbase-shaded-client-project/dependency-info.html +++ b/hbase-archetypes/hbase-shaded-client-project/dependency-info.html @@ -1,5 +1,5 @@ http://www.w3.org/TR/xhtml1/DTD/xhtml1-transitional.dtd";> - + http://www.w3.org/1999/xhtml"; xml:lang="en" lang="en"> @@ -10,7 +10,7 @@ @import url("./css/site.css"); - + @@ -27,7 +27,7 @@ -Last Published: 2017-04-12 +Last Published: 2017-04-13 | Version: 2.0.0-SNAPSHOT Apache HBase - Exemplar for hbase-shaded-client archetype http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7b1830cf/hbase-archetypes/hbase-shaded-client-project/dependency-management.html -- diff --git a/hbase-archetypes/hbase-shaded-client-project/dependency-management.html b/hbase-archetypes/hbase-shaded-client-project/dependency-management.html index 6635ff0..c1373d3 100644 --- a/hbase-archetypes/hbase-shaded-client-project/dependency-management.html +++ b/hbase-archetypes/hbase-shaded-client-project/dependency-management.html @@ -1,5 +1,5 @@ http://www.w3.org/TR/xhtml1/DTD/xhtml1-transitional.dtd";> - + http://www.w3.org/1999/xhtml"; x