Repository: hbase Updated Branches: refs/heads/hbase-12439 a395922ad -> a11091c49
http://git-wip-us.apache.org/repos/asf/hbase/blob/c06a976a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestSnapshotFromClient.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestSnapshotFromClient.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestSnapshotFromClient.java index 1b61147..8317376 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestSnapshotFromClient.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestSnapshotFromClient.java @@ -36,7 +36,7 @@ import org.apache.hadoop.hbase.HBaseTestingUtility; import org.apache.hadoop.hbase.HConstants; import org.apache.hadoop.hbase.TableNotFoundException; import org.apache.hadoop.hbase.master.snapshot.SnapshotManager; -import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription; +import org.apache.hadoop.hbase.protobuf.ProtobufUtil; import org.apache.hadoop.hbase.regionserver.ConstantSizeRegionSplitPolicy; import org.apache.hadoop.hbase.snapshot.SnapshotCreationException; import org.apache.hadoop.hbase.snapshot.SnapshotDoesNotExistException; @@ -212,18 +212,13 @@ public class TestSnapshotFromClient { final String SNAPSHOT_NAME = "offlineTableSnapshot"; byte[] snapshot = Bytes.toBytes(SNAPSHOT_NAME); - SnapshotDescription desc = SnapshotDescription.newBuilder() - .setType(SnapshotDescription.Type.DISABLED) - .setTable(STRING_TABLE_NAME) - .setName(SNAPSHOT_NAME) - .setVersion(SnapshotManifestV1.DESCRIPTOR_VERSION) - .build(); - admin.snapshot(desc); + admin.snapshot(new SnapshotDescription(SNAPSHOT_NAME, STRING_TABLE_NAME, + SnapshotType.DISABLED, null, -1, SnapshotManifestV1.DESCRIPTOR_VERSION)); LOG.debug("Snapshot completed."); // make sure we have the snapshot - List<SnapshotDescription> snapshots = SnapshotTestingUtils.assertOneSnapshotThatMatches(admin, - snapshot, TABLE_NAME); + List<SnapshotDescription> snapshots = + SnapshotTestingUtils.assertOneSnapshotThatMatches(admin, snapshot, TABLE_NAME); // make sure its a valid snapshot FileSystem fs = UTIL.getHBaseCluster().getMaster().getMasterFileSystem().getFileSystem(); @@ -231,9 +226,9 @@ public class TestSnapshotFromClient { LOG.debug("FS state after snapshot:"); FSUtils.logFileSystemState(UTIL.getTestFileSystem(), FSUtils.getRootDir(UTIL.getConfiguration()), LOG); - - SnapshotTestingUtils.confirmSnapshotValid(snapshots.get(0), TABLE_NAME, TEST_FAM, rootDir, - admin, fs); + SnapshotTestingUtils.confirmSnapshotValid( + ProtobufUtil.createHBaseProtosSnapshotDesc(snapshots.get(0)), TABLE_NAME, TEST_FAM, + rootDir, admin, fs); admin.deleteSnapshot(snapshot); snapshots = admin.listSnapshots(); @@ -292,8 +287,8 @@ public class TestSnapshotFromClient { LOG.debug("Snapshot completed."); // make sure we have the snapshot - List<SnapshotDescription> snapshots = SnapshotTestingUtils.assertOneSnapshotThatMatches(admin, - snapshot, TABLE_NAME); + List<SnapshotDescription> snapshots = + SnapshotTestingUtils.assertOneSnapshotThatMatches(admin, snapshot, TABLE_NAME); // make sure its a valid snapshot FileSystem fs = UTIL.getHBaseCluster().getMaster().getMasterFileSystem().getFileSystem(); @@ -304,8 +299,9 @@ public class TestSnapshotFromClient { List<byte[]> emptyCfs = Lists.newArrayList(TEST_FAM); // no file in the region List<byte[]> nonEmptyCfs = Lists.newArrayList(); - SnapshotTestingUtils.confirmSnapshotValid(snapshots.get(0), TABLE_NAME, nonEmptyCfs, emptyCfs, - rootDir, admin, fs); + SnapshotTestingUtils.confirmSnapshotValid( + ProtobufUtil.createHBaseProtosSnapshotDesc(snapshots.get(0)), TABLE_NAME, nonEmptyCfs, + emptyCfs, rootDir, admin, fs); admin.deleteSnapshot(snapshot); snapshots = admin.listSnapshots(); @@ -375,7 +371,8 @@ public class TestSnapshotFromClient { admin.snapshot(Bytes.toBytes(table2Snapshot1), TABLE_NAME); LOG.debug(table2Snapshot1 + " completed."); - List<SnapshotDescription> listTableSnapshots = admin.listTableSnapshots("test.*", "Table1.*"); + List<SnapshotDescription> listTableSnapshots = + admin.listTableSnapshots("test.*", "Table1.*"); List<String> listTableSnapshotNames = new ArrayList<String>(); assertEquals(2, listTableSnapshots.size()); for (SnapshotDescription s : listTableSnapshots) { http://git-wip-us.apache.org/repos/asf/hbase/blob/c06a976a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestSplitOrMergeStatus.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestSplitOrMergeStatus.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestSplitOrMergeStatus.java index e1ce63b..c5ca0b5 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestSplitOrMergeStatus.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestSplitOrMergeStatus.java @@ -78,14 +78,14 @@ public class TestSplitOrMergeStatus { Admin admin = TEST_UTIL.getAdmin(); initSwitchStatus(admin); boolean[] results = admin.setSplitOrMergeEnabled(false, false, - true, Admin.MasterSwitchType.SPLIT); + true, MasterSwitchType.SPLIT); assertEquals(results.length, 1); assertTrue(results[0]); admin.split(t.getName()); int count = waitOnSplitOrMerge(t).size(); assertTrue(orignalCount == count); - results = admin.setSplitOrMergeEnabled(true, false, true, Admin.MasterSwitchType.SPLIT); + results = admin.setSplitOrMergeEnabled(true, false, true, MasterSwitchType.SPLIT); assertEquals(results.length, 1); assertFalse(results[0]); admin.split(t.getName()); @@ -111,7 +111,7 @@ public class TestSplitOrMergeStatus { waitForMergable(admin, name); int orignalCount = locator.getAllRegionLocations().size(); boolean[] results = admin.setSplitOrMergeEnabled(false, false, - true, Admin.MasterSwitchType.MERGE); + true, MasterSwitchType.MERGE); assertEquals(results.length, 1); assertTrue(results[0]); List<HRegionInfo> regions = admin.getTableRegions(t.getName()); @@ -122,7 +122,7 @@ public class TestSplitOrMergeStatus { assertTrue(orignalCount == count); waitForMergable(admin, name); - results = admin.setSplitOrMergeEnabled(true, false, true, Admin.MasterSwitchType.MERGE); + results = admin.setSplitOrMergeEnabled(true, false, true, MasterSwitchType.MERGE); assertEquals(results.length, 1); assertFalse(results[0]); admin.mergeRegions(regions.get(0).getEncodedNameAsBytes(), @@ -136,12 +136,12 @@ public class TestSplitOrMergeStatus { public void testMultiSwitches() throws IOException { Admin admin = TEST_UTIL.getAdmin(); boolean[] switches = admin.setSplitOrMergeEnabled(false, false, true, - Admin.MasterSwitchType.SPLIT, Admin.MasterSwitchType.MERGE); + MasterSwitchType.SPLIT, MasterSwitchType.MERGE); for (boolean s : switches){ assertTrue(s); } - assertFalse(admin.isSplitOrMergeEnabled(Admin.MasterSwitchType.SPLIT)); - assertFalse(admin.isSplitOrMergeEnabled(Admin.MasterSwitchType.MERGE)); + assertFalse(admin.isSplitOrMergeEnabled(MasterSwitchType.SPLIT)); + assertFalse(admin.isSplitOrMergeEnabled(MasterSwitchType.MERGE)); admin.close(); } @@ -149,10 +149,10 @@ public class TestSplitOrMergeStatus { public void testSwitchLock() throws IOException { Admin admin = TEST_UTIL.getAdmin(); admin.setSplitOrMergeEnabled(false, false, false, - Admin.MasterSwitchType.SPLIT, Admin.MasterSwitchType.MERGE); + MasterSwitchType.SPLIT, MasterSwitchType.MERGE); try { admin.setSplitOrMergeEnabled(false, false, true, - Admin.MasterSwitchType.SPLIT, Admin.MasterSwitchType.MERGE); + MasterSwitchType.SPLIT, MasterSwitchType.MERGE); fail(); } catch (IOException e) { LOG.info("", e); @@ -160,7 +160,7 @@ public class TestSplitOrMergeStatus { admin.releaseSplitOrMergeLockAndRollback(); try { admin.setSplitOrMergeEnabled(true, false, true, - Admin.MasterSwitchType.SPLIT, Admin.MasterSwitchType.MERGE); + MasterSwitchType.SPLIT, MasterSwitchType.MERGE); } catch (IOException e) { fail(); } @@ -168,14 +168,14 @@ public class TestSplitOrMergeStatus { } private void initSwitchStatus(Admin admin) throws IOException { - if (!admin.isSplitOrMergeEnabled(Admin.MasterSwitchType.SPLIT)) { - admin.setSplitOrMergeEnabled(true, false, true, Admin.MasterSwitchType.SPLIT); + if (!admin.isSplitOrMergeEnabled(MasterSwitchType.SPLIT)) { + admin.setSplitOrMergeEnabled(true, false, true, MasterSwitchType.SPLIT); } - if (!admin.isSplitOrMergeEnabled(Admin.MasterSwitchType.MERGE)) { - admin.setSplitOrMergeEnabled(true, false, true, Admin.MasterSwitchType.MERGE); + if (!admin.isSplitOrMergeEnabled(MasterSwitchType.MERGE)) { + admin.setSplitOrMergeEnabled(true, false, true, MasterSwitchType.MERGE); } - assertTrue(admin.isSplitOrMergeEnabled(Admin.MasterSwitchType.SPLIT)); - assertTrue(admin.isSplitOrMergeEnabled(Admin.MasterSwitchType.MERGE)); + assertTrue(admin.isSplitOrMergeEnabled(MasterSwitchType.SPLIT)); + assertTrue(admin.isSplitOrMergeEnabled(MasterSwitchType.MERGE)); } private void waitForMergable(Admin admin, TableName t) throws InterruptedException, IOException { http://git-wip-us.apache.org/repos/asf/hbase/blob/c06a976a/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestMasterObserver.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestMasterObserver.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestMasterObserver.java index 65fc0c3..5257cee 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestMasterObserver.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestMasterObserver.java @@ -48,6 +48,7 @@ import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.client.Admin; import org.apache.hadoop.hbase.client.Connection; import org.apache.hadoop.hbase.client.ConnectionFactory; +import org.apache.hadoop.hbase.client.MasterSwitchType; import org.apache.hadoop.hbase.client.RegionLocator; import org.apache.hadoop.hbase.client.Table; import org.apache.hadoop.hbase.master.AssignmentManager; @@ -348,13 +349,13 @@ public class TestMasterObserver { @Override public boolean preSetSplitOrMergeEnabled(final ObserverContext<MasterCoprocessorEnvironment> ctx, - final boolean newValue, final Admin.MasterSwitchType switchType) throws IOException { + final boolean newValue, final MasterSwitchType switchType) throws IOException { return false; } @Override public void postSetSplitOrMergeEnabled(final ObserverContext<MasterCoprocessorEnvironment> ctx, - final boolean newValue, final Admin.MasterSwitchType switchType) throws IOException { + final boolean newValue, final MasterSwitchType switchType) throws IOException { } @Override http://git-wip-us.apache.org/repos/asf/hbase/blob/c06a976a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestDistributedLogSplitting.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestDistributedLogSplitting.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestDistributedLogSplitting.java index cff8db0..4f2385f 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestDistributedLogSplitting.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestDistributedLogSplitting.java @@ -69,6 +69,7 @@ import org.apache.hadoop.hbase.SplitLogCounters; import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.Waiter; import org.apache.hadoop.hbase.client.ClusterConnection; +import org.apache.hadoop.hbase.client.CompactionState; import org.apache.hadoop.hbase.client.ConnectionUtils; import org.apache.hadoop.hbase.client.Delete; import org.apache.hadoop.hbase.client.Get; @@ -87,7 +88,6 @@ import org.apache.hadoop.hbase.exceptions.RegionInRecoveryException; import org.apache.hadoop.hbase.ipc.ServerNotRunningYetException; import org.apache.hadoop.hbase.master.SplitLogManager.TaskBatch; import org.apache.hadoop.hbase.protobuf.ProtobufUtil; -import org.apache.hadoop.hbase.protobuf.generated.AdminProtos.GetRegionInfoResponse.CompactionState; import org.apache.hadoop.hbase.regionserver.HRegion; import org.apache.hadoop.hbase.regionserver.HRegionServer; import org.apache.hadoop.hbase.regionserver.Region; @@ -1421,7 +1421,8 @@ public class TestDistributedLogSplitting { TEST_UTIL.waitFor(30000, 200, new Waiter.Predicate<Exception>() { @Override public boolean evaluate() throws Exception { - return (TEST_UTIL.getHBaseAdmin().getCompactionState(tableName) == CompactionState.NONE); + return (TEST_UTIL.getHBaseAdmin() + .getCompactionState(tableName) == CompactionState.NONE); } }); http://git-wip-us.apache.org/repos/asf/hbase/blob/c06a976a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestWarmupRegion.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestWarmupRegion.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestWarmupRegion.java index 53ee92b..cb7337e 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestWarmupRegion.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestWarmupRegion.java @@ -31,7 +31,8 @@ import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.Waiter; import org.apache.hadoop.hbase.client.Put; import org.apache.hadoop.hbase.client.Table; -import org.apache.hadoop.hbase.protobuf.generated.AdminProtos; +import org.apache.hadoop.hbase.client.Admin; +import org.apache.hadoop.hbase.client.CompactionState; import org.apache.hadoop.hbase.regionserver.HRegion; import org.apache.hadoop.hbase.regionserver.HRegionServer; import org.apache.hadoop.hbase.testclassification.LargeTests; @@ -106,7 +107,7 @@ public class TestWarmupRegion { @Override public boolean evaluate() throws IOException { return TEST_UTIL.getHBaseAdmin().getCompactionState(TABLENAME) == - AdminProtos.GetRegionInfoResponse.CompactionState.NONE; + CompactionState.NONE; } }); http://git-wip-us.apache.org/repos/asf/hbase/blob/c06a976a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/procedure/TestCloneSnapshotProcedure.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/procedure/TestCloneSnapshotProcedure.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/procedure/TestCloneSnapshotProcedure.java index aeafbf8..96f78f7 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/procedure/TestCloneSnapshotProcedure.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/procedure/TestCloneSnapshotProcedure.java @@ -33,7 +33,9 @@ import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.client.Admin; import org.apache.hadoop.hbase.procedure2.ProcedureExecutor; import org.apache.hadoop.hbase.procedure2.ProcedureTestingUtility; -import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription; +import org.apache.hadoop.hbase.protobuf.ProtobufUtil; +import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos; +import org.apache.hadoop.hbase.client.SnapshotDescription; import org.apache.hadoop.hbase.protobuf.generated.MasterProcedureProtos.CloneSnapshotState; import org.apache.hadoop.hbase.snapshot.SnapshotTestingUtils; import org.apache.hadoop.hbase.testclassification.MasterTests; @@ -59,7 +61,7 @@ public class TestCloneSnapshotProcedure { private static long nonceGroup = HConstants.NO_NONCE; private static long nonce = HConstants.NO_NONCE; - private static SnapshotDescription snapshot = null; + private static HBaseProtos.SnapshotDescription snapshot = null; private static void setupConf(Configuration conf) { conf.setInt(MasterProcedureConstants.MASTER_PROCEDURE_THREADS, 1); @@ -99,7 +101,7 @@ public class TestCloneSnapshotProcedure { assertTrue("expected executor to be running", procExec.isRunning()); } - private SnapshotDescription getSnapshot() throws Exception { + private HBaseProtos.SnapshotDescription getSnapshot() throws Exception { if (snapshot == null) { final TableName snapshotTableName = TableName.valueOf("testCloneSnapshot"); long tid = System.currentTimeMillis(); @@ -116,7 +118,7 @@ public class TestCloneSnapshotProcedure { admin.enableTable(snapshotTableName); List<SnapshotDescription> snapshotList = admin.listSnapshots(); - snapshot = snapshotList.get(0); + snapshot = ProtobufUtil.createHBaseProtosSnapshotDesc(snapshotList.get(0)); } return snapshot; } http://git-wip-us.apache.org/repos/asf/hbase/blob/c06a976a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/procedure/TestRestoreSnapshotProcedure.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/procedure/TestRestoreSnapshotProcedure.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/procedure/TestRestoreSnapshotProcedure.java index 44d6988..733dcb9 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/procedure/TestRestoreSnapshotProcedure.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/procedure/TestRestoreSnapshotProcedure.java @@ -35,7 +35,9 @@ import org.apache.hadoop.hbase.TableNotFoundException; import org.apache.hadoop.hbase.client.Admin; import org.apache.hadoop.hbase.procedure2.ProcedureExecutor; import org.apache.hadoop.hbase.procedure2.ProcedureTestingUtility; -import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription; +import org.apache.hadoop.hbase.client.SnapshotDescription; +import org.apache.hadoop.hbase.protobuf.ProtobufUtil; +import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos; import org.apache.hadoop.hbase.protobuf.generated.MasterProcedureProtos.RestoreSnapshotState; import org.apache.hadoop.hbase.snapshot.SnapshotTestingUtils; import org.apache.hadoop.hbase.testclassification.MasterTests; @@ -73,7 +75,7 @@ public class TestRestoreSnapshotProcedure { private static long nonceGroup = HConstants.NO_NONCE; private static long nonce = HConstants.NO_NONCE; - private SnapshotDescription snapshot = null; + private HBaseProtos.SnapshotDescription snapshot = null; private HTableDescriptor snapshotHTD = null; private static void setupConf(Configuration conf) { @@ -141,7 +143,7 @@ public class TestRestoreSnapshotProcedure { admin.snapshot(snapshotName, snapshotTableName); List<SnapshotDescription> snapshotList = admin.listSnapshots(); - snapshot = snapshotList.get(0); + snapshot = ProtobufUtil.createHBaseProtosSnapshotDesc(snapshotList.get(0)); // modify the table HColumnDescriptor columnFamilyDescriptor3 = new HColumnDescriptor(CF3); http://git-wip-us.apache.org/repos/asf/hbase/blob/c06a976a/hbase-server/src/test/java/org/apache/hadoop/hbase/mob/compactions/TestMobCompactor.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/mob/compactions/TestMobCompactor.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/mob/compactions/TestMobCompactor.java index 9922aff..c0ad2dd 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/mob/compactions/TestMobCompactor.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/mob/compactions/TestMobCompactor.java @@ -51,6 +51,8 @@ import org.apache.hadoop.hbase.NamespaceDescriptor; import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.client.Admin; import org.apache.hadoop.hbase.client.BufferedMutator; +import org.apache.hadoop.hbase.client.CompactType; +import org.apache.hadoop.hbase.client.CompactionState; import org.apache.hadoop.hbase.client.Connection; import org.apache.hadoop.hbase.client.ConnectionFactory; import org.apache.hadoop.hbase.client.Delete; @@ -68,7 +70,6 @@ import org.apache.hadoop.hbase.io.hfile.CacheConfig; import org.apache.hadoop.hbase.io.hfile.HFile; import org.apache.hadoop.hbase.mob.MobConstants; import org.apache.hadoop.hbase.mob.MobUtils; -import org.apache.hadoop.hbase.protobuf.generated.AdminProtos.GetRegionInfoResponse.CompactionState; import org.apache.hadoop.hbase.regionserver.BloomType; import org.apache.hadoop.hbase.regionserver.HRegion; import org.apache.hadoop.hbase.regionserver.StoreFile; @@ -356,7 +357,7 @@ public class TestMobCompactor { countFiles(tableName, false, family2)); // do the major mob compaction, it will force all files to compaction - admin.majorCompact(tableName, hcd1.getName(), Admin.CompactType.MOB); + admin.majorCompact(tableName, hcd1.getName(), CompactType.MOB); waitUntilMobCompactionFinished(tableName); assertEquals("After compaction: mob rows count", regionNum * (rowNumPerRegion - delRowNum), @@ -399,7 +400,7 @@ public class TestMobCompactor { Cell cell = result.getColumnLatestCell(hcd1.getName(), Bytes.toBytes(qf1)); assertEquals("Before compaction: mob value of k0", newValue0, Bytes.toString(CellUtil.cloneValue(cell))); - admin.majorCompact(tableName, hcd1.getName(), Admin.CompactType.MOB); + admin.majorCompact(tableName, hcd1.getName(), CompactType.MOB); waitUntilMobCompactionFinished(tableName); // read the latest cell of key0, the cell seqId in bulk loaded file is not reset in the // scanner. The cell that has "new" value is still visible. @@ -449,7 +450,7 @@ public class TestMobCompactor { loadData(admin, bufMut, tableName, new Put[] { put1 }); // now two mob files admin.majorCompact(tableName); waitUntilCompactionFinished(tableName); - admin.majorCompact(tableName, hcd1.getName(), Admin.CompactType.MOB); + admin.majorCompact(tableName, hcd1.getName(), CompactType.MOB); waitUntilMobCompactionFinished(tableName); // read the latest cell of key1. Get get = new Get(key1); @@ -475,12 +476,12 @@ public class TestMobCompactor { private void waitUntilMobCompactionFinished(TableName tableName) throws IOException, InterruptedException { long finished = EnvironmentEdgeManager.currentTime() + 60000; - CompactionState state = admin.getCompactionState(tableName, Admin.CompactType.MOB); + CompactionState state = admin.getCompactionState(tableName, CompactType.MOB); while (EnvironmentEdgeManager.currentTime() < finished) { if (state == CompactionState.NONE) { break; } - state = admin.getCompactionState(tableName, Admin.CompactType.MOB); + state = admin.getCompactionState(tableName, CompactType.MOB); Thread.sleep(10); } assertEquals(CompactionState.NONE, state); http://git-wip-us.apache.org/repos/asf/hbase/blob/c06a976a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestCompactionState.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestCompactionState.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestCompactionState.java index 4715d53..d6302b3 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestCompactionState.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestCompactionState.java @@ -31,9 +31,9 @@ import org.apache.commons.logging.LogFactory; import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.HBaseTestingUtility; import org.apache.hadoop.hbase.client.Admin; +import org.apache.hadoop.hbase.client.CompactionState; import org.apache.hadoop.hbase.client.Put; import org.apache.hadoop.hbase.client.Table; -import org.apache.hadoop.hbase.protobuf.generated.AdminProtos.GetRegionInfoResponse.CompactionState; import org.apache.hadoop.hbase.testclassification.LargeTests; import org.apache.hadoop.hbase.testclassification.VerySlowRegionServerTests; import org.apache.hadoop.hbase.util.Bytes; @@ -164,7 +164,7 @@ public class TestCompactionState { // otherwise, the compaction should have already been done if (expectedState != state) { for (Region region: regions) { - state = region.getCompactionState(); + state = CompactionState.valueOf(region.getCompactionState().toString()); assertEquals(CompactionState.NONE, state); } } else { http://git-wip-us.apache.org/repos/asf/hbase/blob/c06a976a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestTags.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestTags.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestTags.java index 3a9ace2..06dbc37 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestTags.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestTags.java @@ -40,6 +40,7 @@ import org.apache.hadoop.hbase.TagUtil; import org.apache.hadoop.hbase.ArrayBackedTag; import org.apache.hadoop.hbase.client.Admin; import org.apache.hadoop.hbase.client.Append; +import org.apache.hadoop.hbase.client.CompactionState; import org.apache.hadoop.hbase.client.Durability; import org.apache.hadoop.hbase.client.Increment; import org.apache.hadoop.hbase.client.Mutation; @@ -53,7 +54,6 @@ import org.apache.hadoop.hbase.coprocessor.CoprocessorHost; import org.apache.hadoop.hbase.coprocessor.ObserverContext; import org.apache.hadoop.hbase.coprocessor.RegionCoprocessorEnvironment; import org.apache.hadoop.hbase.io.encoding.DataBlockEncoding; -import org.apache.hadoop.hbase.protobuf.generated.AdminProtos.GetRegionInfoResponse.CompactionState; import org.apache.hadoop.hbase.regionserver.wal.WALEdit; import org.apache.hadoop.hbase.testclassification.MediumTests; import org.apache.hadoop.hbase.testclassification.RegionServerTests; http://git-wip-us.apache.org/repos/asf/hbase/blob/c06a976a/hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestAccessController.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestAccessController.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestAccessController.java index 37c42a0..92d7806 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestAccessController.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestAccessController.java @@ -62,6 +62,7 @@ import org.apache.hadoop.hbase.client.ConnectionFactory; import org.apache.hadoop.hbase.client.Delete; import org.apache.hadoop.hbase.client.Get; import org.apache.hadoop.hbase.client.Increment; +import org.apache.hadoop.hbase.client.MasterSwitchType; import org.apache.hadoop.hbase.client.Put; import org.apache.hadoop.hbase.client.RegionLocator; import org.apache.hadoop.hbase.client.Result; @@ -710,7 +711,7 @@ public class TestAccessController extends SecureTestUtil { @Override public Object run() throws Exception { ACCESS_CONTROLLER.preSetSplitOrMergeEnabled(ObserverContext.createAndPrepare(CP_ENV, null), - true, Admin.MasterSwitchType.MERGE); + true, MasterSwitchType.MERGE); return null; } }; http://git-wip-us.apache.org/repos/asf/hbase/blob/c06a976a/hbase-server/src/test/java/org/apache/hadoop/hbase/snapshot/SnapshotTestingUtils.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/snapshot/SnapshotTestingUtils.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/snapshot/SnapshotTestingUtils.java index 666eea3..6615a8f 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/snapshot/SnapshotTestingUtils.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/snapshot/SnapshotTestingUtils.java @@ -48,6 +48,7 @@ import org.apache.hadoop.hbase.TableDescriptor; import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.TableNotEnabledException; import org.apache.hadoop.hbase.client.Admin; +import org.apache.hadoop.hbase.client.SnapshotType; import org.apache.hadoop.hbase.client.BufferedMutator; import org.apache.hadoop.hbase.client.Durability; import org.apache.hadoop.hbase.client.Put; @@ -59,7 +60,8 @@ import org.apache.hadoop.hbase.master.HMaster; import org.apache.hadoop.hbase.master.MasterFileSystem; import org.apache.hadoop.hbase.mob.MobUtils; import org.apache.hadoop.hbase.protobuf.ProtobufUtil; -import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription; +import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos; +import org.apache.hadoop.hbase.client.SnapshotDescription; import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotRegionManifest; import org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsSnapshotDoneRequest; import org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsSnapshotDoneResponse; @@ -120,7 +122,7 @@ public class SnapshotTestingUtils { * Make sure that there is only one snapshot returned from the master */ public static void assertOneSnapshotThatMatches(Admin admin, - SnapshotDescription snapshot) throws IOException { + HBaseProtos.SnapshotDescription snapshot) throws IOException { assertOneSnapshotThatMatches(admin, snapshot.getName(), TableName.valueOf(snapshot.getTable())); } @@ -153,7 +155,7 @@ public class SnapshotTestingUtils { } public static void confirmSnapshotValid(HBaseTestingUtility testUtil, - SnapshotDescription snapshotDescriptor, TableName tableName, byte[] family) + HBaseProtos.SnapshotDescription snapshotDescriptor, TableName tableName, byte[] family) throws IOException { MasterFileSystem mfs = testUtil.getHBaseCluster().getMaster().getMasterFileSystem(); confirmSnapshotValid(snapshotDescriptor, tableName, family, @@ -165,7 +167,7 @@ public class SnapshotTestingUtils { * be in the snapshot. */ public static void confirmSnapshotValid( - SnapshotDescription snapshotDescriptor, TableName tableName, + HBaseProtos.SnapshotDescription snapshotDescriptor, TableName tableName, byte[] testFamily, Path rootDir, Admin admin, FileSystem fs) throws IOException { ArrayList nonEmptyTestFamilies = new ArrayList(1); @@ -178,7 +180,7 @@ public class SnapshotTestingUtils { * Confirm that the snapshot has no references files but only metadata. */ public static void confirmEmptySnapshotValid( - SnapshotDescription snapshotDescriptor, TableName tableName, + HBaseProtos.SnapshotDescription snapshotDescriptor, TableName tableName, byte[] testFamily, Path rootDir, Admin admin, FileSystem fs) throws IOException { ArrayList emptyTestFamilies = new ArrayList(1); @@ -194,7 +196,7 @@ public class SnapshotTestingUtils { * by the MasterSnapshotVerifier, at the end of the snapshot operation. */ public static void confirmSnapshotValid( - SnapshotDescription snapshotDescriptor, TableName tableName, + HBaseProtos.SnapshotDescription snapshotDescriptor, TableName tableName, List<byte[]> nonEmptyTestFamilies, List<byte[]> emptyTestFamilies, Path rootDir, Admin admin, FileSystem fs) throws IOException { final Configuration conf = admin.getConfiguration(); @@ -204,7 +206,7 @@ public class SnapshotTestingUtils { snapshotDescriptor, rootDir); assertTrue(fs.exists(snapshotDir)); - SnapshotDescription desc = SnapshotDescriptionUtils.readSnapshotInfo(fs, snapshotDir); + HBaseProtos.SnapshotDescription desc = SnapshotDescriptionUtils.readSnapshotInfo(fs, snapshotDir); // Extract regions and families with store files final Set<byte[]> snapshotFamilies = new TreeSet<byte[]>(Bytes.BYTES_COMPARATOR); @@ -265,7 +267,7 @@ public class SnapshotTestingUtils { * @throws ServiceException if the snapshot fails */ public static void waitForSnapshotToComplete(HMaster master, - SnapshotDescription snapshot, long sleep) throws ServiceException { + HBaseProtos.SnapshotDescription snapshot, long sleep) throws ServiceException { final IsSnapshotDoneRequest request = IsSnapshotDoneRequest.newBuilder() .setSnapshot(snapshot).build(); IsSnapshotDoneResponse done = IsSnapshotDoneResponse.newBuilder() @@ -286,12 +288,13 @@ public class SnapshotTestingUtils { */ public static void snapshot(Admin admin, final String snapshotName, final String tableName, - SnapshotDescription.Type type, int numTries) throws IOException { + HBaseProtos.SnapshotDescription.Type type, int numTries) throws IOException { int tries = 0; CorruptedSnapshotException lastEx = null; while (tries++ < numTries) { try { - admin.snapshot(snapshotName, TableName.valueOf(tableName), type); + admin.snapshot(new SnapshotDescription(snapshotName, tableName, + SnapshotType.valueOf(type.toString()))); return; } catch (CorruptedSnapshotException cse) { LOG.warn("Got CorruptedSnapshotException", cse); @@ -393,13 +396,14 @@ public class SnapshotTestingUtils { } admin.snapshot(snapshotNameString, tableName); - List<SnapshotDescription> snapshots = SnapshotTestingUtils.assertExistsMatchingSnapshot(admin, - snapshotNameString, tableName); + List<SnapshotDescription> snapshots = + SnapshotTestingUtils.assertExistsMatchingSnapshot(admin, snapshotNameString, tableName); if (snapshots == null || snapshots.size() != 1) { Assert.fail("Incorrect number of snapshots for table " + tableName); } - SnapshotTestingUtils.confirmSnapshotValid(snapshots.get(0), tableName, nonEmptyFamilyNames, + SnapshotTestingUtils.confirmSnapshotValid( + ProtobufUtil.createHBaseProtosSnapshotDesc(snapshots.get(0)), tableName, nonEmptyFamilyNames, emptyFamilyNames, rootDir, admin, fs); } @@ -418,7 +422,8 @@ public class SnapshotTestingUtils { Path snapshotDir = SnapshotDescriptionUtils.getCompletedSnapshotDir(snapshotName, mfs.getRootDir()); - SnapshotDescription snapshotDesc = SnapshotDescriptionUtils.readSnapshotInfo(fs, snapshotDir); + HBaseProtos.SnapshotDescription snapshotDesc = + SnapshotDescriptionUtils.readSnapshotInfo(fs, snapshotDir); final TableName table = TableName.valueOf(snapshotDesc.getTable()); final ArrayList corruptedFiles = new ArrayList(); @@ -467,7 +472,7 @@ public class SnapshotTestingUtils { public static class SnapshotBuilder { private final RegionData[] tableRegions; - private final SnapshotDescription desc; + private final HBaseProtos.SnapshotDescription desc; private final HTableDescriptor htd; private final Configuration conf; private final FileSystem fs; @@ -477,7 +482,7 @@ public class SnapshotTestingUtils { public SnapshotBuilder(final Configuration conf, final FileSystem fs, final Path rootDir, final HTableDescriptor htd, - final SnapshotDescription desc, final RegionData[] tableRegions) + final HBaseProtos.SnapshotDescription desc, final RegionData[] tableRegions) throws IOException { this.fs = fs; this.conf = conf; @@ -495,7 +500,7 @@ public class SnapshotTestingUtils { return this.htd; } - public SnapshotDescription getSnapshotDescription() { + public HBaseProtos.SnapshotDescription getSnapshotDescription() { return this.desc; } @@ -519,7 +524,7 @@ public class SnapshotTestingUtils { .build()); } - private Path[] addRegion(final SnapshotDescription desc) throws IOException { + private Path[] addRegion(final HBaseProtos.SnapshotDescription desc) throws IOException { if (this.snapshotted == tableRegions.length) { throw new UnsupportedOperationException("No more regions in the table"); } @@ -648,7 +653,7 @@ public class SnapshotTestingUtils { HTableDescriptor htd = createHtd(tableName); RegionData[] regions = createTable(htd, numRegions); - SnapshotDescription desc = SnapshotDescription.newBuilder() + HBaseProtos.SnapshotDescription desc = HBaseProtos.SnapshotDescription.newBuilder() .setTable(htd.getNameAsString()) .setName(snapshotName) .setVersion(version) http://git-wip-us.apache.org/repos/asf/hbase/blob/c06a976a/hbase-server/src/test/java/org/apache/hadoop/hbase/snapshot/TestFlushSnapshotFromClient.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/snapshot/TestFlushSnapshotFromClient.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/snapshot/TestFlushSnapshotFromClient.java index 4b988a6..0a933c6 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/snapshot/TestFlushSnapshotFromClient.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/snapshot/TestFlushSnapshotFromClient.java @@ -44,7 +44,9 @@ import org.apache.hadoop.hbase.client.Admin; import org.apache.hadoop.hbase.client.Table; import org.apache.hadoop.hbase.master.HMaster; import org.apache.hadoop.hbase.master.snapshot.SnapshotManager; -import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription; +import org.apache.hadoop.hbase.protobuf.ProtobufUtil; +import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos; +import org.apache.hadoop.hbase.client.SnapshotDescription; import org.apache.hadoop.hbase.regionserver.ConstantSizeRegionSplitPolicy; import org.apache.hadoop.hbase.testclassification.LargeTests; import org.apache.hadoop.hbase.testclassification.RegionServerTests; @@ -147,18 +149,20 @@ public class TestFlushSnapshotFromClient { // take a snapshot of the enabled table String snapshotString = "offlineTableSnapshot"; byte[] snapshot = Bytes.toBytes(snapshotString); - admin.snapshot(snapshotString, TABLE_NAME, SnapshotDescription.Type.FLUSH); + admin.snapshot(snapshotString, TABLE_NAME, + ProtobufUtil.createSnapshotType(HBaseProtos.SnapshotDescription.Type.FLUSH)); LOG.debug("Snapshot completed."); // make sure we have the snapshot - List<SnapshotDescription> snapshots = SnapshotTestingUtils.assertOneSnapshotThatMatches(admin, - snapshot, TABLE_NAME); + List<SnapshotDescription> snapshots = + SnapshotTestingUtils.assertOneSnapshotThatMatches(admin, snapshot, TABLE_NAME); // make sure its a valid snapshot LOG.debug("FS state after snapshot:"); UTIL.getHBaseCluster().getMaster().getMasterFileSystem().logFileSystemState(LOG); - SnapshotTestingUtils.confirmSnapshotValid(UTIL, snapshots.get(0), TABLE_NAME, TEST_FAM); + SnapshotTestingUtils.confirmSnapshotValid(UTIL, + ProtobufUtil.createHBaseProtosSnapshotDesc(snapshots.get(0)), TABLE_NAME, TEST_FAM); } /** @@ -181,18 +185,20 @@ public class TestFlushSnapshotFromClient { // take a snapshot of the enabled table String snapshotString = "skipFlushTableSnapshot"; byte[] snapshot = Bytes.toBytes(snapshotString); - admin.snapshot(snapshotString, TABLE_NAME, SnapshotDescription.Type.SKIPFLUSH); + admin.snapshot(snapshotString, TABLE_NAME, + ProtobufUtil.createSnapshotType(HBaseProtos.SnapshotDescription.Type.SKIPFLUSH)); LOG.debug("Snapshot completed."); // make sure we have the snapshot - List<SnapshotDescription> snapshots = SnapshotTestingUtils.assertOneSnapshotThatMatches(admin, - snapshot, TABLE_NAME); + List<SnapshotDescription> snapshots = + SnapshotTestingUtils.assertOneSnapshotThatMatches(admin, snapshot, TABLE_NAME); // make sure its a valid snapshot LOG.debug("FS state after snapshot:"); UTIL.getHBaseCluster().getMaster().getMasterFileSystem().logFileSystemState(LOG); - SnapshotTestingUtils.confirmSnapshotValid(UTIL, snapshots.get(0), TABLE_NAME, TEST_FAM); + SnapshotTestingUtils.confirmSnapshotValid(UTIL, + ProtobufUtil.createHBaseProtosSnapshotDesc(snapshots.get(0)), TABLE_NAME, TEST_FAM); admin.deleteSnapshot(snapshot); snapshots = admin.listSnapshots(); @@ -234,7 +240,8 @@ public class TestFlushSnapshotFromClient { LOG.debug("FS state after snapshot:"); UTIL.getHBaseCluster().getMaster().getMasterFileSystem().logFileSystemState(LOG); - SnapshotTestingUtils.confirmSnapshotValid(UTIL, snapshots.get(0), TABLE_NAME, TEST_FAM); + SnapshotTestingUtils.confirmSnapshotValid(UTIL, + ProtobufUtil.createHBaseProtosSnapshotDesc(snapshots.get(0)), TABLE_NAME, TEST_FAM); } @Test @@ -258,7 +265,8 @@ public class TestFlushSnapshotFromClient { // snapshot the non-existant table try { - admin.snapshot("fail", tableName, SnapshotDescription.Type.FLUSH); + admin.snapshot("fail", tableName, + ProtobufUtil.createSnapshotType(HBaseProtos.SnapshotDescription.Type.FLUSH)); fail("Snapshot succeeded even though there is not table."); } catch (SnapshotCreationException e) { LOG.info("Correctly failed to snapshot a non-existant table:" + e.getMessage()); @@ -267,13 +275,14 @@ public class TestFlushSnapshotFromClient { @Test public void testAsyncFlushSnapshot() throws Exception { - SnapshotDescription snapshot = SnapshotDescription.newBuilder().setName("asyncSnapshot") - .setTable(TABLE_NAME.getNameAsString()) - .setType(SnapshotDescription.Type.FLUSH) - .build(); + HBaseProtos.SnapshotDescription snapshot = HBaseProtos.SnapshotDescription.newBuilder() + .setName("asyncSnapshot").setTable(TABLE_NAME.getNameAsString()) + .setType(HBaseProtos.SnapshotDescription.Type.FLUSH).build(); // take the snapshot async - admin.takeSnapshotAsync(snapshot); + admin.takeSnapshotAsync( + new SnapshotDescription("asyncSnapshot", TABLE_NAME.getNameAsString(), + ProtobufUtil.createSnapshotType(HBaseProtos.SnapshotDescription.Type.FLUSH))); // constantly loop, looking for the snapshot to complete HMaster master = UTIL.getMiniHBaseCluster().getMaster(); @@ -295,7 +304,8 @@ public class TestFlushSnapshotFromClient { // Take a snapshot String snapshotBeforeMergeName = "snapshotBeforeMerge"; - admin.snapshot(snapshotBeforeMergeName, TABLE_NAME, SnapshotDescription.Type.FLUSH); + admin.snapshot(snapshotBeforeMergeName, TABLE_NAME, + ProtobufUtil.createSnapshotType(HBaseProtos.SnapshotDescription.Type.FLUSH)); // Clone the table TableName cloneBeforeMergeName = TableName.valueOf("cloneBeforeMerge"); @@ -364,7 +374,7 @@ public class TestFlushSnapshotFromClient { // Take a snapshot String snapshotName = "snapshotAfterMerge"; SnapshotTestingUtils.snapshot(admin, snapshotName, TABLE_NAME.getNameAsString(), - SnapshotDescription.Type.FLUSH, 3); + HBaseProtos.SnapshotDescription.Type.FLUSH, 3); // Clone the table TableName cloneName = TableName.valueOf("cloneMerge"); @@ -425,14 +435,16 @@ public class TestFlushSnapshotFromClient { @Override public void run() { try { - LOG.info("Submitting snapshot request: " + ClientSnapshotDescriptionUtils.toString(ss)); + LOG.info("Submitting snapshot request: " + ClientSnapshotDescriptionUtils + .toString(ProtobufUtil.createHBaseProtosSnapshotDesc(ss))); admin.takeSnapshotAsync(ss); } catch (Exception e) { LOG.info("Exception during snapshot request: " + ClientSnapshotDescriptionUtils.toString( - ss) + ProtobufUtil.createHBaseProtosSnapshotDesc(ss)) + ". This is ok, we expect some", e); } - LOG.info("Submitted snapshot request: " + ClientSnapshotDescriptionUtils.toString(ss)); + LOG.info("Submitted snapshot request: " + ClientSnapshotDescriptionUtils + .toString(ProtobufUtil.createHBaseProtosSnapshotDesc(ss))); toBeSubmitted.countDown(); } }; @@ -440,11 +452,15 @@ public class TestFlushSnapshotFromClient { // build descriptions SnapshotDescription[] descs = new SnapshotDescription[ssNum]; for (int i = 0; i < ssNum; i++) { - SnapshotDescription.Builder builder = SnapshotDescription.newBuilder(); - builder.setTable(((i % 2) == 0 ? TABLE_NAME : TABLE2_NAME).getNameAsString()); - builder.setName("ss"+i); - builder.setType(SnapshotDescription.Type.FLUSH); - descs[i] = builder.build(); + HBaseProtos.SnapshotDescription.Builder builder = + HBaseProtos.SnapshotDescription.newBuilder(); + if(i %2 ==0) { + descs[i] = new SnapshotDescription("ss" + i, TABLE_NAME.getNameAsString(), + ProtobufUtil.createSnapshotType(HBaseProtos.SnapshotDescription.Type.FLUSH)); + } else { + descs[i] = new SnapshotDescription("ss" + i, TABLE2_NAME.getNameAsString(), + ProtobufUtil.createSnapshotType(HBaseProtos.SnapshotDescription.Type.FLUSH)); + } } // kick each off its own thread http://git-wip-us.apache.org/repos/asf/hbase/blob/c06a976a/hbase-server/src/test/java/org/apache/hadoop/hbase/snapshot/TestRestoreFlushSnapshotFromClient.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/snapshot/TestRestoreFlushSnapshotFromClient.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/snapshot/TestRestoreFlushSnapshotFromClient.java index 877ee21..04fce5c 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/snapshot/TestRestoreFlushSnapshotFromClient.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/snapshot/TestRestoreFlushSnapshotFromClient.java @@ -28,6 +28,7 @@ import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.client.Admin; import org.apache.hadoop.hbase.client.Table; import org.apache.hadoop.hbase.master.snapshot.SnapshotManager; +import org.apache.hadoop.hbase.protobuf.ProtobufUtil; import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription; import org.apache.hadoop.hbase.regionserver.snapshot.RegionServerSnapshotManager; import org.apache.hadoop.hbase.testclassification.LargeTests; @@ -116,7 +117,7 @@ public class TestRestoreFlushSnapshotFromClient { // take a snapshot admin.snapshot(Bytes.toString(snapshotName0), tableName, - SnapshotDescription.Type.FLUSH); + ProtobufUtil.createSnapshotType(SnapshotDescription.Type.FLUSH)); LOG.info("=== after snapshot with 500 rows"); logFSTree(); @@ -129,7 +130,7 @@ public class TestRestoreFlushSnapshotFromClient { // take a snapshot of the updated table admin.snapshot(Bytes.toString(snapshotName1), tableName, - SnapshotDescription.Type.FLUSH); + ProtobufUtil.createSnapshotType(SnapshotDescription.Type.FLUSH)); LOG.info("=== after snapshot with 1000 rows"); logFSTree(); table.close(); @@ -194,7 +195,8 @@ public class TestRestoreFlushSnapshotFromClient { TableName clonedTableName = TableName.valueOf("clonedtb-" + System.currentTimeMillis()); admin.cloneSnapshot(snapshotName0, clonedTableName); verifyRowCount(UTIL, clonedTableName, snapshot0Rows); - admin.snapshot(Bytes.toString(snapshotName2), clonedTableName, SnapshotDescription.Type.FLUSH); + admin.snapshot(Bytes.toString(snapshotName2), clonedTableName, + ProtobufUtil.createSnapshotType(SnapshotDescription.Type.FLUSH)); UTIL.deleteTable(clonedTableName); admin.cloneSnapshot(snapshotName2, clonedTableName); http://git-wip-us.apache.org/repos/asf/hbase/blob/c06a976a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestHBaseFsckOneRS.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestHBaseFsckOneRS.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestHBaseFsckOneRS.java index dbb23a5..d5869ed 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestHBaseFsckOneRS.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestHBaseFsckOneRS.java @@ -33,12 +33,12 @@ import org.apache.hadoop.hbase.MetaTableAccessor; import org.apache.hadoop.hbase.MiniHBaseCluster; import org.apache.hadoop.hbase.ServerName; import org.apache.hadoop.hbase.TableName; -import org.apache.hadoop.hbase.client.Admin; import org.apache.hadoop.hbase.client.ClusterConnection; import org.apache.hadoop.hbase.client.Connection; import org.apache.hadoop.hbase.client.ConnectionFactory; import org.apache.hadoop.hbase.client.Delete; import org.apache.hadoop.hbase.client.Get; +import org.apache.hadoop.hbase.client.MasterSwitchType; import org.apache.hadoop.hbase.client.Put; import org.apache.hadoop.hbase.client.RegionLocator; import org.apache.hadoop.hbase.client.Result; @@ -1856,9 +1856,9 @@ public class TestHBaseFsckOneRS extends BaseTestHBaseFsck { @Test public void testSplitOrMergeStatWhenHBCKAbort() throws Exception { admin.setSplitOrMergeEnabled(true, false, true, - Admin.MasterSwitchType.SPLIT, Admin.MasterSwitchType.MERGE); - boolean oldSplit = admin.isSplitOrMergeEnabled(Admin.MasterSwitchType.SPLIT); - boolean oldMerge = admin.isSplitOrMergeEnabled(Admin.MasterSwitchType.MERGE); + MasterSwitchType.SPLIT, MasterSwitchType.MERGE); + boolean oldSplit = admin.isSplitOrMergeEnabled(MasterSwitchType.SPLIT); + boolean oldMerge = admin.isSplitOrMergeEnabled(MasterSwitchType.MERGE); assertTrue(oldSplit); assertTrue(oldMerge); @@ -1880,8 +1880,8 @@ public class TestHBaseFsckOneRS extends BaseTestHBaseFsck { spiedHbck.onlineHbck(); spiedHbck.close(); - boolean split = admin.isSplitOrMergeEnabled(Admin.MasterSwitchType.SPLIT); - boolean merge = admin.isSplitOrMergeEnabled(Admin.MasterSwitchType.MERGE); + boolean split = admin.isSplitOrMergeEnabled(MasterSwitchType.SPLIT); + boolean merge = admin.isSplitOrMergeEnabled(MasterSwitchType.MERGE); assertFalse(split); assertFalse(merge); @@ -1892,8 +1892,8 @@ public class TestHBaseFsckOneRS extends BaseTestHBaseFsck { hbck.onlineHbck(); hbck.close(); - split = admin.isSplitOrMergeEnabled(Admin.MasterSwitchType.SPLIT); - merge = admin.isSplitOrMergeEnabled(Admin.MasterSwitchType.MERGE); + split = admin.isSplitOrMergeEnabled(MasterSwitchType.SPLIT); + merge = admin.isSplitOrMergeEnabled(MasterSwitchType.MERGE); assertTrue(split); assertTrue(merge);