http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/MockRegionServerServices.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/MockRegionServerServices.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/MockRegionServerServices.java index 5e2a70f..7740e66 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/MockRegionServerServices.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/MockRegionServerServices.java @@ -63,9 +63,9 @@ import com.google.protobuf.Service; */ public class MockRegionServerServices implements RegionServerServices { protected static final Log LOG = LogFactory.getLog(MockRegionServerServices.class); - private final Map<String, Region> regions = new HashMap<String, Region>(); + private final Map<String, Region> regions = new HashMap<>(); private final ConcurrentSkipListMap<byte[], Boolean> rit = - new ConcurrentSkipListMap<byte[], Boolean>(Bytes.BYTES_COMPARATOR); + new ConcurrentSkipListMap<>(Bytes.BYTES_COMPARATOR); private HFileSystem hfs = null; private final Configuration conf; private ZooKeeperWatcher zkw = null;
http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/MultithreadedTestUtil.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/MultithreadedTestUtil.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/MultithreadedTestUtil.java index 7e251e7..cf07b42 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/MultithreadedTestUtil.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/MultithreadedTestUtil.java @@ -39,7 +39,7 @@ public abstract class MultithreadedTestUtil { private Throwable err = null; private boolean stopped = false; private int threadDoneCount = 0; - private Set<TestThread> testThreads = new HashSet<TestThread>(); + private Set<TestThread> testThreads = new HashSet<>(); public TestContext(Configuration configuration) { this.conf = configuration; http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/PerformanceEvaluation.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/PerformanceEvaluation.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/PerformanceEvaluation.java index b7e4a71..f8345b1 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/PerformanceEvaluation.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/PerformanceEvaluation.java @@ -146,7 +146,7 @@ public class PerformanceEvaluation extends Configured implements Tool { private static final BigDecimal BYTES_PER_MB = BigDecimal.valueOf(1024 * 1024); private static final TestOptions DEFAULT_OPTS = new TestOptions(); - private static Map<String, CmdDescriptor> COMMANDS = new TreeMap<String, CmdDescriptor>(); + private static Map<String, CmdDescriptor> COMMANDS = new TreeMap<>(); private static final Path PERF_EVAL_DIR = new Path("performance_evaluation"); static { @@ -536,7 +536,7 @@ public class PerformanceEvaluation extends Configured implements Tool { Path inputFile = new Path(inputDir, JOB_INPUT_FILENAME); PrintStream out = new PrintStream(fs.create(inputFile)); // Make input random. - Map<Integer, String> m = new TreeMap<Integer, String>(); + Map<Integer, String> m = new TreeMap<>(); Hash h = MurmurHash.getInstance(); int perClientRows = (opts.totalRows / opts.numClientThreads); try { @@ -1311,7 +1311,7 @@ public class PerformanceEvaluation extends Configured implements Tool { protected Pair<byte[], byte[]> generateStartAndStopRows(int maxRange) { int start = this.rand.nextInt(Integer.MAX_VALUE) % opts.totalRows; int stop = start + maxRange; - return new Pair<byte[],byte[]>(format(start), format(stop)); + return new Pair<>(format(start), format(stop)); } @Override @@ -1375,7 +1375,7 @@ public class PerformanceEvaluation extends Configured implements Tool { consistency = options.replicas == DEFAULT_OPTS.replicas ? null : Consistency.TIMELINE; if (opts.multiGet > 0) { LOG.info("MultiGet enabled. Sending GETs in batches of " + opts.multiGet + "."); - this.gets = new ArrayList<Get>(opts.multiGet); + this.gets = new ArrayList<>(opts.multiGet); } } @@ -2207,7 +2207,7 @@ public class PerformanceEvaluation extends Configured implements Tool { } try { - LinkedList<String> argv = new LinkedList<String>(); + LinkedList<String> argv = new LinkedList<>(); argv.addAll(Arrays.asList(args)); TestOptions opts = parseOpts(argv); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/PerformanceEvaluationCommons.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/PerformanceEvaluationCommons.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/PerformanceEvaluationCommons.java index 3809a13..e2350e8 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/PerformanceEvaluationCommons.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/PerformanceEvaluationCommons.java @@ -67,7 +67,7 @@ public class PerformanceEvaluationCommons { public static void concurrentReads(final Runnable r) { final int count = 1; long now = System.currentTimeMillis(); - List<Thread> threads = new ArrayList<Thread>(count); + List<Thread> threads = new ArrayList<>(count); for (int i = 0; i < count; i++) { threads.add(new Thread(r, "concurrentRead-" + i)); } http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/TestCheckTestClasses.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/TestCheckTestClasses.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/TestCheckTestClasses.java index 06b98f7..23ca57f 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/TestCheckTestClasses.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/TestCheckTestClasses.java @@ -39,7 +39,7 @@ public class TestCheckTestClasses { */ @Test public void checkClasses() throws Exception { - List<Class<?>> badClasses = new java.util.ArrayList<Class<?>>(); + List<Class<?>> badClasses = new java.util.ArrayList<>(); ClassTestFinder classFinder = new ClassTestFinder(); for (Class<?> c : classFinder.findClasses(false)) { if (ClassTestFinder.getCategoryAnnotations(c).length == 0) { http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/TestGlobalMemStoreSize.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/TestGlobalMemStoreSize.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/TestGlobalMemStoreSize.java index 5cc7ed9..8d19c1b 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/TestGlobalMemStoreSize.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/TestGlobalMemStoreSize.java @@ -158,7 +158,7 @@ public class TestGlobalMemStoreSize { } private List<HRegionServer> getOnlineRegionServers() { - List<HRegionServer> list = new ArrayList<HRegionServer>(); + List<HRegionServer> list = new ArrayList<>(); for (JVMClusterUtil.RegionServerThread rst : cluster.getRegionServerThreads()) { if (rst.getRegionServer().isOnline()) { http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/TestHDFSBlocksDistribution.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/TestHDFSBlocksDistribution.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/TestHDFSBlocksDistribution.java index 2329fc2..06cfdcf 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/TestHDFSBlocksDistribution.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/TestHDFSBlocksDistribution.java @@ -50,7 +50,7 @@ public class TestHDFSBlocksDistribution { public class MockHDFSBlocksDistribution extends HDFSBlocksDistribution { public Map<String,HostAndWeight> getHostAndWeights() { - HashMap<String, HostAndWeight> map = new HashMap<String, HostAndWeight>(); + HashMap<String, HostAndWeight> map = new HashMap<>(); map.put("test", new HostAndWeight(null, 100)); return map; } http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/TestMetaTableAccessorNoCluster.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/TestMetaTableAccessorNoCluster.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/TestMetaTableAccessorNoCluster.java index 870ebb3..9915f99 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/TestMetaTableAccessorNoCluster.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/TestMetaTableAccessorNoCluster.java @@ -87,7 +87,7 @@ public class TestMetaTableAccessorNoCluster { public void testGetHRegionInfo() throws IOException { assertNull(MetaTableAccessor.getHRegionInfo(new Result())); - List<Cell> kvs = new ArrayList<Cell>(); + List<Cell> kvs = new ArrayList<>(); Result r = Result.create(kvs); assertNull(MetaTableAccessor.getHRegionInfo(r)); @@ -141,7 +141,7 @@ public class TestMetaTableAccessorNoCluster { // show. We will know if they happened or not because we will ask // mockito at the end of this test to verify that scan was indeed // called the wanted number of times. - List<Cell> kvs = new ArrayList<Cell>(); + List<Cell> kvs = new ArrayList<>(); final byte [] rowToVerify = Bytes.toBytes("rowToVerify"); kvs.add(new KeyValue(rowToVerify, HConstants.CATALOG_FAMILY, HConstants.REGIONINFO_QUALIFIER, @@ -152,7 +152,7 @@ public class TestMetaTableAccessorNoCluster { kvs.add(new KeyValue(rowToVerify, HConstants.CATALOG_FAMILY, HConstants.STARTCODE_QUALIFIER, Bytes.toBytes(sn.getStartcode()))); - final List<CellScannable> cellScannables = new ArrayList<CellScannable>(1); + final List<CellScannable> cellScannables = new ArrayList<>(1); cellScannables.add(Result.create(kvs)); final ScanResponse.Builder builder = ScanResponse.newBuilder(); for (CellScannable result : cellScannables) { http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/TestPartialResultsFromClientSide.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/TestPartialResultsFromClientSide.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/TestPartialResultsFromClientSide.java index 57248b6..f3ea814 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/TestPartialResultsFromClientSide.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/TestPartialResultsFromClientSide.java @@ -274,7 +274,7 @@ public class TestPartialResultsFromClientSide { int iterationCount = 0; while (oneShotResult != null && oneShotResult.rawCells() != null) { - List<Cell> aggregatePartialCells = new ArrayList<Cell>(); + List<Cell> aggregatePartialCells = new ArrayList<>(); do { partialResult = partialScanner.next(); assertTrue("Partial Result is null. iteration: " + iterationCount, partialResult != null); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/TestRegionRebalancing.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/TestRegionRebalancing.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/TestRegionRebalancing.java index 03c5524..283d79d 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/TestRegionRebalancing.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/TestRegionRebalancing.java @@ -231,7 +231,7 @@ public class TestRegionRebalancing { } private List<HRegionServer> getOnlineRegionServers() { - List<HRegionServer> list = new ArrayList<HRegionServer>(); + List<HRegionServer> list = new ArrayList<>(); for (JVMClusterUtil.RegionServerThread rst : UTIL.getHBaseCluster().getRegionServerThreads()) { if (rst.getRegionServer().isOnline()) { http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/TestServerSideScanMetricsFromClientSide.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/TestServerSideScanMetricsFromClientSide.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/TestServerSideScanMetricsFromClientSide.java index 953756e..ad63cc8 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/TestServerSideScanMetricsFromClientSide.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/TestServerSideScanMetricsFromClientSide.java @@ -280,7 +280,7 @@ public class TestServerSideScanMetricsFromClientSide { filter = new SingleColumnValueFilter(FAMILIES[0], QUALIFIERS[0], CompareOp.NOT_EQUAL, VALUE); testRowsFilteredMetric(baseScan, filter, ROWS.length); - List<Filter> filters = new ArrayList<Filter>(); + List<Filter> filters = new ArrayList<>(); filters.add(new RowFilter(CompareOp.EQUAL, new BinaryComparator(ROWS[0]))); filters.add(new RowFilter(CompareOp.EQUAL, new BinaryComparator(ROWS[3]))); int numberOfMatchingRowFilters = filters.size(); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/TestStochasticBalancerJmxMetrics.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/TestStochasticBalancerJmxMetrics.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/TestStochasticBalancerJmxMetrics.java index d4f7cdd..c4abd89 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/TestStochasticBalancerJmxMetrics.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/TestStochasticBalancerJmxMetrics.java @@ -222,7 +222,7 @@ public class TestStochasticBalancerJmxMetrics extends BalancerTestBase { target = new ObjectName("Hadoop", pairs); MBeanInfo beanInfo = mb.getMBeanInfo(target); - Set<String> existingAttrs = new HashSet<String>(); + Set<String> existingAttrs = new HashSet<>(); for (MBeanAttributeInfo attrInfo : beanInfo.getAttributes()) { existingAttrs.add(attrInfo.getName()); } @@ -255,7 +255,7 @@ public class TestStochasticBalancerJmxMetrics extends BalancerTestBase { * Given the tables and functions, return metrics names that should exist in JMX */ private Set<String> getExpectedJmxMetrics(String[] tableNames, String[] functionNames) { - Set<String> ret = new HashSet<String>(); + Set<String> ret = new HashSet<>(); for (String tableName : tableNames) { ret.add(StochasticLoadBalancer.composeAttributeName(tableName, "Overall")); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/backup/TestHFileArchiving.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/backup/TestHFileArchiving.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/backup/TestHFileArchiving.java index 284251f..1acb842 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/backup/TestHFileArchiving.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/backup/TestHFileArchiving.java @@ -193,7 +193,7 @@ public class TestHFileArchiving { FileStatus[] regionFiles = FSUtils.listStatus(fs, regionDir, null); Assert.assertNotNull("No files in the region directory", regionFiles); if (LOG.isDebugEnabled()) { - List<Path> files = new ArrayList<Path>(); + List<Path> files = new ArrayList<>(); for (FileStatus file : regionFiles) { files.add(file.getPath()); } @@ -269,7 +269,7 @@ public class TestHFileArchiving { private void assertArchiveFiles(FileSystem fs, List<String> storeFiles, long timeout) throws IOException { long end = System.currentTimeMillis() + timeout; Path archiveDir = HFileArchiveUtil.getArchivePath(UTIL.getConfiguration()); - List<String> archivedFiles = new ArrayList<String>(); + List<String> archivedFiles = new ArrayList<>(); // We have to ensure that the DeleteTableHandler is finished. HBaseAdmin.deleteXXX() can return before all files // are archived. We should fix HBASE-5487 and fix synchronous operations from admin. @@ -434,7 +434,7 @@ public class TestHFileArchiving { return true; } }); - return recurseOnFiles(fs, files, new ArrayList<String>()); + return recurseOnFiles(fs, files, new ArrayList<>()); } /** Recursively lookup all the file names under the file[] array **/ http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/backup/example/TestZooKeeperTableArchiveClient.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/backup/example/TestZooKeeperTableArchiveClient.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/backup/example/TestZooKeeperTableArchiveClient.java index 64139ee..fc56ebd 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/backup/example/TestZooKeeperTableArchiveClient.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/backup/example/TestZooKeeperTableArchiveClient.java @@ -80,7 +80,7 @@ public class TestZooKeeperTableArchiveClient { private static final byte[] TEST_FAM = Bytes.toBytes("fam"); private static final byte[] TABLE_NAME = Bytes.toBytes(STRING_TABLE_NAME); private static ZKTableArchiveClient archivingClient; - private final List<Path> toCleanup = new ArrayList<Path>(); + private final List<Path> toCleanup = new ArrayList<>(); private static ClusterConnection CONNECTION; private static RegionServerServices rss; @@ -178,7 +178,7 @@ public class TestZooKeeperTableArchiveClient { // create the region HColumnDescriptor hcd = new HColumnDescriptor(TEST_FAM); HRegion region = UTIL.createTestRegion(STRING_TABLE_NAME, hcd); - List<Region> regions = new ArrayList<Region>(); + List<Region> regions = new ArrayList<>(); regions.add(region); when(rss.getOnlineRegions()).thenReturn(regions); final CompactedHFilesDischarger compactionCleaner = @@ -231,7 +231,7 @@ public class TestZooKeeperTableArchiveClient { // create the region HColumnDescriptor hcd = new HColumnDescriptor(TEST_FAM); HRegion region = UTIL.createTestRegion(STRING_TABLE_NAME, hcd); - List<Region> regions = new ArrayList<Region>(); + List<Region> regions = new ArrayList<>(); regions.add(region); when(rss.getOnlineRegions()).thenReturn(regions); final CompactedHFilesDischarger compactionCleaner = @@ -241,7 +241,7 @@ public class TestZooKeeperTableArchiveClient { // create the another table that we don't archive hcd = new HColumnDescriptor(TEST_FAM); HRegion otherRegion = UTIL.createTestRegion(otherTable, hcd); - regions = new ArrayList<Region>(); + regions = new ArrayList<>(); regions.add(otherRegion); when(rss.getOnlineRegions()).thenReturn(regions); final CompactedHFilesDischarger compactionCleaner1 = new CompactedHFilesDischarger(100, stop, @@ -388,7 +388,7 @@ public class TestZooKeeperTableArchiveClient { return null; } - List<Path> allFiles = new ArrayList<Path>(); + List<Path> allFiles = new ArrayList<>(); for (FileStatus file : files) { if (file.isDirectory()) { List<Path> subFiles = getAllFiles(fs, file.getPath()); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestAdmin1.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestAdmin1.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestAdmin1.java index f694210..7b69db4 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestAdmin1.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestAdmin1.java @@ -1002,7 +1002,7 @@ public class TestAdmin1 { // insert rows into column families. The number of rows that have values // in a specific column family is decided by rowCounts[familyIndex] for (int index = 0; index < familyNames.length; index++) { - ArrayList<Put> puts = new ArrayList<Put>(rowCounts[index]); + ArrayList<Put> puts = new ArrayList<>(rowCounts[index]); for (int i = 0; i < rowCounts[index]; i++) { byte[] k = Bytes.toBytes(i); Put put = new Put(k); @@ -1143,7 +1143,7 @@ public class TestAdmin1 { } while (oldRegions.size() != 9); //3 regions * 3 replicas // write some data to the table Table ht = TEST_UTIL.getConnection().getTable(tableName); - List<Put> puts = new ArrayList<Put>(); + List<Put> puts = new ArrayList<>(); byte[] qualifier = "c".getBytes(); Put put = new Put(new byte[]{(byte)'1'}); put.addColumn(cf, qualifier, "100".getBytes()); @@ -1295,7 +1295,7 @@ public class TestAdmin1 { byte[] q1 = Bytes.toBytes("q1"); byte[] v1 = Bytes.toBytes("v1"); p.addColumn(Bytes.toBytes(fn), q1, v1); - List<Put> puts = new ArrayList<Put>(2); + List<Put> puts = new ArrayList<>(2); puts.add(p); p = new Put(Bytes.toBytes("rep1_rk")); p.addColumn(Bytes.toBytes(fn1), q1, v1); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestAdmin2.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestAdmin2.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestAdmin2.java index eb15d91..0014401 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestAdmin2.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestAdmin2.java @@ -776,7 +776,7 @@ public class TestAdmin2 { // Drain all region servers. Collection<ServerName> clusterServers = admin.getClusterStatus().getServers(); - drainingServers = new ArrayList<ServerName>(); + drainingServers = new ArrayList<>(); for (ServerName server : clusterServers) { drainingServers.add(server); } http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestAvoidCellReferencesIntoShippedBlocks.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestAvoidCellReferencesIntoShippedBlocks.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestAvoidCellReferencesIntoShippedBlocks.java index ab09c5e..e7d7f0a 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestAvoidCellReferencesIntoShippedBlocks.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestAvoidCellReferencesIntoShippedBlocks.java @@ -231,7 +231,7 @@ public class TestAvoidCellReferencesIntoShippedBlocks { } catch (InterruptedException e) { } } - List<BlockCacheKey> cacheList = new ArrayList<BlockCacheKey>(); + List<BlockCacheKey> cacheList = new ArrayList<>(); Iterator<CachedBlock> iterator = cache.iterator(); // evict all the blocks while (iterator.hasNext()) { @@ -379,7 +379,7 @@ public class TestAvoidCellReferencesIntoShippedBlocks { Thread evictorThread = new Thread() { @Override public void run() { - List<BlockCacheKey> cacheList = new ArrayList<BlockCacheKey>(); + List<BlockCacheKey> cacheList = new ArrayList<>(); Iterator<CachedBlock> iterator = cache.iterator(); // evict all the blocks while (iterator.hasNext()) { @@ -416,7 +416,7 @@ public class TestAvoidCellReferencesIntoShippedBlocks { } assertEquals("Count the rows", count, 2); iterator = cache.iterator(); - List<BlockCacheKey> newCacheList = new ArrayList<BlockCacheKey>(); + List<BlockCacheKey> newCacheList = new ArrayList<>(); while (iterator.hasNext()) { CachedBlock next = iterator.next(); BlockCacheKey cacheKey = new BlockCacheKey(next.getFilename(), next.getOffset()); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestBlockEvictionFromClient.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestBlockEvictionFromClient.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestBlockEvictionFromClient.java index f453662..023095f 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestBlockEvictionFromClient.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestBlockEvictionFromClient.java @@ -1329,7 +1329,7 @@ public class TestBlockEvictionFromClient { private static class MultiGetThread extends Thread { private final Table table; - private final List<Get> gets = new ArrayList<Get>(); + private final List<Get> gets = new ArrayList<>(); public MultiGetThread(Table table) { this.table = table; } @@ -1565,7 +1565,7 @@ public class TestBlockEvictionFromClient { static final AtomicInteger countOfGets = new AtomicInteger(0); static final AtomicBoolean waitForGets = new AtomicBoolean(false); static final AtomicBoolean throwException = new AtomicBoolean(false); - private static final AtomicReference<CountDownLatch> cdl = new AtomicReference<CountDownLatch>( + private static final AtomicReference<CountDownLatch> cdl = new AtomicReference<>( new CountDownLatch(0)); @Override http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestClientOperationInterrupt.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestClientOperationInterrupt.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestClientOperationInterrupt.java index c1cb0a6..62ceca3 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestClientOperationInterrupt.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestClientOperationInterrupt.java @@ -96,7 +96,7 @@ public class TestClientOperationInterrupt { final AtomicInteger badEx = new AtomicInteger(0); final AtomicInteger noInt = new AtomicInteger(0); final AtomicInteger done = new AtomicInteger(0); - List<Thread> threads = new ArrayList<Thread>(); + List<Thread> threads = new ArrayList<>(); final int nbThread = 100; http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestConnectionUtils.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestConnectionUtils.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestConnectionUtils.java index 69729f0..c3e4a28 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestConnectionUtils.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestConnectionUtils.java @@ -42,7 +42,7 @@ public class TestConnectionUtils { retries[i] = ConnectionUtils.getPauseTime(baseTime, 0); } - Set<Long> retyTimeSet = new TreeSet<Long>(); + Set<Long> retyTimeSet = new TreeSet<>(); for (long l : retries) { /*make sure that there is some jitter but only 1%*/ assertTrue(l >= baseTime); @@ -62,7 +62,7 @@ public class TestConnectionUtils { long minTimeExpected = (long) (basePause * 0.75f); int testTries = 100; - Set<Long> timeSet = new TreeSet<Long>(); + Set<Long> timeSet = new TreeSet<>(); for (int i = 0; i < testTries; i++) { long withJitter = ConnectionUtils.addJitter(basePause, 0.5f); assertTrue(withJitter >= minTimeExpected); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFastFail.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFastFail.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFastFail.java index 2aed0ff..465bdfb 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFastFail.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFastFail.java @@ -163,7 +163,7 @@ public class TestFastFail { final AtomicInteger numBlockedWorkers = new AtomicInteger(0); final AtomicInteger numPreemptiveFastFailExceptions = new AtomicInteger(0); - List<Future<Boolean>> futures = new ArrayList<Future<Boolean>>(); + List<Future<Boolean>> futures = new ArrayList<>(); for (int i = 0; i < nThreads; i++) { futures.add(service.submit(new Callable<Boolean>() { /** http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide.java index aab586d..b1a0d3c 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide.java @@ -190,7 +190,7 @@ public class TestFromClientSide { @Test public void testDuplicateAppend() throws Exception { HTableDescriptor hdt = TEST_UTIL.createTableDescriptor(name.getMethodName()); - Map<String, String> kvs = new HashMap<String, String>(); + Map<String, String> kvs = new HashMap<>(); kvs.put(HConnectionTestingUtility.SleepAtFirstRpcCall.SLEEP_TIME_CONF_KEY, "2000"); hdt.addCoprocessor(HConnectionTestingUtility.SleepAtFirstRpcCall.class.getName(), null, 1, kvs); TEST_UTIL.createTable(hdt, new byte[][] { ROW }).close(); @@ -2299,7 +2299,7 @@ public class TestFromClientSide { result = ht.get(get); assertTrue(result.size() == 1); } - ArrayList<Delete> deletes = new ArrayList<Delete>(); + ArrayList<Delete> deletes = new ArrayList<>(); for (int i = 0; i < 10; i++) { byte [] bytes = Bytes.toBytes(i); delete = new Delete(bytes); @@ -4707,7 +4707,7 @@ public class TestFromClientSide { final Object waitLock = new Object(); ExecutorService executorService = Executors.newFixedThreadPool(numVersions); - final AtomicReference<AssertionError> error = new AtomicReference<AssertionError>(null); + final AtomicReference<AssertionError> error = new AtomicReference<>(null); for (int versions = numVersions; versions < numVersions * 2; versions++) { final int versionsCopy = versions; executorService.submit(new Callable<Void>() { @@ -5315,7 +5315,7 @@ public class TestFromClientSide { private List<HRegionLocation> getRegionsInRange(TableName tableName, byte[] startKey, byte[] endKey) throws IOException { - List<HRegionLocation> regionsInRange = new ArrayList<HRegionLocation>(); + List<HRegionLocation> regionsInRange = new ArrayList<>(); byte[] currentKey = startKey; final boolean endKeyIsEndOfTable = Bytes.equals(endKey, HConstants.EMPTY_END_ROW); try (RegionLocator r = TEST_UTIL.getConnection().getRegionLocator(tableName);) { @@ -6237,7 +6237,7 @@ public class TestFromClientSide { HRegionLocator locator = (HRegionLocator) admin.getConnection().getRegionLocator(htd.getTableName()); for (int regionReplication = 1; regionReplication < 4; regionReplication++) { - List<RegionLocations> regionLocations = new ArrayList<RegionLocations>(); + List<RegionLocations> regionLocations = new ArrayList<>(); // mock region locations coming from meta with multiple replicas for (HRegionInfo region : regions) { http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide3.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide3.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide3.java index 3680822..7f44a2a 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide3.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide3.java @@ -338,7 +338,7 @@ public class TestFromClientSide3 { put.addColumn(FAMILY, QUALIFIER, VALUE); table.put(put); - List<Get> gets = new ArrayList<Get>(); + List<Get> gets = new ArrayList<>(); gets.add(new Get(ROW)); gets.add(null); gets.add(new Get(ANOTHERROW)); @@ -432,7 +432,7 @@ public class TestFromClientSide3 { put.addColumn(FAMILY, QUALIFIER, VALUE); table.put (put); - List<Get> gets = new ArrayList<Get>(); + List<Get> gets = new ArrayList<>(); gets.add(new Get(ANOTHERROW)); gets.add(new Get(Bytes.add(ROW, new byte[] { 0x00 }))); gets.add(new Get(ROW)); @@ -450,7 +450,7 @@ public class TestFromClientSide3 { put.addColumn(FAMILY, QUALIFIER, VALUE); table.put(put); - gets = new ArrayList<Get>(); + gets = new ArrayList<>(); gets.add(new Get(new byte[] { 0x00 })); gets.add(new Get(new byte[] { 0x00, 0x00 })); results = table.existsAll(gets); @@ -462,7 +462,7 @@ public class TestFromClientSide3 { put.addColumn(FAMILY, QUALIFIER, VALUE); table.put(put); - gets = new ArrayList<Get>(); + gets = new ArrayList<>(); gets.add(new Get(new byte[] { (byte) 0xff })); gets.add(new Get(new byte[] { (byte) 0xff, (byte) 0xff })); gets.add(new Get(new byte[] { (byte) 0xff, (byte) 0xff, (byte) 0xff })); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestHBaseAdminNoCluster.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestHBaseAdminNoCluster.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestHBaseAdminNoCluster.java index 8d93a0a..1eb83d9 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestHBaseAdminNoCluster.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestHBaseAdminNoCluster.java @@ -170,7 +170,7 @@ public class TestHBaseAdminNoCluster { testMasterOperationIsRetried(new MethodCaller() { @Override public void call(Admin admin) throws Exception { - admin.getTableDescriptorsByTableName(new ArrayList<TableName>()); + admin.getTableDescriptorsByTableName(new ArrayList<>()); } @Override public void verify(MasterKeepAliveConnection masterAdmin, int count) throws Exception { http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestHCM.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestHCM.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestHCM.java index 8475828..70be7fa 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestHCM.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestHCM.java @@ -244,7 +244,7 @@ public class TestHCM { public void testClusterConnection() throws IOException { ThreadPoolExecutor otherPool = new ThreadPoolExecutor(1, 1, 5, TimeUnit.SECONDS, - new SynchronousQueue<Runnable>(), + new SynchronousQueue<>(), Threads.newDaemonThreadFactory("test-hcm")); Connection con1 = ConnectionFactory.createConnection(TEST_UTIL.getConfiguration()); @@ -748,7 +748,7 @@ public class TestHCM { // 4 steps: ready=0; doGets=1; mustStop=2; stopped=3 final AtomicInteger step = new AtomicInteger(0); - final AtomicReference<Throwable> failed = new AtomicReference<Throwable>(null); + final AtomicReference<Throwable> failed = new AtomicReference<>(null); Thread t = new Thread("testConnectionCloseThread") { @Override public void run() { http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestHTableMultiplexer.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestHTableMultiplexer.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestHTableMultiplexer.java index 26764d3..5c47de0 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestHTableMultiplexer.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestHTableMultiplexer.java @@ -131,7 +131,7 @@ public class TestHTableMultiplexer { } // MultiPut case - List<Put> multiput = new ArrayList<Put>(); + List<Put> multiput = new ArrayList<>(); for (int i = 0; i < NUM_REGIONS; i++) { byte [] row = endRows[i]; if (row == null || row.length <= 0) continue; http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestIllegalTableDescriptor.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestIllegalTableDescriptor.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestIllegalTableDescriptor.java index 9a2aa3d..999760d 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestIllegalTableDescriptor.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestIllegalTableDescriptor.java @@ -200,7 +200,7 @@ public class TestIllegalTableDescriptor { } private static class ListAppender extends AppenderSkeleton { - private final List<String> messages = new ArrayList<String>(); + private final List<String> messages = new ArrayList<>(); @Override protected void append(LoggingEvent event) { http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestIncrementsFromClientSide.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestIncrementsFromClientSide.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestIncrementsFromClientSide.java index 3d7e1dc..4406812 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestIncrementsFromClientSide.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestIncrementsFromClientSide.java @@ -94,7 +94,7 @@ public class TestIncrementsFromClientSide { @Test public void testDuplicateIncrement() throws Exception { HTableDescriptor hdt = TEST_UTIL.createTableDescriptor(TableName.valueOf(name.getMethodName())); - Map<String, String> kvs = new HashMap<String, String>(); + Map<String, String> kvs = new HashMap<>(); kvs.put(HConnectionTestingUtility.SleepAtFirstRpcCall.SLEEP_TIME_CONF_KEY, "2000"); hdt.addCoprocessor(HConnectionTestingUtility.SleepAtFirstRpcCall.class.getName(), null, 1, kvs); TEST_UTIL.createTable(hdt, new byte[][] { ROW }).close(); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestIntraRowPagination.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestIntraRowPagination.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestIntraRowPagination.java index 1f6dc98..43a2e77 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestIntraRowPagination.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestIntraRowPagination.java @@ -68,7 +68,7 @@ public class TestIntraRowPagination { Result result; boolean toLog = true; - List<Cell> kvListExp = new ArrayList<Cell>(); + List<Cell> kvListExp = new ArrayList<>(); int storeOffset = 1; int storeLimit = 3; @@ -91,8 +91,8 @@ public class TestIntraRowPagination { scan.setRowOffsetPerColumnFamily(storeOffset); scan.setMaxResultsPerColumnFamily(storeLimit); RegionScanner scanner = region.getScanner(scan); - List<Cell> kvListScan = new ArrayList<Cell>(); - List<Cell> results = new ArrayList<Cell>(); + List<Cell> kvListScan = new ArrayList<>(); + List<Cell> results = new ArrayList<>(); while (scanner.next(results) || !results.isEmpty()) { kvListScan.addAll(results); results.clear(); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestMultiParallel.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestMultiParallel.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestMultiParallel.java index 1209d25..a3bcc76 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestMultiParallel.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestMultiParallel.java @@ -110,7 +110,7 @@ public class TestMultiParallel { // not a multiple of the number of regions int numKeys = (int) ((float) starterKeys.length * 10.33F); - List<byte[]> keys = new ArrayList<byte[]>(); + List<byte[]> keys = new ArrayList<>(); for (int i = 0; i < numKeys; i++) { int kIdx = i % starterKeys.length; byte[] k = starterKeys[kIdx]; @@ -155,7 +155,7 @@ public class TestMultiParallel { try (Table t = connection.getTable(TEST_TABLE, executor)) { List<Put> puts = constructPutRequests(); // creates a Put for every region t.batch(puts, null); - HashSet<ServerName> regionservers = new HashSet<ServerName>(); + HashSet<ServerName> regionservers = new HashSet<>(); try (RegionLocator locator = connection.getRegionLocator(TEST_TABLE)) { for (Row r : puts) { HRegionLocation location = locator.getRegionLocation(r.getRow()); @@ -180,7 +180,7 @@ public class TestMultiParallel { table.batch(puts, null); // create a list of gets and run it - List<Row> gets = new ArrayList<Row>(); + List<Row> gets = new ArrayList<>(); for (byte[] k : KEYS) { Get get = new Get(k); get.addColumn(BYTES_FAMILY, QUALIFIER); @@ -190,7 +190,7 @@ public class TestMultiParallel { table.batch(gets, multiRes); // Same gets using individual call API - List<Result> singleRes = new ArrayList<Result>(); + List<Result> singleRes = new ArrayList<>(); for (Row get : gets) { singleRes.add(table.get((Get) get)); } @@ -214,7 +214,7 @@ public class TestMultiParallel { LOG.info("test=testBadFam"); Table table = UTIL.getConnection().getTable(TEST_TABLE); - List<Row> actions = new ArrayList<Row>(); + List<Row> actions = new ArrayList<>(); Put p = new Put(Bytes.toBytes("row1")); p.addColumn(Bytes.toBytes("bad_family"), Bytes.toBytes("qual"), Bytes.toBytes("value")); actions.add(p); @@ -368,7 +368,7 @@ public class TestMultiParallel { validateSizeAndEmpty(results, KEYS.length); // Deletes - List<Row> deletes = new ArrayList<Row>(); + List<Row> deletes = new ArrayList<>(); for (int i = 0; i < KEYS.length; i++) { Delete delete = new Delete(KEYS[i]); delete.addFamily(BYTES_FAMILY); @@ -399,7 +399,7 @@ public class TestMultiParallel { validateSizeAndEmpty(results, KEYS.length); // Deletes - ArrayList<Delete> deletes = new ArrayList<Delete>(); + ArrayList<Delete> deletes = new ArrayList<>(); for (int i = 0; i < KEYS.length; i++) { Delete delete = new Delete(KEYS[i]); delete.addFamily(BYTES_FAMILY); @@ -422,7 +422,7 @@ public class TestMultiParallel { LOG.info("test=testBatchWithManyColsInOneRowGetAndPut"); Table table = UTIL.getConnection().getTable(TEST_TABLE); - List<Row> puts = new ArrayList<Row>(); + List<Row> puts = new ArrayList<>(); for (int i = 0; i < 100; i++) { Put put = new Put(ONE_ROW); byte[] qual = Bytes.toBytes("column" + i); @@ -436,7 +436,7 @@ public class TestMultiParallel { validateSizeAndEmpty(results, 100); // get the data back and validate that it is correct - List<Row> gets = new ArrayList<Row>(); + List<Row> gets = new ArrayList<>(); for (int i = 0; i < 100; i++) { Get get = new Get(ONE_ROW); byte[] qual = Bytes.toBytes("column" + i); @@ -478,7 +478,7 @@ public class TestMultiParallel { Append a = new Append(ONE_ROW); a.add(BYTES_FAMILY, QUAL1, Bytes.toBytes("def")); a.add(BYTES_FAMILY, QUAL4, Bytes.toBytes("xyz")); - List<Row> actions = new ArrayList<Row>(); + List<Row> actions = new ArrayList<>(); actions.add(inc); actions.add(a); @@ -604,7 +604,7 @@ public class TestMultiParallel { // Batch: get, get, put(new col), delete, get, get of put, get of deleted, // put - List<Row> actions = new ArrayList<Row>(); + List<Row> actions = new ArrayList<>(); byte[] qual2 = Bytes.toBytes("qual2"); byte[] val2 = Bytes.toBytes("putvalue2"); @@ -693,7 +693,7 @@ public class TestMultiParallel { private void validateLoadedData(Table table) throws IOException { // get the data back and validate that it is correct LOG.info("Validating data on " + table); - List<Get> gets = new ArrayList<Get>(); + List<Get> gets = new ArrayList<>(); for (byte[] k : KEYS) { Get get = new Get(k); get.addColumn(BYTES_FAMILY, QUALIFIER); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestReplicaWithCluster.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestReplicaWithCluster.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestReplicaWithCluster.java index 22e88da..becb2eb 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestReplicaWithCluster.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestReplicaWithCluster.java @@ -377,12 +377,12 @@ public class TestReplicaWithCluster { final int numRows = 10; final byte[] qual = Bytes.toBytes("qual"); final byte[] val = Bytes.toBytes("val"); - final List<Pair<byte[], String>> famPaths = new ArrayList<Pair<byte[], String>>(); + final List<Pair<byte[], String>> famPaths = new ArrayList<>(); for (HColumnDescriptor col : hdt.getColumnFamilies()) { Path hfile = new Path(dir, col.getNameAsString()); TestHRegionServerBulkLoad.createHFile(HTU.getTestFileSystem(), hfile, col.getName(), qual, val, numRows); - famPaths.add(new Pair<byte[], String>(col.getName(), hfile.toString())); + famPaths.add(new Pair<>(col.getName(), hfile.toString())); } // bulk load HFiles http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestReplicasClient.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestReplicasClient.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestReplicasClient.java index 7cbb5ad..7b4442b 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestReplicasClient.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestReplicasClient.java @@ -103,7 +103,7 @@ public class TestReplicasClient { static final AtomicBoolean slowDownNext = new AtomicBoolean(false); static final AtomicInteger countOfNext = new AtomicInteger(0); private static final AtomicReference<CountDownLatch> cdl = - new AtomicReference<CountDownLatch>(new CountDownLatch(0)); + new AtomicReference<>(new CountDownLatch(0)); Random r = new Random(); public SlowMeCopro() { } @@ -530,7 +530,7 @@ public class TestReplicasClient { public void testCancelOfMultiGet() throws Exception { openRegion(hriSecondary); try { - List<Put> puts = new ArrayList<Put>(2); + List<Put> puts = new ArrayList<>(2); byte[] b1 = Bytes.toBytes("testCancelOfMultiGet" + 0); Put p = new Put(b1); p.addColumn(f, b1, b1); @@ -552,7 +552,7 @@ public class TestReplicasClient { // Make primary slowdown SlowMeCopro.getCdl().set(new CountDownLatch(1)); - List<Get> gets = new ArrayList<Get>(); + List<Get> gets = new ArrayList<>(); Get g = new Get(b1); g.setCheckExistenceOnly(true); g.setConsistency(Consistency.TIMELINE); @@ -762,7 +762,7 @@ public class TestReplicasClient { Iterator<Result> iter = scanner.iterator(); // Maps of row keys that we have seen so far - HashMap<String, Boolean> map = new HashMap<String, Boolean>(); + HashMap<String, Boolean> map = new HashMap<>(); // Tracked metrics int rowCount = 0; http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestRestoreSnapshotFromClient.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestRestoreSnapshotFromClient.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestRestoreSnapshotFromClient.java index ec6e020..3190fb9 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestRestoreSnapshotFromClient.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestRestoreSnapshotFromClient.java @@ -304,7 +304,7 @@ public class TestRestoreSnapshotFromClient { private Set<String> getFamiliesFromFS(final TableName tableName) throws IOException { MasterFileSystem mfs = TEST_UTIL.getMiniHBaseCluster().getMaster().getMasterFileSystem(); - Set<String> families = new HashSet<String>(); + Set<String> families = new HashSet<>(); Path tableDir = FSUtils.getTableDir(mfs.getRootDir(), tableName); for (Path regionDir: FSUtils.getRegionDirs(mfs.getFileSystem(), tableDir)) { for (Path familyDir: FSUtils.getFamilyDirs(mfs.getFileSystem(), regionDir)) { http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestScannersFromClientSide.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestScannersFromClientSide.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestScannersFromClientSide.java index 42fecfb..6f40093 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestScannersFromClientSide.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestScannersFromClientSide.java @@ -151,7 +151,7 @@ public class TestScannersFromClientSide { scanner = ht.getScanner(scan); // c4:4, c5:5, c6:6, c7:7 - kvListExp = new ArrayList<Cell>(); + kvListExp = new ArrayList<>(); kvListExp.add(new KeyValue(ROW, FAMILY, QUALIFIERS[4], 4, VALUE)); kvListExp.add(new KeyValue(ROW, FAMILY, QUALIFIERS[5], 5, VALUE)); kvListExp.add(new KeyValue(ROW, FAMILY, QUALIFIERS[6], 6, VALUE)); @@ -166,14 +166,14 @@ public class TestScannersFromClientSide { scanner = ht.getScanner(scan); // First batch: c4:4, c5:5 - kvListExp = new ArrayList<Cell>(); + kvListExp = new ArrayList<>(); kvListExp.add(new KeyValue(ROW, FAMILY, QUALIFIERS[4], 4, VALUE)); kvListExp.add(new KeyValue(ROW, FAMILY, QUALIFIERS[5], 5, VALUE)); result = scanner.next(); verifyResult(result, kvListExp, toLog, "Testing first batch of scan"); // Second batch: c6:6, c7:7 - kvListExp = new ArrayList<Cell>(); + kvListExp = new ArrayList<>(); kvListExp.add(new KeyValue(ROW, FAMILY, QUALIFIERS[6], 6, VALUE)); kvListExp.add(new KeyValue(ROW, FAMILY, QUALIFIERS[7], 7, VALUE)); result = scanner.next(); @@ -204,7 +204,7 @@ public class TestScannersFromClientSide { byte[] cellValue = Bytes.createMaxByteArray(cellSize); Put put; - List<Put> puts = new ArrayList<Put>(); + List<Put> puts = new ArrayList<>(); for (int row = 0; row < ROWS.length; row++) { put = new Put(ROWS[row]); for (int qual = 0; qual < QUALIFIERS.length; qual++) { @@ -244,7 +244,7 @@ public class TestScannersFromClientSide { Table ht = TEST_UTIL.createTable(tableName, FAMILY); Put put; - List<Put> puts = new ArrayList<Put>(); + List<Put> puts = new ArrayList<>(); for (int row = 0; row < ROWS.length; row++) { put = new Put(ROWS[row]); for (int qual = 0; qual < QUALIFIERS.length; qual++) { @@ -327,7 +327,7 @@ public class TestScannersFromClientSide { boolean toLog = true; List<Cell> kvListExp; - kvListExp = new ArrayList<Cell>(); + kvListExp = new ArrayList<>(); // Insert one CF for row[0] put = new Put(ROW); for (int i=0; i < 10; i++) { @@ -344,7 +344,7 @@ public class TestScannersFromClientSide { get = new Get(ROW); get.setMaxResultsPerColumnFamily(2); result = ht.get(get); - kvListExp = new ArrayList<Cell>(); + kvListExp = new ArrayList<>(); kvListExp.add(new KeyValue(ROW, FAMILIES[0], QUALIFIERS[0], 1, VALUE)); kvListExp.add(new KeyValue(ROW, FAMILIES[0], QUALIFIERS[1], 1, VALUE)); verifyResult(result, kvListExp, toLog, "Testing basic setMaxResults"); @@ -355,7 +355,7 @@ public class TestScannersFromClientSide { get.setFilter(new ColumnRangeFilter(QUALIFIERS[2], true, QUALIFIERS[5], true)); result = ht.get(get); - kvListExp = new ArrayList<Cell>(); + kvListExp = new ArrayList<>(); kvListExp.add(new KeyValue(ROW, FAMILIES[0], QUALIFIERS[2], 1, VALUE)); kvListExp.add(new KeyValue(ROW, FAMILIES[0], QUALIFIERS[3], 1, VALUE)); kvListExp.add(new KeyValue(ROW, FAMILIES[0], QUALIFIERS[4], 1, VALUE)); @@ -383,7 +383,7 @@ public class TestScannersFromClientSide { get.addFamily(FAMILIES[1]); get.addFamily(FAMILIES[2]); result = ht.get(get); - kvListExp = new ArrayList<Cell>(); + kvListExp = new ArrayList<>(); //Exp: CF1:q0, ..., q9, CF2: q0, q1, q10, q11, ..., q19 for (int i=0; i < 10; i++) { kvListExp.add(new KeyValue(ROW, FAMILIES[1], QUALIFIERS[i], 1, VALUE)); @@ -401,7 +401,7 @@ public class TestScannersFromClientSide { get.setMaxResultsPerColumnFamily(3); get.setFilter(new ColumnRangeFilter(QUALIFIERS[2], true, null, true)); result = ht.get(get); - kvListExp = new ArrayList<Cell>(); + kvListExp = new ArrayList<>(); for (int i=2; i < 5; i++) { kvListExp.add(new KeyValue(ROW, FAMILIES[0], QUALIFIERS[i], 1, VALUE)); } @@ -417,7 +417,7 @@ public class TestScannersFromClientSide { get.setMaxResultsPerColumnFamily(7); get.setFilter(new ColumnPrefixFilter(QUALIFIERS[1])); result = ht.get(get); - kvListExp = new ArrayList<Cell>(); + kvListExp = new ArrayList<>(); kvListExp.add(new KeyValue(ROW, FAMILIES[0], QUALIFIERS[1], 1, VALUE)); kvListExp.add(new KeyValue(ROW, FAMILIES[1], QUALIFIERS[1], 1, VALUE)); kvListExp.add(new KeyValue(ROW, FAMILIES[2], QUALIFIERS[1], 1, VALUE)); @@ -448,7 +448,7 @@ public class TestScannersFromClientSide { boolean toLog = true; List<Cell> kvListExp, kvListScan; - kvListExp = new ArrayList<Cell>(); + kvListExp = new ArrayList<>(); for (int r=0; r < ROWS.length; r++) { put = new Put(ROWS[r]); @@ -467,7 +467,7 @@ public class TestScannersFromClientSide { scan = new Scan(); scan.setMaxResultsPerColumnFamily(4); ResultScanner scanner = ht.getScanner(scan); - kvListScan = new ArrayList<Cell>(); + kvListScan = new ArrayList<>(); while ((result = scanner.next()) != null) { for (Cell kv : result.listCells()) { kvListScan.add(kv); @@ -498,7 +498,7 @@ public class TestScannersFromClientSide { List<Cell> kvListExp; // Insert one CF for row - kvListExp = new ArrayList<Cell>(); + kvListExp = new ArrayList<>(); put = new Put(ROW); for (int i=0; i < 10; i++) { KeyValue kv = new KeyValue(ROW, FAMILIES[0], QUALIFIERS[i], 1, VALUE); @@ -519,7 +519,7 @@ public class TestScannersFromClientSide { get = new Get(ROW); get.setRowOffsetPerColumnFamily(20); result = ht.get(get); - kvListExp = new ArrayList<Cell>(); + kvListExp = new ArrayList<>(); verifyResult(result, kvListExp, toLog, "Testing offset > #kvs"); //offset + maxResultPerCF @@ -527,7 +527,7 @@ public class TestScannersFromClientSide { get.setRowOffsetPerColumnFamily(4); get.setMaxResultsPerColumnFamily(5); result = ht.get(get); - kvListExp = new ArrayList<Cell>(); + kvListExp = new ArrayList<>(); for (int i=4; i < 9; i++) { kvListExp.add(new KeyValue(ROW, FAMILIES[0], QUALIFIERS[i], 1, VALUE)); } @@ -540,7 +540,7 @@ public class TestScannersFromClientSide { get.setFilter(new ColumnRangeFilter(QUALIFIERS[2], true, QUALIFIERS[5], true)); result = ht.get(get); - kvListExp = new ArrayList<Cell>(); + kvListExp = new ArrayList<>(); kvListExp.add(new KeyValue(ROW, FAMILIES[0], QUALIFIERS[3], 1, VALUE)); kvListExp.add(new KeyValue(ROW, FAMILIES[0], QUALIFIERS[4], 1, VALUE)); kvListExp.add(new KeyValue(ROW, FAMILIES[0], QUALIFIERS[5], 1, VALUE)); @@ -563,7 +563,7 @@ public class TestScannersFromClientSide { get.addFamily(FAMILIES[1]); get.addFamily(FAMILIES[2]); result = ht.get(get); - kvListExp = new ArrayList<Cell>(); + kvListExp = new ArrayList<>(); //Exp: CF1:q4, q5, CF2: q4, q5 kvListExp.add(new KeyValue(ROW, FAMILIES[1], QUALIFIERS[4], 1, VALUE)); kvListExp.add(new KeyValue(ROW, FAMILIES[1], QUALIFIERS[5], 1, VALUE)); @@ -644,7 +644,7 @@ public class TestScannersFromClientSide { } // c0:0, c1:1 - kvListExp = new ArrayList<Cell>(); + kvListExp = new ArrayList<>(); kvListExp.add(new KeyValue(ROW, FAMILY, QUALIFIERS[0], 0, VALUE)); kvListExp.add(new KeyValue(ROW, FAMILY, QUALIFIERS[1], 1, VALUE)); result = scanner.next(); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestSnapshotFromClient.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestSnapshotFromClient.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestSnapshotFromClient.java index 3d8ee55..f2f3b26 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestSnapshotFromClient.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestSnapshotFromClient.java @@ -337,7 +337,7 @@ public class TestSnapshotFromClient { LOG.debug(table2Snapshot1 + " completed."); List<SnapshotDescription> listTableSnapshots = admin.listTableSnapshots("test.*", ".*"); - List<String> listTableSnapshotNames = new ArrayList<String>(); + List<String> listTableSnapshotNames = new ArrayList<>(); assertEquals(3, listTableSnapshots.size()); for (SnapshotDescription s : listTableSnapshots) { listTableSnapshotNames.add(s.getName()); @@ -379,7 +379,7 @@ public class TestSnapshotFromClient { List<SnapshotDescription> listTableSnapshots = admin.listTableSnapshots("test.*", "Table1.*"); - List<String> listTableSnapshotNames = new ArrayList<String>(); + List<String> listTableSnapshotNames = new ArrayList<>(); assertEquals(2, listTableSnapshots.size()); for (SnapshotDescription s : listTableSnapshots) { listTableSnapshotNames.add(s.getName()); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestSnapshotMetadata.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestSnapshotMetadata.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestSnapshotMetadata.java index 55086b5..99c4340 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestSnapshotMetadata.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestSnapshotMetadata.java @@ -193,7 +193,7 @@ public class TestSnapshotMetadata { final byte[] snapshotName = Bytes.toBytes(snapshotNameAsString); // restore the snapshot into a cloned table and examine the output - List<byte[]> familiesList = new ArrayList<byte[]>(); + List<byte[]> familiesList = new ArrayList<>(); Collections.addAll(familiesList, families); // Create a snapshot in which all families are empty @@ -262,8 +262,8 @@ public class TestSnapshotMetadata { // populate it with data final byte[] familyForUpdate = BLOCKSIZE_FAM; - List<byte[]> familiesWithDataList = new ArrayList<byte[]>(); - List<byte[]> emptyFamiliesList = new ArrayList<byte[]>(); + List<byte[]> familiesWithDataList = new ArrayList<>(); + List<byte[]> emptyFamiliesList = new ArrayList<>(); if (addData) { Table original = UTIL.getConnection().getTable(originalTableName); UTIL.loadTable(original, familyForUpdate); // family arbitrarily chosen http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestTimestampsFilter.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestTimestampsFilter.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestTimestampsFilter.java index 5c487d7..0a1fafe 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestTimestampsFilter.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestTimestampsFilter.java @@ -145,7 +145,7 @@ public class TestTimestampsFilter { // Request an empty list of versions using the Timestamps filter; // Should return none. - kvs = getNVersions(ht, FAMILY, 2, 2, new ArrayList<Long>()); + kvs = getNVersions(ht, FAMILY, 2, 2, new ArrayList<>()); assertEquals(0, kvs == null? 0: kvs.length); // @@ -192,7 +192,7 @@ public class TestTimestampsFilter { p.addColumn(FAMILY, Bytes.toBytes("column4"), (long) 3, Bytes.toBytes("value4-3")); ht.put(p); - ArrayList<Long> timestamps = new ArrayList<Long>(); + ArrayList<Long> timestamps = new ArrayList<>(); timestamps.add(new Long(3)); TimestampsFilter filter = new TimestampsFilter(timestamps); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/client/replication/TestReplicationAdmin.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/replication/TestReplicationAdmin.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/replication/TestReplicationAdmin.java index 481f311..f092a48 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/replication/TestReplicationAdmin.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/replication/TestReplicationAdmin.java @@ -255,7 +255,7 @@ public class TestReplicationAdmin { // append table column family: f1 of t3 to replication tableCFs.clear(); - tableCFs.put(tableName3, new ArrayList<String>()); + tableCFs.put(tableName3, new ArrayList<>()); tableCFs.get(tableName3).add("f1"); admin.appendPeerTableCFs(ID_ONE, tableCFs); result = ReplicationSerDeHelper.parseTableCFsFromConfig(admin.getPeerTableCFs(ID_ONE)); @@ -269,7 +269,7 @@ public class TestReplicationAdmin { assertEquals("f1", result.get(tableName3).get(0)); tableCFs.clear(); - tableCFs.put(tableName4, new ArrayList<String>()); + tableCFs.put(tableName4, new ArrayList<>()); tableCFs.get(tableName4).add("f1"); tableCFs.get(tableName4).add("f2"); admin.appendPeerTableCFs(ID_ONE, tableCFs); @@ -289,10 +289,10 @@ public class TestReplicationAdmin { // append "table5" => [], then append "table5" => ["f1"] tableCFs.clear(); - tableCFs.put(tableName5, new ArrayList<String>()); + tableCFs.put(tableName5, new ArrayList<>()); admin.appendPeerTableCFs(ID_ONE, tableCFs); tableCFs.clear(); - tableCFs.put(tableName5, new ArrayList<String>()); + tableCFs.put(tableName5, new ArrayList<>()); tableCFs.get(tableName5).add("f1"); admin.appendPeerTableCFs(ID_ONE, tableCFs); result = ReplicationSerDeHelper.parseTableCFsFromConfig(admin.getPeerTableCFs(ID_ONE)); @@ -303,11 +303,11 @@ public class TestReplicationAdmin { // append "table6" => ["f1"], then append "table6" => [] tableCFs.clear(); - tableCFs.put(tableName6, new ArrayList<String>()); + tableCFs.put(tableName6, new ArrayList<>()); tableCFs.get(tableName6).add("f1"); admin.appendPeerTableCFs(ID_ONE, tableCFs); tableCFs.clear(); - tableCFs.put(tableName6, new ArrayList<String>()); + tableCFs.put(tableName6, new ArrayList<>()); admin.appendPeerTableCFs(ID_ONE, tableCFs); result = ReplicationSerDeHelper.parseTableCFsFromConfig(admin.getPeerTableCFs(ID_ONE)); assertEquals(6, result.size()); @@ -339,7 +339,7 @@ public class TestReplicationAdmin { tableCFs.clear(); tableCFs.put(tableName1, null); - tableCFs.put(tableName2, new ArrayList<String>()); + tableCFs.put(tableName2, new ArrayList<>()); tableCFs.get(tableName2).add("cf1"); admin.setPeerTableCFs(ID_ONE, tableCFs); try { @@ -360,7 +360,7 @@ public class TestReplicationAdmin { try { tableCFs.clear(); - tableCFs.put(tableName1, new ArrayList<String>()); + tableCFs.put(tableName1, new ArrayList<>()); tableCFs.get(tableName1).add("f1"); admin.removePeerTableCFs(ID_ONE, tableCFs); assertTrue(false); @@ -382,13 +382,13 @@ public class TestReplicationAdmin { } catch (ReplicationException e) { } tableCFs.clear(); - tableCFs.put(tableName2, new ArrayList<String>()); + tableCFs.put(tableName2, new ArrayList<>()); tableCFs.get(tableName2).add("cf1"); admin.removePeerTableCFs(ID_ONE, tableCFs); assertNull(admin.getPeerTableCFs(ID_ONE)); tableCFs.clear(); - tableCFs.put(tableName4, new ArrayList<String>()); + tableCFs.put(tableName4, new ArrayList<>()); admin.setPeerTableCFs(ID_ONE, tableCFs); admin.removePeerTableCFs(ID_ONE, tableCFs); assertNull(admin.getPeerTableCFs(ID_ONE)); @@ -407,7 +407,7 @@ public class TestReplicationAdmin { admin.peerAdded(ID_ONE); rpc = admin.getPeerConfig(ID_ONE); - Set<String> namespaces = new HashSet<String>(); + Set<String> namespaces = new HashSet<>(); namespaces.add(ns1); namespaces.add(ns2); rpc.setNamespaces(namespaces); @@ -448,7 +448,7 @@ public class TestReplicationAdmin { admin.updatePeerConfig(ID_ONE, rpc); rpc = admin.getPeerConfig(ID_ONE); Map<TableName, List<String>> tableCfs = new HashMap<>(); - tableCfs.put(tableName1, new ArrayList<String>()); + tableCfs.put(tableName1, new ArrayList<>()); rpc.setTableCFsMap(tableCfs); try { admin.updatePeerConfig(ID_ONE, rpc); @@ -460,7 +460,7 @@ public class TestReplicationAdmin { rpc = admin.getPeerConfig(ID_ONE); tableCfs.clear(); - tableCfs.put(tableName2, new ArrayList<String>()); + tableCfs.put(tableName2, new ArrayList<>()); rpc.setTableCFsMap(tableCfs); admin.updatePeerConfig(ID_ONE, rpc); rpc = admin.getPeerConfig(ID_ONE); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/client/replication/TestReplicationAdminWithClusters.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/replication/TestReplicationAdminWithClusters.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/replication/TestReplicationAdminWithClusters.java index 24889ad..b44ecbf 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/replication/TestReplicationAdminWithClusters.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/replication/TestReplicationAdminWithClusters.java @@ -183,8 +183,7 @@ public class TestReplicationAdminWithClusters extends TestReplicationBase { } assertFalse("Table should not exists in the peer cluster", admin2.isTableAvailable(TestReplicationBase.tableName)); - Map<TableName, ? extends Collection<String>> tableCfs = - new HashMap<TableName, Collection<String>>(); + Map<TableName, ? extends Collection<String>> tableCfs = new HashMap<>(); tableCfs.put(tableName, null); try { adminExt.setPeerTableCFs(peerId, tableCfs); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/constraint/TestConstraints.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/constraint/TestConstraints.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/constraint/TestConstraints.java index acc3fca..12a229d 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/constraint/TestConstraints.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/constraint/TestConstraints.java @@ -75,11 +75,9 @@ public class TestConstraints { @Test public void testReadWriteWithConf() throws Throwable { HTableDescriptor desc = new HTableDescriptor(TableName.valueOf(name.getMethodName())); - Constraints.add( - desc, - new Pair<Class<? extends Constraint>, Configuration>( - CheckConfigurationConstraint.class, CheckConfigurationConstraint - .getConfiguration())); + Constraints.add(desc, + new Pair<>(CheckConfigurationConstraint.class, + CheckConfigurationConstraint.getConfiguration())); List<? extends Constraint> c = Constraints.getConstraints(desc, this .getClass().getClassLoader()); @@ -88,7 +86,7 @@ public class TestConstraints { assertEquals(CheckConfigurationConstraint.class, c.get(0).getClass()); // check to make sure that we overwrite configurations - Constraints.add(desc, new Pair<Class<? extends Constraint>, Configuration>( + Constraints.add(desc, new Pair<>( CheckConfigurationConstraint.class, new Configuration(false))); try { http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestCoprocessorInterface.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestCoprocessorInterface.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestCoprocessorInterface.java index 21d9861..422c54b 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestCoprocessorInterface.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestCoprocessorInterface.java @@ -349,7 +349,7 @@ public class TestCoprocessorInterface { RegionScanner scanner = region.getCoprocessorHost().postScannerOpen(s, region.getScanner(s)); assertTrue(scanner instanceof CustomScanner); // this would throw an exception before HBASE-4197 - scanner.next(new ArrayList<Cell>()); + scanner.next(new ArrayList<>()); HBaseTestingUtility.closeRegionAndWAL(region); Coprocessor c = region.getCoprocessorHost(). http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestHTableWrapper.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestHTableWrapper.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestHTableWrapper.java index 04d8c8c..9f20ba2 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestHTableWrapper.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestHTableWrapper.java @@ -221,7 +221,7 @@ public class TestHTableWrapper { // multiple deletes: Delete[] deletes = new Delete[] { new Delete(ROW_D), new Delete(ROW_E) }; - hTableInterface.delete(new ArrayList<Delete>(Arrays.asList(deletes))); + hTableInterface.delete(new ArrayList<>(Arrays.asList(deletes))); checkRowsValues(new byte[][] { ROW_D, ROW_E }, new byte[][] { null, null }); } http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestOpenTableInCoprocessor.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestOpenTableInCoprocessor.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestOpenTableInCoprocessor.java index b75fc79..1ed0008 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestOpenTableInCoprocessor.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestOpenTableInCoprocessor.java @@ -92,7 +92,7 @@ public class TestOpenTableInCoprocessor { long keepAliveTime = 60; ThreadPoolExecutor pool = new ThreadPoolExecutor(1, maxThreads, keepAliveTime, TimeUnit.SECONDS, - new SynchronousQueue<Runnable>(), Threads.newDaemonThreadFactory("hbase-table")); + new SynchronousQueue<>(), Threads.newDaemonThreadFactory("hbase-table")); pool.allowCoreThreadTimeOut(true); return pool; } http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionObserverBypass.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionObserverBypass.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionObserverBypass.java index fb87ff6..63d7544 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionObserverBypass.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionObserverBypass.java @@ -112,7 +112,7 @@ public class TestRegionObserverBypass { EnvironmentEdgeManagerTestHelper.injectEdge(new IncrementingEnvironmentEdge()); Table t = util.getConnection().getTable(tableName); - List<Put> puts = new ArrayList<Put>(); + List<Put> puts = new ArrayList<>(); Put p = new Put(row1); p.addColumn(dummy, dummy, dummy); puts.add(p); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionObserverInterface.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionObserverInterface.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionObserverInterface.java index 2e0db44..7b4cc40 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionObserverInterface.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionObserverInterface.java @@ -424,7 +424,7 @@ public class TestRegionObserverInterface { @Override public boolean next(List<Cell> results, ScannerContext scannerContext) throws IOException { - List<Cell> internalResults = new ArrayList<Cell>(); + List<Cell> internalResults = new ArrayList<>(); boolean hasMore; do { hasMore = scanner.next(internalResults, scannerContext); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestWALObserver.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestWALObserver.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestWALObserver.java index 2ab91c1..5fb5421 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestWALObserver.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestWALObserver.java @@ -181,8 +181,7 @@ public class TestWALObserver { HRegionInfo hri = createBasic3FamilyHRegionInfo(Bytes.toString(TEST_TABLE)); final HTableDescriptor htd = createBasic3FamilyHTD(Bytes .toString(TEST_TABLE)); - NavigableMap<byte[], Integer> scopes = new TreeMap<byte[], Integer>( - Bytes.BYTES_COMPARATOR); + NavigableMap<byte[], Integer> scopes = new TreeMap<>(Bytes.BYTES_COMPARATOR); for(byte[] fam : htd.getFamiliesKeys()) { scopes.put(fam, 0); } @@ -272,8 +271,7 @@ public class TestWALObserver { final HRegionInfo hri = createBasic3FamilyHRegionInfo(Bytes.toString(TEST_TABLE)); final HTableDescriptor htd = createBasic3FamilyHTD(Bytes.toString(TEST_TABLE)); final MultiVersionConcurrencyControl mvcc = new MultiVersionConcurrencyControl(); - NavigableMap<byte[], Integer> scopes = new TreeMap<byte[], Integer>( - Bytes.BYTES_COMPARATOR); + NavigableMap<byte[], Integer> scopes = new TreeMap<>(Bytes.BYTES_COMPARATOR); for(byte[] fam : htd.getFamiliesKeys()) { scopes.put(fam, 0); } @@ -328,8 +326,7 @@ public class TestWALObserver { WALEdit edit = new WALEdit(); long now = EnvironmentEdgeManager.currentTime(); final int countPerFamily = 1000; - NavigableMap<byte[], Integer> scopes = new TreeMap<byte[], Integer>( - Bytes.BYTES_COMPARATOR); + NavigableMap<byte[], Integer> scopes = new TreeMap<>(Bytes.BYTES_COMPARATOR); for (HColumnDescriptor hcd : htd.getFamilies()) { scopes.put(hcd.getName(), 0); }