http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestStoreScanner.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestStoreScanner.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestStoreScanner.java index 52efe63..ccbf067 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestStoreScanner.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestStoreScanner.java @@ -173,7 +173,7 @@ public class TestStoreScanner { * @return */ NavigableSet<byte[]> getCols(String ...strCols) { - NavigableSet<byte[]> cols = new TreeSet<byte[]>(Bytes.BYTES_COMPARATOR); + NavigableSet<byte[]> cols = new TreeSet<>(Bytes.BYTES_COMPARATOR); for (String col : strCols) { byte[] bytes = Bytes.toBytes(col); cols.add(bytes); @@ -189,7 +189,7 @@ public class TestStoreScanner { Scan scan = new Scan(get); CellGridStoreScanner scanner = new CellGridStoreScanner(scan, this.scanInfo, this.scanType); try { - List<Cell> results = new ArrayList<Cell>(); + List<Cell> results = new ArrayList<>(); while (scanner.next(results)) { continue; } @@ -213,7 +213,7 @@ public class TestStoreScanner { Scan scan = new Scan(get); CellGridStoreScanner scanner = new CellGridStoreScanner(scan, this.scanInfo, this.scanType); try { - List<Cell> results = new ArrayList<Cell>(); + List<Cell> results = new ArrayList<>(); while (scanner.next(results)) { continue; } @@ -242,7 +242,7 @@ public class TestStoreScanner { scan.addColumn(CF, ONE); CellGridStoreScanner scanner = new CellGridStoreScanner(scan, this.scanInfo, this.scanType); try { - List<Cell> results = new ArrayList<Cell>(); + List<Cell> results = new ArrayList<>(); while (scanner.next(results)) { continue; } @@ -276,7 +276,7 @@ public class TestStoreScanner { Scan scan = new Scan(get); CellGridStoreScanner scanner = new CellGridStoreScanner(scan, this.scanInfo, this.scanType); try { - List<Cell> results = new ArrayList<Cell>(); + List<Cell> results = new ArrayList<>(); // For a Get there should be no more next's after the first call. Assert.assertEquals(false, scanner.next(results)); // Should be one result only. @@ -307,7 +307,7 @@ public class TestStoreScanner { Scan scan = new Scan(get); CellGridStoreScanner scanner = new CellGridStoreScanner(scan, this.scanInfo, this.scanType); try { - List<Cell> results = new ArrayList<Cell>(); + List<Cell> results = new ArrayList<>(); // For a Get there should be no more next's after the first call. Assert.assertEquals(false, scanner.next(results)); // Should be one result only. @@ -341,7 +341,7 @@ public class TestStoreScanner { List<Cell> results = null; try (StoreScanner scan = new StoreScanner(scanSpec, scanInfo, scanType, getCols("a"), scanners)) { - results = new ArrayList<Cell>(); + results = new ArrayList<>(); Assert.assertEquals(true, scan.next(results)); Assert.assertEquals(5, results.size()); Assert.assertEquals(kvs[kvs.length - 1], results.get(0)); @@ -352,7 +352,7 @@ public class TestStoreScanner { scanSpec.setMaxVersions(); try (StoreScanner scan = new StoreScanner(scanSpec, scanInfo, scanType, getCols("a"), scanners)) { - results = new ArrayList<Cell>(); + results = new ArrayList<>(); Assert.assertEquals(true, scan.next(results)); Assert.assertEquals(2, results.size()); } @@ -362,7 +362,7 @@ public class TestStoreScanner { scanSpec.setMaxVersions(); try (StoreScanner scan = new StoreScanner(scanSpec, scanInfo, scanType, getCols("a"), scanners)) { - results = new ArrayList<Cell>(); + results = new ArrayList<>(); Assert.assertEquals(true, scan.next(results)); Assert.assertEquals(1, results.size()); } @@ -373,7 +373,7 @@ public class TestStoreScanner { scanSpec.setMaxVersions(3); try (StoreScanner scan = new StoreScanner(scanSpec, scanInfo, scanType, getCols("a"), scanners)) { - results = new ArrayList<Cell>(); + results = new ArrayList<>(); Assert.assertEquals(true, scan.next(results)); Assert.assertEquals(3, results.size()); } @@ -395,7 +395,7 @@ public class TestStoreScanner { // this only uses maxVersions (default=1) and TimeRange (default=all) try (StoreScanner scan = new StoreScanner(scanSpec, scanInfo, scanType, getCols("a"), scanners)) { - List<Cell> results = new ArrayList<Cell>(); + List<Cell> results = new ArrayList<>(); Assert.assertEquals(true, scan.next(results)); Assert.assertEquals(1, results.size()); Assert.assertEquals(kvs[0], results.get(0)); @@ -423,7 +423,7 @@ public class TestStoreScanner { // this only uses maxVersions (default=1) and TimeRange (default=all) try (StoreScanner scan = new StoreScanner(scanSpec, scanInfo, scanType, getCols("a"), scanners)) { - List<Cell> results = new ArrayList<Cell>(); + List<Cell> results = new ArrayList<>(); scan.next(results); Assert.assertEquals(1, results.size()); Assert.assertEquals(kvs[0], results.get(0)); @@ -451,7 +451,7 @@ public class TestStoreScanner { Scan scanSpec = new Scan(Bytes.toBytes("R1")); try (StoreScanner scan = new StoreScanner(scanSpec, scanInfo, scanType, getCols("a"), scanners)) { - List<Cell> results = new ArrayList<Cell>(); + List<Cell> results = new ArrayList<>(); Assert.assertFalse(scan.next(results)); Assert.assertEquals(0, results.size()); } @@ -472,7 +472,7 @@ public class TestStoreScanner { Scan scanSpec = new Scan(Bytes.toBytes("R1")); try (StoreScanner scan = new StoreScanner(scanSpec, scanInfo, scanType, getCols("a"), scanners)) { - List<Cell> results = new ArrayList<Cell>(); + List<Cell> results = new ArrayList<>(); Assert.assertEquals(true, scan.next(results)); Assert.assertEquals(0, results.size()); @@ -499,7 +499,7 @@ public class TestStoreScanner { try (StoreScanner scan = new StoreScanner(new Scan(Bytes.toBytes("R1")), scanInfo, scanType, getCols("a"), scanners)) { - List<Cell> results = new ArrayList<Cell>(); + List<Cell> results = new ArrayList<>(); // the two put at ts=now will be masked by the 1 delete, and // since the scan default returns 1 version we'll return the newest // key, which is kvs[2], now-100. @@ -525,7 +525,7 @@ public class TestStoreScanner { Scan scanSpec = new Scan(Bytes.toBytes("R1")).setMaxVersions(2); try (StoreScanner scan = new StoreScanner(scanSpec, scanInfo, scanType, getCols("a"), scanners)) { - List<Cell> results = new ArrayList<Cell>(); + List<Cell> results = new ArrayList<>(); Assert.assertEquals(true, scan.next(results)); Assert.assertEquals(2, results.size()); Assert.assertEquals(kvs2[1], results.get(0)); @@ -543,7 +543,7 @@ public class TestStoreScanner { List<KeyValueScanner> scanners = scanFixture(kvs); try (StoreScanner scan = new StoreScanner(new Scan(Bytes.toBytes("R1")), scanInfo, scanType, null, scanners)) { - List<Cell> results = new ArrayList<Cell>(); + List<Cell> results = new ArrayList<>(); Assert.assertEquals(true, scan.next(results)); Assert.assertEquals(2, results.size()); Assert.assertEquals(kvs[0], results.get(0)); @@ -574,7 +574,7 @@ public class TestStoreScanner { List<KeyValueScanner> scanners = scanFixture(kvs); try (StoreScanner scan = new StoreScanner(new Scan().setMaxVersions(2), scanInfo, scanType, null, scanners)) { - List<Cell> results = new ArrayList<Cell>(); + List<Cell> results = new ArrayList<>(); Assert.assertEquals(true, scan.next(results)); Assert.assertEquals(5, results.size()); Assert.assertEquals(kvs[0], results.get(0)); @@ -605,7 +605,7 @@ public class TestStoreScanner { try (StoreScanner scan = new StoreScanner(new Scan().setMaxVersions(Integer.MAX_VALUE), scanInfo, scanType, null, scanners)) { - List<Cell> results = new ArrayList<Cell>(); + List<Cell> results = new ArrayList<>(); Assert.assertEquals(true, scan.next(results)); Assert.assertEquals(0, results.size()); Assert.assertEquals(true, scan.next(results)); @@ -627,7 +627,7 @@ public class TestStoreScanner { List<KeyValueScanner> scanners = scanFixture(kvs); try (StoreScanner scan = new StoreScanner(new Scan(), scanInfo, scanType, null, scanners)) { - List<Cell> results = new ArrayList<Cell>(); + List<Cell> results = new ArrayList<>(); Assert.assertEquals(true, scan.next(results)); Assert.assertEquals(1, results.size()); Assert.assertEquals(kvs[3], results.get(0)); @@ -652,7 +652,7 @@ public class TestStoreScanner { List<KeyValueScanner> scanners = scanFixture(kvs); try (StoreScanner scan = new StoreScanner(new Scan(), scanInfo, scanType, getCols("a", "d"), scanners)) { - List<Cell> results = new ArrayList<Cell>(); + List<Cell> results = new ArrayList<>(); Assert.assertEquals(true, scan.next(results)); Assert.assertEquals(2, results.size()); Assert.assertEquals(kvs[0], results.get(0)); @@ -692,7 +692,7 @@ public class TestStoreScanner { CellComparator.COMPARATOR); ScanType scanType = ScanType.USER_SCAN; try (StoreScanner scanner = new StoreScanner(scan, scanInfo, scanType, null, scanners)) { - List<Cell> results = new ArrayList<Cell>(); + List<Cell> results = new ArrayList<>(); Assert.assertEquals(true, scanner.next(results)); Assert.assertEquals(2, results.size()); Assert.assertEquals(kvs[1], results.get(0)); @@ -720,9 +720,9 @@ public class TestStoreScanner { // normally cause an NPE because scan.store is null. So as long as we get through these // two calls we are good and the bug was quashed. - scan.updateReaders(new ArrayList<StoreFile>()); + scan.updateReaders(new ArrayList<>()); - scan.updateReaders(new ArrayList<StoreFile>()); + scan.updateReaders(new ArrayList<>()); scan.peek(); } @@ -767,7 +767,7 @@ public class TestStoreScanner { try (StoreScanner scanner = new StoreScanner(scan, scanInfo, scanType, null, scanners)) { - List<Cell> results = new ArrayList<Cell>(); + List<Cell> results = new ArrayList<>(); Assert.assertEquals(true, scanner.next(results)); Assert.assertEquals(1, results.size()); Assert.assertEquals(kvs[1], results.get(0)); @@ -833,8 +833,8 @@ public class TestStoreScanner { new StoreScanner(scan, scanInfo, ScanType.COMPACT_DROP_DELETES, null, scanners, HConstants.OLDEST_TIMESTAMP)) { - List<Cell> results = new ArrayList<Cell>(); - results = new ArrayList<Cell>(); + List<Cell> results = new ArrayList<>(); + results = new ArrayList<>(); Assert.assertEquals(true, scanner.next(results)); Assert.assertEquals(kvs[0], results.get(0)); Assert.assertEquals(kvs[2], results.get(1));
http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestStripeStoreEngine.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestStripeStoreEngine.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestStripeStoreEngine.java index 3e3eef9..b2739e1 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestStripeStoreEngine.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestStripeStoreEngine.java @@ -78,7 +78,7 @@ public class TestStripeStoreEngine { mockCompactor.compact(any(CompactionRequest.class), anyInt(), anyLong(), any(byte[].class), any(byte[].class), any(byte[].class), any(byte[].class), any(ThroughputController.class), any(User.class))) - .thenReturn(new ArrayList<Path>()); + .thenReturn(new ArrayList<>()); // Produce 3 L0 files. StoreFile sf = createFile(); @@ -118,6 +118,6 @@ public class TestStripeStoreEngine { } private static ArrayList<StoreFile> al(StoreFile... sfs) { - return new ArrayList<StoreFile>(Arrays.asList(sfs)); + return new ArrayList<>(Arrays.asList(sfs)); } } http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestStripeStoreFileManager.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestStripeStoreFileManager.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestStripeStoreFileManager.java index c533257..a6ce270 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestStripeStoreFileManager.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestStripeStoreFileManager.java @@ -115,7 +115,7 @@ public class TestStripeStoreFileManager { } private static ArrayList<StoreFile> dumpIterator(Iterator<StoreFile> iter) { - ArrayList<StoreFile> result = new ArrayList<StoreFile>(); + ArrayList<StoreFile> result = new ArrayList<>(); for (; iter.hasNext(); result.add(iter.next())); return result; } @@ -233,7 +233,7 @@ public class TestStripeStoreFileManager { private void verifySplitPointScenario(int splitPointAfter, boolean shouldSplitStripe, float splitRatioToVerify, int... sizes) throws Exception { assertTrue(sizes.length > 1); - ArrayList<StoreFile> sfs = new ArrayList<StoreFile>(); + ArrayList<StoreFile> sfs = new ArrayList<>(); for (int sizeIx = 0; sizeIx < sizes.length; ++sizeIx) { byte[] startKey = (sizeIx == 0) ? OPEN_KEY : Bytes.toBytes(sizeIx - 1); byte[] endKey = (sizeIx == sizes.length - 1) ? OPEN_KEY : Bytes.toBytes(sizeIx); @@ -525,7 +525,7 @@ public class TestStripeStoreFileManager { sfm.insertNewFiles(al(createFile())); } for (int i = 0; i < filesInStripe; ++i) { - ArrayList<StoreFile> stripe = new ArrayList<StoreFile>(); + ArrayList<StoreFile> stripe = new ArrayList<>(); for (int j = 0; j < stripes; ++j) { stripe.add(createFile( (j == 0) ? OPEN_KEY : keys[j - 1], (j == stripes - 1) ? OPEN_KEY : keys[j])); @@ -597,7 +597,7 @@ public class TestStripeStoreFileManager { } private static StripeStoreFileManager createManager() throws Exception { - return createManager(new ArrayList<StoreFile>()); + return createManager(new ArrayList<>()); } private static StripeStoreFileManager createManager(ArrayList<StoreFile> sfs) throws Exception { @@ -615,11 +615,11 @@ public class TestStripeStoreFileManager { } private static ArrayList<StoreFile> al(StoreFile... sfs) { - return new ArrayList<StoreFile>(Arrays.asList(sfs)); + return new ArrayList<>(Arrays.asList(sfs)); } private static ArrayList<StoreFile> flattenLists(ArrayList<StoreFile>... sfls) { - ArrayList<StoreFile> result = new ArrayList<StoreFile>(); + ArrayList<StoreFile> result = new ArrayList<>(); for (ArrayList<StoreFile> sfl : sfls) { result.addAll(sfl); } http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestTags.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestTags.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestTags.java index 4d531ac..40eebb6 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestTags.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestTags.java @@ -421,7 +421,7 @@ public class TestTags { assertEquals(5L, Bytes.toLong(kv.getValueArray(), kv.getValueOffset(), kv.getValueLength())); assertEquals(2, tags.size()); // We cannot assume the ordering of tags - List<String> tagValues = new ArrayList<String>(); + List<String> tagValues = new ArrayList<>(); for (Tag tag: tags) { tagValues.add(Bytes.toString(TagUtil.cloneValue(tag))); } @@ -557,7 +557,7 @@ public class TestTags { private void updateMutationAddingTags(final Mutation m) { byte[] attribute = m.getAttribute("visibility"); byte[] cf = null; - List<Cell> updatedCells = new ArrayList<Cell>(); + List<Cell> updatedCells = new ArrayList<>(); if (attribute != null) { for (List<? extends Cell> edits : m.getFamilyCellMap().values()) { for (Cell cell : edits) { @@ -566,7 +566,7 @@ public class TestTags { cf = CellUtil.cloneFamily(kv); } Tag tag = new ArrayBackedTag((byte) 1, attribute); - List<Tag> tagList = new ArrayList<Tag>(); + List<Tag> tagList = new ArrayList<>(); tagList.add(tag); KeyValue newKV = new KeyValue(CellUtil.cloneRow(kv), 0, kv.getRowLength(), http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestWALLockup.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestWALLockup.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestWALLockup.java index 4821c74..4f247b0 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestWALLockup.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestWALLockup.java @@ -218,7 +218,7 @@ public class TestWALLockup { HTableDescriptor htd = new HTableDescriptor(TableName.META_TABLE_NAME); final HRegion region = initHRegion(tableName, null, null, dodgyWAL); byte [] bytes = Bytes.toBytes(getName()); - NavigableMap<byte[], Integer> scopes = new TreeMap<byte[], Integer>( + NavigableMap<byte[], Integer> scopes = new TreeMap<>( Bytes.BYTES_COMPARATOR); scopes.put(COLUMN_FAMILY_BYTES, 0); MultiVersionConcurrencyControl mvcc = new MultiVersionConcurrencyControl(); @@ -398,7 +398,7 @@ public class TestWALLockup { HTableDescriptor htd = new HTableDescriptor(TableName.META_TABLE_NAME); final HRegion region = initHRegion(tableName, null, null, dodgyWAL1); byte[] bytes = Bytes.toBytes(getName()); - NavigableMap<byte[], Integer> scopes = new TreeMap<byte[], Integer>( + NavigableMap<byte[], Integer> scopes = new TreeMap<>( Bytes.BYTES_COMPARATOR); scopes.put(COLUMN_FAMILY_BYTES, 0); MultiVersionConcurrencyControl mvcc = new MultiVersionConcurrencyControl(); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestWideScanner.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestWideScanner.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestWideScanner.java index f598a8d..cdf84d2 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestWideScanner.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestWideScanner.java @@ -94,7 +94,7 @@ public class TestWideScanner extends HBaseTestCase { try { this.r = createNewHRegion(TESTTABLEDESC, null, null); int inserted = addWideContent(this.r); - List<Cell> results = new ArrayList<Cell>(); + List<Cell> results = new ArrayList<>(); Scan scan = new Scan(); scan.addFamily(A); scan.addFamily(B); @@ -130,7 +130,7 @@ public class TestWideScanner extends HBaseTestCase { ((HRegion.RegionScannerImpl)s).storeHeap.getHeap().iterator(); while (scanners.hasNext()) { StoreScanner ss = (StoreScanner)scanners.next(); - ss.updateReaders(new ArrayList<StoreFile>()); + ss.updateReaders(new ArrayList<>()); } } while (more); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/ConstantSizeFileListGenerator.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/ConstantSizeFileListGenerator.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/ConstantSizeFileListGenerator.java index 68d57af..5014b41 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/ConstantSizeFileListGenerator.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/ConstantSizeFileListGenerator.java @@ -49,7 +49,7 @@ class ConstantSizeFileListGenerator extends StoreFileListGenerator { @Override public List<StoreFile> next() { count += 1; - ArrayList<StoreFile> files = new ArrayList<StoreFile>(NUM_FILES_GEN); + ArrayList<StoreFile> files = new ArrayList<>(NUM_FILES_GEN); for (int i = 0; i < NUM_FILES_GEN; i++) { files.add(createMockStoreFile(FILESIZE)); } http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/EverythingPolicy.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/EverythingPolicy.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/EverythingPolicy.java index 9a4bb8e..46bb639 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/EverythingPolicy.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/EverythingPolicy.java @@ -46,9 +46,9 @@ public class EverythingPolicy extends RatioBasedCompactionPolicy { final boolean mayUseOffPeak, final boolean mayBeStuck) throws IOException { if (candidates.size() < comConf.getMinFilesToCompact()) { - return new ArrayList<StoreFile>(0); + return new ArrayList<>(0); } - return new ArrayList<StoreFile>(candidates); + return new ArrayList<>(candidates); } } http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/GaussianFileListGenerator.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/GaussianFileListGenerator.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/GaussianFileListGenerator.java index a19e9ad..fb8c30a 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/GaussianFileListGenerator.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/GaussianFileListGenerator.java @@ -47,7 +47,7 @@ class GaussianFileListGenerator extends StoreFileListGenerator { @Override public List<StoreFile> next() { count += 1; - ArrayList<StoreFile> files = new ArrayList<StoreFile>(NUM_FILES_GEN); + ArrayList<StoreFile> files = new ArrayList<>(NUM_FILES_GEN); for (int i = 0; i < NUM_FILES_GEN; i++) { files.add(createMockStoreFile( (int) Math.ceil(Math.max(0, gen.nextNormalizedDouble() * 32 + 32))) http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/MockStoreFileGenerator.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/MockStoreFileGenerator.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/MockStoreFileGenerator.java index 663714a..cb97d27 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/MockStoreFileGenerator.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/MockStoreFileGenerator.java @@ -46,7 +46,7 @@ class MockStoreFileGenerator { } protected List<StoreFile> createStoreFileList(final int[] fs) { - List<StoreFile> storeFiles = new LinkedList<StoreFile>(); + List<StoreFile> storeFiles = new LinkedList<>(); for (int fileSize : fs) { storeFiles.add(createMockStoreFile(fileSize)); } http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/PerfTestCompactionPolicies.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/PerfTestCompactionPolicies.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/PerfTestCompactionPolicies.java index 3fcd3fe..0a84fe9 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/PerfTestCompactionPolicies.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/PerfTestCompactionPolicies.java @@ -82,7 +82,7 @@ public class PerfTestCompactionPolicies extends MockStoreFileGenerator { int[] minFilesValues = new int[] {3}; float[] ratioValues = new float[] {1.2f}; - List<Object[]> params = new ArrayList<Object[]>( + List<Object[]> params = new ArrayList<>( maxFileValues.length * minFilesValues.length * fileListGenClasses.length @@ -152,7 +152,7 @@ public class PerfTestCompactionPolicies extends MockStoreFileGenerator { public final void testSelection() throws Exception { long fileDiff = 0; for (List<StoreFile> storeFileList : generator) { - List<StoreFile> currentFiles = new ArrayList<StoreFile>(18); + List<StoreFile> currentFiles = new ArrayList<>(18); for (StoreFile file : storeFileList) { currentFiles.add(file); currentFiles = runIteration(currentFiles); @@ -175,16 +175,16 @@ public class PerfTestCompactionPolicies extends MockStoreFileGenerator { private List<StoreFile> runIteration(List<StoreFile> startingStoreFiles) throws IOException { - List<StoreFile> storeFiles = new ArrayList<StoreFile>(startingStoreFiles); + List<StoreFile> storeFiles = new ArrayList<>(startingStoreFiles); CompactionRequest req = cp.selectCompaction( - storeFiles, new ArrayList<StoreFile>(), false, false, false); + storeFiles, new ArrayList<>(), false, false, false); long newFileSize = 0; Collection<StoreFile> filesToCompact = req.getFiles(); if (!filesToCompact.isEmpty()) { - storeFiles = new ArrayList<StoreFile>(storeFiles); + storeFiles = new ArrayList<>(storeFiles); storeFiles.removeAll(filesToCompact); for (StoreFile storeFile : filesToCompact) { http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/SemiConstantSizeFileListGenerator.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/SemiConstantSizeFileListGenerator.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/SemiConstantSizeFileListGenerator.java index ed4531a..5fe47f3 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/SemiConstantSizeFileListGenerator.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/SemiConstantSizeFileListGenerator.java @@ -42,7 +42,7 @@ class SemiConstantSizeFileListGenerator extends StoreFileListGenerator { @Override public List<StoreFile> next() { count += 1; - ArrayList<StoreFile> files = new ArrayList<StoreFile>(NUM_FILES_GEN); + ArrayList<StoreFile> files = new ArrayList<>(NUM_FILES_GEN); for (int i = 0; i < NUM_FILES_GEN; i++) { files.add(createMockStoreFile(random.nextInt(5) + 30)); } http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/SinusoidalFileListGenerator.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/SinusoidalFileListGenerator.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/SinusoidalFileListGenerator.java index 6afbb2f..f5f36ac 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/SinusoidalFileListGenerator.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/SinusoidalFileListGenerator.java @@ -43,7 +43,7 @@ class SinusoidalFileListGenerator extends StoreFileListGenerator { @Override public List<StoreFile> next() { count += 1; - ArrayList<StoreFile> files = new ArrayList<StoreFile>(NUM_FILES_GEN); + ArrayList<StoreFile> files = new ArrayList<>(NUM_FILES_GEN); for (int x = 0; x < NUM_FILES_GEN; x++) { int fileSize = (int) Math.abs(64 * Math.sin((Math.PI * x) / 50.0)) + 1; files.add(createMockStoreFile(fileSize)); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/SpikyFileListGenerator.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/SpikyFileListGenerator.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/SpikyFileListGenerator.java index ebaa711..5201eb7 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/SpikyFileListGenerator.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/SpikyFileListGenerator.java @@ -43,7 +43,7 @@ class SpikyFileListGenerator extends StoreFileListGenerator { @Override public List<StoreFile> next() { count += 1; - ArrayList<StoreFile> files = new ArrayList<StoreFile>(NUM_FILES_GEN); + ArrayList<StoreFile> files = new ArrayList<>(NUM_FILES_GEN); for (int x = 0; x < NUM_FILES_GEN; x++) { int fileSize = random.nextInt(5) + 1; if ( x % 10 == 0) { http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/TestCompactedHFilesDischarger.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/TestCompactedHFilesDischarger.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/TestCompactedHFilesDischarger.java index aa5a20e..08fc7bf 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/TestCompactedHFilesDischarger.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/TestCompactedHFilesDischarger.java @@ -77,7 +77,7 @@ public class TestCompactedHFilesDischarger { Path path = testUtil.getDataTestDir(getClass().getSimpleName()); region = HBaseTestingUtility.createRegionAndWAL(info, path, testUtil.getConfiguration(), htd); rss = mock(RegionServerServices.class); - List<Region> regions = new ArrayList<Region>(1); + List<Region> regions = new ArrayList<>(1); regions.add(region); when(rss.getOnlineRegions()).thenReturn(regions); } @@ -379,7 +379,7 @@ public class TestCompactedHFilesDischarger { RegionScanner resScanner = null; try { resScanner = region.getScanner(scan); - List<Cell> results = new ArrayList<Cell>(); + List<Cell> results = new ArrayList<>(); boolean next = resScanner.next(results); try { counter.incrementAndGet(); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/TestCompactor.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/TestCompactor.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/TestCompactor.java index 89f61d0..dff6919 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/TestCompactor.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/TestCompactor.java @@ -82,12 +82,12 @@ public class TestCompactor { public static class StoreFileWritersCapture implements Answer<StoreFileWriter>, StripeMultiFileWriter.WriterFactory { public static class Writer { - public ArrayList<KeyValue> kvs = new ArrayList<KeyValue>(); - public TreeMap<byte[], byte[]> data = new TreeMap<byte[], byte[]>(Bytes.BYTES_COMPARATOR); + public ArrayList<KeyValue> kvs = new ArrayList<>(); + public TreeMap<byte[], byte[]> data = new TreeMap<>(Bytes.BYTES_COMPARATOR); public boolean hasMetadata; } - private List<Writer> writers = new ArrayList<Writer>(); + private List<Writer> writers = new ArrayList<>(); @Override public StoreFileWriter createWriter() throws IOException { @@ -192,7 +192,7 @@ public class TestCompactor { private final ArrayList<KeyValue> kvs; public Scanner(KeyValue... kvs) { - this.kvs = new ArrayList<KeyValue>(Arrays.asList(kvs)); + this.kvs = new ArrayList<>(Arrays.asList(kvs)); } @Override http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/TestDateTieredCompactor.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/TestDateTieredCompactor.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/TestDateTieredCompactor.java index 38d9f99..e590639 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/TestDateTieredCompactor.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/TestDateTieredCompactor.java @@ -156,7 +156,7 @@ public class TestDateTieredCompactor { StoreFileWritersCapture writers = new StoreFileWritersCapture(); CompactionRequest request = createDummyRequest(); DateTieredCompactor dtc = createCompactor(writers, new KeyValue[0], - new ArrayList<StoreFile>(request.getFiles())); + new ArrayList<>(request.getFiles())); List<Path> paths = dtc.compact(request, Arrays.asList(Long.MIN_VALUE, Long.MAX_VALUE), NoLimitThroughputController.INSTANCE, null); assertEquals(1, paths.size()); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/TestStripeCompactionPolicy.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/TestStripeCompactionPolicy.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/TestStripeCompactionPolicy.java index 5fadee8..f2d00b3 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/TestStripeCompactionPolicy.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/TestStripeCompactionPolicy.java @@ -415,7 +415,7 @@ public class TestStripeCompactionPolicy { private static StripeCompactionPolicy.StripeInformationProvider createStripesWithFiles( List<StoreFile>... stripeFiles) throws Exception { return createStripesWithFiles(createBoundaries(stripeFiles.length), - Lists.newArrayList(stripeFiles), new ArrayList<StoreFile>()); + Lists.newArrayList(stripeFiles), new ArrayList<>()); } @Test @@ -433,7 +433,7 @@ public class TestStripeCompactionPolicy { verifySingleStripeCompaction(policy, si, 0, false); // Unless there are enough to cause L0 compaction. si = createStripesWithSizes(6, 2, stripes); - ConcatenatedLists<StoreFile> sfs = new ConcatenatedLists<StoreFile>(); + ConcatenatedLists<StoreFile> sfs = new ConcatenatedLists<>(); sfs.addSublist(si.getLevel0Files()); sfs.addSublist(si.getStripes().get(0)); verifyCompaction( @@ -446,7 +446,7 @@ public class TestStripeCompactionPolicy { // if all files of stripe aren't selected, delete must not be dropped. stripes = new Long[][] { new Long[] { 100L, 3L, 2L, 2L, 2L }, new Long[] { 6L } }; si = createStripesWithSizes(0, 0, stripes); - List<StoreFile> compact_file = new ArrayList<StoreFile>(); + List<StoreFile> compact_file = new ArrayList<>(); Iterator<StoreFile> iter = si.getStripes().get(0).listIterator(1); while (iter.hasNext()) { compact_file.add(iter.next()); @@ -472,7 +472,7 @@ public class TestStripeCompactionPolicy { } private static ArrayList<StoreFile> al(StoreFile... sfs) { - return new ArrayList<StoreFile>(Arrays.asList(sfs)); + return new ArrayList<>(Arrays.asList(sfs)); } private void verifyMergeCompatcion(StripeCompactionPolicy policy, StripeInformationProvider si, @@ -619,7 +619,7 @@ public class TestStripeCompactionPolicy { private static List<StoreFile> getAllFiles( StripeInformationProvider si, int fromStripe, int toStripe) { - ArrayList<StoreFile> expected = new ArrayList<StoreFile>(); + ArrayList<StoreFile> expected = new ArrayList<>(); for (int i = fromStripe; i <= toStripe; ++i) { expected.addAll(si.getStripes().get(i)); } @@ -633,11 +633,11 @@ public class TestStripeCompactionPolicy { */ private static StripeInformationProvider createStripes( int l0Count, byte[]... boundaries) throws Exception { - List<Long> l0Sizes = new ArrayList<Long>(); + List<Long> l0Sizes = new ArrayList<>(); for (int i = 0; i < l0Count; ++i) { l0Sizes.add(5L); } - List<List<Long>> sizes = new ArrayList<List<Long>>(); + List<List<Long>> sizes = new ArrayList<>(); for (int i = 0; i <= boundaries.length; ++i) { sizes.add(Arrays.asList(Long.valueOf(5))); } @@ -651,11 +651,11 @@ public class TestStripeCompactionPolicy { */ private static StripeInformationProvider createStripesL0Only( int l0Count, long l0Size) throws Exception { - List<Long> l0Sizes = new ArrayList<Long>(); + List<Long> l0Sizes = new ArrayList<>(); for (int i = 0; i < l0Count; ++i) { l0Sizes.add(l0Size); } - return createStripes(null, new ArrayList<List<Long>>(), l0Sizes); + return createStripes(null, new ArrayList<>(), l0Sizes); } /** @@ -666,7 +666,7 @@ public class TestStripeCompactionPolicy { */ private static StripeInformationProvider createStripesWithSizes( int l0Count, long l0Size, Long[]... sizes) throws Exception { - ArrayList<List<Long>> sizeList = new ArrayList<List<Long>>(sizes.length); + ArrayList<List<Long>> sizeList = new ArrayList<>(sizes.length); for (Long[] size : sizes) { sizeList.add(Arrays.asList(size)); } @@ -676,7 +676,7 @@ public class TestStripeCompactionPolicy { private static StripeInformationProvider createStripesWithSizes( int l0Count, long l0Size, List<List<Long>> sizes) throws Exception { List<byte[]> boundaries = createBoundaries(sizes.size()); - List<Long> l0Sizes = new ArrayList<Long>(); + List<Long> l0Sizes = new ArrayList<>(); for (int i = 0; i < l0Count; ++i) { l0Sizes.add(l0Size); } @@ -686,22 +686,22 @@ public class TestStripeCompactionPolicy { private static List<byte[]> createBoundaries(int stripeCount) { byte[][] keys = new byte[][] { KEY_A, KEY_B, KEY_C, KEY_D, KEY_E }; assert stripeCount <= keys.length + 1; - List<byte[]> boundaries = new ArrayList<byte[]>(); + List<byte[]> boundaries = new ArrayList<>(); boundaries.addAll(Arrays.asList(keys).subList(0, stripeCount - 1)); return boundaries; } private static StripeInformationProvider createStripes(List<byte[]> boundaries, List<List<Long>> stripeSizes, List<Long> l0Sizes) throws Exception { - List<List<StoreFile>> stripeFiles = new ArrayList<List<StoreFile>>(stripeSizes.size()); + List<List<StoreFile>> stripeFiles = new ArrayList<>(stripeSizes.size()); for (List<Long> sizes : stripeSizes) { - List<StoreFile> sfs = new ArrayList<StoreFile>(sizes.size()); + List<StoreFile> sfs = new ArrayList<>(sizes.size()); for (Long size : sizes) { sfs.add(createFile(size)); } stripeFiles.add(sfs); } - List<StoreFile> l0Files = new ArrayList<StoreFile>(); + List<StoreFile> l0Files = new ArrayList<>(); for (Long size : l0Sizes) { l0Files.add(createFile(size)); } @@ -713,8 +713,8 @@ public class TestStripeCompactionPolicy { */ private static StripeInformationProvider createStripesWithFiles(List<byte[]> boundaries, List<List<StoreFile>> stripeFiles, List<StoreFile> l0Files) throws Exception { - ArrayList<ImmutableList<StoreFile>> stripes = new ArrayList<ImmutableList<StoreFile>>(); - ArrayList<byte[]> boundariesList = new ArrayList<byte[]>(); + ArrayList<ImmutableList<StoreFile>> stripes = new ArrayList<>(); + ArrayList<byte[]> boundariesList = new ArrayList<>(); StripeInformationProvider si = mock(StripeInformationProvider.class); if (!stripeFiles.isEmpty()) { assert stripeFiles.size() == (boundaries.size() + 1); @@ -731,7 +731,7 @@ public class TestStripeCompactionPolicy { when(si.getEndRow(eq(i))).thenReturn(endKey); } } - ConcatenatedLists<StoreFile> sfs = new ConcatenatedLists<StoreFile>(); + ConcatenatedLists<StoreFile> sfs = new ConcatenatedLists<>(); sfs.addAllSublists(stripes); sfs.addSublist(l0Files); when(si.getStorefiles()).thenReturn(sfs); @@ -803,7 +803,7 @@ public class TestStripeCompactionPolicy { private final ArrayList<KeyValue> kvs; public Scanner(KeyValue... kvs) { - this.kvs = new ArrayList<KeyValue>(Arrays.asList(kvs)); + this.kvs = new ArrayList<>(Arrays.asList(kvs)); } @Override http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/TestStripeCompactor.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/TestStripeCompactor.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/TestStripeCompactor.java index 4b82940..088c958 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/TestStripeCompactor.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/TestStripeCompactor.java @@ -177,7 +177,7 @@ public class TestStripeCompactor { null, NoLimitThroughputController.INSTANCE, null); assertEquals(output.length, paths.size()); writers.verifyKvs(output, true, true); - List<byte[]> boundaries = new ArrayList<byte[]>(output.length + 2); + List<byte[]> boundaries = new ArrayList<>(output.length + 2); boundaries.add(left); for (int i = 1; i < output.length; ++i) { boundaries.add(CellUtil.cloneRow(output[i][0])); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/querymatcher/TestCompactionScanQueryMatcher.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/querymatcher/TestCompactionScanQueryMatcher.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/querymatcher/TestCompactionScanQueryMatcher.java index 055fe1c..af8c27d 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/querymatcher/TestCompactionScanQueryMatcher.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/querymatcher/TestCompactionScanQueryMatcher.java @@ -79,8 +79,7 @@ public class TestCompactionScanQueryMatcher extends AbstractTestScanQueryMatcher CompactionScanQueryMatcher qm = CompactionScanQueryMatcher.create(scanInfo, ScanType.COMPACT_RETAIN_DELETES, Long.MAX_VALUE, HConstants.OLDEST_TIMESTAMP, HConstants.OLDEST_TIMESTAMP, now, from, to, null); - List<ScanQueryMatcher.MatchCode> actual = new ArrayList<ScanQueryMatcher.MatchCode>( - rows.length); + List<ScanQueryMatcher.MatchCode> actual = new ArrayList<>(rows.length); byte[] prevRow = null; for (byte[] row : rows) { if (prevRow == null || !Bytes.equals(prevRow, row)) { http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/querymatcher/TestExplicitColumnTracker.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/querymatcher/TestExplicitColumnTracker.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/querymatcher/TestExplicitColumnTracker.java index 3480597..4e07f80 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/querymatcher/TestExplicitColumnTracker.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/querymatcher/TestExplicitColumnTracker.java @@ -49,7 +49,7 @@ public class TestExplicitColumnTracker { ColumnTracker exp = new ExplicitColumnTracker(trackColumns, 0, maxVersions, Long.MIN_VALUE); // Initialize result - List<ScanQueryMatcher.MatchCode> result = new ArrayList<ScanQueryMatcher.MatchCode>(scannerColumns.size()); + List<ScanQueryMatcher.MatchCode> result = new ArrayList<>(scannerColumns.size()); long timestamp = 0; // "Match" @@ -67,11 +67,11 @@ public class TestExplicitColumnTracker { @Test public void testGetSingleVersion() throws IOException { // Create tracker - TreeSet<byte[]> columns = new TreeSet<byte[]>(Bytes.BYTES_COMPARATOR); + TreeSet<byte[]> columns = new TreeSet<>(Bytes.BYTES_COMPARATOR); // Looking for every other columns.add(col2); columns.add(col4); - List<MatchCode> expected = new ArrayList<ScanQueryMatcher.MatchCode>(5); + List<MatchCode> expected = new ArrayList<>(5); expected.add(ScanQueryMatcher.MatchCode.SEEK_NEXT_COL); // col1 expected.add(ScanQueryMatcher.MatchCode.INCLUDE_AND_SEEK_NEXT_COL); // col2 expected.add(ScanQueryMatcher.MatchCode.SEEK_NEXT_COL); // col3 @@ -80,7 +80,7 @@ public class TestExplicitColumnTracker { int maxVersions = 1; // Create "Scanner" - List<byte[]> scanner = new ArrayList<byte[]>(5); + List<byte[]> scanner = new ArrayList<>(5); scanner.add(col1); scanner.add(col2); scanner.add(col3); @@ -93,12 +93,12 @@ public class TestExplicitColumnTracker { @Test public void testGetMultiVersion() throws IOException { // Create tracker - TreeSet<byte[]> columns = new TreeSet<byte[]>(Bytes.BYTES_COMPARATOR); + TreeSet<byte[]> columns = new TreeSet<>(Bytes.BYTES_COMPARATOR); // Looking for every other columns.add(col2); columns.add(col4); - List<ScanQueryMatcher.MatchCode> expected = new ArrayList<ScanQueryMatcher.MatchCode>(15); + List<ScanQueryMatcher.MatchCode> expected = new ArrayList<>(15); expected.add(ScanQueryMatcher.MatchCode.SEEK_NEXT_COL); expected.add(ScanQueryMatcher.MatchCode.SEEK_NEXT_COL); expected.add(ScanQueryMatcher.MatchCode.SEEK_NEXT_COL); @@ -121,7 +121,7 @@ public class TestExplicitColumnTracker { int maxVersions = 2; // Create "Scanner" - List<byte[]> scanner = new ArrayList<byte[]>(15); + List<byte[]> scanner = new ArrayList<>(15); scanner.add(col1); scanner.add(col1); scanner.add(col1); @@ -148,7 +148,7 @@ public class TestExplicitColumnTracker { @Test public void testStackOverflow() throws IOException { int maxVersions = 1; - TreeSet<byte[]> columns = new TreeSet<byte[]>(Bytes.BYTES_COMPARATOR); + TreeSet<byte[]> columns = new TreeSet<>(Bytes.BYTES_COMPARATOR); for (int i = 0; i < 100000; i++) { columns.add(Bytes.toBytes("col" + i)); } @@ -173,7 +173,7 @@ public class TestExplicitColumnTracker { */ @Test public void testInfiniteLoop() throws IOException { - TreeSet<byte[]> columns = new TreeSet<byte[]>(Bytes.BYTES_COMPARATOR); + TreeSet<byte[]> columns = new TreeSet<>(Bytes.BYTES_COMPARATOR); columns.addAll(Arrays.asList(new byte[][] { col2, col3, col5 })); List<byte[]> scanner = Arrays.<byte[]> asList(new byte[][] { col1, col4 }); List<ScanQueryMatcher.MatchCode> expected = Arrays.<ScanQueryMatcher.MatchCode> asList( http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/querymatcher/TestScanWildcardColumnTracker.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/querymatcher/TestScanWildcardColumnTracker.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/querymatcher/TestScanWildcardColumnTracker.java index 2852947..6d6e58e 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/querymatcher/TestScanWildcardColumnTracker.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/querymatcher/TestScanWildcardColumnTracker.java @@ -44,20 +44,20 @@ public class TestScanWildcardColumnTracker { ScanWildcardColumnTracker tracker = new ScanWildcardColumnTracker(0, VERSIONS, Long.MIN_VALUE); // Create list of qualifiers - List<byte[]> qualifiers = new ArrayList<byte[]>(4); + List<byte[]> qualifiers = new ArrayList<>(4); qualifiers.add(Bytes.toBytes("qualifier1")); qualifiers.add(Bytes.toBytes("qualifier2")); qualifiers.add(Bytes.toBytes("qualifier3")); qualifiers.add(Bytes.toBytes("qualifier4")); // Setting up expected result - List<MatchCode> expected = new ArrayList<MatchCode>(4); + List<MatchCode> expected = new ArrayList<>(4); expected.add(ScanQueryMatcher.MatchCode.INCLUDE); expected.add(ScanQueryMatcher.MatchCode.INCLUDE); expected.add(ScanQueryMatcher.MatchCode.INCLUDE); expected.add(ScanQueryMatcher.MatchCode.INCLUDE); - List<ScanQueryMatcher.MatchCode> actual = new ArrayList<MatchCode>(qualifiers.size()); + List<ScanQueryMatcher.MatchCode> actual = new ArrayList<>(qualifiers.size()); for (byte[] qualifier : qualifiers) { ScanQueryMatcher.MatchCode mc = ScanQueryMatcher.checkColumn(tracker, qualifier, 0, @@ -76,20 +76,20 @@ public class TestScanWildcardColumnTracker { ScanWildcardColumnTracker tracker = new ScanWildcardColumnTracker(0, VERSIONS, Long.MIN_VALUE); // Create list of qualifiers - List<byte[]> qualifiers = new ArrayList<byte[]>(4); + List<byte[]> qualifiers = new ArrayList<>(4); qualifiers.add(Bytes.toBytes("qualifier1")); qualifiers.add(Bytes.toBytes("qualifier1")); qualifiers.add(Bytes.toBytes("qualifier1")); qualifiers.add(Bytes.toBytes("qualifier2")); // Setting up expected result - List<ScanQueryMatcher.MatchCode> expected = new ArrayList<MatchCode>(4); + List<ScanQueryMatcher.MatchCode> expected = new ArrayList<>(4); expected.add(ScanQueryMatcher.MatchCode.INCLUDE); expected.add(ScanQueryMatcher.MatchCode.INCLUDE); expected.add(ScanQueryMatcher.MatchCode.SEEK_NEXT_COL); expected.add(ScanQueryMatcher.MatchCode.INCLUDE); - List<MatchCode> actual = new ArrayList<ScanQueryMatcher.MatchCode>(qualifiers.size()); + List<MatchCode> actual = new ArrayList<>(qualifiers.size()); long timestamp = 0; for (byte[] qualifier : qualifiers) { @@ -109,7 +109,7 @@ public class TestScanWildcardColumnTracker { ScanWildcardColumnTracker tracker = new ScanWildcardColumnTracker(0, VERSIONS, Long.MIN_VALUE); // Create list of qualifiers - List<byte[]> qualifiers = new ArrayList<byte[]>(2); + List<byte[]> qualifiers = new ArrayList<>(2); qualifiers.add(Bytes.toBytes("qualifier2")); qualifiers.add(Bytes.toBytes("qualifier1")); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/querymatcher/TestUserScanQueryMatcher.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/querymatcher/TestUserScanQueryMatcher.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/querymatcher/TestUserScanQueryMatcher.java index 0831404..b4e4311 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/querymatcher/TestUserScanQueryMatcher.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/querymatcher/TestUserScanQueryMatcher.java @@ -69,7 +69,7 @@ public class TestUserScanQueryMatcher extends AbstractTestScanQueryMatcher { // of just byte [] // Expected result - List<MatchCode> expected = new ArrayList<ScanQueryMatcher.MatchCode>(6); + List<MatchCode> expected = new ArrayList<>(6); expected.add(ScanQueryMatcher.MatchCode.SEEK_NEXT_COL); expected.add(ScanQueryMatcher.MatchCode.INCLUDE_AND_SEEK_NEXT_COL); expected.add(ScanQueryMatcher.MatchCode.SEEK_NEXT_COL); @@ -83,7 +83,7 @@ public class TestUserScanQueryMatcher extends AbstractTestScanQueryMatcher { new ScanInfo(this.conf, fam2, 0, 1, ttl, KeepDeletedCells.FALSE, 0, rowComparator), get.getFamilyMap().get(fam2), now - ttl, now, null); - List<KeyValue> memstore = new ArrayList<KeyValue>(6); + List<KeyValue> memstore = new ArrayList<>(6); memstore.add(new KeyValue(row1, fam2, col1, 1, data)); memstore.add(new KeyValue(row1, fam2, col2, 1, data)); memstore.add(new KeyValue(row1, fam2, col3, 1, data)); @@ -92,7 +92,7 @@ public class TestUserScanQueryMatcher extends AbstractTestScanQueryMatcher { memstore.add(new KeyValue(row2, fam1, col1, data)); - List<ScanQueryMatcher.MatchCode> actual = new ArrayList<ScanQueryMatcher.MatchCode>(memstore.size()); + List<ScanQueryMatcher.MatchCode> actual = new ArrayList<>(memstore.size()); KeyValue k = memstore.get(0); qm.setToNewRow(k); @@ -113,7 +113,7 @@ public class TestUserScanQueryMatcher extends AbstractTestScanQueryMatcher { // of just byte [] // Expected result - List<MatchCode> expected = new ArrayList<ScanQueryMatcher.MatchCode>(6); + List<MatchCode> expected = new ArrayList<>(6); expected.add(ScanQueryMatcher.MatchCode.INCLUDE); expected.add(ScanQueryMatcher.MatchCode.INCLUDE); expected.add(ScanQueryMatcher.MatchCode.INCLUDE); @@ -126,7 +126,7 @@ public class TestUserScanQueryMatcher extends AbstractTestScanQueryMatcher { new ScanInfo(this.conf, fam2, 0, 1, ttl, KeepDeletedCells.FALSE, 0, rowComparator), null, now - ttl, now, null); - List<KeyValue> memstore = new ArrayList<KeyValue>(6); + List<KeyValue> memstore = new ArrayList<>(6); memstore.add(new KeyValue(row1, fam2, col1, 1, data)); memstore.add(new KeyValue(row1, fam2, col2, 1, data)); memstore.add(new KeyValue(row1, fam2, col3, 1, data)); @@ -134,7 +134,7 @@ public class TestUserScanQueryMatcher extends AbstractTestScanQueryMatcher { memstore.add(new KeyValue(row1, fam2, col5, 1, data)); memstore.add(new KeyValue(row2, fam1, col1, 1, data)); - List<ScanQueryMatcher.MatchCode> actual = new ArrayList<ScanQueryMatcher.MatchCode>(memstore.size()); + List<ScanQueryMatcher.MatchCode> actual = new ArrayList<>(memstore.size()); KeyValue k = memstore.get(0); qm.setToNewRow(k); @@ -181,7 +181,7 @@ public class TestUserScanQueryMatcher extends AbstractTestScanQueryMatcher { KeyValue k = kvs[0]; qm.setToNewRow(k); - List<MatchCode> actual = new ArrayList<MatchCode>(kvs.length); + List<MatchCode> actual = new ArrayList<>(kvs.length); for (KeyValue kv : kvs) { actual.add(qm.match(kv)); } @@ -222,7 +222,7 @@ public class TestUserScanQueryMatcher extends AbstractTestScanQueryMatcher { KeyValue k = kvs[0]; qm.setToNewRow(k); - List<ScanQueryMatcher.MatchCode> actual = new ArrayList<ScanQueryMatcher.MatchCode>(kvs.length); + List<ScanQueryMatcher.MatchCode> actual = new ArrayList<>(kvs.length); for (KeyValue kv : kvs) { actual.add(qm.match(kv)); } http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/AbstractTestFSWAL.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/AbstractTestFSWAL.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/AbstractTestFSWAL.java index fb0b514..0be7b31 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/AbstractTestFSWAL.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/AbstractTestFSWAL.java @@ -255,11 +255,11 @@ public abstract class AbstractTestFSWAL { new HRegionInfo(t2.getTableName(), HConstants.EMPTY_START_ROW, HConstants.EMPTY_END_ROW); // add edits and roll the wal MultiVersionConcurrencyControl mvcc = new MultiVersionConcurrencyControl(); - NavigableMap<byte[], Integer> scopes1 = new TreeMap<byte[], Integer>(Bytes.BYTES_COMPARATOR); + NavigableMap<byte[], Integer> scopes1 = new TreeMap<>(Bytes.BYTES_COMPARATOR); for (byte[] fam : t1.getFamiliesKeys()) { scopes1.put(fam, 0); } - NavigableMap<byte[], Integer> scopes2 = new TreeMap<byte[], Integer>(Bytes.BYTES_COMPARATOR); + NavigableMap<byte[], Integer> scopes2 = new TreeMap<>(Bytes.BYTES_COMPARATOR); for (byte[] fam : t2.getFamiliesKeys()) { scopes2.put(fam, 0); } @@ -362,7 +362,7 @@ public abstract class AbstractTestFSWAL { HBaseTestingUtility.closeRegionAndWAL(r); final int countPerFamily = 10; final AtomicBoolean goslow = new AtomicBoolean(false); - NavigableMap<byte[], Integer> scopes = new TreeMap<byte[], Integer>(Bytes.BYTES_COMPARATOR); + NavigableMap<byte[], Integer> scopes = new TreeMap<>(Bytes.BYTES_COMPARATOR); for (byte[] fam : htd.getFamiliesKeys()) { scopes.put(fam, 0); } @@ -402,7 +402,7 @@ public abstract class AbstractTestFSWAL { } } // Add any old cluster id. - List<UUID> clusterIds = new ArrayList<UUID>(1); + List<UUID> clusterIds = new ArrayList<>(1); clusterIds.add(UUID.randomUUID()); // Now make appends run slow. goslow.set(true); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/AbstractTestLogRollPeriod.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/AbstractTestLogRollPeriod.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/AbstractTestLogRollPeriod.java index f70bcc8..04a4bbc 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/AbstractTestLogRollPeriod.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/AbstractTestLogRollPeriod.java @@ -130,7 +130,7 @@ public abstract class AbstractTestLogRollPeriod { private void checkMinLogRolls(final WAL log, final int minRolls) throws Exception { - final List<Path> paths = new ArrayList<Path>(); + final List<Path> paths = new ArrayList<>(); log.registerWALActionsListener(new WALActionsListener.Base() { @Override public void postLogRoll(Path oldFile, Path newFile) { http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/AbstractTestWALReplay.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/AbstractTestWALReplay.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/AbstractTestWALReplay.java index 90eacf0..237d24a 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/AbstractTestWALReplay.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/AbstractTestWALReplay.java @@ -299,8 +299,7 @@ public abstract class AbstractTestWALReplay { // Add 1k to each family. final int countPerFamily = 1000; - NavigableMap<byte[], Integer> scopes = new TreeMap<byte[], Integer>( - Bytes.BYTES_COMPARATOR); + NavigableMap<byte[], Integer> scopes = new TreeMap<>(Bytes.BYTES_COMPARATOR); for(byte[] fam : htd.getFamiliesKeys()) { scopes.put(fam, 0); } @@ -366,7 +365,7 @@ public abstract class AbstractTestWALReplay { Path f = new Path(basedir, "hfile"); HFileTestUtil.createHFile(this.conf, fs, f, family, family, Bytes.toBytes(""), Bytes.toBytes("z"), 10); - List<Pair<byte[], String>> hfs = new ArrayList<Pair<byte[], String>>(1); + List<Pair<byte[], String>> hfs = new ArrayList<>(1); hfs.add(Pair.newPair(family, f.toString())); region.bulkLoadHFiles(hfs, true, null); @@ -434,7 +433,7 @@ public abstract class AbstractTestWALReplay { region.put((new Put(row)).addColumn(family, family, family)); wal.sync(); - List <Pair<byte[],String>> hfs= new ArrayList<Pair<byte[],String>>(1); + List <Pair<byte[],String>> hfs= new ArrayList<>(1); for (int i = 0; i < 3; i++) { Path f = new Path(basedir, "hfile"+i); HFileTestUtil.createHFile(this.conf, fs, f, family, family, Bytes.toBytes(i + "00"), @@ -700,8 +699,7 @@ public abstract class AbstractTestWALReplay { HRegion region = HRegion.openHRegion(this.hbaseRootDir, hri, htd, wal, customConf, rsServices, null); int writtenRowCount = 10; - List<HColumnDescriptor> families = new ArrayList<HColumnDescriptor>( - htd.getFamilies()); + List<HColumnDescriptor> families = new ArrayList<>(htd.getFamilies()); for (int i = 0; i < writtenRowCount; i++) { Put put = new Put(Bytes.toBytes(tableName + Integer.toString(i))); put.addColumn(families.get(i % families.size()).getName(), Bytes.toBytes("q"), @@ -759,7 +757,7 @@ public abstract class AbstractTestWALReplay { private int getScannedCount(RegionScanner scanner) throws IOException { int scannedCount = 0; - List<Cell> results = new ArrayList<Cell>(); + List<Cell> results = new ArrayList<>(); while (true) { boolean existMore = scanner.next(results); if (!results.isEmpty()) @@ -794,9 +792,8 @@ public abstract class AbstractTestWALReplay { // Add 1k to each family. final int countPerFamily = 1000; - Set<byte[]> familyNames = new HashSet<byte[]>(); - NavigableMap<byte[], Integer> scopes = new TreeMap<byte[], Integer>( - Bytes.BYTES_COMPARATOR); + Set<byte[]> familyNames = new HashSet<>(); + NavigableMap<byte[], Integer> scopes = new TreeMap<>(Bytes.BYTES_COMPARATOR); for(byte[] fam : htd.getFamiliesKeys()) { scopes.put(fam, 0); } @@ -1048,7 +1045,7 @@ public abstract class AbstractTestWALReplay { deleteDir(basedir); final HTableDescriptor htd = createBasic1FamilyHTD(tableName); - NavigableMap<byte[], Integer> scopes = new TreeMap<byte[], Integer>(Bytes.BYTES_COMPARATOR); + NavigableMap<byte[], Integer> scopes = new TreeMap<>(Bytes.BYTES_COMPARATOR); for (byte[] fam : htd.getFamiliesKeys()) { scopes.put(fam, 0); } @@ -1196,7 +1193,7 @@ public abstract class AbstractTestWALReplay { static List<Put> addRegionEdits(final byte[] rowName, final byte[] family, final int count, EnvironmentEdge ee, final Region r, final String qualifierPrefix) throws IOException { - List<Put> puts = new ArrayList<Put>(); + List<Put> puts = new ArrayList<>(); for (int j = 0; j < count; j++) { byte[] qualifier = Bytes.toBytes(qualifierPrefix + Integer.toString(j)); Put p = new Put(rowName); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/FaultyProtobufLogReader.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/FaultyProtobufLogReader.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/FaultyProtobufLogReader.java index c654c16..1af21d2 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/FaultyProtobufLogReader.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/FaultyProtobufLogReader.java @@ -31,7 +31,7 @@ public class FaultyProtobufLogReader extends ProtobufLogReader { BEGINNING, MIDDLE, END, NONE } - Queue<Entry> nextQueue = new LinkedList<Entry>(); + Queue<Entry> nextQueue = new LinkedList<>(); int numberOfFileEntries = 0; FailureType getFailureType() { http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestFSHLog.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestFSHLog.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestFSHLog.java index 9e546e6..8847c4c 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestFSHLog.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestFSHLog.java @@ -102,7 +102,7 @@ public class TestFSHLog extends AbstractTestFSWAL { syncRunnerIndexField.set(ringBufferEventHandler, Integer.MAX_VALUE - 1); HTableDescriptor htd = new HTableDescriptor(TableName.valueOf(this.name.getMethodName())).addFamily(new HColumnDescriptor("row")); - NavigableMap<byte[], Integer> scopes = new TreeMap<byte[], Integer>(Bytes.BYTES_COMPARATOR); + NavigableMap<byte[], Integer> scopes = new TreeMap<>(Bytes.BYTES_COMPARATOR); for (byte[] fam : htd.getFamiliesKeys()) { scopes.put(fam, 0); } http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestKeyValueCompression.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestKeyValueCompression.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestKeyValueCompression.java index 104f897..4a256a6 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestKeyValueCompression.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestKeyValueCompression.java @@ -107,7 +107,7 @@ public class TestKeyValueCompression { byte[] cf = Bytes.toBytes("myCF"); byte[] q = Bytes.toBytes("myQualifier"); byte[] value = Bytes.toBytes("myValue"); - List<Tag> tags = new ArrayList<Tag>(noOfTags); + List<Tag> tags = new ArrayList<>(noOfTags); for (int i = 1; i <= noOfTags; i++) { tags.add(new ArrayBackedTag((byte) i, Bytes.toBytes("tagValue" + i))); } http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestLogRollAbort.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestLogRollAbort.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestLogRollAbort.java index 22395c8..ccb00c7 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestLogRollAbort.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestLogRollAbort.java @@ -199,8 +199,7 @@ public class TestLogRollAbort { kvs.add(new KeyValue(Bytes.toBytes(i), tableName.getName(), tableName.getName())); HTableDescriptor htd = new HTableDescriptor(tableName); htd.addFamily(new HColumnDescriptor("column")); - NavigableMap<byte[], Integer> scopes = new TreeMap<byte[], Integer>( - Bytes.BYTES_COMPARATOR); + NavigableMap<byte[], Integer> scopes = new TreeMap<>(Bytes.BYTES_COMPARATOR); for(byte[] fam : htd.getFamiliesKeys()) { scopes.put(fam, 0); } http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestLogRolling.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestLogRolling.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestLogRolling.java index 19c534e..5bc4c9b 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestLogRolling.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestLogRolling.java @@ -245,8 +245,8 @@ public class TestLogRolling extends AbstractTestLogRolling { server = TEST_UTIL.getRSForFirstRegionInTable(desc.getTableName()); HRegionInfo region = server.getOnlineRegions(desc.getTableName()).get(0).getRegionInfo(); final WAL log = server.getWAL(region); - final List<Path> paths = new ArrayList<Path>(1); - final List<Integer> preLogRolledCalled = new ArrayList<Integer>(); + final List<Path> paths = new ArrayList<>(1); + final List<Integer> preLogRolledCalled = new ArrayList<>(); paths.add(AbstractFSWALProvider.getCurrentFileName(log)); log.registerWALActionsListener(new WALActionsListener.Base() { @@ -307,7 +307,7 @@ public class TestLogRolling extends AbstractTestLogRolling { preLogRolledCalled.size() >= 1); // read back the data written - Set<String> loggedRows = new HashSet<String>(); + Set<String> loggedRows = new HashSet<>(); FSUtils fsUtils = FSUtils.getInstance(fs, TEST_UTIL.getConfiguration()); for (Path p : paths) { LOG.debug("recovering lease for " + p); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestLogRollingNoCluster.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestLogRollingNoCluster.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestLogRollingNoCluster.java index 7412128..d3d582c 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestLogRollingNoCluster.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestLogRollingNoCluster.java @@ -157,8 +157,7 @@ public class TestLogRollingNoCluster { edit.add(new KeyValue(bytes, bytes, bytes, now, EMPTY_1K_ARRAY)); final HRegionInfo hri = HRegionInfo.FIRST_META_REGIONINFO; final HTableDescriptor htd = TEST_UTIL.getMetaTableDescriptor(); - NavigableMap<byte[], Integer> scopes = new TreeMap<byte[], Integer>( - Bytes.BYTES_COMPARATOR); + NavigableMap<byte[], Integer> scopes = new TreeMap<>(Bytes.BYTES_COMPARATOR); for(byte[] fam : htd.getFamiliesKeys()) { scopes.put(fam, 0); } http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestSequenceIdAccounting.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestSequenceIdAccounting.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestSequenceIdAccounting.java index 9fd0cb1..9f5acbd 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestSequenceIdAccounting.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestSequenceIdAccounting.java @@ -38,7 +38,7 @@ public class TestSequenceIdAccounting { private static final byte [] FAMILY_NAME = Bytes.toBytes("cf"); private static final Set<byte[]> FAMILIES; static { - FAMILIES = new HashSet<byte[]>(); + FAMILIES = new HashSet<>(); FAMILIES.add(FAMILY_NAME); } @@ -46,7 +46,7 @@ public class TestSequenceIdAccounting { public void testStartCacheFlush() { SequenceIdAccounting sida = new SequenceIdAccounting(); sida.getOrCreateLowestSequenceIds(ENCODED_REGION_NAME); - Map<byte[], Long> m = new HashMap<byte[], Long>(); + Map<byte[], Long> m = new HashMap<>(); m.put(ENCODED_REGION_NAME, HConstants.NO_SEQNUM); assertEquals(HConstants.NO_SEQNUM, (long)sida.startCacheFlush(ENCODED_REGION_NAME, FAMILIES)); sida.completeCacheFlush(ENCODED_REGION_NAME); @@ -57,7 +57,7 @@ public class TestSequenceIdAccounting { sida.completeCacheFlush(ENCODED_REGION_NAME); long currentSequenceId = sequenceid; sida.update(ENCODED_REGION_NAME, FAMILIES, sequenceid, true); - final Set<byte[]> otherFamily = new HashSet<byte[]>(1); + final Set<byte[]> otherFamily = new HashSet<>(1); otherFamily.add(Bytes.toBytes("otherCf")); sida.update(ENCODED_REGION_NAME, FAMILIES, ++sequenceid, true); // Should return oldest sequence id in the region. @@ -69,7 +69,7 @@ public class TestSequenceIdAccounting { public void testAreAllLower() { SequenceIdAccounting sida = new SequenceIdAccounting(); sida.getOrCreateLowestSequenceIds(ENCODED_REGION_NAME); - Map<byte[], Long> m = new HashMap<byte[], Long>(); + Map<byte[], Long> m = new HashMap<>(); m.put(ENCODED_REGION_NAME, HConstants.NO_SEQNUM); assertTrue(sida.areAllLower(m)); long sequenceid = 1; @@ -117,7 +117,7 @@ public class TestSequenceIdAccounting { public void testFindLower() { SequenceIdAccounting sida = new SequenceIdAccounting(); sida.getOrCreateLowestSequenceIds(ENCODED_REGION_NAME); - Map<byte[], Long> m = new HashMap<byte[], Long>(); + Map<byte[], Long> m = new HashMap<>(); m.put(ENCODED_REGION_NAME, HConstants.NO_SEQNUM); long sequenceid = 1; sida.update(ENCODED_REGION_NAME, FAMILIES, sequenceid, true); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestWALActionsListener.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestWALActionsListener.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestWALActionsListener.java index 9ac9f0e..9f9e2df 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestWALActionsListener.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestWALActionsListener.java @@ -95,7 +95,7 @@ public class TestWALActionsListener { @Test public void testActionListener() throws Exception { DummyWALActionsListener observer = new DummyWALActionsListener(); - List<WALActionsListener> list = new ArrayList<WALActionsListener>(1); + List<WALActionsListener> list = new ArrayList<>(1); list.add(observer); final WALFactory wals = new WALFactory(conf, list, "testActionListener"); DummyWALActionsListener laterobserver = new DummyWALActionsListener(); @@ -110,8 +110,7 @@ public class TestWALActionsListener { edit.add(kv); HTableDescriptor htd = new HTableDescriptor(TableName.valueOf(SOME_BYTES)); htd.addFamily(new HColumnDescriptor(b)); - NavigableMap<byte[], Integer> scopes = new TreeMap<byte[], Integer>( - Bytes.BYTES_COMPARATOR); + NavigableMap<byte[], Integer> scopes = new TreeMap<>(Bytes.BYTES_COMPARATOR); for(byte[] fam : htd.getFamiliesKeys()) { scopes.put(fam, 0); } http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestWALCellCodecWithCompression.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestWALCellCodecWithCompression.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestWALCellCodecWithCompression.java index c4329b8..8a246be 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestWALCellCodecWithCompression.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestWALCellCodecWithCompression.java @@ -102,7 +102,7 @@ public class TestWALCellCodecWithCompression { byte[] cf = Bytes.toBytes("myCF"); byte[] q = Bytes.toBytes("myQualifier"); byte[] value = Bytes.toBytes("myValue"); - List<Tag> tags = new ArrayList<Tag>(noOfTags); + List<Tag> tags = new ArrayList<>(noOfTags); for (int i = 1; i <= noOfTags; i++) { tags.add(new ArrayBackedTag((byte) i, Bytes.toBytes("tagValue" + i))); } @@ -114,7 +114,7 @@ public class TestWALCellCodecWithCompression { byte[] cf = Bytes.toBytes("myCF"); byte[] q = Bytes.toBytes("myQualifier"); byte[] value = Bytes.toBytes("myValue"); - List<Tag> tags = new ArrayList<Tag>(noOfTags); + List<Tag> tags = new ArrayList<>(noOfTags); for (int i = 1; i <= noOfTags; i++) { tags.add(new ArrayBackedTag((byte) i, Bytes.toBytes("tagValue" + i))); } http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestNamespaceReplication.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestNamespaceReplication.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestNamespaceReplication.java index 3814562..e296f87 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestNamespaceReplication.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestNamespaceReplication.java @@ -165,7 +165,7 @@ public class TestNamespaceReplication extends TestReplicationBase { namespaces.add(ns2); rpc.setNamespaces(namespaces); Map<TableName, List<String>> tableCfs = new HashMap<>(); - tableCfs.put(tabAName, new ArrayList<String>()); + tableCfs.put(tabAName, new ArrayList<>()); tableCfs.get(tabAName).add("f1"); rpc.setTableCFsMap(tableCfs); admin.updatePeerConfig("2", rpc); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestPerTableCFReplication.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestPerTableCFReplication.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestPerTableCFReplication.java index a7a4cd8..abf2db3 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestPerTableCFReplication.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestPerTableCFReplication.java @@ -283,7 +283,7 @@ public class TestPerTableCFReplication { // 1. null or empty string, result should be null assertNull(ReplicationSerDeHelper.convert(tabCFsMap)); - tabCFsMap = new HashMap<TableName, List<String>>(); + tabCFsMap = new HashMap<>(); tableCFs = ReplicationSerDeHelper.convert(tabCFsMap); assertEquals(0, tableCFs.length); @@ -301,7 +301,7 @@ public class TestPerTableCFReplication { assertEquals(0, tableCFs[0].getFamiliesCount()); tabCFsMap.clear(); - tabCFsMap.put(tableName2, new ArrayList<String>()); + tabCFsMap.put(tableName2, new ArrayList<>()); tabCFsMap.get(tableName2).add("cf1"); tableCFs = ReplicationSerDeHelper.convert(tabCFsMap); assertEquals(1, tableCFs.length); // only one table @@ -311,7 +311,7 @@ public class TestPerTableCFReplication { assertEquals("cf1", tableCFs[0].getFamilies(0).toStringUtf8()); tabCFsMap.clear(); - tabCFsMap.put(tableName3, new ArrayList<String>()); + tabCFsMap.put(tableName3, new ArrayList<>()); tabCFsMap.get(tableName3).add("cf1"); tabCFsMap.get(tableName3).add("cf3"); tableCFs = ReplicationSerDeHelper.convert(tabCFsMap); @@ -324,9 +324,9 @@ public class TestPerTableCFReplication { tabCFsMap.clear(); tabCFsMap.put(tableName1, null); - tabCFsMap.put(tableName2, new ArrayList<String>()); + tabCFsMap.put(tableName2, new ArrayList<>()); tabCFsMap.get(tableName2).add("cf1"); - tabCFsMap.put(tableName3, new ArrayList<String>()); + tabCFsMap.put(tableName3, new ArrayList<>()); tabCFsMap.get(tableName3).add("cf1"); tabCFsMap.get(tableName3).add("cf3"); @@ -406,7 +406,7 @@ public class TestPerTableCFReplication { rpc2.setClusterKey(utility2.getClusterKey()); Map<TableName, List<String>> tableCFs = new HashMap<>(); tableCFs.put(tabCName, null); - tableCFs.put(tabBName, new ArrayList<String>()); + tableCFs.put(tabBName, new ArrayList<>()); tableCFs.get(tabBName).add("f1"); tableCFs.get(tabBName).add("f3"); replicationAdmin.addPeer("2", rpc2, tableCFs); @@ -415,7 +415,7 @@ public class TestPerTableCFReplication { rpc3.setClusterKey(utility3.getClusterKey()); tableCFs.clear(); tableCFs.put(tabAName, null); - tableCFs.put(tabBName, new ArrayList<String>()); + tableCFs.put(tabBName, new ArrayList<>()); tableCFs.get(tabBName).add("f1"); tableCFs.get(tabBName).add("f2"); replicationAdmin.addPeer("3", rpc3, tableCFs); @@ -462,17 +462,17 @@ public class TestPerTableCFReplication { // B. change peers' replicable table-cf config tableCFs.clear(); - tableCFs.put(tabAName, new ArrayList<String>()); + tableCFs.put(tabAName, new ArrayList<>()); tableCFs.get(tabAName).add("f1"); tableCFs.get(tabAName).add("f2"); - tableCFs.put(tabCName, new ArrayList<String>()); + tableCFs.put(tabCName, new ArrayList<>()); tableCFs.get(tabCName).add("f2"); tableCFs.get(tabCName).add("f3"); replicationAdmin.setPeerTableCFs("2", tableCFs); tableCFs.clear(); tableCFs.put(tabBName, null); - tableCFs.put(tabCName, new ArrayList<String>()); + tableCFs.put(tabCName, new ArrayList<>()); tableCFs.get(tabCName).add("f3"); replicationAdmin.setPeerTableCFs("3", tableCFs); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestReplicationBase.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestReplicationBase.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestReplicationBase.java index 474039b..caad544 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestReplicationBase.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestReplicationBase.java @@ -142,8 +142,7 @@ public class TestReplicationBase { table.addFamily(fam); fam = new HColumnDescriptor(noRepfamName); table.addFamily(fam); - scopes = new TreeMap<byte[], Integer>( - Bytes.BYTES_COMPARATOR); + scopes = new TreeMap<>(Bytes.BYTES_COMPARATOR); for(HColumnDescriptor f : table.getColumnFamilies()) { scopes.put(f.getName(), f.getScope()); } http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestReplicationEndpoint.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestReplicationEndpoint.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestReplicationEndpoint.java index 5e8d569..4925aab 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestReplicationEndpoint.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestReplicationEndpoint.java @@ -106,7 +106,7 @@ public class TestReplicationEndpoint extends TestReplicationBase { @Override public String explainFailure() throws Exception { - List<String> logRollInProgressRsList = new ArrayList<String>(); + List<String> logRollInProgressRsList = new ArrayList<>(); for (RegionServerThread rs : rsThreads) { if (!rs.getRegionServer().walRollRequestFinished()) { logRollInProgressRsList.add(rs.getRegionServer().toString()); @@ -462,7 +462,7 @@ public class TestReplicationEndpoint extends TestReplicationBase { public static class ReplicationEndpointReturningFalse extends ReplicationEndpointForTest { static int COUNT = 10; - static AtomicReference<Exception> ex = new AtomicReference<Exception>(null); + static AtomicReference<Exception> ex = new AtomicReference<>(null); static AtomicBoolean replicated = new AtomicBoolean(false); @Override public boolean replicate(ReplicateContext replicateContext) { @@ -483,7 +483,7 @@ public class TestReplicationEndpoint extends TestReplicationBase { // return a WALEntry filter which only accepts "row", but not other rows public static class ReplicationEndpointWithWALEntryFilter extends ReplicationEndpointForTest { - static AtomicReference<Exception> ex = new AtomicReference<Exception>(null); + static AtomicReference<Exception> ex = new AtomicReference<>(null); @Override public boolean replicate(ReplicateContext replicateContext) { http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestReplicationSmallTests.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestReplicationSmallTests.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestReplicationSmallTests.java index 9536f9f..1c5a994 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestReplicationSmallTests.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestReplicationSmallTests.java @@ -413,7 +413,7 @@ public class TestReplicationSmallTests extends TestReplicationBase { @Test(timeout=300000) public void testLoading() throws Exception { LOG.info("Writing out rows to table1 in testLoading"); - List<Put> puts = new ArrayList<Put>(NB_ROWS_IN_BIG_BATCH); + List<Put> puts = new ArrayList<>(NB_ROWS_IN_BIG_BATCH); for (int i = 0; i < NB_ROWS_IN_BIG_BATCH; i++) { Put put = new Put(Bytes.toBytes(i)); put.addColumn(famName, row, row); @@ -519,8 +519,7 @@ public class TestReplicationSmallTests extends TestReplicationBase { fam.setMaxVersions(100); fam.setScope(HConstants.REPLICATION_SCOPE_GLOBAL); table.addFamily(fam); - scopes = new TreeMap<byte[], Integer>( - Bytes.BYTES_COMPARATOR); + scopes = new TreeMap<>(Bytes.BYTES_COMPARATOR); for (HColumnDescriptor f : table.getColumnFamilies()) { scopes.put(f.getName(), f.getScope()); } @@ -818,7 +817,7 @@ public class TestReplicationSmallTests extends TestReplicationBase { HRegion region = utility1.getMiniHBaseCluster().getRegions(tableName).get(0); HRegionInfo hri = region.getRegionInfo(); - NavigableMap<byte[], Integer> scopes = new TreeMap<byte[], Integer>(Bytes.BYTES_COMPARATOR); + NavigableMap<byte[], Integer> scopes = new TreeMap<>(Bytes.BYTES_COMPARATOR); for (byte[] fam : htable1.getTableDescriptor().getFamiliesKeys()) { scopes.put(fam, 1); }