Repository: hbase Updated Branches: refs/heads/branch-2 1e5fc1ed6 -> b1269ec57
http://git-wip-us.apache.org/repos/asf/hbase/blob/b1269ec5/hbase-server/src/test/java/org/apache/hadoop/hbase/util/MultiThreadedAction.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/MultiThreadedAction.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/MultiThreadedAction.java index 5098e0b..410dd0c 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/MultiThreadedAction.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/MultiThreadedAction.java @@ -101,7 +101,7 @@ public abstract class MultiThreadedAction { @Override public byte[] getDeterministicUniqueKey(long keyBase) { - return LoadTestKVGenerator.md5PrefixedKey(keyBase).getBytes(); + return Bytes.toBytes(LoadTestKVGenerator.md5PrefixedKey(keyBase)); } @Override @@ -114,7 +114,7 @@ public abstract class MultiThreadedAction { int numColumns = minColumnsPerKey + random.nextInt(maxColumnsPerKey - minColumnsPerKey + 1); byte[][] columns = new byte[numColumns][]; for (int i = 0; i < numColumns; ++i) { - columns[i] = Integer.toString(i).getBytes(); + columns[i] = Bytes.toBytes(Integer.toString(i)); } return columns; } http://git-wip-us.apache.org/repos/asf/hbase/blob/b1269ec5/hbase-server/src/test/java/org/apache/hadoop/hbase/util/MultiThreadedReader.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/MultiThreadedReader.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/MultiThreadedReader.java index 447cca8..6864366 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/MultiThreadedReader.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/MultiThreadedReader.java @@ -201,8 +201,7 @@ public class MultiThreadedReader extends MultiThreadedAction "to read " + k + " is out of range (startKey=" + startKey + ", endKey=" + endKey + ")"); } - if (k % numThreads != readerId || - writer != null && writer.failedToWriteKey(k)) { + if (k % numThreads != readerId || (writer != null && writer.failedToWriteKey(k))) { // Skip keys that this thread should not read, as well as the keys // that we know the writer failed to write. continue; http://git-wip-us.apache.org/repos/asf/hbase/blob/b1269ec5/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestBoundedPriorityBlockingQueue.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestBoundedPriorityBlockingQueue.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestBoundedPriorityBlockingQueue.java index 7112d50..7746bea 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestBoundedPriorityBlockingQueue.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestBoundedPriorityBlockingQueue.java @@ -42,7 +42,7 @@ import org.junit.experimental.categories.Category; public class TestBoundedPriorityBlockingQueue { private final static int CAPACITY = 16; - class TestObject { + static class TestObject { private final int priority; private final int seqId; @@ -60,7 +60,7 @@ public class TestBoundedPriorityBlockingQueue { } } - class TestObjectComparator implements Comparator<TestObject> { + static class TestObjectComparator implements Comparator<TestObject> { public TestObjectComparator() {} @Override @@ -208,6 +208,7 @@ public class TestBoundedPriorityBlockingQueue { final CyclicBarrier threadsStarted = new CyclicBarrier(2); ExecutorService executor = Executors.newFixedThreadPool(2); executor.execute(new Runnable() { + @Override public void run() { try { assertNull(queue.poll(1000, TimeUnit.MILLISECONDS)); @@ -221,6 +222,7 @@ public class TestBoundedPriorityBlockingQueue { }); executor.execute(new Runnable() { + @Override public void run() { try { threadsStarted.await(); http://git-wip-us.apache.org/repos/asf/hbase/blob/b1269ec5/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestByteBuffUtils.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestByteBuffUtils.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestByteBuffUtils.java index 4c6990e..ecc6611 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestByteBuffUtils.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestByteBuffUtils.java @@ -36,7 +36,7 @@ public class TestByteBuffUtils { ByteBuffer bb2 = ByteBuffer.allocate(50); MultiByteBuff src = new MultiByteBuff(bb1, bb2); for (int i = 0; i < 7; i++) { - src.putLong(8l); + src.putLong(8L); } src.put((byte) 1); src.put((byte) 1); @@ -58,7 +58,7 @@ public class TestByteBuffUtils { bb3 = ByteBuffer.allocate(100); SingleByteBuff sbb = new SingleByteBuff(bb3); for (int i = 0; i < 7; i++) { - sbb.putLong(8l); + sbb.putLong(8L); } sbb.put((byte) 1); sbb.put((byte) 1); http://git-wip-us.apache.org/repos/asf/hbase/blob/b1269ec5/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestFSHDFSUtils.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestFSHDFSUtils.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestFSHDFSUtils.java index 8f503e0..a554e99 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestFSHDFSUtils.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestFSHDFSUtils.java @@ -161,11 +161,12 @@ public class TestFSHDFSUtils { /** * Version of DFS that has HDFS-4525 in it. */ - class IsFileClosedDistributedFileSystem extends DistributedFileSystem { + static class IsFileClosedDistributedFileSystem extends DistributedFileSystem { /** * Close status of a file. Copied over from HDFS-4525 * @return true if file is already closed **/ + @Override public boolean isFileClosed(Path f) throws IOException{ return false; } http://git-wip-us.apache.org/repos/asf/hbase/blob/b1269ec5/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestFSVisitor.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestFSVisitor.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestFSVisitor.java index cb23a0b..be302d5 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestFSVisitor.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestFSVisitor.java @@ -77,6 +77,7 @@ public class TestFSVisitor { final Set<String> families = new HashSet<>(); final Set<String> hfiles = new HashSet<>(); FSVisitor.visitTableStoreFiles(fs, tableDir, new FSVisitor.StoreFileVisitor() { + @Override public void storeFile(final String region, final String family, final String hfileName) throws IOException { regions.add(region); @@ -84,9 +85,9 @@ public class TestFSVisitor { hfiles.add(hfileName); } }); - assertEquals(tableRegions, regions); - assertEquals(tableFamilies, families); - assertEquals(tableHFiles, hfiles); + assertEquals(regions, tableRegions); + assertEquals(families, tableFamilies); + assertEquals(hfiles, tableHFiles); } /* http://git-wip-us.apache.org/repos/asf/hbase/blob/b1269ec5/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestHBaseFsckEncryption.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestHBaseFsckEncryption.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestHBaseFsckEncryption.java index 641f66e..31f1909 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestHBaseFsckEncryption.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestHBaseFsckEncryption.java @@ -131,12 +131,12 @@ public class TestHBaseFsckEncryption { // Insure HBck doesn't consider them corrupt HBaseFsck res = HbckTestingUtil.doHFileQuarantine(conf, htd.getTableName()); - assertEquals(res.getRetCode(), 0); + assertEquals(0, res.getRetCode()); HFileCorruptionChecker hfcc = res.getHFilecorruptionChecker(); - assertEquals(hfcc.getCorrupted().size(), 0); - assertEquals(hfcc.getFailures().size(), 0); - assertEquals(hfcc.getQuarantined().size(), 0); - assertEquals(hfcc.getMissing().size(), 0); + assertEquals(0, hfcc.getCorrupted().size()); + assertEquals(0, hfcc.getFailures().size()); + assertEquals(0, hfcc.getQuarantined().size()); + assertEquals(0, hfcc.getMissing().size()); } private List<Path> findStorefilePaths(TableName tableName) throws Exception { http://git-wip-us.apache.org/repos/asf/hbase/blob/b1269ec5/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestHBaseFsckMOB.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestHBaseFsckMOB.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestHBaseFsckMOB.java index 3661207..2548ecf 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestHBaseFsckMOB.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestHBaseFsckMOB.java @@ -118,18 +118,18 @@ public class TestHBaseFsckMOB extends BaseTestHBaseFsck { // A corrupt mob file doesn't abort the start of regions, so we can enable the table. admin.enableTable(table); HBaseFsck res = HbckTestingUtil.doHFileQuarantine(conf, table); - assertEquals(res.getRetCode(), 0); + assertEquals(0, res.getRetCode()); HFileCorruptionChecker hfcc = res.getHFilecorruptionChecker(); - assertEquals(hfcc.getHFilesChecked(), 4); - assertEquals(hfcc.getCorrupted().size(), 0); - assertEquals(hfcc.getFailures().size(), 0); - assertEquals(hfcc.getQuarantined().size(), 0); - assertEquals(hfcc.getMissing().size(), 0); - assertEquals(hfcc.getMobFilesChecked(), 5); - assertEquals(hfcc.getCorruptedMobFiles().size(), 1); - assertEquals(hfcc.getFailureMobFiles().size(), 0); - assertEquals(hfcc.getQuarantinedMobFiles().size(), 1); - assertEquals(hfcc.getMissedMobFiles().size(), 0); + assertEquals(4, hfcc.getHFilesChecked()); + assertEquals(0, hfcc.getCorrupted().size()); + assertEquals(0, hfcc.getFailures().size()); + assertEquals(0, hfcc.getQuarantined().size()); + assertEquals(0, hfcc.getMissing().size()); + assertEquals(5, hfcc.getMobFilesChecked()); + assertEquals(1, hfcc.getCorruptedMobFiles().size()); + assertEquals(0, hfcc.getFailureMobFiles().size()); + assertEquals(1, hfcc.getQuarantinedMobFiles().size()); + assertEquals(0, hfcc.getMissedMobFiles().size()); String quarantinedMobFile = hfcc.getQuarantinedMobFiles().iterator().next().getName(); assertEquals(corruptMobFile, quarantinedMobFile); } finally { http://git-wip-us.apache.org/repos/asf/hbase/blob/b1269ec5/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestJSONMetricUtil.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestJSONMetricUtil.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestJSONMetricUtil.java index 13c6df5..bb68898 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestJSONMetricUtil.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestJSONMetricUtil.java @@ -86,7 +86,7 @@ public class TestJSONMetricUtil { Hashtable<String, String> properties = JSONMetricUtil.buldKeyValueTable(keys, values); ObjectName testObject = JSONMetricUtil.buildObjectName(JSONMetricUtil.JAVA_LANG_DOMAIN, properties); - assertEquals(testObject.getDomain(), JSONMetricUtil.JAVA_LANG_DOMAIN); + assertEquals(JSONMetricUtil.JAVA_LANG_DOMAIN, testObject.getDomain()); assertEquals(testObject.getKeyPropertyList(), properties); } http://git-wip-us.apache.org/repos/asf/hbase/blob/b1269ec5/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestMiniClusterLoadParallel.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestMiniClusterLoadParallel.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestMiniClusterLoadParallel.java index 7b1cd2d..e6b05e9 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestMiniClusterLoadParallel.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestMiniClusterLoadParallel.java @@ -40,6 +40,7 @@ public class TestMiniClusterLoadParallel super(isMultiPut, encoding); } + @Override @Test(timeout=TIMEOUT_MS) public void loadTest() throws Exception { prepareForLoadTest(); http://git-wip-us.apache.org/repos/asf/hbase/blob/b1269ec5/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestRegionSplitCalculator.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestRegionSplitCalculator.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestRegionSplitCalculator.java index fd86beb..eee3030 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestRegionSplitCalculator.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestRegionSplitCalculator.java @@ -68,6 +68,7 @@ public class TestRegionSplitCalculator { return end; } + @Override public String toString() { return "[" + Bytes.toString(start) + ", " + Bytes.toString(end) + "]"; } @@ -135,8 +136,7 @@ public class TestRegionSplitCalculator { LOG.info("Standard"); String res = dump(sc.getSplits(), regions); checkDepths(sc.getSplits(), regions, 1, 1, 1, 0); - assertEquals(res, "A:\t[A, B]\t\n" + "B:\t[B, C]\t\n" + "C:\t[C, D]\t\n" - + "D:\t\n"); + assertEquals("A:\t[A, B]\t\n" + "B:\t[B, C]\t\n" + "C:\t[C, D]\t\nD:\t\n", res); } @Test http://git-wip-us.apache.org/repos/asf/hbase/blob/b1269ec5/hbase-server/src/test/java/org/apache/hadoop/hbase/util/test/LoadTestDataGenerator.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/test/LoadTestDataGenerator.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/test/LoadTestDataGenerator.java index 3650706..1aab1f8 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/test/LoadTestDataGenerator.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/test/LoadTestDataGenerator.java @@ -20,10 +20,11 @@ import java.io.IOException; import java.util.Random; import java.util.Set; -import org.apache.yetus.audience.InterfaceAudience; +import org.apache.hadoop.hbase.util.Bytes; import org.apache.hadoop.hbase.client.Get; import org.apache.hadoop.hbase.client.Mutation; import org.apache.hadoop.hbase.util.LoadTestKVGenerator; +import org.apache.yetus.audience.InterfaceAudience; /** * A generator of random data (keys/cfs/columns/values) for load testing. @@ -35,11 +36,11 @@ public abstract class LoadTestDataGenerator { // The mutate info column stores information // about update done to this column family this row. - public final static byte[] MUTATE_INFO = "mutate_info".getBytes(); + public final static byte[] MUTATE_INFO = Bytes.toBytes("mutate_info"); // The increment column always has a long value, // which can be incremented later on during updates. - public final static byte[] INCREMENT = "increment".getBytes(); + public final static byte[] INCREMENT = Bytes.toBytes("increment"); protected String[] args; http://git-wip-us.apache.org/repos/asf/hbase/blob/b1269ec5/hbase-server/src/test/java/org/apache/hadoop/hbase/wal/TestWALFactory.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/wal/TestWALFactory.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/wal/TestWALFactory.java index f0f35e7..ceb43d5 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/wal/TestWALFactory.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/wal/TestWALFactory.java @@ -476,7 +476,7 @@ public class TestWALFactory { reader.close(); // Reset the lease period - setLeasePeriod.invoke(cluster, new Object[]{new Long(60000), new Long(3600000)}); + setLeasePeriod.invoke(cluster, new Object[]{ 60000L, 3600000L }); } /** http://git-wip-us.apache.org/repos/asf/hbase/blob/b1269ec5/hbase-server/src/test/java/org/apache/hadoop/hbase/wal/TestWALSplit.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/wal/TestWALSplit.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/wal/TestWALSplit.java index a1206aa..dded506 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/wal/TestWALSplit.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/wal/TestWALSplit.java @@ -131,9 +131,9 @@ public class TestWALSplit { private static final String FILENAME_BEING_SPLIT = "testfile"; private static final TableName TABLE_NAME = TableName.valueOf("t1"); - private static final byte[] FAMILY = "f1".getBytes(); - private static final byte[] QUALIFIER = "q1".getBytes(); - private static final byte[] VALUE = "v1".getBytes(); + private static final byte[] FAMILY = Bytes.toBytes("f1"); + private static final byte[] QUALIFIER = Bytes.toBytes("q1"); + private static final byte[] VALUE = Bytes.toBytes("v1"); private static final String WAL_FILE_PREFIX = "wal.dat."; private static List<String> REGIONS = new ArrayList<>(); private static final String HBASE_SKIP_ERRORS = "hbase.hlog.split.skip.errors"; @@ -341,7 +341,7 @@ public class TestWALSplit { while (!stop.get()) { try { long seq = appendEntry(writer, TABLE_NAME, regionBytes, - ("r" + editsCount.get()).getBytes(), regionBytes, QUALIFIER, VALUE, 0); + Bytes.toBytes("r" + editsCount.get()), regionBytes, QUALIFIER, VALUE, 0); long count = editsCount.incrementAndGet(); LOG.info(getName() + " sync count=" + count + ", seq=" + seq); try { @@ -407,7 +407,7 @@ public class TestWALSplit { TableName.META_TABLE_NAME, 1, now, HConstants.DEFAULT_CLUSTER_ID), new WALEdit()); Path parent = WALSplitter.getRegionDirRecoveredEditsDir(regiondir); - assertEquals(parent.getName(), HConstants.RECOVERED_EDITS_DIR); + assertEquals(HConstants.RECOVERED_EDITS_DIR, parent.getName()); fs.createNewFile(parent); // create a recovered.edits file Path p = WALSplitter.getRegionSplitEditsPath(fs, entry, HBASEDIR, @@ -588,8 +588,8 @@ public class TestWALSplit { archivedLogs.add(log.getPath().getName()); } LOG.debug(archived.toString()); - assertEquals(failureType.name() + ": expected to find all of our wals corrupt.", - walDirContents, archivedLogs); + assertEquals(failureType.name() + ": expected to find all of our wals corrupt.", archivedLogs, + walDirContents); } } @@ -687,7 +687,7 @@ public class TestWALSplit { // should not have stored the EOF files as corrupt FileStatus[] archivedLogs = fs.listStatus(CORRUPTDIR); - assertEquals(archivedLogs.length, 0); + assertEquals(0, archivedLogs.length); } @@ -749,7 +749,7 @@ public class TestWALSplit { InstrumentedLogWriter.activateFailure = false; appendEntry(writer, TABLE_NAME, Bytes.toBytes(region), - ("r" + 999).getBytes(), FAMILY, QUALIFIER, VALUE, 0); + Bytes.toBytes("r" + 999), FAMILY, QUALIFIER, VALUE, 0); writer.close(); try { @@ -1206,8 +1206,8 @@ public class TestWALSplit { int prefix = 0; for (String region : REGIONS) { String row_key = region + prefix++ + i + j; - appendEntry(ws[i], TABLE_NAME, region.getBytes(), row_key.getBytes(), FAMILY, QUALIFIER, - VALUE, seq++); + appendEntry(ws[i], TABLE_NAME, Bytes.toBytes(region), Bytes.toBytes(row_key), FAMILY, + QUALIFIER, VALUE, seq++); if (numRegionEventsAdded < regionEvents) { numRegionEventsAdded ++; @@ -1233,7 +1233,7 @@ public class TestWALSplit { Path tdir = FSUtils.getTableDir(rootdir, table); @SuppressWarnings("deprecation") Path editsdir = WALSplitter.getRegionDirRecoveredEditsDir(HRegion.getRegionDir(tdir, - Bytes.toString(region.getBytes()))); + Bytes.toString(Bytes.toBytes(region)))); FileStatus[] files = fs.listStatus(editsdir, new PathFilter() { @Override public boolean accept(Path p) { @@ -1260,46 +1260,46 @@ public class TestWALSplit { in.close(); switch (corruption) { - case APPEND_GARBAGE: - fs.delete(path, false); - out = fs.create(path); - out.write(corrupted_bytes); - out.write("-----".getBytes()); - closeOrFlush(close, out); - break; - - case INSERT_GARBAGE_ON_FIRST_LINE: - fs.delete(path, false); - out = fs.create(path); - out.write(0); - out.write(corrupted_bytes); - closeOrFlush(close, out); - break; - - case INSERT_GARBAGE_IN_THE_MIDDLE: - fs.delete(path, false); - out = fs.create(path); - int middle = (int) Math.floor(corrupted_bytes.length / 2); - out.write(corrupted_bytes, 0, middle); - out.write(0); - out.write(corrupted_bytes, middle, corrupted_bytes.length - middle); - closeOrFlush(close, out); - break; - - case TRUNCATE: - fs.delete(path, false); - out = fs.create(path); - out.write(corrupted_bytes, 0, fileSize - - (32 + ProtobufLogReader.PB_WAL_COMPLETE_MAGIC.length + Bytes.SIZEOF_INT)); - closeOrFlush(close, out); - break; - - case TRUNCATE_TRAILER: - fs.delete(path, false); - out = fs.create(path); - out.write(corrupted_bytes, 0, fileSize - Bytes.SIZEOF_INT);// trailer is truncated. - closeOrFlush(close, out); - break; + case APPEND_GARBAGE: + fs.delete(path, false); + out = fs.create(path); + out.write(corrupted_bytes); + out.write(Bytes.toBytes("-----")); + closeOrFlush(close, out); + break; + + case INSERT_GARBAGE_ON_FIRST_LINE: + fs.delete(path, false); + out = fs.create(path); + out.write(0); + out.write(corrupted_bytes); + closeOrFlush(close, out); + break; + + case INSERT_GARBAGE_IN_THE_MIDDLE: + fs.delete(path, false); + out = fs.create(path); + int middle = (int) Math.floor(corrupted_bytes.length / 2); + out.write(corrupted_bytes, 0, middle); + out.write(0); + out.write(corrupted_bytes, middle, corrupted_bytes.length - middle); + closeOrFlush(close, out); + break; + + case TRUNCATE: + fs.delete(path, false); + out = fs.create(path); + out.write(corrupted_bytes, 0, fileSize + - (32 + ProtobufLogReader.PB_WAL_COMPLETE_MAGIC.length + Bytes.SIZEOF_INT)); + closeOrFlush(close, out); + break; + + case TRUNCATE_TRAILER: + fs.delete(path, false); + out = fs.create(path); + out.write(corrupted_bytes, 0, fileSize - Bytes.SIZEOF_INT);// trailer is truncated. + closeOrFlush(close, out); + break; } } @@ -1360,14 +1360,14 @@ public class TestWALSplit { WALProtos.RegionEventDescriptor regionOpenDesc = ProtobufUtil.toRegionEventDescriptor( WALProtos.RegionEventDescriptor.EventType.REGION_OPEN, TABLE_NAME.toBytes(), - region.getBytes(), - String.valueOf(region.hashCode()).getBytes(), + Bytes.toBytes(region), + Bytes.toBytes(String.valueOf(region.hashCode())), 1, ServerName.parseServerName("ServerName:9099"), ImmutableMap.<byte[], List<Path>>of()); final long time = EnvironmentEdgeManager.currentTime(); - KeyValue kv = new KeyValue(region.getBytes(), WALEdit.METAFAMILY, WALEdit.REGION_EVENT, + KeyValue kv = new KeyValue(Bytes.toBytes(region), WALEdit.METAFAMILY, WALEdit.REGION_EVENT, time, regionOpenDesc.toByteArray()); - final WALKeyImpl walKey = new WALKeyImpl(region.getBytes(), TABLE_NAME, 1, time, + final WALKeyImpl walKey = new WALKeyImpl(Bytes.toBytes(region), TABLE_NAME, 1, time, HConstants.DEFAULT_CLUSTER_ID); w.append( new Entry(walKey, new WALEdit().add(kv))); http://git-wip-us.apache.org/repos/asf/hbase/blob/b1269ec5/hbase-server/src/test/java/org/apache/hadoop/hbase/wal/TestWALSplitBoundedLogWriterCreation.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/wal/TestWALSplitBoundedLogWriterCreation.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/wal/TestWALSplitBoundedLogWriterCreation.java index 844cb3a..400d12b 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/wal/TestWALSplitBoundedLogWriterCreation.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/wal/TestWALSplitBoundedLogWriterCreation.java @@ -35,6 +35,7 @@ public class TestWALSplitBoundedLogWriterCreation extends TestWALSplit{ /** * The logic of this test has conflict with the limit writers split logic, skip this test */ + @Override @Test(timeout=300000) @Ignore public void testThreadingSlowWriterSmallBuffer() throws Exception { http://git-wip-us.apache.org/repos/asf/hbase/blob/b1269ec5/hbase-server/src/test/java/org/apache/hadoop/hbase/zookeeper/TestZooKeeperACL.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/zookeeper/TestZooKeeperACL.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/zookeeper/TestZooKeeperACL.java index 4e67b91..3e9e650 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/zookeeper/TestZooKeeperACL.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/zookeeper/TestZooKeeperACL.java @@ -129,10 +129,10 @@ public class TestZooKeeperACL { List<ACL> acls = zkw.getRecoverableZooKeeper().getZooKeeper() .getACL("/hbase", new Stat()); - assertEquals(acls.size(),1); - assertEquals(acls.get(0).getId().getScheme(),"sasl"); - assertEquals(acls.get(0).getId().getId(),"hbase"); - assertEquals(acls.get(0).getPerms(), ZooDefs.Perms.ALL); + assertEquals(1, acls.size()); + assertEquals("sasl", acls.get(0).getId().getScheme()); + assertEquals("hbase", acls.get(0).getId().getId()); + assertEquals(ZooDefs.Perms.ALL, acls.get(0).getPerms()); } /** @@ -148,20 +148,20 @@ public class TestZooKeeperACL { List<ACL> acls = zkw.getRecoverableZooKeeper().getZooKeeper() .getACL("/hbase/root-region-server", new Stat()); - assertEquals(acls.size(),2); + assertEquals(2, acls.size()); boolean foundWorldReadableAcl = false; boolean foundHBaseOwnerAcl = false; for(int i = 0; i < 2; i++) { if (acls.get(i).getId().getScheme().equals("world") == true) { - assertEquals(acls.get(0).getId().getId(),"anyone"); - assertEquals(acls.get(0).getPerms(), ZooDefs.Perms.READ); + assertEquals("anyone", acls.get(0).getId().getId()); + assertEquals(ZooDefs.Perms.READ, acls.get(0).getPerms()); foundWorldReadableAcl = true; } else { if (acls.get(i).getId().getScheme().equals("sasl") == true) { - assertEquals(acls.get(1).getId().getId(),"hbase"); - assertEquals(acls.get(1).getId().getScheme(),"sasl"); + assertEquals("hbase", acls.get(1).getId().getId()); + assertEquals("sasl", acls.get(1).getId().getScheme()); foundHBaseOwnerAcl = true; } else { // error: should not get here: test fails. assertTrue(false); @@ -185,19 +185,19 @@ public class TestZooKeeperACL { List<ACL> acls = zkw.getRecoverableZooKeeper().getZooKeeper() .getACL("/hbase/master", new Stat()); - assertEquals(acls.size(),2); + assertEquals(2, acls.size()); boolean foundWorldReadableAcl = false; boolean foundHBaseOwnerAcl = false; for(int i = 0; i < 2; i++) { if (acls.get(i).getId().getScheme().equals("world") == true) { - assertEquals(acls.get(0).getId().getId(),"anyone"); - assertEquals(acls.get(0).getPerms(), ZooDefs.Perms.READ); + assertEquals("anyone", acls.get(0).getId().getId()); + assertEquals(ZooDefs.Perms.READ, acls.get(0).getPerms()); foundWorldReadableAcl = true; } else { if (acls.get(i).getId().getScheme().equals("sasl") == true) { - assertEquals(acls.get(1).getId().getId(),"hbase"); - assertEquals(acls.get(1).getId().getScheme(),"sasl"); + assertEquals("hbase", acls.get(1).getId().getId()); + assertEquals("sasl", acls.get(1).getId().getScheme()); foundHBaseOwnerAcl = true; } else { // error: should not get here: test fails. assertTrue(false); @@ -221,19 +221,19 @@ public class TestZooKeeperACL { List<ACL> acls = zkw.getRecoverableZooKeeper().getZooKeeper() .getACL("/hbase/hbaseid", new Stat()); - assertEquals(acls.size(),2); + assertEquals(2, acls.size()); boolean foundWorldReadableAcl = false; boolean foundHBaseOwnerAcl = false; for(int i = 0; i < 2; i++) { if (acls.get(i).getId().getScheme().equals("world") == true) { - assertEquals(acls.get(0).getId().getId(),"anyone"); - assertEquals(acls.get(0).getPerms(), ZooDefs.Perms.READ); + assertEquals("anyone", acls.get(0).getId().getId()); + assertEquals(ZooDefs.Perms.READ, acls.get(0).getPerms()); foundWorldReadableAcl = true; } else { if (acls.get(i).getId().getScheme().equals("sasl") == true) { - assertEquals(acls.get(1).getId().getId(),"hbase"); - assertEquals(acls.get(1).getId().getScheme(),"sasl"); + assertEquals("hbase", acls.get(1).getId().getId()); + assertEquals("sasl", acls.get(1).getId().getScheme()); foundHBaseOwnerAcl = true; } else { // error: should not get here: test fails. assertTrue(false); @@ -257,10 +257,10 @@ public class TestZooKeeperACL { ZKUtil.createWithParents(zkw, "/testACLNode"); List<ACL> acls = zkw.getRecoverableZooKeeper().getZooKeeper() .getACL("/testACLNode", new Stat()); - assertEquals(acls.size(),1); - assertEquals(acls.get(0).getId().getScheme(),"sasl"); - assertEquals(acls.get(0).getId().getId(),"hbase"); - assertEquals(acls.get(0).getPerms(), ZooDefs.Perms.ALL); + assertEquals(1, acls.size()); + assertEquals("sasl", acls.get(0).getId().getScheme()); + assertEquals("hbase", acls.get(0).getId().getId()); + assertEquals(ZooDefs.Perms.ALL, acls.get(0).getPerms()); } /** @@ -281,7 +281,7 @@ public class TestZooKeeperACL { saslConfFile.getAbsolutePath()); testJaasConfig = ZKUtil.isSecureZooKeeper(new Configuration(TEST_UTIL.getConfiguration())); - assertEquals(testJaasConfig, false); + assertEquals(false, testJaasConfig); saslConfFile.delete(); } @@ -295,13 +295,13 @@ public class TestZooKeeperACL { Configuration config = new Configuration(HBaseConfiguration.create()); boolean testJaasConfig = ZKUtil.isSecureZooKeeper(config); - assertEquals(testJaasConfig, false); + assertEquals(false, testJaasConfig); // Now set authentication scheme to Kerberos still it should return false // because no configuration set config.set("hbase.security.authentication", "kerberos"); testJaasConfig = ZKUtil.isSecureZooKeeper(config); - assertEquals(testJaasConfig, false); + assertEquals(false, testJaasConfig); // Now set programmatic options related to security config.set(HConstants.ZK_CLIENT_KEYTAB_FILE, "/dummy/file");