Repository: hadoop Updated Branches: refs/heads/HDFS-7285 fbf7e81ca -> 1d37a8812
http://git-wip-us.apache.org/repos/asf/hadoop/blob/1d37a881/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestEncryptionZones.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestEncryptionZones.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestEncryptionZones.java index 1f317b8..6e2ec5e 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestEncryptionZones.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestEncryptionZones.java @@ -737,7 +737,7 @@ public class TestEncryptionZones { version, new byte[suite.getAlgorithmBlockSize()], new byte[suite.getAlgorithmBlockSize()], "fakeKey", "fakeVersion"), - (byte) 0, null, 0)) + (byte) 0, null)) .when(mcp) .create(anyString(), (FsPermission) anyObject(), anyString(), (EnumSetWritable<CreateFlag>) anyObject(), anyBoolean(), http://git-wip-us.apache.org/repos/asf/hadoop/blob/1d37a881/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestErasureCodingZones.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestErasureCodingZones.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestErasureCodingZones.java index 1a10ebf..a878501 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestErasureCodingZones.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestErasureCodingZones.java @@ -22,10 +22,10 @@ import org.apache.hadoop.fs.CommonConfigurationKeys; import org.apache.hadoop.fs.Path; import org.apache.hadoop.fs.permission.FsPermission; import org.apache.hadoop.hdfs.protocol.HdfsFileStatus; -import org.apache.hadoop.hdfs.server.namenode.ErasureCodingSchemaManager; +import org.apache.hadoop.hdfs.server.namenode.ErasureCodingPolicyManager; import org.apache.hadoop.hdfs.server.namenode.FSNamesystem; import org.apache.hadoop.hdfs.server.namenode.INode; -import org.apache.hadoop.io.erasurecode.ECSchema; +import org.apache.hadoop.hdfs.protocol.ErasureCodingPolicy; import org.junit.After; import org.junit.Before; import org.junit.Test; @@ -65,7 +65,7 @@ public class TestErasureCodingZones { fs.mkdir(testDir, FsPermission.getDirDefault()); /* Normal creation of an erasure coding zone */ - fs.getClient().createErasureCodingZone(testDir.toString(), null, 0); + fs.getClient().createErasureCodingZone(testDir.toString(), null); /* Verify files under the zone are striped */ final Path ECFilePath = new Path(testDir, "foo"); @@ -78,7 +78,7 @@ public class TestErasureCodingZones { fs.mkdir(notEmpty, FsPermission.getDirDefault()); fs.create(new Path(notEmpty, "foo")); try { - fs.getClient().createErasureCodingZone(notEmpty.toString(), null, 0); + fs.getClient().createErasureCodingZone(notEmpty.toString(), null); fail("Erasure coding zone on non-empty dir"); } catch (IOException e) { assertExceptionContains("erasure coding zone for a non-empty directory", e); @@ -88,10 +88,10 @@ public class TestErasureCodingZones { final Path zone1 = new Path("/zone1"); final Path zone2 = new Path(zone1, "zone2"); fs.mkdir(zone1, FsPermission.getDirDefault()); - fs.getClient().createErasureCodingZone(zone1.toString(), null, 0); + fs.getClient().createErasureCodingZone(zone1.toString(), null); fs.mkdir(zone2, FsPermission.getDirDefault()); try { - fs.getClient().createErasureCodingZone(zone2.toString(), null, 0); + fs.getClient().createErasureCodingZone(zone2.toString(), null); fail("Nested erasure coding zones"); } catch (IOException e) { assertExceptionContains("already in an erasure coding zone", e); @@ -101,7 +101,7 @@ public class TestErasureCodingZones { final Path fPath = new Path("/file"); fs.create(fPath); try { - fs.getClient().createErasureCodingZone(fPath.toString(), null, 0); + fs.getClient().createErasureCodingZone(fPath.toString(), null); fail("Erasure coding zone on file"); } catch (IOException e) { assertExceptionContains("erasure coding zone for a file", e); @@ -114,8 +114,8 @@ public class TestErasureCodingZones { final Path dstECDir = new Path("/dstEC"); fs.mkdir(srcECDir, FsPermission.getDirDefault()); fs.mkdir(dstECDir, FsPermission.getDirDefault()); - fs.getClient().createErasureCodingZone(srcECDir.toString(), null, 0); - fs.getClient().createErasureCodingZone(dstECDir.toString(), null, 0); + fs.getClient().createErasureCodingZone(srcECDir.toString(), null); + fs.getClient().createErasureCodingZone(dstECDir.toString(), null); final Path srcFile = new Path(srcECDir, "foo"); fs.create(srcFile); @@ -160,7 +160,7 @@ public class TestErasureCodingZones { public void testReplication() throws IOException { final Path testDir = new Path("/ec"); fs.mkdir(testDir, FsPermission.getDirDefault()); - fs.createErasureCodingZone(testDir, null, 0); + fs.createErasureCodingZone(testDir, null); final Path fooFile = new Path(testDir, "foo"); // create ec file with replication=0 fs.create(fooFile, FsPermission.getFileDefault(), true, @@ -177,47 +177,47 @@ public class TestErasureCodingZones { } @Test - public void testGetErasureCodingInfoWithSystemDefaultSchema() throws Exception { + public void testGetErasureCodingInfoWithSystemDefaultECPolicy() throws Exception { String src = "/ec"; final Path ecDir = new Path(src); fs.mkdir(ecDir, FsPermission.getDirDefault()); // dir ECInfo before creating ec zone - assertNull(fs.getClient().getFileInfo(src).getECSchema()); + assertNull(fs.getClient().getFileInfo(src).getErasureCodingPolicy()); // dir ECInfo after creating ec zone - fs.getClient().createErasureCodingZone(src, null, 0); //Default one will be used. - ECSchema sysDefaultSchema = ErasureCodingSchemaManager.getSystemDefaultSchema(); - verifyErasureCodingInfo(src, sysDefaultSchema); + fs.getClient().createErasureCodingZone(src, null); //Default one will be used. + ErasureCodingPolicy sysDefaultECPolicy = ErasureCodingPolicyManager.getSystemDefaultPolicy(); + verifyErasureCodingInfo(src, sysDefaultECPolicy); fs.create(new Path(ecDir, "child1")).close(); // verify for the files in ec zone - verifyErasureCodingInfo(src + "/child1", sysDefaultSchema); + verifyErasureCodingInfo(src + "/child1", sysDefaultECPolicy); } @Test public void testGetErasureCodingInfo() throws Exception { - ECSchema[] sysSchemas = ErasureCodingSchemaManager.getSystemSchemas(); - assertTrue("System schemas should be of only 1 for now", - sysSchemas.length == 1); + ErasureCodingPolicy[] sysECPolicies = ErasureCodingPolicyManager.getSystemPolices(); + assertTrue("System ecPolicies should be of only 1 for now", + sysECPolicies.length == 1); - ECSchema usingSchema = sysSchemas[0]; + ErasureCodingPolicy usingECPolicy = sysECPolicies[0]; String src = "/ec2"; final Path ecDir = new Path(src); fs.mkdir(ecDir, FsPermission.getDirDefault()); // dir ECInfo before creating ec zone - assertNull(fs.getClient().getFileInfo(src).getECSchema()); + assertNull(fs.getClient().getFileInfo(src).getErasureCodingPolicy()); // dir ECInfo after creating ec zone - fs.getClient().createErasureCodingZone(src, usingSchema, 0); - verifyErasureCodingInfo(src, usingSchema); + fs.getClient().createErasureCodingZone(src, usingECPolicy); + verifyErasureCodingInfo(src, usingECPolicy); fs.create(new Path(ecDir, "child1")).close(); // verify for the files in ec zone - verifyErasureCodingInfo(src + "/child1", usingSchema); + verifyErasureCodingInfo(src + "/child1", usingECPolicy); } private void verifyErasureCodingInfo( - String src, ECSchema usingSchema) throws IOException { + String src, ErasureCodingPolicy usingECPolicy) throws IOException { HdfsFileStatus hdfsFileStatus = fs.getClient().getFileInfo(src); - ECSchema schema = hdfsFileStatus.getECSchema(); - assertNotNull(schema); - assertEquals("Actually used schema should be equal with target schema", - usingSchema, schema); + ErasureCodingPolicy ecPolicy = hdfsFileStatus.getErasureCodingPolicy(); + assertNotNull(ecPolicy); + assertEquals("Actually used ecPolicy should be equal with target ecPolicy", + usingECPolicy, ecPolicy); } } http://git-wip-us.apache.org/repos/asf/hadoop/blob/1d37a881/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestFileStatusWithECPolicy.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestFileStatusWithECPolicy.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestFileStatusWithECPolicy.java new file mode 100644 index 0000000..4610ced --- /dev/null +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestFileStatusWithECPolicy.java @@ -0,0 +1,65 @@ +package org.apache.hadoop.hdfs; + +import static org.junit.Assert.assertNotNull; +import static org.junit.Assert.assertNull; +import static org.junit.Assert.assertTrue; + +import java.io.IOException; + +import org.apache.hadoop.conf.Configuration; +import org.apache.hadoop.fs.Path; +import org.apache.hadoop.fs.permission.FsPermission; +import org.apache.hadoop.hdfs.server.namenode.ErasureCodingPolicyManager; +import org.apache.hadoop.hdfs.protocol.ErasureCodingPolicy; +import org.junit.After; +import org.junit.Before; +import org.junit.Test; + +public class TestFileStatusWithECPolicy { + private MiniDFSCluster cluster; + private DistributedFileSystem fs; + private DFSClient client; + + @Before + public void before() throws IOException { + cluster = + new MiniDFSCluster.Builder(new Configuration()).numDataNodes(1).build(); + cluster.waitActive(); + fs = cluster.getFileSystem(); + client = fs.getClient(); + } + + @After + public void after() { + if (cluster != null) { + cluster.shutdown(); + } + } + + @Test + public void testFileStatusWithECPolicy() throws Exception { + // test directory not in EC zone + final Path dir = new Path("/foo"); + assertTrue(fs.mkdir(dir, FsPermission.getDirDefault())); + assertNull(client.getFileInfo(dir.toString()).getErasureCodingPolicy()); + // test file not in EC zone + final Path file = new Path(dir, "foo"); + fs.create(file).close(); + assertNull(client.getFileInfo(file.toString()).getErasureCodingPolicy()); + fs.delete(file, true); + + final ErasureCodingPolicy ecPolicy1 = ErasureCodingPolicyManager.getSystemDefaultPolicy(); + // create EC zone on dir + fs.createErasureCodingZone(dir, ecPolicy1); + final ErasureCodingPolicy ecPolicy2 = client.getFileInfo(dir.toUri().getPath()).getErasureCodingPolicy(); + assertNotNull(ecPolicy2); + assertTrue(ecPolicy1.equals(ecPolicy2)); + + // test file in EC zone + fs.create(file).close(); + final ErasureCodingPolicy ecPolicy3 = + fs.getClient().getFileInfo(file.toUri().getPath()).getErasureCodingPolicy(); + assertNotNull(ecPolicy3); + assertTrue(ecPolicy1.equals(ecPolicy3)); + } +} http://git-wip-us.apache.org/repos/asf/hadoop/blob/1d37a881/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestFileStatusWithECschema.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestFileStatusWithECschema.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestFileStatusWithECschema.java deleted file mode 100644 index 3c400b7..0000000 --- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestFileStatusWithECschema.java +++ /dev/null @@ -1,65 +0,0 @@ -package org.apache.hadoop.hdfs; - -import static org.junit.Assert.assertNotNull; -import static org.junit.Assert.assertNull; -import static org.junit.Assert.assertTrue; - -import java.io.IOException; - -import org.apache.hadoop.conf.Configuration; -import org.apache.hadoop.fs.Path; -import org.apache.hadoop.fs.permission.FsPermission; -import org.apache.hadoop.hdfs.server.namenode.ErasureCodingSchemaManager; -import org.apache.hadoop.io.erasurecode.ECSchema; -import org.junit.After; -import org.junit.Before; -import org.junit.Test; - -public class TestFileStatusWithECschema { - private MiniDFSCluster cluster; - private DistributedFileSystem fs; - private DFSClient client; - - @Before - public void before() throws IOException { - cluster = - new MiniDFSCluster.Builder(new Configuration()).numDataNodes(1).build(); - cluster.waitActive(); - fs = cluster.getFileSystem(); - client = fs.getClient(); - } - - @After - public void after() { - if (cluster != null) { - cluster.shutdown(); - } - } - - @Test - public void testFileStatusWithECschema() throws Exception { - // test directory not in EC zone - final Path dir = new Path("/foo"); - assertTrue(fs.mkdir(dir, FsPermission.getDirDefault())); - assertNull(client.getFileInfo(dir.toString()).getECSchema()); - // test file not in EC zone - final Path file = new Path(dir, "foo"); - fs.create(file).close(); - assertNull(client.getFileInfo(file.toString()).getECSchema()); - fs.delete(file, true); - - final ECSchema schema1 = ErasureCodingSchemaManager.getSystemDefaultSchema(); - // create EC zone on dir - fs.createErasureCodingZone(dir, schema1, 0); - final ECSchema schame2 = client.getFileInfo(dir.toUri().getPath()).getECSchema(); - assertNotNull(schame2); - assertTrue(schema1.equals(schame2)); - - // test file in EC zone - fs.create(file).close(); - final ECSchema schame3 = - fs.getClient().getFileInfo(file.toUri().getPath()).getECSchema(); - assertNotNull(schame3); - assertTrue(schema1.equals(schame3)); - } -} http://git-wip-us.apache.org/repos/asf/hadoop/blob/1d37a881/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestLease.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestLease.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestLease.java index d0cd335..b77ff3a 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestLease.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestLease.java @@ -354,12 +354,12 @@ public class TestLease { Mockito.doReturn( new HdfsFileStatus(0, false, 1, 1024, 0, 0, new FsPermission( (short) 777), "owner", "group", new byte[0], new byte[0], - 1010, 0, null, (byte) 0, null, 0)).when(mcp).getFileInfo(anyString()); + 1010, 0, null, (byte) 0, null)).when(mcp).getFileInfo(anyString()); Mockito .doReturn( new HdfsFileStatus(0, false, 1, 1024, 0, 0, new FsPermission( (short) 777), "owner", "group", new byte[0], new byte[0], - 1010, 0, null, (byte) 0, null, 0)) + 1010, 0, null, (byte) 0, null)) .when(mcp) .create(anyString(), (FsPermission) anyObject(), anyString(), (EnumSetWritable<CreateFlag>) anyObject(), anyBoolean(), http://git-wip-us.apache.org/repos/asf/hadoop/blob/1d37a881/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestReadStripedFileWithDecoding.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestReadStripedFileWithDecoding.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestReadStripedFileWithDecoding.java index 1719d3f..cb2ec11 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestReadStripedFileWithDecoding.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestReadStripedFileWithDecoding.java @@ -64,8 +64,7 @@ public class TestReadStripedFileWithDecoding { public void setup() throws IOException { cluster = new MiniDFSCluster.Builder(new HdfsConfiguration()) .numDataNodes(numDNs).build(); - cluster.getFileSystem().getClient().createErasureCodingZone("/", - null, cellSize); + cluster.getFileSystem().getClient().createErasureCodingZone("/", null); fs = cluster.getFileSystem(); } http://git-wip-us.apache.org/repos/asf/hadoop/blob/1d37a881/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestReadStripedFileWithMissingBlocks.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestReadStripedFileWithMissingBlocks.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestReadStripedFileWithMissingBlocks.java index 4c2438d..38256ab 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestReadStripedFileWithMissingBlocks.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestReadStripedFileWithMissingBlocks.java @@ -52,8 +52,7 @@ public class TestReadStripedFileWithMissingBlocks { public void setup() throws IOException { conf.setLong(DFSConfigKeys.DFS_BLOCK_SIZE_KEY, blockSize); cluster = new MiniDFSCluster.Builder(conf).numDataNodes(numDNs).build(); - cluster.getFileSystem().getClient().createErasureCodingZone("/", - null, cellSize); + cluster.getFileSystem().getClient().createErasureCodingZone("/", null); fs = cluster.getFileSystem(); } http://git-wip-us.apache.org/repos/asf/hadoop/blob/1d37a881/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestRecoverStripedFile.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestRecoverStripedFile.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestRecoverStripedFile.java index 9285fd7..5c17359 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestRecoverStripedFile.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestRecoverStripedFile.java @@ -78,7 +78,7 @@ public class TestRecoverStripedFile { cluster.waitActive(); fs = cluster.getFileSystem(); - fs.getClient().createErasureCodingZone("/", null, 0); + fs.getClient().createErasureCodingZone("/", null); List<DataNode> datanodes = cluster.getDataNodes(); for (int i = 0; i < dnNum; i++) { http://git-wip-us.apache.org/repos/asf/hadoop/blob/1d37a881/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestSafeModeWithStripedFile.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestSafeModeWithStripedFile.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestSafeModeWithStripedFile.java index 6f0bc71..f577ddb 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestSafeModeWithStripedFile.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestSafeModeWithStripedFile.java @@ -54,8 +54,7 @@ public class TestSafeModeWithStripedFile { conf = new HdfsConfiguration(); conf.setLong(DFSConfigKeys.DFS_BLOCK_SIZE_KEY, blockSize); cluster = new MiniDFSCluster.Builder(conf).numDataNodes(numDNs).build(); - cluster.getFileSystem().getClient().createErasureCodingZone("/", - null, cellSize); + cluster.getFileSystem().getClient().createErasureCodingZone("/", null); cluster.waitActive(); } http://git-wip-us.apache.org/repos/asf/hadoop/blob/1d37a881/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestWriteReadStripedFile.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestWriteReadStripedFile.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestWriteReadStripedFile.java index 089a134..810edb2 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestWriteReadStripedFile.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestWriteReadStripedFile.java @@ -57,8 +57,7 @@ public class TestWriteReadStripedFile { public void setup() throws IOException { conf.setLong(DFSConfigKeys.DFS_BLOCK_SIZE_KEY, blockSize); cluster = new MiniDFSCluster.Builder(conf).numDataNodes(numDNs).build(); - cluster.getFileSystem().getClient().createErasureCodingZone("/", - null, cellSize); + cluster.getFileSystem().getClient().createErasureCodingZone("/", null); fs = cluster.getFileSystem(); } http://git-wip-us.apache.org/repos/asf/hadoop/blob/1d37a881/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestWriteStripedFileWithFailure.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestWriteStripedFileWithFailure.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestWriteStripedFileWithFailure.java index 3679c5f..deffbcc 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestWriteStripedFileWithFailure.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestWriteStripedFileWithFailure.java @@ -48,8 +48,7 @@ public class TestWriteStripedFileWithFailure { public void setup() throws IOException { conf.setLong(DFSConfigKeys.DFS_BLOCK_SIZE_KEY, blockSize); cluster = new MiniDFSCluster.Builder(conf).numDataNodes(numDNs).build(); - cluster.getFileSystem().getClient().createErasureCodingZone("/", - null, cellSize); + cluster.getFileSystem().getClient().createErasureCodingZone("/", null); fs = cluster.getFileSystem(); } http://git-wip-us.apache.org/repos/asf/hadoop/blob/1d37a881/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/protocolPB/TestPBHelper.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/protocolPB/TestPBHelper.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/protocolPB/TestPBHelper.java index 3675e63..6942ac8 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/protocolPB/TestPBHelper.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/protocolPB/TestPBHelper.java @@ -71,7 +71,7 @@ import org.apache.hadoop.hdfs.server.common.HdfsServerConstants.NamenodeRole; import org.apache.hadoop.hdfs.server.common.HdfsServerConstants.NodeType; import org.apache.hadoop.hdfs.server.common.StorageInfo; import org.apache.hadoop.hdfs.server.namenode.CheckpointSignature; -import org.apache.hadoop.hdfs.server.namenode.ErasureCodingSchemaManager; +import org.apache.hadoop.hdfs.server.namenode.ErasureCodingPolicyManager; import org.apache.hadoop.hdfs.server.protocol.BlockCommand; import org.apache.hadoop.hdfs.server.protocol.BlockECRecoveryCommand.BlockECRecoveryInfo; import org.apache.hadoop.hdfs.server.protocol.BlockRecoveryCommand; @@ -88,7 +88,7 @@ import org.apache.hadoop.hdfs.server.protocol.NamespaceInfo; import org.apache.hadoop.hdfs.server.protocol.RemoteEditLog; import org.apache.hadoop.hdfs.server.protocol.RemoteEditLogManifest; import org.apache.hadoop.io.Text; -import org.apache.hadoop.io.erasurecode.ECSchema; +import org.apache.hadoop.hdfs.protocol.ErasureCodingPolicy; import org.apache.hadoop.security.proto.SecurityProtos.TokenProto; import org.apache.hadoop.security.token.Token; import org.apache.hadoop.util.DataChecksum; @@ -682,8 +682,7 @@ public class TestPBHelper { short[] liveBlkIndices0 = new short[2]; BlockECRecoveryInfo blkECRecoveryInfo0 = new BlockECRecoveryInfo( new ExtendedBlock("bp1", 1234), dnInfos0, targetDnInfos0, - liveBlkIndices0, ErasureCodingSchemaManager.getSystemDefaultSchema(), - 64 * 1024); + liveBlkIndices0, ErasureCodingPolicyManager.getSystemDefaultPolicy()); DatanodeInfo[] dnInfos1 = new DatanodeInfo[] { DFSTestUtil.getLocalDatanodeInfo(), DFSTestUtil.getLocalDatanodeInfo() }; DatanodeStorageInfo targetDnInfos_2 = BlockManagerTestUtil @@ -697,8 +696,7 @@ public class TestPBHelper { short[] liveBlkIndices1 = new short[2]; BlockECRecoveryInfo blkECRecoveryInfo1 = new BlockECRecoveryInfo( new ExtendedBlock("bp2", 3256), dnInfos1, targetDnInfos1, - liveBlkIndices1, ErasureCodingSchemaManager.getSystemDefaultSchema(), - 64 * 1024); + liveBlkIndices1, ErasureCodingPolicyManager.getSystemDefaultPolicy()); List<BlockECRecoveryInfo> blkRecoveryInfosList = new ArrayList<BlockECRecoveryInfo>(); blkRecoveryInfosList.add(blkECRecoveryInfo0); blkRecoveryInfosList.add(blkECRecoveryInfo1); @@ -740,18 +738,18 @@ public class TestPBHelper { assertEquals(liveBlockIndices1[i], liveBlockIndices2[i]); } - ECSchema ecSchema1 = blkECRecoveryInfo1.getECSchema(); - ECSchema ecSchema2 = blkECRecoveryInfo2.getECSchema(); - // Compare ECSchemas same as default ECSchema as we used system default - // ECSchema used in this test - compareECSchemas(ErasureCodingSchemaManager.getSystemDefaultSchema(), ecSchema1); - compareECSchemas(ErasureCodingSchemaManager.getSystemDefaultSchema(), ecSchema2); - } - - private void compareECSchemas(ECSchema ecSchema1, ECSchema ecSchema2) { - assertEquals(ecSchema1.getSchemaName(), ecSchema2.getSchemaName()); - assertEquals(ecSchema1.getNumDataUnits(), ecSchema2.getNumDataUnits()); - assertEquals(ecSchema1.getNumParityUnits(), ecSchema2.getNumParityUnits()); + ErasureCodingPolicy ecPolicy1 = blkECRecoveryInfo1.getErasureCodingPolicy(); + ErasureCodingPolicy ecPolicy2 = blkECRecoveryInfo2.getErasureCodingPolicy(); + // Compare ECPolicies same as default ECPolicy as we used system default + // ECPolicy used in this test + compareECPolicies(ErasureCodingPolicyManager.getSystemDefaultPolicy(), ecPolicy1); + compareECPolicies(ErasureCodingPolicyManager.getSystemDefaultPolicy(), ecPolicy2); + } + + private void compareECPolicies(ErasureCodingPolicy ecPolicy1, ErasureCodingPolicy ecPolicy2) { + assertEquals(ecPolicy1.getName(), ecPolicy2.getName()); + assertEquals(ecPolicy1.getNumDataUnits(), ecPolicy2.getNumDataUnits()); + assertEquals(ecPolicy1.getNumParityUnits(), ecPolicy2.getNumParityUnits()); } private void assertDnInfosEqual(DatanodeInfo[] dnInfos1, http://git-wip-us.apache.org/repos/asf/hadoop/blob/1d37a881/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/balancer/TestBalancer.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/balancer/TestBalancer.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/balancer/TestBalancer.java index 8239e5f..99b460a 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/balancer/TestBalancer.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/balancer/TestBalancer.java @@ -1503,7 +1503,7 @@ public class TestBalancer { cluster.waitActive(); client = NameNodeProxies.createProxy(conf, cluster.getFileSystem(0).getUri(), ClientProtocol.class).getProxy(); - client.createErasureCodingZone("/", null, 0); + client.createErasureCodingZone("/", null); long totalCapacity = sum(capacities); http://git-wip-us.apache.org/repos/asf/hadoop/blob/1d37a881/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/blockmanagement/TestBlockInfoStriped.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/blockmanagement/TestBlockInfoStriped.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/blockmanagement/TestBlockInfoStriped.java index 6788770..1b23600 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/blockmanagement/TestBlockInfoStriped.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/blockmanagement/TestBlockInfoStriped.java @@ -19,10 +19,9 @@ package org.apache.hadoop.hdfs.server.blockmanagement; import org.apache.hadoop.hdfs.DFSTestUtil; import org.apache.hadoop.hdfs.protocol.Block; -import org.apache.hadoop.hdfs.protocol.HdfsConstants; import org.apache.hadoop.hdfs.server.blockmanagement.DatanodeStorageInfo.AddBlockResult; -import org.apache.hadoop.hdfs.server.namenode.ErasureCodingSchemaManager; -import org.apache.hadoop.io.erasurecode.ECSchema; +import org.apache.hadoop.hdfs.server.namenode.ErasureCodingPolicyManager; +import org.apache.hadoop.hdfs.protocol.ErasureCodingPolicy; import org.junit.Assert; import org.junit.Test; import org.mockito.internal.util.reflection.Whitebox; @@ -45,11 +44,10 @@ public class TestBlockInfoStriped { private static final int TOTAL_NUM_BLOCKS = NUM_DATA_BLOCKS + NUM_PARITY_BLOCKS; private static final long BASE_ID = -1600; private static final Block baseBlock = new Block(BASE_ID); - private static final ECSchema testSchema - = ErasureCodingSchemaManager.getSystemDefaultSchema(); - private static final int cellSize = HdfsConstants.BLOCK_STRIPED_CELL_SIZE; + private static final ErasureCodingPolicy testECPolicy + = ErasureCodingPolicyManager.getSystemDefaultPolicy(); private final BlockInfoStriped info = new BlockInfoStriped(baseBlock, - testSchema, cellSize); + testECPolicy); private Block[] createReportedBlocks(int num) { Block[] blocks = new Block[num]; @@ -237,7 +235,7 @@ public class TestBlockInfoStriped { ByteArrayOutputStream byteStream = new ByteArrayOutputStream(); DataOutput out = new DataOutputStream(byteStream); BlockInfoStriped blk = new BlockInfoStriped(new Block(blkID, numBytes, - generationStamp), testSchema, cellSize); + generationStamp), testECPolicy); try { blk.write(out); http://git-wip-us.apache.org/repos/asf/hadoop/blob/1d37a881/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/blockmanagement/TestBlockTokenWithDFSStriped.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/blockmanagement/TestBlockTokenWithDFSStriped.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/blockmanagement/TestBlockTokenWithDFSStriped.java index f985f54..2202b34 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/blockmanagement/TestBlockTokenWithDFSStriped.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/blockmanagement/TestBlockTokenWithDFSStriped.java @@ -56,7 +56,7 @@ public class TestBlockTokenWithDFSStriped extends TestBlockTokenWithDFS { conf = getConf(); cluster = new MiniDFSCluster.Builder(conf).numDataNodes(numDNs).build(); cluster.getFileSystem().getClient() - .createErasureCodingZone("/", null, cellSize); + .createErasureCodingZone("/", null); try { cluster.waitActive(); doTestRead(conf, cluster, true); http://git-wip-us.apache.org/repos/asf/hadoop/blob/1d37a881/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/blockmanagement/TestSequentialBlockGroupId.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/blockmanagement/TestSequentialBlockGroupId.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/blockmanagement/TestSequentialBlockGroupId.java index 2f2356f..2e084fc 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/blockmanagement/TestSequentialBlockGroupId.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/blockmanagement/TestSequentialBlockGroupId.java @@ -86,7 +86,7 @@ public class TestSequentialBlockGroupId { .getBlockGroupIdGenerator(); fs.mkdirs(eczone); cluster.getFileSystem().getClient() - .createErasureCodingZone("/eczone", null, cellSize); + .createErasureCodingZone("/eczone", null); } @After http://git-wip-us.apache.org/repos/asf/hadoop/blob/1d37a881/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/blockmanagement/TestUnderReplicatedBlockQueues.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/blockmanagement/TestUnderReplicatedBlockQueues.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/blockmanagement/TestUnderReplicatedBlockQueues.java index 0f419ef..7cd2e19 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/blockmanagement/TestUnderReplicatedBlockQueues.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/blockmanagement/TestUnderReplicatedBlockQueues.java @@ -19,9 +19,8 @@ package org.apache.hadoop.hdfs.server.blockmanagement; import org.apache.hadoop.hdfs.protocol.Block; -import org.apache.hadoop.hdfs.protocol.HdfsConstants; -import org.apache.hadoop.hdfs.server.namenode.ErasureCodingSchemaManager; -import org.apache.hadoop.io.erasurecode.ECSchema; +import org.apache.hadoop.hdfs.server.namenode.ErasureCodingPolicyManager; +import org.apache.hadoop.hdfs.protocol.ErasureCodingPolicy; import org.junit.Test; import static org.junit.Assert.assertEquals; @@ -31,17 +30,15 @@ import static org.junit.Assert.fail; public class TestUnderReplicatedBlockQueues { - private final ECSchema ecSchema = - ErasureCodingSchemaManager.getSystemDefaultSchema(); - private final int CELLSIZE = HdfsConstants.BLOCK_STRIPED_CELL_SIZE; + private final ErasureCodingPolicy ecPolicy = + ErasureCodingPolicyManager.getSystemDefaultPolicy(); private BlockInfo genBlockInfo(long id) { return new BlockInfoContiguous(new Block(id), (short) 3); } private BlockInfo genStripedBlockInfo(long id, long numBytes) { - BlockInfoStriped sblk = new BlockInfoStriped(new Block(id), ecSchema, - CELLSIZE); + BlockInfoStriped sblk = new BlockInfoStriped(new Block(id), ecPolicy); sblk.setNumBytes(numBytes); return sblk; } @@ -101,8 +98,8 @@ public class TestUnderReplicatedBlockQueues { @Test public void testStripedBlockPriorities() throws Throwable { - int dataBlkNum = ecSchema.getNumDataUnits(); - int parityBlkNUm = ecSchema.getNumParityUnits(); + int dataBlkNum = ecPolicy.getNumDataUnits(); + int parityBlkNUm = ecPolicy.getNumParityUnits(); doTestStripedBlockPriorities(1, parityBlkNUm); doTestStripedBlockPriorities(dataBlkNum, parityBlkNUm); } @@ -110,7 +107,7 @@ public class TestUnderReplicatedBlockQueues { private void doTestStripedBlockPriorities(int dataBlkNum, int parityBlkNum) throws Throwable { int groupSize = dataBlkNum + parityBlkNum; - long numBytes = CELLSIZE * dataBlkNum; + long numBytes = ecPolicy.getCellSize() * dataBlkNum; UnderReplicatedBlocks queues = new UnderReplicatedBlocks(); // add a striped block which been left NUM_DATA_BLOCKS internal blocks http://git-wip-us.apache.org/repos/asf/hadoop/blob/1d37a881/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/mover/TestMover.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/mover/TestMover.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/mover/TestMover.java index 29e8d24..261d397 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/mover/TestMover.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/mover/TestMover.java @@ -426,7 +426,7 @@ public class TestMover { client.setStoragePolicy(barDir, HdfsServerConstants.HOT_STORAGE_POLICY_NAME); // set "/bar" directory with EC zone. - client.createErasureCodingZone(barDir, null, 0); + client.createErasureCodingZone(barDir, null); // write file to barDir final String fooFile = "/bar/foo"; http://git-wip-us.apache.org/repos/asf/hadoop/blob/1d37a881/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestAddOverReplicatedStripedBlocks.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestAddOverReplicatedStripedBlocks.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestAddOverReplicatedStripedBlocks.java index 337911d..7d06a9b 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestAddOverReplicatedStripedBlocks.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestAddOverReplicatedStripedBlocks.java @@ -70,7 +70,7 @@ public class TestAddOverReplicatedStripedBlocks { cluster.waitActive(); fs = cluster.getFileSystem(); fs.mkdirs(dirPath); - fs.getClient().createErasureCodingZone(dirPath.toString(), null, CELLSIZE); + fs.getClient().createErasureCodingZone(dirPath.toString(), null); } @After @@ -180,7 +180,7 @@ public class TestAddOverReplicatedStripedBlocks { long groupId = bg.getBlock().getBlockId(); Block blk = new Block(groupId, BLOCK_SIZE, gs); BlockInfoStriped blockInfo = new BlockInfoStriped(blk, - ErasureCodingSchemaManager.getSystemDefaultSchema(), CELLSIZE); + ErasureCodingPolicyManager.getSystemDefaultPolicy()); for (int i = 0; i < GROUP_SIZE; i++) { blk.setBlockId(groupId + i); cluster.injectBlocks(i, Arrays.asList(blk), bpid); http://git-wip-us.apache.org/repos/asf/hadoop/blob/1d37a881/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestAddStripedBlocks.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestAddStripedBlocks.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestAddStripedBlocks.java index dd7086b..4ae10db 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestAddStripedBlocks.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestAddStripedBlocks.java @@ -75,7 +75,7 @@ public class TestAddStripedBlocks { .numDataNodes(GROUP_SIZE).build(); cluster.waitActive(); dfs = cluster.getFileSystem(); - dfs.getClient().createErasureCodingZone("/", null, 0); + dfs.getClient().createErasureCodingZone("/", null); } @After http://git-wip-us.apache.org/repos/asf/hadoop/blob/1d37a881/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestFSEditLogLoader.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestFSEditLogLoader.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestFSEditLogLoader.java index 5180f13..aefafd0 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestFSEditLogLoader.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestFSEditLogLoader.java @@ -45,7 +45,7 @@ import org.apache.hadoop.hdfs.DFSConfigKeys; import org.apache.hadoop.hdfs.DFSTestUtil; import org.apache.hadoop.hdfs.HdfsConfiguration; import org.apache.hadoop.hdfs.MiniDFSCluster; -import org.apache.hadoop.hdfs.protocol.ErasureCodingZone; +import org.apache.hadoop.hdfs.protocol.ErasureCodingPolicy; import org.apache.hadoop.hdfs.server.blockmanagement.BlockInfo; import org.apache.hadoop.hdfs.server.common.HdfsServerConstants; import org.apache.hadoop.hdfs.DistributedFileSystem; @@ -57,7 +57,6 @@ import org.apache.hadoop.hdfs.server.namenode.NNStorage.NameNodeDirType; import org.apache.hadoop.hdfs.server.blockmanagement.BlockInfoStriped; import org.apache.hadoop.hdfs.protocol.Block; import org.apache.hadoop.io.IOUtils; -import org.apache.hadoop.io.erasurecode.ECSchema; import org.apache.hadoop.test.PathUtils; import org.apache.log4j.Level; import org.junit.Test; @@ -76,8 +75,8 @@ public class TestFSEditLogLoader { private static final int NUM_DATA_NODES = 0; - private static final ECSchema testSchema - = ErasureCodingSchemaManager.getSystemDefaultSchema(); + private static final ErasureCodingPolicy testECPolicy + = ErasureCodingPolicyManager.getSystemDefaultPolicy(); @Test public void testDisplayRecentEditLogOpCodes() throws IOException { @@ -450,11 +449,10 @@ public class TestFSEditLogLoader { long timestamp = 1426222918; short blockNum = HdfsConstants.NUM_DATA_BLOCKS; short parityNum = HdfsConstants.NUM_PARITY_BLOCKS; - int cellSize = HdfsConstants.BLOCK_STRIPED_CELL_SIZE; //set the storage policy of the directory fs.mkdir(new Path(testDir), new FsPermission("755")); - fs.getClient().getNamenode().createErasureCodingZone(testDir, null, 0); + fs.getClient().getNamenode().createErasureCodingZone(testDir, null); // Create a file with striped block Path p = new Path(testFilePath); @@ -466,7 +464,7 @@ public class TestFSEditLogLoader { // Add a striped block to the file BlockInfoStriped stripedBlk = new BlockInfoStriped( - new Block(blkId, blkNumBytes, timestamp), testSchema, cellSize); + new Block(blkId, blkNumBytes, timestamp), testECPolicy); INodeFile file = (INodeFile)fns.getFSDirectory().getINode(testFilePath); file.toUnderConstruction(clientName, clientMachine); file.addBlock(stripedBlk); @@ -491,7 +489,6 @@ public class TestFSEditLogLoader { assertEquals(timestamp, blks[0].getGenerationStamp()); assertEquals(blockNum, ((BlockInfoStriped)blks[0]).getDataBlockNum()); assertEquals(parityNum, ((BlockInfoStriped)blks[0]).getParityBlockNum()); - assertEquals(cellSize, ((BlockInfoStriped)blks[0]).getCellSize()); cluster.shutdown(); cluster = null; @@ -524,17 +521,16 @@ public class TestFSEditLogLoader { long timestamp = 1426222918; short blockNum = HdfsConstants.NUM_DATA_BLOCKS; short parityNum = HdfsConstants.NUM_PARITY_BLOCKS; - int cellSize = HdfsConstants.BLOCK_STRIPED_CELL_SIZE; //set the storage policy of the directory fs.mkdir(new Path(testDir), new FsPermission("755")); - fs.getClient().getNamenode().createErasureCodingZone(testDir, null, 0); + fs.getClient().getNamenode().createErasureCodingZone(testDir, null); //create a file with striped blocks Path p = new Path(testFilePath); DFSTestUtil.createFile(fs, p, 0, (short) 1, 1); BlockInfoStriped stripedBlk = new BlockInfoStriped( - new Block(blkId, blkNumBytes, timestamp), testSchema, cellSize); + new Block(blkId, blkNumBytes, timestamp), testECPolicy); INodeFile file = (INodeFile)fns.getFSDirectory().getINode(testFilePath); file.toUnderConstruction(clientName, clientMachine); file.addBlock(stripedBlk); @@ -573,7 +569,6 @@ public class TestFSEditLogLoader { assertEquals(newTimestamp, blks[0].getGenerationStamp()); assertEquals(blockNum, ((BlockInfoStriped)blks[0]).getDataBlockNum()); assertEquals(parityNum, ((BlockInfoStriped)blks[0]).getParityBlockNum()); - assertEquals(cellSize, ((BlockInfoStriped)blks[0]).getCellSize()); cluster.shutdown(); cluster = null; http://git-wip-us.apache.org/repos/asf/hadoop/blob/1d37a881/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestFSImage.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestFSImage.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestFSImage.java index a144922..d1c7600 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestFSImage.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestFSImage.java @@ -28,11 +28,11 @@ import java.io.IOException; import java.util.EnumSet; import org.apache.hadoop.hdfs.protocol.Block; +import org.apache.hadoop.hdfs.protocol.ErasureCodingPolicy; import org.apache.hadoop.hdfs.server.blockmanagement.BlockInfo; import org.apache.hadoop.hdfs.server.blockmanagement.BlockInfoContiguous; import org.apache.hadoop.hdfs.server.blockmanagement.BlockInfoStriped; import org.apache.hadoop.hdfs.server.common.HdfsServerConstants.StartupOption; -import org.apache.hadoop.io.erasurecode.ECSchema; import org.junit.Assert; import org.apache.hadoop.fs.permission.PermissionStatus; @@ -68,9 +68,8 @@ public class TestFSImage { private static final String HADOOP_2_7_ZER0_BLOCK_SIZE_TGZ = "image-with-zero-block-size.tar.gz"; - private static final ECSchema testSchema - = ErasureCodingSchemaManager.getSystemDefaultSchema(); - private static final int cellSize = HdfsConstants.BLOCK_STRIPED_CELL_SIZE; + private static final ErasureCodingPolicy testECPolicy + = ErasureCodingPolicyManager.getSystemDefaultPolicy(); @Test public void testPersist() throws IOException { @@ -141,7 +140,7 @@ public class TestFSImage { private void testSaveAndLoadStripedINodeFile(FSNamesystem fsn, Configuration conf, boolean isUC) throws IOException{ // contruct a INode with StripedBlock for saving and loading - fsn.createErasureCodingZone("/", null, 0, false); + fsn.createErasureCodingZone("/", null, false); long id = 123456789; byte[] name = "testSaveAndLoadInodeFile_testfile".getBytes(); PermissionStatus permissionStatus = new PermissionStatus("testuser_a", @@ -162,7 +161,7 @@ public class TestFSImage { for (int i = 0; i < stripedBlks.length; i++) { stripedBlks[i] = new BlockInfoStriped( new Block(stripedBlkId + i, preferredBlockSize, timestamp), - testSchema, cellSize); + testECPolicy); file.addBlock(stripedBlks[i]); } @@ -386,7 +385,7 @@ public class TestFSImage { .build(); cluster.waitActive(); DistributedFileSystem fs = cluster.getFileSystem(); - fs.getClient().getNamenode().createErasureCodingZone("/", null, 0); + fs.getClient().getNamenode().createErasureCodingZone("/", null); Path file = new Path("/striped"); FSDataOutputStream out = fs.create(file); byte[] bytes = DFSTestUtil.generateSequentialBytes(0, BLOCK_SIZE); http://git-wip-us.apache.org/repos/asf/hadoop/blob/1d37a881/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestFsck.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestFsck.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestFsck.java index 7e2b41d..e6f505e 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestFsck.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestFsck.java @@ -1202,7 +1202,7 @@ public class TestFsck { HdfsFileStatus file = new HdfsFileStatus(length, isDir, blockReplication, blockSize, modTime, accessTime, perms, owner, group, symlink, - path, fileId, numChildren, null, storagePolicy, null, 0); + path, fileId, numChildren, null, storagePolicy, null); Result replRes = new ReplicationResult(conf); Result ecRes = new ErasureCodingResult(conf); @@ -1644,8 +1644,8 @@ public class TestFsck { final long precision = 1L; conf.setLong(DFSConfigKeys.DFS_NAMENODE_ACCESSTIME_PRECISION_KEY, precision); conf.setLong(DFSConfigKeys.DFS_BLOCKREPORT_INTERVAL_MSEC_KEY, 10000L); - int totalSize = ErasureCodingSchemaManager.getSystemDefaultSchema().getNumDataUnits() - + ErasureCodingSchemaManager.getSystemDefaultSchema().getNumParityUnits(); + int totalSize = ErasureCodingPolicyManager.getSystemDefaultPolicy().getNumDataUnits() + + ErasureCodingPolicyManager.getSystemDefaultPolicy().getNumParityUnits(); cluster = new MiniDFSCluster.Builder(conf).numDataNodes(totalSize).build(); fs = cluster.getFileSystem(); http://git-wip-us.apache.org/repos/asf/hadoop/blob/1d37a881/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestQuotaWithStripedBlocks.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestQuotaWithStripedBlocks.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestQuotaWithStripedBlocks.java index 9c585a4..b1bb191 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestQuotaWithStripedBlocks.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestQuotaWithStripedBlocks.java @@ -26,11 +26,11 @@ import org.apache.hadoop.hdfs.DFSTestUtil; import org.apache.hadoop.hdfs.DistributedFileSystem; import org.apache.hadoop.hdfs.MiniDFSCluster; import org.apache.hadoop.hdfs.protocol.Block; +import org.apache.hadoop.hdfs.protocol.ErasureCodingPolicy; import org.apache.hadoop.hdfs.protocol.ExtendedBlock; import org.apache.hadoop.hdfs.protocol.HdfsConstants; import org.apache.hadoop.hdfs.server.common.HdfsServerConstants; import org.apache.hadoop.io.IOUtils; -import org.apache.hadoop.io.erasurecode.ECSchema; import org.junit.After; import org.junit.Assert; import org.junit.Before; @@ -44,10 +44,10 @@ import java.io.IOException; public class TestQuotaWithStripedBlocks { private static final int BLOCK_SIZE = 1024 * 1024; private static final long DISK_QUOTA = BLOCK_SIZE * 10; - private static final ECSchema ecSchema = - ErasureCodingSchemaManager.getSystemDefaultSchema(); - private static final int NUM_DATA_BLOCKS = ecSchema.getNumDataUnits(); - private static final int NUM_PARITY_BLOCKS = ecSchema.getNumParityUnits(); + private static final ErasureCodingPolicy ecPolicy = + ErasureCodingPolicyManager.getSystemDefaultPolicy(); + private static final int NUM_DATA_BLOCKS = ecPolicy.getNumDataUnits(); + private static final int NUM_PARITY_BLOCKS = ecPolicy.getNumParityUnits(); private static final int GROUP_SIZE = NUM_DATA_BLOCKS + NUM_PARITY_BLOCKS; private static final Path ecDir = new Path("/ec"); @@ -66,7 +66,7 @@ public class TestQuotaWithStripedBlocks { dfs = cluster.getFileSystem(); dfs.mkdirs(ecDir); - dfs.getClient().createErasureCodingZone(ecDir.toString(), ecSchema, 0); + dfs.getClient().createErasureCodingZone(ecDir.toString(), ecPolicy); dfs.setQuota(ecDir, Long.MAX_VALUE - 1, DISK_QUOTA); dfs.setQuotaByStorageType(ecDir, StorageType.DISK, DISK_QUOTA); dfs.setStoragePolicy(ecDir, HdfsServerConstants.HOT_STORAGE_POLICY_NAME); http://git-wip-us.apache.org/repos/asf/hadoop/blob/1d37a881/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestStripedINodeFile.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestStripedINodeFile.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestStripedINodeFile.java index a9b2048..7684c1d 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestStripedINodeFile.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestStripedINodeFile.java @@ -35,13 +35,13 @@ import org.apache.hadoop.hdfs.DistributedFileSystem; import org.apache.hadoop.hdfs.MiniDFSCluster; import org.apache.hadoop.hdfs.protocol.Block; import org.apache.hadoop.hdfs.protocol.BlockStoragePolicy; +import org.apache.hadoop.hdfs.protocol.ErasureCodingPolicy; import org.apache.hadoop.hdfs.protocol.HdfsConstants; import org.apache.hadoop.hdfs.server.blockmanagement.BlockInfo; import org.apache.hadoop.hdfs.server.blockmanagement.BlockInfoStriped; import org.apache.hadoop.hdfs.server.blockmanagement.BlockInfoUnderConstructionStriped; import org.apache.hadoop.hdfs.server.blockmanagement.BlockStoragePolicySuite; import org.apache.hadoop.hdfs.server.common.HdfsServerConstants; -import org.apache.hadoop.io.erasurecode.ECSchema; import org.junit.Test; @@ -59,9 +59,8 @@ public class TestStripedINodeFile { private final BlockStoragePolicy defaultPolicy = defaultSuite.getDefaultPolicy(); - private static final ECSchema testSchema - = ErasureCodingSchemaManager.getSystemDefaultSchema(); - private static final int cellSize = HdfsConstants.BLOCK_STRIPED_CELL_SIZE; + private static final ErasureCodingPolicy testECPolicy + = ErasureCodingPolicyManager.getSystemDefaultPolicy(); private static INodeFile createStripedINodeFile() { return new INodeFile(HdfsConstants.GRANDFATHER_INODE_ID, null, perm, 0L, 0L, @@ -79,7 +78,7 @@ public class TestStripedINodeFile { public void testBlockStripedTotalBlockCount() { Block blk = new Block(1); BlockInfoStriped blockInfoStriped - = new BlockInfoStriped(blk, testSchema, cellSize); + = new BlockInfoStriped(blk, testECPolicy); assertEquals(9, blockInfoStriped.getTotalBlockNum()); } @@ -89,7 +88,7 @@ public class TestStripedINodeFile { INodeFile inf = createStripedINodeFile(); Block blk = new Block(1); BlockInfoStriped blockInfoStriped - = new BlockInfoStriped(blk, testSchema, cellSize); + = new BlockInfoStriped(blk, testECPolicy); inf.addBlock(blockInfoStriped); assertEquals(1, inf.getBlocks().length); } @@ -100,7 +99,7 @@ public class TestStripedINodeFile { INodeFile inf = createStripedINodeFile(); Block blk = new Block(1); BlockInfoStriped blockInfoStriped - = new BlockInfoStriped(blk, testSchema, cellSize); + = new BlockInfoStriped(blk, testECPolicy); blockInfoStriped.setNumBytes(1); inf.addBlock(blockInfoStriped); // 0. Calculate the total bytes per stripes <Num Bytes per Stripes> @@ -125,11 +124,11 @@ public class TestStripedINodeFile { INodeFile inf = createStripedINodeFile(); Block blk1 = new Block(1); BlockInfoStriped blockInfoStriped1 - = new BlockInfoStriped(blk1, testSchema, cellSize); + = new BlockInfoStriped(blk1, testECPolicy); blockInfoStriped1.setNumBytes(1); Block blk2 = new Block(2); BlockInfoStriped blockInfoStriped2 - = new BlockInfoStriped(blk2, testSchema, cellSize); + = new BlockInfoStriped(blk2, testECPolicy); blockInfoStriped2.setNumBytes(1); inf.addBlock(blockInfoStriped1); inf.addBlock(blockInfoStriped2); @@ -144,7 +143,7 @@ public class TestStripedINodeFile { INodeFile inf = createStripedINodeFile(); Block blk = new Block(1); BlockInfoStriped blockInfoStriped - = new BlockInfoStriped(blk, testSchema, cellSize); + = new BlockInfoStriped(blk, testECPolicy); blockInfoStriped.setNumBytes(100); inf.addBlock(blockInfoStriped); // Compute file size should return actual data @@ -159,7 +158,7 @@ public class TestStripedINodeFile { INodeFile inf = createStripedINodeFile(); Block blk = new Block(1); BlockInfoUnderConstructionStriped bInfoUCStriped - = new BlockInfoUnderConstructionStriped(blk, testSchema, cellSize); + = new BlockInfoUnderConstructionStriped(blk, testECPolicy); bInfoUCStriped.setNumBytes(100); inf.addBlock(bInfoUCStriped); assertEquals(100, inf.computeFileSize()); @@ -172,7 +171,7 @@ public class TestStripedINodeFile { INodeFile inf = createStripedINodeFile(); Block blk = new Block(1); BlockInfoStriped blockInfoStriped - = new BlockInfoStriped(blk, testSchema, cellSize); + = new BlockInfoStriped(blk, testECPolicy); blockInfoStriped.setNumBytes(100); inf.addBlock(blockInfoStriped); @@ -193,7 +192,7 @@ public class TestStripedINodeFile { INodeFile inf = createStripedINodeFile(); Block blk = new Block(1); BlockInfoUnderConstructionStriped bInfoUCStriped - = new BlockInfoUnderConstructionStriped(blk, testSchema, cellSize); + = new BlockInfoUnderConstructionStriped(blk, testECPolicy); bInfoUCStriped.setNumBytes(100); inf.addBlock(bInfoUCStriped); @@ -235,7 +234,7 @@ public class TestStripedINodeFile { dfs.mkdirs(zone); // create erasure zone - dfs.createErasureCodingZone(zone, null, 0); + dfs.createErasureCodingZone(zone, null); DFSTestUtil.createFile(dfs, zoneFile, len, (short) 1, 0xFEED); DFSTestUtil.createFile(dfs, contiguousFile, len, (short) 1, 0xFEED); final FSDirectory fsd = fsn.getFSDirectory(); http://git-wip-us.apache.org/repos/asf/hadoop/blob/1d37a881/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/tools/offlineImageViewer/TestOfflineImageViewerWithStripedBlocks.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/tools/offlineImageViewer/TestOfflineImageViewerWithStripedBlocks.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/tools/offlineImageViewer/TestOfflineImageViewerWithStripedBlocks.java index f4efbcf..7bfaab6 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/tools/offlineImageViewer/TestOfflineImageViewerWithStripedBlocks.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/tools/offlineImageViewer/TestOfflineImageViewerWithStripedBlocks.java @@ -60,7 +60,7 @@ public class TestOfflineImageViewerWithStripedBlocks { conf.setLong(DFSConfigKeys.DFS_BLOCK_SIZE_KEY, blockSize); cluster = new MiniDFSCluster.Builder(conf).numDataNodes(numDNs).build(); cluster.waitActive(); - cluster.getFileSystem().getClient().createErasureCodingZone("/", null, 0); + cluster.getFileSystem().getClient().createErasureCodingZone("/", null); fs = cluster.getFileSystem(); Path eczone = new Path("/eczone"); fs.mkdirs(eczone); http://git-wip-us.apache.org/repos/asf/hadoop/blob/1d37a881/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/util/TestStripedBlockUtil.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/util/TestStripedBlockUtil.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/util/TestStripedBlockUtil.java index 5d85073..e61ac07 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/util/TestStripedBlockUtil.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/util/TestStripedBlockUtil.java @@ -29,8 +29,8 @@ import org.apache.hadoop.hdfs.protocol.LocatedStripedBlock; import org.apache.hadoop.hdfs.server.blockmanagement.BlockIdManager; import static org.apache.hadoop.hdfs.util.StripedBlockUtil.*; -import org.apache.hadoop.hdfs.server.namenode.ErasureCodingSchemaManager; -import org.apache.hadoop.io.erasurecode.ECSchema; +import org.apache.hadoop.hdfs.server.namenode.ErasureCodingPolicyManager; +import org.apache.hadoop.hdfs.protocol.ErasureCodingPolicy; import org.junit.Before; import org.junit.Test; @@ -84,8 +84,8 @@ public class TestStripedBlockUtil { private final int FULL_STRIPE_SIZE = DATA_BLK_NUM * CELLSIZE; /** number of full stripes in a full block group */ private final int BLK_GROUP_STRIPE_NUM = 16; - private final ECSchema SCEHMA = ErasureCodingSchemaManager. - getSystemDefaultSchema(); + private final ErasureCodingPolicy ECPOLICY = ErasureCodingPolicyManager. + getSystemDefaultPolicy(); private final Random random = new Random(); private int[] blockGroupSizes; @@ -152,7 +152,7 @@ public class TestStripedBlockUtil { int done = 0; while (done < bgSize) { Preconditions.checkState(done % CELLSIZE == 0); - StripingCell cell = new StripingCell(SCEHMA, CELLSIZE, done / CELLSIZE, 0); + StripingCell cell = new StripingCell(ECPOLICY, CELLSIZE, done / CELLSIZE, 0); int idxInStripe = cell.idxInStripe; int size = Math.min(CELLSIZE, bgSize - done); for (int i = 0; i < size; i++) { @@ -245,7 +245,7 @@ public class TestStripedBlockUtil { if (brStart + brSize > bgSize) { continue; } - AlignedStripe[] stripes = divideByteRangeIntoStripes(SCEHMA, + AlignedStripe[] stripes = divideByteRangeIntoStripes(ECPOLICY, CELLSIZE, blockGroup, brStart, brStart + brSize - 1, assembled, 0); for (AlignedStripe stripe : stripes) { http://git-wip-us.apache.org/repos/asf/hadoop/blob/1d37a881/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/web/TestJsonUtil.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/web/TestJsonUtil.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/web/TestJsonUtil.java index 303d063..8947c5b 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/web/TestJsonUtil.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/web/TestJsonUtil.java @@ -65,7 +65,7 @@ public class TestJsonUtil { final HdfsFileStatus status = new HdfsFileStatus(1001L, false, 3, 1L << 26, now, now + 10, new FsPermission((short) 0644), "user", "group", DFSUtil.string2Bytes("bar"), DFSUtil.string2Bytes("foo"), - HdfsConstants.GRANDFATHER_INODE_ID, 0, null, (byte) 0, null, 0); + HdfsConstants.GRANDFATHER_INODE_ID, 0, null, (byte) 0, null); final FileStatus fstatus = toFileStatus(status, parent); System.out.println("status = " + status); System.out.println("fstatus = " + fstatus); http://git-wip-us.apache.org/repos/asf/hadoop/blob/1d37a881/hadoop-hdfs-project/hadoop-hdfs/src/test/resources/testErasureCodingConf.xml ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/resources/testErasureCodingConf.xml b/hadoop-hdfs-project/hadoop-hdfs/src/test/resources/testErasureCodingConf.xml index 70020d5..5e60658 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/test/resources/testErasureCodingConf.xml +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/resources/testErasureCodingConf.xml @@ -57,11 +57,11 @@ <comparators> <comparator> <type>RegexpComparator</type> - <expected-output>^[ \t]*Create a zone to encode files using a specified schema( )*</expected-output> + <expected-output>^[ \t]*Create a zone to encode files using a specified policy( )*</expected-output> </comparator> <comparator> <type>RegexpComparator</type> - <expected-output>^-createZone \[-s <schemaName>\] \[-c <cellSize>\] <path>(.)*</expected-output> + <expected-output>^-createZone \[-s <policyName>\] <path>(.)*</expected-output> </comparator> </comparators> </test> @@ -86,20 +86,20 @@ </test> <test> - <description>help: listSchemas command</description> + <description>help: listPolicies command</description> <test-commands> - <ec-admin-command>-fs NAMENODE -help listSchemas</ec-admin-command> + <ec-admin-command>-fs NAMENODE -help listPolicies</ec-admin-command> </test-commands> <cleanup-commands> </cleanup-commands> <comparators> <comparator> <type>SubstringComparator</type> - <expected-output>Get the list of ECSchemas supported</expected-output> + <expected-output>Get the list of erasure coding policies supported</expected-output> </comparator> <comparator> <type>RegexpComparator</type> - <expected-output>^-listSchemas (.)*</expected-output> + <expected-output>^-listPolicies (.)*</expected-output> </comparator> </comparators> </test> @@ -109,7 +109,7 @@ <description>createZone : create a zone to encode files</description> <test-commands> <command>-fs NAMENODE -mkdir /eczone</command> - <ec-admin-command>-fs NAMENODE -createZone -s RS-6-3 /eczone</ec-admin-command> + <ec-admin-command>-fs NAMENODE -createZone -s RS-6-3-64k /eczone</ec-admin-command> </test-commands> <cleanup-commands> <command>-fs NAMENODE -rmdir /eczone</command> @@ -141,7 +141,7 @@ </test> <test> - <description>createZone : default schema</description> + <description>createZone : default policy</description> <test-commands> <command>-fs NAMENODE -mkdir /eczone</command> <ec-admin-command>-fs NAMENODE -createZone /eczone</ec-admin-command> @@ -153,7 +153,7 @@ <comparators> <comparator> <type>SubstringComparator</type> - <expected-output>Dir: /eczone, Schema: ECSchema=[Name=RS-6-3</expected-output> + <expected-output>Dir: /eczone, Policy: ErasureCodingPolicy=[Name=RS-6-3-64k</expected-output> </comparator> </comparators> </test> @@ -179,7 +179,7 @@ <description>getZone : get information about the EC zone at specified path</description> <test-commands> <command>-fs NAMENODE -mkdir /eczone</command> - <ec-admin-command>-fs NAMENODE -createZone -s RS-6-3 /eczone</ec-admin-command> + <ec-admin-command>-fs NAMENODE -createZone -s RS-6-3-64k /eczone</ec-admin-command> <ec-admin-command>-fs NAMENODE -getZone /eczone</ec-admin-command> </test-commands> <cleanup-commands> @@ -188,7 +188,7 @@ <comparators> <comparator> <type>SubstringComparator</type> - <expected-output>Dir: /eczone, Schema: ECSchema=[Name=RS-6-3</expected-output> + <expected-output>Dir: /eczone, Policy: ErasureCodingPolicy=[Name=RS-6-3-64k</expected-output> </comparator> </comparators> </test> @@ -197,7 +197,7 @@ <description>getZone : get EC zone at specified file path</description> <test-commands> <command>-fs NAMENODE -mkdir /eczone</command> - <ec-admin-command>-fs NAMENODE -createZone -s RS-6-3 /eczone</ec-admin-command> + <ec-admin-command>-fs NAMENODE -createZone -s RS-6-3-64k /eczone</ec-admin-command> <command>-fs NAMENODE -touchz /eczone/ecfile</command> <ec-admin-command>-fs NAMENODE -getZone /eczone/ecfile</ec-admin-command> </test-commands> @@ -208,15 +208,15 @@ <comparators> <comparator> <type>SubstringComparator</type> - <expected-output>Dir: /eczone, Schema: ECSchema=[Name=RS-6-3</expected-output> + <expected-output>Dir: /eczone, Policy: ErasureCodingPolicy=[Name=RS-6-3-64k</expected-output> </comparator> </comparators> </test> <test> - <description>listSchemas : get the list of ECSchemas supported</description> + <description>listPolicies : get the list of ECPolicies supported</description> <test-commands> - <ec-admin-command>-fs NAMENODE -listSchemas</ec-admin-command> + <ec-admin-command>-fs NAMENODE -listPolicies</ec-admin-command> </test-commands> <cleanup-commands> </cleanup-commands> @@ -247,7 +247,7 @@ </test> <test> - <description>createZone : illegal parameters - schema name is missing</description> + <description>createZone : illegal parameters - policy name is missing</description> <test-commands> <command>-fs NAMENODE -mkdir /eczone</command> <ec-admin-command>-fs NAMENODE -createZone -s</ec-admin-command> @@ -281,10 +281,10 @@ </test> <test> - <description>createZone : illegal parameters - invalidschema</description> + <description>createZone : illegal parameters - invalidpolicy</description> <test-commands> <command>-fs NAMENODE -mkdir /eczone</command> - <ec-admin-command>-fs NAMENODE -createZone -s invalidschema /eczone</ec-admin-command> + <ec-admin-command>-fs NAMENODE -createZone -s invalidpolicy /eczone</ec-admin-command> </test-commands> <cleanup-commands> <command>-fs NAMENODE -rmdir /eczone</command> @@ -292,7 +292,7 @@ <comparators> <comparator> <type>SubstringComparator</type> - <expected-output>Schema 'invalidschema' does not match any of the supported schemas. Please select any one of [RS-6-3]</expected-output> + <expected-output>Policy 'invalidpolicy' does not match any of the supported policies. Please select any one of [RS-6-3-64k]</expected-output> </comparator> </comparators> </test> @@ -359,16 +359,16 @@ </test> <test> - <description>listSchemas : illegal parameters - too many parameters</description> + <description>listPolicies : illegal parameters - too many parameters</description> <test-commands> - <ec-admin-command>-fs NAMENODE -listSchemas /eczone</ec-admin-command> + <ec-admin-command>-fs NAMENODE -listPolicies /eczone</ec-admin-command> </test-commands> <cleanup-commands> </cleanup-commands> <comparators> <comparator> <type>SubstringComparator</type> - <expected-output>-listSchemas: Too many parameters</expected-output> + <expected-output>-listPolicies: Too many parameters</expected-output> </comparator> </comparators> </test>