Repository: hadoop Updated Branches: refs/heads/YARN-6592 0e66d31e2 -> bf2a8ccc0 (forced update)
http://git-wip-us.apache.org/repos/asf/hadoop/blob/98f5ed5a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/TestProvidedImpl.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/TestProvidedImpl.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/TestProvidedImpl.java index 8782e71..40d77f7a 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/TestProvidedImpl.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/TestProvidedImpl.java @@ -52,11 +52,12 @@ import org.apache.hadoop.fs.FileSystemTestHelper; import org.apache.hadoop.fs.Path; import org.apache.hadoop.fs.StorageType; import org.apache.hadoop.hdfs.DFSConfigKeys; +import org.apache.hadoop.hdfs.protocol.Block; import org.apache.hadoop.hdfs.protocol.ExtendedBlock; import org.apache.hadoop.hdfs.protocol.HdfsConstants; import org.apache.hadoop.hdfs.server.common.FileRegion; -import org.apache.hadoop.hdfs.server.common.FileRegionProvider; import org.apache.hadoop.hdfs.server.common.Storage; +import org.apache.hadoop.hdfs.server.common.blockaliasmap.BlockAliasMap; import org.apache.hadoop.hdfs.server.datanode.BlockScanner; import org.apache.hadoop.hdfs.server.datanode.DNConf; import org.apache.hadoop.hdfs.server.datanode.DataNode; @@ -168,49 +169,66 @@ public class TestProvidedImpl { } /** - * A simple FileRegion provider for tests. + * A simple FileRegion BlockAliasMap for tests. */ - public static class TestFileRegionProvider - extends FileRegionProvider implements Configurable { + public static class TestFileRegionBlockAliasMap + extends BlockAliasMap<FileRegion> { private Configuration conf; private int minId; private int numBlocks; private Iterator<FileRegion> suppliedIterator; - TestFileRegionProvider() { + TestFileRegionBlockAliasMap() { this(null, MIN_BLK_ID, NUM_PROVIDED_BLKS); } - TestFileRegionProvider(Iterator<FileRegion> iterator, int minId, - int numBlocks) { + TestFileRegionBlockAliasMap(Iterator<FileRegion> iterator, int minId, + int numBlocks) { this.suppliedIterator = iterator; this.minId = minId; this.numBlocks = numBlocks; } @Override - public Iterator<FileRegion> iterator() { - if (suppliedIterator == null) { - return new TestFileRegionIterator(providedBasePath, minId, numBlocks); - } else { - return suppliedIterator; - } - } + public Reader<FileRegion> getReader(Reader.Options opts) + throws IOException { + + BlockAliasMap.Reader<FileRegion> reader = + new BlockAliasMap.Reader<FileRegion>() { + @Override + public Iterator<FileRegion> iterator() { + if (suppliedIterator == null) { + return new TestFileRegionIterator(providedBasePath, minId, + numBlocks); + } else { + return suppliedIterator; + } + } - @Override - public void setConf(Configuration conf) { - this.conf = conf; + @Override + public void close() throws IOException { + + } + + @Override + public FileRegion resolve(Block ident) throws IOException { + return null; + } + }; + return reader; } @Override - public Configuration getConf() { - return conf; + public Writer<FileRegion> getWriter(Writer.Options opts) + throws IOException { + // not implemented + return null; } @Override - public void refresh() { - //do nothing! + public void refresh() throws IOException { + // do nothing! } public void setMinBlkId(int minId) { @@ -359,8 +377,8 @@ public class TestProvidedImpl { new ShortCircuitRegistry(conf); when(datanode.getShortCircuitRegistry()).thenReturn(shortCircuitRegistry); - conf.setClass(DFSConfigKeys.DFS_PROVIDER_CLASS, - TestFileRegionProvider.class, FileRegionProvider.class); + this.conf.setClass(DFSConfigKeys.DFS_PROVIDED_ALIASMAP_CLASS, + TestFileRegionBlockAliasMap.class, BlockAliasMap.class); conf.setClass(DFSConfigKeys.DFS_PROVIDER_DF_CLASS, TestProvidedVolumeDF.class, ProvidedVolumeDF.class); @@ -496,12 +514,13 @@ public class TestProvidedImpl { conf.setInt(DFSConfigKeys.DFS_DATANODE_DIRECTORYSCAN_THREADS_KEY, 1); for (int i = 0; i < providedVolumes.size(); i++) { ProvidedVolumeImpl vol = (ProvidedVolumeImpl) providedVolumes.get(i); - TestFileRegionProvider provider = (TestFileRegionProvider) - vol.getFileRegionProvider(BLOCK_POOL_IDS[CHOSEN_BP_ID]); + TestFileRegionBlockAliasMap testBlockFormat = + (TestFileRegionBlockAliasMap) vol + .getBlockFormat(BLOCK_POOL_IDS[CHOSEN_BP_ID]); //equivalent to two new blocks appearing - provider.setBlockCount(NUM_PROVIDED_BLKS + 2); + testBlockFormat.setBlockCount(NUM_PROVIDED_BLKS + 2); //equivalent to deleting the first block - provider.setMinBlkId(MIN_BLK_ID + 1); + testBlockFormat.setMinBlkId(MIN_BLK_ID + 1); DirectoryScanner scanner = new DirectoryScanner(datanode, dataset, conf); scanner.reconcile(); @@ -525,7 +544,7 @@ public class TestProvidedImpl { for (int i = 0; i < providedVolumes.size(); i++) { ProvidedVolumeImpl vol = (ProvidedVolumeImpl) providedVolumes.get(i); vol.setFileRegionProvider(BLOCK_POOL_IDS[CHOSEN_BP_ID], - new TestFileRegionProvider(fileRegionIterator, minBlockId, + new TestFileRegionBlockAliasMap(fileRegionIterator, minBlockId, numBlocks)); ReplicaMap volumeMap = new ReplicaMap(new AutoCloseableLock()); vol.getVolumeMap(BLOCK_POOL_IDS[CHOSEN_BP_ID], volumeMap, null); http://git-wip-us.apache.org/repos/asf/hadoop/blob/98f5ed5a/hadoop-tools/hadoop-fs2img/src/main/java/org/apache/hadoop/hdfs/server/namenode/FileSystemImage.java ---------------------------------------------------------------------- diff --git a/hadoop-tools/hadoop-fs2img/src/main/java/org/apache/hadoop/hdfs/server/namenode/FileSystemImage.java b/hadoop-tools/hadoop-fs2img/src/main/java/org/apache/hadoop/hdfs/server/namenode/FileSystemImage.java index e1e85c1..2e57c9f 100644 --- a/hadoop-tools/hadoop-fs2img/src/main/java/org/apache/hadoop/hdfs/server/namenode/FileSystemImage.java +++ b/hadoop-tools/hadoop-fs2img/src/main/java/org/apache/hadoop/hdfs/server/namenode/FileSystemImage.java @@ -29,7 +29,7 @@ import org.apache.commons.cli.PosixParser; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; -import org.apache.hadoop.hdfs.server.common.BlockFormat; +import org.apache.hadoop.hdfs.server.common.blockaliasmap.BlockAliasMap; import org.apache.hadoop.util.ReflectionUtils; import org.apache.hadoop.util.Tool; import org.apache.hadoop.util.ToolRunner; @@ -103,7 +103,7 @@ public class FileSystemImage implements Tool { break; case "b": opts.blocks( - Class.forName(o.getValue()).asSubclass(BlockFormat.class)); + Class.forName(o.getValue()).asSubclass(BlockAliasMap.class)); break; case "i": opts.blockIds( http://git-wip-us.apache.org/repos/asf/hadoop/blob/98f5ed5a/hadoop-tools/hadoop-fs2img/src/main/java/org/apache/hadoop/hdfs/server/namenode/ImageWriter.java ---------------------------------------------------------------------- diff --git a/hadoop-tools/hadoop-fs2img/src/main/java/org/apache/hadoop/hdfs/server/namenode/ImageWriter.java b/hadoop-tools/hadoop-fs2img/src/main/java/org/apache/hadoop/hdfs/server/namenode/ImageWriter.java index a3603a1..ea1888a 100644 --- a/hadoop-tools/hadoop-fs2img/src/main/java/org/apache/hadoop/hdfs/server/namenode/ImageWriter.java +++ b/hadoop-tools/hadoop-fs2img/src/main/java/org/apache/hadoop/hdfs/server/namenode/ImageWriter.java @@ -44,8 +44,8 @@ import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.LocalFileSystem; import org.apache.hadoop.fs.Path; import org.apache.hadoop.hdfs.DFSConfigKeys; -import org.apache.hadoop.hdfs.server.common.BlockFormat; import org.apache.hadoop.hdfs.server.common.FileRegion; +import org.apache.hadoop.hdfs.server.common.blockaliasmap.BlockAliasMap; import org.apache.hadoop.hdfs.server.namenode.FSImageFormatProtobuf.SectionName; import org.apache.hadoop.hdfs.server.namenode.FsImageProto.CacheManagerSection; import org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary; @@ -88,7 +88,7 @@ public class ImageWriter implements Closeable { private final long startBlock; private final long startInode; private final UGIResolver ugis; - private final BlockFormat.Writer<FileRegion> blocks; + private final BlockAliasMap.Writer<FileRegion> blocks; private final BlockResolver blockIds; private final Map<Long, DirEntry.Builder> dircache; private final TrackedOutputStream<DigestOutputStream> raw; @@ -155,8 +155,8 @@ public class ImageWriter implements Closeable { ugis = null == opts.ugis ? ReflectionUtils.newInstance(opts.ugisClass, opts.getConf()) : opts.ugis; - BlockFormat<FileRegion> fmt = null == opts.blocks - ? ReflectionUtils.newInstance(opts.blockFormatClass, opts.getConf()) + BlockAliasMap<FileRegion> fmt = null == opts.blocks + ? ReflectionUtils.newInstance(opts.aliasMap, opts.getConf()) : opts.blocks; blocks = fmt.getWriter(null); blockIds = null == opts.blockIds @@ -509,10 +509,10 @@ public class ImageWriter implements Closeable { private long startInode; private UGIResolver ugis; private Class<? extends UGIResolver> ugisClass; - private BlockFormat<FileRegion> blocks; + private BlockAliasMap<FileRegion> blocks; @SuppressWarnings("rawtypes") - private Class<? extends BlockFormat> blockFormatClass; + private Class<? extends BlockAliasMap> aliasMap; private BlockResolver blockIds; private Class<? extends BlockResolver> blockIdsClass; private FSImageCompression compress = @@ -524,7 +524,6 @@ public class ImageWriter implements Closeable { @Override public void setConf(Configuration conf) { this.conf = conf; - //long lastTxn = conf.getLong(LAST_TXN, 0L); String def = new File("hdfs/name").toURI().toString(); outdir = new Path(conf.get(DFSConfigKeys.DFS_NAMENODE_NAME_DIR_KEY, def)); startBlock = conf.getLong(FixedBlockResolver.START_BLOCK, (1L << 30) + 1); @@ -532,9 +531,9 @@ public class ImageWriter implements Closeable { maxdircache = conf.getInt(CACHE_ENTRY, 100); ugisClass = conf.getClass(UGI_CLASS, SingleUGIResolver.class, UGIResolver.class); - blockFormatClass = conf.getClass( - DFSConfigKeys.DFS_PROVIDER_BLK_FORMAT_CLASS, - NullBlockFormat.class, BlockFormat.class); + aliasMap = conf.getClass( + DFSConfigKeys.DFS_PROVIDED_ALIASMAP_CLASS, + NullBlockAliasMap.class, BlockAliasMap.class); blockIdsClass = conf.getClass(BLOCK_RESOLVER_CLASS, FixedBlockResolver.class, BlockResolver.class); } @@ -584,14 +583,14 @@ public class ImageWriter implements Closeable { return this; } - public Options blocks(BlockFormat<FileRegion> blocks) { + public Options blocks(BlockAliasMap<FileRegion> blocks) { this.blocks = blocks; return this; } @SuppressWarnings("rawtypes") - public Options blocks(Class<? extends BlockFormat> blocksClass) { - this.blockFormatClass = blocksClass; + public Options blocks(Class<? extends BlockAliasMap> blocksClass) { + this.aliasMap = blocksClass; return this; } http://git-wip-us.apache.org/repos/asf/hadoop/blob/98f5ed5a/hadoop-tools/hadoop-fs2img/src/main/java/org/apache/hadoop/hdfs/server/namenode/NullBlockAliasMap.java ---------------------------------------------------------------------- diff --git a/hadoop-tools/hadoop-fs2img/src/main/java/org/apache/hadoop/hdfs/server/namenode/NullBlockAliasMap.java b/hadoop-tools/hadoop-fs2img/src/main/java/org/apache/hadoop/hdfs/server/namenode/NullBlockAliasMap.java new file mode 100644 index 0000000..4cdf473 --- /dev/null +++ b/hadoop-tools/hadoop-fs2img/src/main/java/org/apache/hadoop/hdfs/server/namenode/NullBlockAliasMap.java @@ -0,0 +1,86 @@ +/** + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ +package org.apache.hadoop.hdfs.server.namenode; + +import java.io.IOException; +import java.util.Iterator; +import java.util.NoSuchElementException; + +import org.apache.hadoop.hdfs.protocol.Block; +import org.apache.hadoop.hdfs.server.common.FileRegion; +import org.apache.hadoop.hdfs.server.common.blockaliasmap.BlockAliasMap; + +/** + * Null sink for region information emitted from FSImage. + */ +public class NullBlockAliasMap extends BlockAliasMap<FileRegion> { + + @Override + public Reader<FileRegion> getReader(Reader.Options opts) throws IOException { + return new Reader<FileRegion>() { + @Override + public Iterator<FileRegion> iterator() { + return new Iterator<FileRegion>() { + @Override + public boolean hasNext() { + return false; + } + @Override + public FileRegion next() { + throw new NoSuchElementException(); + } + @Override + public void remove() { + throw new UnsupportedOperationException(); + } + }; + } + + @Override + public void close() throws IOException { + // do nothing + } + + @Override + public FileRegion resolve(Block ident) throws IOException { + throw new UnsupportedOperationException(); + } + }; + } + + @Override + public Writer<FileRegion> getWriter(Writer.Options opts) throws IOException { + return new Writer<FileRegion>() { + @Override + public void store(FileRegion token) throws IOException { + // do nothing + } + + @Override + public void close() throws IOException { + // do nothing + } + }; + } + + @Override + public void refresh() throws IOException { + // do nothing + } + +} http://git-wip-us.apache.org/repos/asf/hadoop/blob/98f5ed5a/hadoop-tools/hadoop-fs2img/src/main/java/org/apache/hadoop/hdfs/server/namenode/NullBlockFormat.java ---------------------------------------------------------------------- diff --git a/hadoop-tools/hadoop-fs2img/src/main/java/org/apache/hadoop/hdfs/server/namenode/NullBlockFormat.java b/hadoop-tools/hadoop-fs2img/src/main/java/org/apache/hadoop/hdfs/server/namenode/NullBlockFormat.java deleted file mode 100644 index aabdf74..0000000 --- a/hadoop-tools/hadoop-fs2img/src/main/java/org/apache/hadoop/hdfs/server/namenode/NullBlockFormat.java +++ /dev/null @@ -1,87 +0,0 @@ -/** - * Licensed to the Apache Software Foundation (ASF) under one - * or more contributor license agreements. See the NOTICE file - * distributed with this work for additional information - * regarding copyright ownership. The ASF licenses this file - * to you under the Apache License, Version 2.0 (the - * "License"); you may not use this file except in compliance - * with the License. You may obtain a copy of the License at - * - * http://www.apache.org/licenses/LICENSE-2.0 - * - * Unless required by applicable law or agreed to in writing, software - * distributed under the License is distributed on an "AS IS" BASIS, - * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. - * See the License for the specific language governing permissions and - * limitations under the License. - */ -package org.apache.hadoop.hdfs.server.namenode; - -import java.io.IOException; -import java.util.Iterator; -import java.util.NoSuchElementException; - -import org.apache.hadoop.hdfs.protocol.Block; -import org.apache.hadoop.hdfs.server.common.BlockFormat; -import org.apache.hadoop.hdfs.server.common.BlockFormat.Reader.Options; -import org.apache.hadoop.hdfs.server.common.FileRegion; - -/** - * Null sink for region information emitted from FSImage. - */ -public class NullBlockFormat extends BlockFormat<FileRegion> { - - @Override - public Reader<FileRegion> getReader(Options opts) throws IOException { - return new Reader<FileRegion>() { - @Override - public Iterator<FileRegion> iterator() { - return new Iterator<FileRegion>() { - @Override - public boolean hasNext() { - return false; - } - @Override - public FileRegion next() { - throw new NoSuchElementException(); - } - @Override - public void remove() { - throw new UnsupportedOperationException(); - } - }; - } - - @Override - public void close() throws IOException { - // do nothing - } - - @Override - public FileRegion resolve(Block ident) throws IOException { - throw new UnsupportedOperationException(); - } - }; - } - - @Override - public Writer<FileRegion> getWriter(Writer.Options opts) throws IOException { - return new Writer<FileRegion>() { - @Override - public void store(FileRegion token) throws IOException { - // do nothing - } - - @Override - public void close() throws IOException { - // do nothing - } - }; - } - - @Override - public void refresh() throws IOException { - // do nothing - } - -} http://git-wip-us.apache.org/repos/asf/hadoop/blob/98f5ed5a/hadoop-tools/hadoop-fs2img/src/main/java/org/apache/hadoop/hdfs/server/namenode/TreePath.java ---------------------------------------------------------------------- diff --git a/hadoop-tools/hadoop-fs2img/src/main/java/org/apache/hadoop/hdfs/server/namenode/TreePath.java b/hadoop-tools/hadoop-fs2img/src/main/java/org/apache/hadoop/hdfs/server/namenode/TreePath.java index 14e6bed..d327363 100644 --- a/hadoop-tools/hadoop-fs2img/src/main/java/org/apache/hadoop/hdfs/server/namenode/TreePath.java +++ b/hadoop-tools/hadoop-fs2img/src/main/java/org/apache/hadoop/hdfs/server/namenode/TreePath.java @@ -24,8 +24,8 @@ import com.google.protobuf.ByteString; import org.apache.hadoop.fs.FileStatus; import org.apache.hadoop.hdfs.protocol.HdfsConstants; import org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto; -import org.apache.hadoop.hdfs.server.common.BlockFormat; import org.apache.hadoop.hdfs.server.common.FileRegion; +import org.apache.hadoop.hdfs.server.common.blockaliasmap.BlockAliasMap; import org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode; import org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory; import org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile; @@ -70,7 +70,7 @@ public class TreePath { } public INode toINode(UGIResolver ugi, BlockResolver blk, - BlockFormat.Writer<FileRegion> out, String blockPoolID) + BlockAliasMap.Writer<FileRegion> out, String blockPoolID) throws IOException { if (stat.isFile()) { return toFile(ugi, blk, out, blockPoolID); @@ -101,14 +101,14 @@ public class TreePath { void writeBlock(long blockId, long offset, long length, long genStamp, String blockPoolID, - BlockFormat.Writer<FileRegion> out) throws IOException { + BlockAliasMap.Writer<FileRegion> out) throws IOException { FileStatus s = getFileStatus(); out.store(new FileRegion(blockId, s.getPath(), offset, length, blockPoolID, genStamp)); } INode toFile(UGIResolver ugi, BlockResolver blk, - BlockFormat.Writer<FileRegion> out, String blockPoolID) + BlockAliasMap.Writer<FileRegion> out, String blockPoolID) throws IOException { final FileStatus s = getFileStatus(); // TODO should this store resolver's user/group? http://git-wip-us.apache.org/repos/asf/hadoop/blob/98f5ed5a/hadoop-tools/hadoop-fs2img/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestNameNodeProvidedImplementation.java ---------------------------------------------------------------------- diff --git a/hadoop-tools/hadoop-fs2img/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestNameNodeProvidedImplementation.java b/hadoop-tools/hadoop-fs2img/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestNameNodeProvidedImplementation.java index d622b9e..2170baa 100644 --- a/hadoop-tools/hadoop-fs2img/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestNameNodeProvidedImplementation.java +++ b/hadoop-tools/hadoop-fs2img/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestNameNodeProvidedImplementation.java @@ -44,13 +44,9 @@ import org.apache.hadoop.hdfs.MiniDFSCluster; import org.apache.hadoop.hdfs.protocol.DatanodeInfo; import org.apache.hadoop.hdfs.protocol.LocatedBlock; import org.apache.hadoop.hdfs.protocol.LocatedBlocks; -import org.apache.hadoop.hdfs.server.blockmanagement.BlockFormatProvider; import org.apache.hadoop.hdfs.server.blockmanagement.BlockManagerTestUtil; -import org.apache.hadoop.hdfs.server.blockmanagement.BlockProvider; -import org.apache.hadoop.hdfs.server.common.BlockFormat; -import org.apache.hadoop.hdfs.server.common.FileRegionProvider; -import org.apache.hadoop.hdfs.server.common.TextFileRegionFormat; -import org.apache.hadoop.hdfs.server.common.TextFileRegionProvider; +import org.apache.hadoop.hdfs.server.common.blockaliasmap.BlockAliasMap; +import org.apache.hadoop.hdfs.server.common.blockaliasmap.impl.TextFileRegionAliasMap; import org.apache.hadoop.hdfs.server.datanode.DataNode; import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_NAMENODE_NAME_DIR_KEY; @@ -103,18 +99,13 @@ public class TestNameNodeProvidedImplementation { DFSConfigKeys.DFS_PROVIDER_STORAGEUUID_DEFAULT); conf.setBoolean(DFSConfigKeys.DFS_NAMENODE_PROVIDED_ENABLED, true); - conf.setClass(DFSConfigKeys.DFS_NAMENODE_BLOCK_PROVIDER_CLASS, - BlockFormatProvider.class, BlockProvider.class); - conf.setClass(DFSConfigKeys.DFS_PROVIDER_CLASS, - TextFileRegionProvider.class, FileRegionProvider.class); - conf.setClass(DFSConfigKeys.DFS_PROVIDER_BLK_FORMAT_CLASS, - TextFileRegionFormat.class, BlockFormat.class); - - conf.set(DFSConfigKeys.DFS_PROVIDED_BLOCK_MAP_WRITE_PATH, + conf.setClass(DFSConfigKeys.DFS_PROVIDED_ALIASMAP_CLASS, + TextFileRegionAliasMap.class, BlockAliasMap.class); + conf.set(DFSConfigKeys.DFS_PROVIDED_ALIASMAP_TEXT_WRITE_PATH, BLOCKFILE.toString()); - conf.set(DFSConfigKeys.DFS_PROVIDED_BLOCK_MAP_READ_PATH, + conf.set(DFSConfigKeys.DFS_PROVIDED_ALIASMAP_TEXT_READ_PATH, BLOCKFILE.toString()); - conf.set(DFSConfigKeys.DFS_PROVIDED_BLOCK_MAP_DELIMITER, ","); + conf.set(DFSConfigKeys.DFS_PROVIDED_ALIASMAP_TEXT_DELIMITER, ","); conf.set(MiniDFSCluster.HDFS_MINIDFS_BASEDIR_PROVIDED, new File(NAMEPATH.toUri()).toString()); @@ -167,7 +158,7 @@ public class TestNameNodeProvidedImplementation { ImageWriter.Options opts = ImageWriter.defaults(); opts.setConf(conf); opts.output(out.toString()) - .blocks(TextFileRegionFormat.class) + .blocks(TextFileRegionAliasMap.class) .blockIds(blockIdsClass); try (ImageWriter w = new ImageWriter(opts)) { for (TreePath e : t) { --------------------------------------------------------------------- To unsubscribe, e-mail: common-commits-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-commits-h...@hadoop.apache.org