This is an automated email from the ASF dual-hosted git repository. tmarquardt pushed a commit to branch branch-2 in repository https://gitbox.apache.org/repos/asf/hadoop.git
commit 50953c586b1a96e46119fdb6a3366309da389985 Author: Da Zhou <da.z...@microsoft.com> AuthorDate: Wed May 8 17:20:46 2019 +0100 HADOOP-16269. ABFS: add listFileStatus with StartFrom. Author: Da Zhou --- .../fs/azurebfs/AzureBlobFileSystemStore.java | 107 ++++++++++++++- .../fs/azurebfs/constants/AbfsHttpConstants.java | 9 ++ .../org/apache/hadoop/fs/azurebfs/utils/CRC64.java | 60 ++++++++ .../fs/azurebfs/AbstractAbfsIntegrationTest.java | 7 +- ...zureBlobFileSystemStoreListStatusWithRange.java | 151 +++++++++++++++++++++ .../apache/hadoop/fs/azurebfs/TestAbfsCrc64.java | 38 ++++++ 6 files changed, 363 insertions(+), 9 deletions(-) diff --git a/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/AzureBlobFileSystemStore.java b/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/AzureBlobFileSystemStore.java index 4d11563..f36153d 100644 --- a/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/AzureBlobFileSystemStore.java +++ b/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/AzureBlobFileSystemStore.java @@ -31,6 +31,7 @@ import java.nio.charset.CharacterCodingException; import java.nio.charset.Charset; import java.nio.charset.CharsetDecoder; import java.nio.charset.CharsetEncoder; +import java.nio.charset.StandardCharsets; import java.text.ParseException; import java.text.SimpleDateFormat; import java.util.ArrayList; @@ -46,6 +47,7 @@ import java.util.Set; import com.google.common.annotations.VisibleForTesting; import com.google.common.base.Preconditions; +import com.google.common.base.Strings; import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.classification.InterfaceStability; @@ -79,6 +81,7 @@ import org.apache.hadoop.fs.azurebfs.services.AuthType; import org.apache.hadoop.fs.azurebfs.services.ExponentialRetryPolicy; import org.apache.hadoop.fs.azurebfs.services.SharedKeyCredentials; import org.apache.hadoop.fs.azurebfs.utils.Base64; +import org.apache.hadoop.fs.azurebfs.utils.CRC64; import org.apache.hadoop.fs.azurebfs.utils.UriUtils; import org.apache.hadoop.fs.permission.AclEntry; import org.apache.hadoop.fs.permission.AclStatus; @@ -89,7 +92,17 @@ import org.apache.http.client.utils.URIBuilder; import org.slf4j.Logger; import org.slf4j.LoggerFactory; +import static org.apache.hadoop.fs.azurebfs.constants.AbfsHttpConstants.CHAR_EQUALS; +import static org.apache.hadoop.fs.azurebfs.constants.AbfsHttpConstants.CHAR_FORWARD_SLASH; +import static org.apache.hadoop.fs.azurebfs.constants.AbfsHttpConstants.CHAR_HYPHEN; +import static org.apache.hadoop.fs.azurebfs.constants.AbfsHttpConstants.CHAR_PLUS; +import static org.apache.hadoop.fs.azurebfs.constants.AbfsHttpConstants.CHAR_STAR; +import static org.apache.hadoop.fs.azurebfs.constants.AbfsHttpConstants.CHAR_UNDERSCORE; +import static org.apache.hadoop.fs.azurebfs.constants.AbfsHttpConstants.ROOT_PATH; +import static org.apache.hadoop.fs.azurebfs.constants.AbfsHttpConstants.SINGLE_WHITE_SPACE; +import static org.apache.hadoop.fs.azurebfs.constants.AbfsHttpConstants.TOKEN_VERSION; import static org.apache.hadoop.fs.azurebfs.constants.ConfigurationKeys.AZURE_ABFS_ENDPOINT; + /** * Provides the bridging logic between Hadoop's abstract filesystem and Azure Storage. */ @@ -103,6 +116,7 @@ public class AzureBlobFileSystemStore { private String userName; private String primaryUserGroup; private static final String DATE_TIME_PATTERN = "E, dd MMM yyyy HH:mm:ss 'GMT'"; + private static final String TOKEN_DATE_PATTERN = "yyyy-MM-dd'T'HH:mm:ss.SSSSSSS'Z'"; private static final String XMS_PROPERTIES_ENCODING = "ISO-8859-1"; private static final int LIST_MAX_RESULTS = 500; @@ -514,15 +528,43 @@ public class AzureBlobFileSystemStore { eTag); } + /** + * @param path The list path. + * @return the entries in the path. + * */ public FileStatus[] listStatus(final Path path) throws IOException { - LOG.debug("listStatus filesystem: {} path: {}", + return listStatus(path, null); + } + + /** + * @param path Path the list path. + * @param startFrom the entry name that list results should start with. + * For example, if folder "/folder" contains four files: "afile", "bfile", "hfile", "ifile". + * Then listStatus(Path("/folder"), "hfile") will return "/folder/hfile" and "folder/ifile" + * Notice that if startFrom is a non-existent entry name, then the list response contains + * all entries after this non-existent entry in lexical order: + * listStatus(Path("/folder"), "cfile") will return "/folder/hfile" and "/folder/ifile". + * + * @return the entries in the path start from "startFrom" in lexical order. + * */ + @InterfaceStability.Unstable + public FileStatus[] listStatus(final Path path, final String startFrom) throws IOException { + LOG.debug("listStatus filesystem: {} path: {}, startFrom: {}", client.getFileSystem(), - path); + path, + startFrom); - String relativePath = path.isRoot() ? AbfsHttpConstants.EMPTY_STRING : getRelativePath(path); + final String relativePath = path.isRoot() ? AbfsHttpConstants.EMPTY_STRING : getRelativePath(path); String continuation = null; - ArrayList<FileStatus> fileStatuses = new ArrayList<>(); + // generate continuation token if a valid startFrom is provided. + if (startFrom != null && !startFrom.isEmpty()) { + continuation = getIsNamespaceEnabled() + ? generateContinuationTokenForXns(startFrom) + : generateContinuationTokenForNonXns(path.isRoot() ? ROOT_PATH : relativePath, startFrom); + } + + ArrayList<FileStatus> fileStatuses = new ArrayList<>(); do { AbfsRestOperation op = client.listPath(relativePath, false, LIST_MAX_RESULTS, continuation); continuation = op.getResult().getResponseHeader(HttpHeaderConfigurations.X_MS_CONTINUATION); @@ -575,6 +617,61 @@ public class AzureBlobFileSystemStore { return fileStatuses.toArray(new FileStatus[fileStatuses.size()]); } + // generate continuation token for xns account + private String generateContinuationTokenForXns(final String firstEntryName) { + Preconditions.checkArgument(!Strings.isNullOrEmpty(firstEntryName) + && !firstEntryName.startsWith(AbfsHttpConstants.ROOT_PATH), + "startFrom must be a dir/file name and it can not be a full path"); + + StringBuilder sb = new StringBuilder(); + sb.append(firstEntryName).append("#$").append("0"); + + CRC64 crc64 = new CRC64(); + StringBuilder token = new StringBuilder(); + token.append(crc64.compute(sb.toString().getBytes(StandardCharsets.UTF_8))) + .append(SINGLE_WHITE_SPACE) + .append("0") + .append(SINGLE_WHITE_SPACE) + .append(firstEntryName); + + return Base64.encode(token.toString().getBytes(StandardCharsets.UTF_8)); + } + + // generate continuation token for non-xns account + private String generateContinuationTokenForNonXns(final String path, final String firstEntryName) { + Preconditions.checkArgument(!Strings.isNullOrEmpty(firstEntryName) + && !firstEntryName.startsWith(AbfsHttpConstants.ROOT_PATH), + "startFrom must be a dir/file name and it can not be a full path"); + + // Notice: non-xns continuation token requires full path (first "/" is not included) for startFrom + final String startFrom = (path.isEmpty() || path.equals(ROOT_PATH)) + ? firstEntryName + : path + ROOT_PATH + firstEntryName; + + SimpleDateFormat simpleDateFormat = new SimpleDateFormat(TOKEN_DATE_PATTERN, Locale.US); + String date = simpleDateFormat.format(new Date()); + String token = String.format("%06d!%s!%06d!%s!%06d!%s!", + path.length(), path, startFrom.length(), startFrom, date.length(), date); + String base64EncodedToken = Base64.encode(token.getBytes(StandardCharsets.UTF_8)); + + StringBuilder encodedTokenBuilder = new StringBuilder(base64EncodedToken.length() + 5); + encodedTokenBuilder.append(String.format("%s!%d!", TOKEN_VERSION, base64EncodedToken.length())); + + for (int i = 0; i < base64EncodedToken.length(); i++) { + char current = base64EncodedToken.charAt(i); + if (CHAR_FORWARD_SLASH == current) { + current = CHAR_UNDERSCORE; + } else if (CHAR_PLUS == current) { + current = CHAR_STAR; + } else if (CHAR_EQUALS == current) { + current = CHAR_HYPHEN; + } + encodedTokenBuilder.append(current); + } + + return encodedTokenBuilder.toString(); + } + public void setOwner(final Path path, final String owner, final String group) throws AzureBlobFileSystemException { if (!getIsNamespaceEnabled()) { @@ -990,7 +1087,7 @@ public class AzureBlobFileSystemStore { FileStatus other = (FileStatus) obj; - if (!other.equals(this)) {// compare the path + if (!this.getPath().equals(other.getPath())) {// compare the path return false; } diff --git a/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/constants/AbfsHttpConstants.java b/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/constants/AbfsHttpConstants.java index 1f35854..e85c7f0 100644 --- a/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/constants/AbfsHttpConstants.java +++ b/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/constants/AbfsHttpConstants.java @@ -39,6 +39,7 @@ public final class AbfsHttpConstants { public static final String GET_ACCESS_CONTROL = "getAccessControl"; public static final String GET_STATUS = "getStatus"; public static final String DEFAULT_TIMEOUT = "90"; + public static final String TOKEN_VERSION = "2"; public static final String JAVA_VERSION = "java.version"; public static final String OS_NAME = "os.name"; @@ -91,5 +92,13 @@ public final class AbfsHttpConstants { public static final String PERMISSION_FORMAT = "%04d"; public static final String SUPER_USER = "$superuser"; + public static final char CHAR_FORWARD_SLASH = '/'; + public static final char CHAR_EXCLAMATION_POINT = '!'; + public static final char CHAR_UNDERSCORE = '_'; + public static final char CHAR_HYPHEN = '-'; + public static final char CHAR_EQUALS = '='; + public static final char CHAR_STAR = '*'; + public static final char CHAR_PLUS = '+'; + private AbfsHttpConstants() {} } diff --git a/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/utils/CRC64.java b/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/utils/CRC64.java new file mode 100644 index 0000000..9790744 --- /dev/null +++ b/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/utils/CRC64.java @@ -0,0 +1,60 @@ +/** + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ +package org.apache.hadoop.fs.azurebfs.utils; + +/** + * CRC64 implementation for AzureBlobFileSystem. + */ +public class CRC64 { + + private static final long POLY = 0x9a6c9329ac4bc9b5L; + private static final int TABLE_LENGTH = 256; + private static final long[] TABLE = new long[TABLE_LENGTH]; + + private long value = -1; + + /** + * @param input byte arrays. + * @return long value of the CRC-64 checksum of the data. + * */ + public long compute(byte[] input) { + init(); + for (int i = 0; i < input.length; i++) { + value = TABLE[(input[i] ^ (int) value) & 0xFF] ^ (value >>> 8); + } + return ~value; + } + + /* + * Initialize a table constructed from POLY (0x9a6c9329ac4bc9b5L). + * */ + private void init() { + value = -1; + for (int n = 0; n < TABLE_LENGTH; ++n) { + long crc = n; + for (int i = 0; i < 8; ++i) { + if ((crc & 1) == 1) { + crc = (crc >>> 1) ^ POLY; + } else { + crc >>>= 1; + } + } + TABLE[n] = crc; + } + } +} diff --git a/hadoop-tools/hadoop-azure/src/test/java/org/apache/hadoop/fs/azurebfs/AbstractAbfsIntegrationTest.java b/hadoop-tools/hadoop-azure/src/test/java/org/apache/hadoop/fs/azurebfs/AbstractAbfsIntegrationTest.java index cb9549d..04be4f4 100644 --- a/hadoop-tools/hadoop-azure/src/test/java/org/apache/hadoop/fs/azurebfs/AbstractAbfsIntegrationTest.java +++ b/hadoop-tools/hadoop-azure/src/test/java/org/apache/hadoop/fs/azurebfs/AbstractAbfsIntegrationTest.java @@ -24,7 +24,6 @@ import java.util.Hashtable; import java.util.UUID; import java.util.concurrent.Callable; -import com.google.common.base.Preconditions; import org.junit.After; import org.junit.Before; import org.slf4j.Logger; @@ -210,9 +209,9 @@ public abstract class AbstractAbfsIntegrationTest extends * @throws IOException failure during create/init. */ public AzureBlobFileSystem createFileSystem() throws IOException { - Preconditions.checkState(abfs == null, - "existing ABFS instance exists: %s", abfs); - abfs = (AzureBlobFileSystem) FileSystem.newInstance(rawConfig); + if (abfs == null) { + abfs = (AzureBlobFileSystem) FileSystem.newInstance(rawConfig); + } return abfs; } diff --git a/hadoop-tools/hadoop-azure/src/test/java/org/apache/hadoop/fs/azurebfs/ITestAzureBlobFileSystemStoreListStatusWithRange.java b/hadoop-tools/hadoop-azure/src/test/java/org/apache/hadoop/fs/azurebfs/ITestAzureBlobFileSystemStoreListStatusWithRange.java new file mode 100644 index 0000000..849bb6b --- /dev/null +++ b/hadoop-tools/hadoop-azure/src/test/java/org/apache/hadoop/fs/azurebfs/ITestAzureBlobFileSystemStoreListStatusWithRange.java @@ -0,0 +1,151 @@ +/** + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ +package org.apache.hadoop.fs.azurebfs; + +import java.io.IOException; +import java.util.Arrays; + +import org.junit.Assert; +import org.junit.Test; +import org.junit.runner.RunWith; +import org.junit.runners.Parameterized; + +import org.apache.hadoop.fs.contract.ContractTestUtils; +import org.apache.hadoop.fs.FileStatus; +import org.apache.hadoop.fs.Path; + +/** + * Test AzureBlobFileSystemStore listStatus with startFrom. + * */ +@RunWith(Parameterized.class) +public class ITestAzureBlobFileSystemStoreListStatusWithRange extends + AbstractAbfsIntegrationTest { + private static final boolean SUCCEED = true; + private static final boolean FAIL = false; + private static final String[] SORTED_ENTRY_NAMES = {"1_folder", "A0", "D01", "a+", "c0", "name5"}; + + private AzureBlobFileSystemStore store; + private AzureBlobFileSystem fs; + + @Parameterized.Parameter + public String path; + + /** + * A valid startFrom for listFileStatus with range is a non-fully qualified dir/file name + * */ + @Parameterized.Parameter(1) + public String startFrom; + + @Parameterized.Parameter(2) + public int expectedStartIndexInArray; + + @Parameterized.Parameter(3) + public boolean expectedResult; + + @Parameterized.Parameters(name = "Testing path \"{0}\", startFrom: \"{1}\", Expecting result : {3}") // Test path + public static Iterable<Object[]> params() { + return Arrays.asList( + new Object[][]{ + // case 0: list in root, without range + {"/", null, 0, SUCCEED}, + + // case 1: list in the root, start from the second file + {"/", SORTED_ENTRY_NAMES[1], 1, SUCCEED}, + + // case 2: list in the root, invalid startFrom + {"/", "/", -1, FAIL}, + + // case 3: list in non-root level, valid startFrom : dir name + {"/" + SORTED_ENTRY_NAMES[2], SORTED_ENTRY_NAMES[1], 1, SUCCEED}, + + // case 4: list in non-root level, valid startFrom : file name + {"/" + SORTED_ENTRY_NAMES[2], SORTED_ENTRY_NAMES[2], 2, SUCCEED}, + + // case 5: list in non root level, invalid startFrom + {"/" + SORTED_ENTRY_NAMES[2], "/" + SORTED_ENTRY_NAMES[3], -1, FAIL}, + + // case 6: list using non existent startFrom, startFrom is smaller than the entries in lexical order + // expecting return all entries + {"/" + SORTED_ENTRY_NAMES[2], "0-non-existent", 0, SUCCEED}, + + // case 7: list using non existent startFrom, startFrom is larger than the entries in lexical order + // expecting return 0 entries + {"/" + SORTED_ENTRY_NAMES[2], "z-non-existent", -1, SUCCEED}, + + // case 8: list using non existent startFrom, startFrom is in the range + {"/" + SORTED_ENTRY_NAMES[2], "A1", 2, SUCCEED} + }); + } + + public ITestAzureBlobFileSystemStoreListStatusWithRange() throws Exception { + super(); + if (this.getFileSystem() == null) { + super.createFileSystem(); + } + fs = this.getFileSystem(); + store = fs.getAbfsStore(); + prepareTestFiles(); + // Sort the names for verification, ABFS service should return the results in order. + Arrays.sort(SORTED_ENTRY_NAMES); + } + + @Test + public void testListWithRange() throws IOException { + try { + FileStatus[] listResult = store.listStatus(new Path(path), startFrom); + if (!expectedResult) { + Assert.fail("Excepting failure with IllegalArgumentException"); + } + verifyFileStatus(listResult, new Path(path), expectedStartIndexInArray); + } catch (IllegalArgumentException ex) { + if (expectedResult) { + Assert.fail("Excepting success"); + } + } + } + + // compare the file status + private void verifyFileStatus(FileStatus[] listResult, Path parentPath, int startIndexInSortedName) throws IOException { + if (startIndexInSortedName == -1) { + Assert.assertEquals("Expected empty FileStatus array", 0, listResult.length); + return; + } + + FileStatus[] allFileStatuses = fs.listStatus(parentPath); + Assert.assertEquals("number of dir/file doesn't match", + SORTED_ENTRY_NAMES.length, allFileStatuses.length); + int indexInResult = 0; + for (int index = startIndexInSortedName; index < SORTED_ENTRY_NAMES.length; index++) { + Assert.assertEquals("fileStatus doesn't match", allFileStatuses[index], listResult[indexInResult++]); + } + } + + private void prepareTestFiles() throws IOException { + final AzureBlobFileSystem fs = getFileSystem(); + // created 2 level file structures + for (String levelOneFolder : SORTED_ENTRY_NAMES) { + Path levelOnePath = new Path("/" + levelOneFolder); + Assert.assertTrue(fs.mkdirs(levelOnePath)); + for (String fileName : SORTED_ENTRY_NAMES) { + Path filePath = new Path(levelOnePath, fileName); + ContractTestUtils.touch(fs, filePath); + ContractTestUtils.assertIsFile(fs, filePath); + } + } + } +} diff --git a/hadoop-tools/hadoop-azure/src/test/java/org/apache/hadoop/fs/azurebfs/TestAbfsCrc64.java b/hadoop-tools/hadoop-azure/src/test/java/org/apache/hadoop/fs/azurebfs/TestAbfsCrc64.java new file mode 100644 index 0000000..ab39750 --- /dev/null +++ b/hadoop-tools/hadoop-azure/src/test/java/org/apache/hadoop/fs/azurebfs/TestAbfsCrc64.java @@ -0,0 +1,38 @@ +/** + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ +package org.apache.hadoop.fs.azurebfs; + +import org.junit.Assert; +import org.junit.Test; + +import org.apache.hadoop.fs.azurebfs.utils.CRC64; +/** + * Test for Crc64 in AzureBlobFileSystem, notice that ABFS CRC64 has its own polynomial. + * */ +public class TestAbfsCrc64 { + + @Test + public void tesCrc64Compute() { + CRC64 crc64 = new CRC64(); + final String[] testStr = {"#$", "dir_2_ac83abee", "dir_42_976df1f5"}; + final String[] expected = {"f91f7e6a837dbfa8", "203f9fefc38ae97b", "cc0d56eafe58a855"}; + for (int i = 0; i < testStr.length; i++) { + Assert.assertEquals(expected[i], Long.toHexString(crc64.compute(testStr[i].getBytes()))); + } + } +} --------------------------------------------------------------------- To unsubscribe, e-mail: common-commits-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-commits-h...@hadoop.apache.org