[jira] [Created] (HDFS-17197) Show file replication when listing corrupt files.
Shuyan Zhang created HDFS-17197: --- Summary: Show file replication when listing corrupt files. Key: HDFS-17197 URL: https://issues.apache.org/jira/browse/HDFS-17197 Project: Hadoop HDFS Issue Type: Improvement Reporter: Shuyan Zhang Files with different replication have different reliability guarantees. We need to pay attention to corrupted files with a specified replication greater than or equal to 3. So, when listing corrupt files, it would be useful to display the corresponding replication of the files. -- This message was sent by Atlassian Jira (v8.20.10#820010) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-17197) Show file replication when listing corrupt files.
[ https://issues.apache.org/jira/browse/HDFS-17197?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17766068#comment-17766068 ] ASF GitHub Bot commented on HDFS-17197: --- zhangshuyan0 opened a new pull request, #6095: URL: https://github.com/apache/hadoop/pull/6095 ### Description of PR Files with different replication have different reliability guarantees. We need to pay attention to corrupted files with a specified replication greater than or equal to 3. So, when listing corrupt files, it would be useful to display the corresponding replication of the files. > Show file replication when listing corrupt files. > - > > Key: HDFS-17197 > URL: https://issues.apache.org/jira/browse/HDFS-17197 > Project: Hadoop HDFS > Issue Type: Improvement >Reporter: Shuyan Zhang >Priority: Major > > Files with different replication have different reliability guarantees. We > need to pay attention to corrupted files with a specified replication greater > than or equal to 3. So, when listing corrupt files, it would be useful to > display the corresponding replication of the files. -- This message was sent by Atlassian Jira (v8.20.10#820010) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-17197) Show file replication when listing corrupt files.
[ https://issues.apache.org/jira/browse/HDFS-17197?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] ASF GitHub Bot updated HDFS-17197: -- Labels: pull-request-available (was: ) > Show file replication when listing corrupt files. > - > > Key: HDFS-17197 > URL: https://issues.apache.org/jira/browse/HDFS-17197 > Project: Hadoop HDFS > Issue Type: Improvement >Reporter: Shuyan Zhang >Priority: Major > Labels: pull-request-available > > Files with different replication have different reliability guarantees. We > need to pay attention to corrupted files with a specified replication greater > than or equal to 3. So, when listing corrupt files, it would be useful to > display the corresponding replication of the files. -- This message was sent by Atlassian Jira (v8.20.10#820010) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-17194) Enhance the log message for striped block recovery
[ https://issues.apache.org/jira/browse/HDFS-17194?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17766069#comment-17766069 ] ASF GitHub Bot commented on HDFS-17194: --- hadoop-yetus commented on PR #6094: URL: https://github.com/apache/hadoop/pull/6094#issuecomment-1722441749 :broken_heart: **-1 overall** | Vote | Subsystem | Runtime | Logfile | Comment | |::|--:|:|::|:---:| | +0 :ok: | reexec | 0m 26s | | Docker mode activated. | _ Prechecks _ | | +1 :green_heart: | dupname | 0m 0s | | No case conflicting files found. | | +0 :ok: | codespell | 0m 1s | | codespell was not available. | | +0 :ok: | detsecrets | 0m 1s | | detect-secrets was not available. | | +1 :green_heart: | @author | 0m 0s | | The patch does not contain any @author tags. | | -1 :x: | test4tests | 0m 0s | | The patch doesn't appear to include any new or modified tests. Please justify why no new tests are needed for this patch. Also please list what manual steps were performed to verify this patch. | _ trunk Compile Tests _ | | +1 :green_heart: | mvninstall | 31m 44s | | trunk passed | | +1 :green_heart: | compile | 0m 53s | | trunk passed with JDK Ubuntu-11.0.20+8-post-Ubuntu-1ubuntu120.04 | | +1 :green_heart: | compile | 0m 46s | | trunk passed with JDK Private Build-1.8.0_382-8u382-ga-1~20.04.1-b05 | | +1 :green_heart: | checkstyle | 0m 46s | | trunk passed | | +1 :green_heart: | mvnsite | 0m 54s | | trunk passed | | +1 :green_heart: | javadoc | 0m 50s | | trunk passed with JDK Ubuntu-11.0.20+8-post-Ubuntu-1ubuntu120.04 | | +1 :green_heart: | javadoc | 1m 9s | | trunk passed with JDK Private Build-1.8.0_382-8u382-ga-1~20.04.1-b05 | | +1 :green_heart: | spotbugs | 1m 56s | | trunk passed | | +1 :green_heart: | shadedclient | 21m 53s | | branch has no errors when building and testing our client artifacts. | _ Patch Compile Tests _ | | +1 :green_heart: | mvninstall | 0m 44s | | the patch passed | | +1 :green_heart: | compile | 0m 45s | | the patch passed with JDK Ubuntu-11.0.20+8-post-Ubuntu-1ubuntu120.04 | | +1 :green_heart: | javac | 0m 45s | | the patch passed | | +1 :green_heart: | compile | 0m 40s | | the patch passed with JDK Private Build-1.8.0_382-8u382-ga-1~20.04.1-b05 | | +1 :green_heart: | javac | 0m 40s | | the patch passed | | +1 :green_heart: | blanks | 0m 0s | | The patch has no blanks issues. | | +1 :green_heart: | checkstyle | 0m 34s | | the patch passed | | +1 :green_heart: | mvnsite | 0m 42s | | the patch passed | | +1 :green_heart: | javadoc | 0m 38s | | the patch passed with JDK Ubuntu-11.0.20+8-post-Ubuntu-1ubuntu120.04 | | +1 :green_heart: | javadoc | 1m 3s | | the patch passed with JDK Private Build-1.8.0_382-8u382-ga-1~20.04.1-b05 | | +1 :green_heart: | spotbugs | 1m 50s | | the patch passed | | +1 :green_heart: | shadedclient | 21m 44s | | patch has no errors when building and testing our client artifacts. | _ Other Tests _ | | +1 :green_heart: | unit | 192m 34s | | hadoop-hdfs in the patch passed. | | +1 :green_heart: | asflicense | 0m 41s | | The patch does not generate ASF License warnings. | | | | 284m 29s | | | | Subsystem | Report/Notes | |--:|:-| | Docker | ClientAPI=1.43 ServerAPI=1.43 base: https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-6094/2/artifact/out/Dockerfile | | GITHUB PR | https://github.com/apache/hadoop/pull/6094 | | Optional Tests | dupname asflicense compile javac javadoc mvninstall mvnsite unit shadedclient spotbugs checkstyle codespell detsecrets | | uname | Linux c43a5ab2308b 4.15.0-213-generic #224-Ubuntu SMP Mon Jun 19 13:30:12 UTC 2023 x86_64 x86_64 x86_64 GNU/Linux | | Build tool | maven | | Personality | dev-support/bin/hadoop.sh | | git revision | trunk / 040ca48885a1a7965a868fd6bcc2ee0c06f97db5 | | Default Java | Private Build-1.8.0_382-8u382-ga-1~20.04.1-b05 | | Multi-JDK versions | /usr/lib/jvm/java-11-openjdk-amd64:Ubuntu-11.0.20+8-post-Ubuntu-1ubuntu120.04 /usr/lib/jvm/java-8-openjdk-amd64:Private Build-1.8.0_382-8u382-ga-1~20.04.1-b05 | | Test Results | https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-6094/2/testReport/ | | Max. process+thread count | 3552 (vs. ulimit of 5500) | | modules | C: hadoop-hdfs-project/hadoop-hdfs U: hadoop-hdfs-project/hadoop-hdfs | | Console output | https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-6094/2/console | | versions | git=2.25.1 maven=3.6.3 spotbugs=4.2.2 | | Powered by | Apache Yetus 0.14.0 https://yetus.apache.org |
[jira] [Commented] (HDFS-17194) Enhance the log message for striped block recovery
[ https://issues.apache.org/jira/browse/HDFS-17194?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17766073#comment-17766073 ] ASF GitHub Bot commented on HDFS-17194: --- hadoop-yetus commented on PR #6094: URL: https://github.com/apache/hadoop/pull/6094#issuecomment-1722453433 :broken_heart: **-1 overall** | Vote | Subsystem | Runtime | Logfile | Comment | |::|--:|:|::|:---:| | +0 :ok: | reexec | 12m 51s | | Docker mode activated. | _ Prechecks _ | | +1 :green_heart: | dupname | 0m 0s | | No case conflicting files found. | | +0 :ok: | codespell | 0m 1s | | codespell was not available. | | +0 :ok: | detsecrets | 0m 1s | | detect-secrets was not available. | | +1 :green_heart: | @author | 0m 0s | | The patch does not contain any @author tags. | | -1 :x: | test4tests | 0m 0s | | The patch doesn't appear to include any new or modified tests. Please justify why no new tests are needed for this patch. Also please list what manual steps were performed to verify this patch. | _ trunk Compile Tests _ | | +1 :green_heart: | mvninstall | 43m 47s | | trunk passed | | +1 :green_heart: | compile | 1m 30s | | trunk passed with JDK Ubuntu-11.0.20.1+1-post-Ubuntu-0ubuntu120.04 | | +1 :green_heart: | compile | 1m 21s | | trunk passed with JDK Private Build-1.8.0_382-8u382-ga-1~20.04.1-b05 | | +1 :green_heart: | checkstyle | 1m 12s | | trunk passed | | +1 :green_heart: | mvnsite | 1m 29s | | trunk passed | | +1 :green_heart: | javadoc | 1m 12s | | trunk passed with JDK Ubuntu-11.0.20.1+1-post-Ubuntu-0ubuntu120.04 | | +1 :green_heart: | javadoc | 1m 41s | | trunk passed with JDK Private Build-1.8.0_382-8u382-ga-1~20.04.1-b05 | | +1 :green_heart: | spotbugs | 3m 19s | | trunk passed | | +1 :green_heart: | shadedclient | 36m 15s | | branch has no errors when building and testing our client artifacts. | _ Patch Compile Tests _ | | +1 :green_heart: | mvninstall | 1m 14s | | the patch passed | | +1 :green_heart: | compile | 1m 16s | | the patch passed with JDK Ubuntu-11.0.20.1+1-post-Ubuntu-0ubuntu120.04 | | +1 :green_heart: | javac | 1m 16s | | the patch passed | | +1 :green_heart: | compile | 1m 10s | | the patch passed with JDK Private Build-1.8.0_382-8u382-ga-1~20.04.1-b05 | | +1 :green_heart: | javac | 1m 10s | | the patch passed | | +1 :green_heart: | blanks | 0m 1s | | The patch has no blanks issues. | | +1 :green_heart: | checkstyle | 0m 59s | | the patch passed | | +1 :green_heart: | mvnsite | 1m 16s | | the patch passed | | +1 :green_heart: | javadoc | 0m 57s | | the patch passed with JDK Ubuntu-11.0.20.1+1-post-Ubuntu-0ubuntu120.04 | | +1 :green_heart: | javadoc | 1m 29s | | the patch passed with JDK Private Build-1.8.0_382-8u382-ga-1~20.04.1-b05 | | +1 :green_heart: | spotbugs | 3m 11s | | the patch passed | | +1 :green_heart: | shadedclient | 35m 51s | | patch has no errors when building and testing our client artifacts. | _ Other Tests _ | | -1 :x: | unit | 224m 41s | [/patch-unit-hadoop-hdfs-project_hadoop-hdfs.txt](https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-6094/1/artifact/out/patch-unit-hadoop-hdfs-project_hadoop-hdfs.txt) | hadoop-hdfs in the patch passed. | | +1 :green_heart: | asflicense | 0m 57s | | The patch does not generate ASF License warnings. | | | | 378m 48s | | | | Reason | Tests | |---:|:--| | Failed junit tests | hadoop.hdfs.server.datanode.TestDirectoryScanner | | Subsystem | Report/Notes | |--:|:-| | Docker | ClientAPI=1.43 ServerAPI=1.43 base: https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-6094/1/artifact/out/Dockerfile | | GITHUB PR | https://github.com/apache/hadoop/pull/6094 | | Optional Tests | dupname asflicense compile javac javadoc mvninstall mvnsite unit shadedclient spotbugs checkstyle codespell detsecrets | | uname | Linux de2d5fc27e47 4.15.0-213-generic #224-Ubuntu SMP Mon Jun 19 13:30:12 UTC 2023 x86_64 x86_64 x86_64 GNU/Linux | | Build tool | maven | | Personality | dev-support/bin/hadoop.sh | | git revision | trunk / bdc7fbe8d5245afb7b94fc3050c74f2d83824dca | | Default Java | Private Build-1.8.0_382-8u382-ga-1~20.04.1-b05 | | Multi-JDK versions | /usr/lib/jvm/java-11-openjdk-amd64:Ubuntu-11.0.20.1+1-post-Ubuntu-0ubuntu120.04 /usr/lib/jvm/java-8-openjdk-amd64:Private Build-1.8.0_382-8u382-ga-1~20.04.1-b05 | | Test Results | https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-6094/1/testReport/ | | Max. process+thread count | 3442 (vs. ulimit of 5500) | |
[jira] [Commented] (HDFS-17197) Show file replication when listing corrupt files.
[ https://issues.apache.org/jira/browse/HDFS-17197?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17766101#comment-17766101 ] ASF GitHub Bot commented on HDFS-17197: --- hadoop-yetus commented on PR #6095: URL: https://github.com/apache/hadoop/pull/6095#issuecomment-1722504685 :broken_heart: **-1 overall** | Vote | Subsystem | Runtime | Logfile | Comment | |::|--:|:|::|:---:| | +0 :ok: | reexec | 0m 37s | | Docker mode activated. | _ Prechecks _ | | +1 :green_heart: | dupname | 0m 0s | | No case conflicting files found. | | +0 :ok: | codespell | 0m 1s | | codespell was not available. | | +0 :ok: | detsecrets | 0m 1s | | detect-secrets was not available. | | +1 :green_heart: | @author | 0m 0s | | The patch does not contain any @author tags. | | -1 :x: | test4tests | 0m 0s | | The patch doesn't appear to include any new or modified tests. Please justify why no new tests are needed for this patch. Also please list what manual steps were performed to verify this patch. | _ trunk Compile Tests _ | | +1 :green_heart: | mvninstall | 44m 7s | | trunk passed | | +1 :green_heart: | compile | 1m 24s | | trunk passed with JDK Ubuntu-11.0.20.1+1-post-Ubuntu-0ubuntu120.04 | | +1 :green_heart: | compile | 1m 20s | | trunk passed with JDK Private Build-1.8.0_382-8u382-ga-1~20.04.1-b05 | | +1 :green_heart: | checkstyle | 1m 14s | | trunk passed | | +1 :green_heart: | mvnsite | 1m 29s | | trunk passed | | +1 :green_heart: | javadoc | 1m 10s | | trunk passed with JDK Ubuntu-11.0.20.1+1-post-Ubuntu-0ubuntu120.04 | | +1 :green_heart: | javadoc | 1m 41s | | trunk passed with JDK Private Build-1.8.0_382-8u382-ga-1~20.04.1-b05 | | +1 :green_heart: | spotbugs | 3m 21s | | trunk passed | | +1 :green_heart: | shadedclient | 35m 54s | | branch has no errors when building and testing our client artifacts. | _ Patch Compile Tests _ | | +1 :green_heart: | mvninstall | 1m 12s | | the patch passed | | +1 :green_heart: | compile | 1m 15s | | the patch passed with JDK Ubuntu-11.0.20.1+1-post-Ubuntu-0ubuntu120.04 | | +1 :green_heart: | javac | 1m 15s | | the patch passed | | +1 :green_heart: | compile | 1m 9s | | the patch passed with JDK Private Build-1.8.0_382-8u382-ga-1~20.04.1-b05 | | +1 :green_heart: | javac | 1m 9s | | the patch passed | | +1 :green_heart: | blanks | 0m 0s | | The patch has no blanks issues. | | -0 :warning: | checkstyle | 1m 1s | [/results-checkstyle-hadoop-hdfs-project_hadoop-hdfs.txt](https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-6095/1/artifact/out/results-checkstyle-hadoop-hdfs-project_hadoop-hdfs.txt) | hadoop-hdfs-project/hadoop-hdfs: The patch generated 2 new + 110 unchanged - 1 fixed = 112 total (was 111) | | +1 :green_heart: | mvnsite | 1m 17s | | the patch passed | | +1 :green_heart: | javadoc | 0m 57s | | the patch passed with JDK Ubuntu-11.0.20.1+1-post-Ubuntu-0ubuntu120.04 | | +1 :green_heart: | javadoc | 1m 31s | | the patch passed with JDK Private Build-1.8.0_382-8u382-ga-1~20.04.1-b05 | | +1 :green_heart: | spotbugs | 3m 16s | | the patch passed | | +1 :green_heart: | shadedclient | 36m 14s | | patch has no errors when building and testing our client artifacts. | _ Other Tests _ | | +1 :green_heart: | unit | 218m 41s | | hadoop-hdfs in the patch passed. | | +1 :green_heart: | asflicense | 0m 56s | | The patch does not generate ASF License warnings. | | | | 360m 38s | | | | Subsystem | Report/Notes | |--:|:-| | Docker | ClientAPI=1.43 ServerAPI=1.43 base: https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-6095/1/artifact/out/Dockerfile | | GITHUB PR | https://github.com/apache/hadoop/pull/6095 | | Optional Tests | dupname asflicense compile javac javadoc mvninstall mvnsite unit shadedclient spotbugs checkstyle codespell detsecrets | | uname | Linux 846959e2d4ab 4.15.0-212-generic #223-Ubuntu SMP Tue May 23 13:09:22 UTC 2023 x86_64 x86_64 x86_64 GNU/Linux | | Build tool | maven | | Personality | dev-support/bin/hadoop.sh | | git revision | trunk / b6957d74806004b85674016af809acd72c32c42c | | Default Java | Private Build-1.8.0_382-8u382-ga-1~20.04.1-b05 | | Multi-JDK versions | /usr/lib/jvm/java-11-openjdk-amd64:Ubuntu-11.0.20.1+1-post-Ubuntu-0ubuntu120.04 /usr/lib/jvm/java-8-openjdk-amd64:Private Build-1.8.0_382-8u382-ga-1~20.04.1-b05 | | Test Results | https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-6095/1/testReport/ | | Max. process+thread count | 3337 (vs. ulimit of 5500) | | modules | C: ha
[jira] [Commented] (HDFS-17190) EC: Fix bug of OIV processing XAttr.
[ https://issues.apache.org/jira/browse/HDFS-17190?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17766178#comment-17766178 ] ASF GitHub Bot commented on HDFS-17190: --- Hexiaoqiao merged PR #6067: URL: https://github.com/apache/hadoop/pull/6067 > EC: Fix bug of OIV processing XAttr. > > > Key: HDFS-17190 > URL: https://issues.apache.org/jira/browse/HDFS-17190 > Project: Hadoop HDFS > Issue Type: Bug >Reporter: Shuyan Zhang >Priority: Major > Labels: pull-request-available > > When we need to use OIV to print EC information for a directory, > `PBImageTextWriter#getErasureCodingPolicyName` will be called. Currently, > this method uses `XATTR_ERASURECODING_POLICY.contains(xattr.getName())` to > filter and obtain EC XAttr, which is very dangerous. If we have an XAttr > whose name happens to be a substring of `hdfs.erasurecoding.policy`, then > `getErasureCodingPolicyName` will return the wrong result. Our internal > production environment has customized some XAttrs, and this bug caused errors > in the parsing results of OIV when using `-ec` option. -- This message was sent by Atlassian Jira (v8.20.10#820010) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-17190) EC: Fix bug of OIV processing XAttr.
[ https://issues.apache.org/jira/browse/HDFS-17190?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17766180#comment-17766180 ] ASF GitHub Bot commented on HDFS-17190: --- Hexiaoqiao commented on PR #6067: URL: https://github.com/apache/hadoop/pull/6067#issuecomment-1722678846 Committed to trunk. Thanks @zhangshuyan0 . > EC: Fix bug of OIV processing XAttr. > > > Key: HDFS-17190 > URL: https://issues.apache.org/jira/browse/HDFS-17190 > Project: Hadoop HDFS > Issue Type: Bug >Reporter: Shuyan Zhang >Priority: Major > Labels: pull-request-available > > When we need to use OIV to print EC information for a directory, > `PBImageTextWriter#getErasureCodingPolicyName` will be called. Currently, > this method uses `XATTR_ERASURECODING_POLICY.contains(xattr.getName())` to > filter and obtain EC XAttr, which is very dangerous. If we have an XAttr > whose name happens to be a substring of `hdfs.erasurecoding.policy`, then > `getErasureCodingPolicyName` will return the wrong result. Our internal > production environment has customized some XAttrs, and this bug caused errors > in the parsing results of OIV when using `-ec` option. -- This message was sent by Atlassian Jira (v8.20.10#820010) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Resolved] (HDFS-17190) EC: Fix bug of OIV processing XAttr.
[ https://issues.apache.org/jira/browse/HDFS-17190?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Xiaoqiao He resolved HDFS-17190. Fix Version/s: 3.4.0 Hadoop Flags: Reviewed Assignee: Shuyan Zhang Resolution: Fixed > EC: Fix bug of OIV processing XAttr. > > > Key: HDFS-17190 > URL: https://issues.apache.org/jira/browse/HDFS-17190 > Project: Hadoop HDFS > Issue Type: Bug >Reporter: Shuyan Zhang >Assignee: Shuyan Zhang >Priority: Major > Labels: pull-request-available > Fix For: 3.4.0 > > > When we need to use OIV to print EC information for a directory, > `PBImageTextWriter#getErasureCodingPolicyName` will be called. Currently, > this method uses `XATTR_ERASURECODING_POLICY.contains(xattr.getName())` to > filter and obtain EC XAttr, which is very dangerous. If we have an XAttr > whose name happens to be a substring of `hdfs.erasurecoding.policy`, then > `getErasureCodingPolicyName` will return the wrong result. Our internal > production environment has customized some XAttrs, and this bug caused errors > in the parsing results of OIV when using `-ec` option. -- This message was sent by Atlassian Jira (v8.20.10#820010) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-17138) RBF: We changed the hadoop.security.auth_to_local configuration of one router, the other routers stopped working
[ https://issues.apache.org/jira/browse/HDFS-17138?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17766181#comment-17766181 ] ASF GitHub Bot commented on HDFS-17138: --- zhangxiping1 commented on PR #5921: URL: https://github.com/apache/hadoop/pull/5921#issuecomment-1722681250 ![image](https://github.com/apache/hadoop/assets/40832063/4ab3c82a-2afc-4830-8674-06a1c32c10a7) window 10 environment, the test case runs successfully ,try to build again > RBF: We changed the hadoop.security.auth_to_local configuration of one > router, the other routers stopped working > > > Key: HDFS-17138 > URL: https://issues.apache.org/jira/browse/HDFS-17138 > Project: Hadoop HDFS > Issue Type: Bug > Environment: hadoop 3.3.0 >Reporter: Xiping Zhang >Assignee: Xiping Zhang >Priority: Major > Labels: pull-request-available > Attachments: image-2023-08-02-16-20-34-454.png, > image-2023-08-03-10-32-03-457.png > > > other routers error log: > !image-2023-08-02-16-20-34-454.png! -- This message was sent by Atlassian Jira (v8.20.10#820010) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-17197) Show file replication when listing corrupt files.
[ https://issues.apache.org/jira/browse/HDFS-17197?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17766183#comment-17766183 ] ASF GitHub Bot commented on HDFS-17197: --- Hexiaoqiao commented on code in PR #6095: URL: https://github.com/apache/hadoop/pull/6095#discussion_r1328207545 ## hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSNamesystem.java: ## @@ -6195,7 +6197,11 @@ Collection listCorruptFileBlocks(String path, if (inode != null) { String src = inode.getFullPathName(); if (isParentEntry(src, path)) { -corruptFiles.add(new CorruptFileBlockInfo(src, blk)); +int repl = 0; +if (inode.isFile()) { + repl = inode.asFile().getFileReplication(); Review Comment: What will be return if this inode is one EC file here? ## hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSNamesystem.java: ## @@ -6131,15 +6131,17 @@ void releaseBackupNode(NamenodeRegistration registration) static class CorruptFileBlockInfo { final String path; final Block block; +private final int replication; -public CorruptFileBlockInfo(String p, Block b) { +CorruptFileBlockInfo(String p, Block b, int r) { path = p; block = b; + replication = r; } @Override public String toString() { - return block.getBlockName() + "\t" + path; + return block.getBlockName() + "\t" + replication + "\t" + path; Review Comment: I am a little worried about the compatibility while upstream dependency with this string but we change the format here. > Show file replication when listing corrupt files. > - > > Key: HDFS-17197 > URL: https://issues.apache.org/jira/browse/HDFS-17197 > Project: Hadoop HDFS > Issue Type: Improvement >Reporter: Shuyan Zhang >Priority: Major > Labels: pull-request-available > > Files with different replication have different reliability guarantees. We > need to pay attention to corrupted files with a specified replication greater > than or equal to 3. So, when listing corrupt files, it would be useful to > display the corresponding replication of the files. -- This message was sent by Atlassian Jira (v8.20.10#820010) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-17129) mis-order of ibr and fbr on datanode
[ https://issues.apache.org/jira/browse/HDFS-17129?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17766185#comment-17766185 ] ASF GitHub Bot commented on HDFS-17129: --- LiuGuH commented on PR #5892: URL: https://github.com/apache/hadoop/pull/5892#issuecomment-1722705134 https://github.com/apache/hadoop/pull/6082 is the same code as this. In that is a clean build and Test Results are Successful. > mis-order of ibr and fbr on datanode > - > > Key: HDFS-17129 > URL: https://issues.apache.org/jira/browse/HDFS-17129 > Project: Hadoop HDFS > Issue Type: Bug >Affects Versions: 3.4.0 > Environment: hdfs3.4.0 >Reporter: liuguanghua >Assignee: liuguanghua >Priority: Major > Labels: pull-request-available > > HDFS-16016 , provide new thread to handler IBR. That is a greate improvement. > But it maybe casue the mis-order of ibr and fbr -- This message was sent by Atlassian Jira (v8.20.10#820010) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Created] (HDFS-17198) RBF: fix bug of getRepresentativeQuorum
Jian Zhang created HDFS-17198: - Summary: RBF: fix bug of getRepresentativeQuorum Key: HDFS-17198 URL: https://issues.apache.org/jira/browse/HDFS-17198 Project: Hadoop HDFS Issue Type: Bug Reporter: Jian Zhang h2. *Bug description* In the original implementation, when each router reports nn status at different times, the nn status is the status reported by majority routers, for example: router1 -> nn0:active dateModified:1 router2 -> nn0:active dateModified:2 router3 -> nn0:active dateModified:3 router0 -> nn0:standby dateModified:4 Then, the status of nn0 is active, because majority routers report that nn0 is active. If majority routers report nn status at the same time, for example: (record1) router1 -> nn0:active dateModified:1 (record2) router2 -> nn0:active dateModified:1 (record3) router3 -> nn0:active dateModified:1 (record4) router0 -> nn0:standbydateModified:2 Then the state of nn0 is standby, but We expect the status of nn0 is active This bug is because the above record is put into the Treeset in the method getRepresentativeQuorum. Since record1,2,3 have the same dateModified, there will only be one record in the final treeset of this method, so this method thinks that this nn is standby, because record4 newer h2. *How to reproduce* Running my unit test testRegistrationMajorityQuorumEqDateModified, but using the original code -- This message was sent by Atlassian Jira (v8.20.10#820010) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Assigned] (HDFS-17198) RBF: fix bug of getRepresentativeQuorum
[ https://issues.apache.org/jira/browse/HDFS-17198?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jian Zhang reassigned HDFS-17198: - Assignee: Jian Zhang > RBF: fix bug of getRepresentativeQuorum > --- > > Key: HDFS-17198 > URL: https://issues.apache.org/jira/browse/HDFS-17198 > Project: Hadoop HDFS > Issue Type: Bug >Reporter: Jian Zhang >Assignee: Jian Zhang >Priority: Major > > h2. *Bug description* > In the original implementation, when each router reports nn status at > different times, the nn status is the status reported by majority routers, > for example: > router1 -> nn0:active dateModified:1 > router2 -> nn0:active dateModified:2 > router3 -> nn0:active dateModified:3 > router0 -> nn0:standby dateModified:4 > Then, the status of nn0 is active, because majority routers report that nn0 > is active. > If majority routers report nn status at the same time, for example: > (record1) router1 -> nn0:active dateModified:1 > (record2) router2 -> nn0:active dateModified:1 > (record3) router3 -> nn0:active dateModified:1 > (record4) router0 -> nn0:standbydateModified:2 > Then the state of nn0 is standby, but We expect the status of nn0 is active > This bug is because the above record is put into the Treeset in the method > getRepresentativeQuorum. Since record1,2,3 have the same dateModified, there > will only be one record in the final treeset of this method, so this method > thinks that this nn is standby, because record4 newer > h2. *How to reproduce* > Running my unit test testRegistrationMajorityQuorumEqDateModified, but using > the original code -- This message was sent by Atlassian Jira (v8.20.10#820010) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-17198) RBF: fix bug of getRepresentativeQuorum when records have same dateModified
[ https://issues.apache.org/jira/browse/HDFS-17198?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jian Zhang updated HDFS-17198: -- Summary: RBF: fix bug of getRepresentativeQuorum when records have same dateModified (was: RBF: fix bug of getRepresentativeQuorum) > RBF: fix bug of getRepresentativeQuorum when records have same dateModified > --- > > Key: HDFS-17198 > URL: https://issues.apache.org/jira/browse/HDFS-17198 > Project: Hadoop HDFS > Issue Type: Bug >Reporter: Jian Zhang >Assignee: Jian Zhang >Priority: Major > > h2. *Bug description* > In the original implementation, when each router reports nn status at > different times, the nn status is the status reported by majority routers, > for example: > router1 -> nn0:active dateModified:1 > router2 -> nn0:active dateModified:2 > router3 -> nn0:active dateModified:3 > router0 -> nn0:standby dateModified:4 > Then, the status of nn0 is active, because majority routers report that nn0 > is active. > If majority routers report nn status at the same time, for example: > (record1) router1 -> nn0:active dateModified:1 > (record2) router2 -> nn0:active dateModified:1 > (record3) router3 -> nn0:active dateModified:1 > (record4) router0 -> nn0:standbydateModified:2 > Then the state of nn0 is standby, but We expect the status of nn0 is active > This bug is because the above record is put into the Treeset in the method > getRepresentativeQuorum. Since record1,2,3 have the same dateModified, there > will only be one record in the final treeset of this method, so this method > thinks that this nn is standby, because record4 newer > h2. *How to reproduce* > Running my unit test testRegistrationMajorityQuorumEqDateModified, but using > the original code -- This message was sent by Atlassian Jira (v8.20.10#820010) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-17198) RBF: fix bug of getRepresentativeQuorum when records have same dateModified
[ https://issues.apache.org/jira/browse/HDFS-17198?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jian Zhang updated HDFS-17198: -- Attachment: HDFS-17198.v001.patch Status: Patch Available (was: Open) > RBF: fix bug of getRepresentativeQuorum when records have same dateModified > --- > > Key: HDFS-17198 > URL: https://issues.apache.org/jira/browse/HDFS-17198 > Project: Hadoop HDFS > Issue Type: Bug >Reporter: Jian Zhang >Assignee: Jian Zhang >Priority: Major > Attachments: HDFS-17198.v001.patch > > > h2. *Bug description* > In the original implementation, when each router reports nn status at > different times, the nn status is the status reported by majority routers, > for example: > router1 -> nn0:active dateModified:1 > router2 -> nn0:active dateModified:2 > router3 -> nn0:active dateModified:3 > router0 -> nn0:standby dateModified:4 > Then, the status of nn0 is active, because majority routers report that nn0 > is active. > If majority routers report nn status at the same time, for example: > (record1) router1 -> nn0:active dateModified:1 > (record2) router2 -> nn0:active dateModified:1 > (record3) router3 -> nn0:active dateModified:1 > (record4) router0 -> nn0:standbydateModified:2 > Then the state of nn0 is standby, but We expect the status of nn0 is active > This bug is because the above record is put into the Treeset in the method > getRepresentativeQuorum. Since record1,2,3 have the same dateModified, there > will only be one record in the final treeset of this method, so this method > thinks that this nn is standby, because record4 newer > h2. *How to reproduce* > Running my unit test testRegistrationMajorityQuorumEqDateModified, but using > the original code -- This message was sent by Atlassian Jira (v8.20.10#820010) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-16000) HDFS : Rename performance optimization
[ https://issues.apache.org/jira/browse/HDFS-16000?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17766194#comment-17766194 ] ASF GitHub Bot commented on HDFS-16000: --- Hexiaoqiao commented on code in PR #2964: URL: https://github.com/apache/hadoop/pull/2964#discussion_r1328244599 ## hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSDirectory.java: ## @@ -1468,6 +1475,30 @@ static Collection normalizePaths(Collection paths, return normalized; } + /** + * Get the first Node that sets Quota. + */ + static INode getFirstSetQuotaParentNode(INodesInPath iip) { +for (int i = iip.length() - 1; i > 0; i--) { + INode currNode = iip.getINode(i); + if (currNode == null) { Review Comment: Will it meet null here? if it is not expected, we should throw exception IMO. ## hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSDirRenameOp.java: ## @@ -470,17 +475,53 @@ static RenameResult unprotectedRenameTo(FSDirectory fsd, } } finally { if (undoRemoveSrc) { -tx.restoreSource(); +tx.restoreSource(srcStoragePolicyCounts); } if (undoRemoveDst) { // Rename failed - restore dst -tx.restoreDst(bsps); +tx.restoreDst(bsps, dstStoragePolicyCounts); } } NameNode.stateChangeLog.warn("DIR* FSDirectory.unprotectedRenameTo: " + "failed to rename " + src + " to " + dst); throw new IOException("rename from " + src + " to " + dst + " failed."); } + /* + * Calculate QuotaCounts based on parent directory and storage policy + * 1. If the storage policy of src and dst are different, + * calculate the QuotaCounts of src and dst respectively. + * 2. If all parent nodes of src and dst are not set with Quota, + * there is no need to calculate QuotaCount. + * 3. if parent nodes of src and dst have Quota configured, + * the QuotaCount is calculated once using the storage policy of src. + * */ + private static void computeQuotaCounts( + QuotaCounts srcStoragePolicyCounts, + QuotaCounts dstStoragePolicyCounts, + INodesInPath srcIIP, + INodesInPath dstIIP, + BlockStoragePolicySuite bsps, + RenameOperation tx) { +INode dstParent = dstIIP.getINode(-2); +INode srcParentNode = FSDirectory. +getFirstSetQuotaParentNode(srcIIP); +INode srcInode = srcIIP.getLastINode(); +INode dstParentNode = FSDirectory. +getFirstSetQuotaParentNode(dstIIP); +byte srcStoragePolicyID = FSDirectory.getStoragePolicyId(srcInode); +byte dstStoragePolicyID = FSDirectory.getStoragePolicyId(dstParent); +if (srcStoragePolicyID != dstStoragePolicyID) { + srcStoragePolicyCounts.add(srcIIP.getLastINode(). + computeQuotaUsage(bsps)); + dstStoragePolicyCounts.add(srcIIP.getLastINode() Review Comment: IIUC, this result will be used for the next verify and storage used addition/subtraction for src and dst inode, right? But I am confused if it will meet some issues here, given directory /a/b (whose storage policy is HDD), /c/d (whose storage policy is SSD), when rename from /a/b/r1 (let's assume 1GB used) to /c/d/r2, then total HDD storage used will decrease 1GB, and SSD storage used increase 1GB, this will be different with fact? ## hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSDirRenameOp.java: ## @@ -470,17 +475,53 @@ static RenameResult unprotectedRenameTo(FSDirectory fsd, } } finally { if (undoRemoveSrc) { -tx.restoreSource(); +tx.restoreSource(srcStoragePolicyCounts); } if (undoRemoveDst) { // Rename failed - restore dst -tx.restoreDst(bsps); +tx.restoreDst(bsps, dstStoragePolicyCounts); } } NameNode.stateChangeLog.warn("DIR* FSDirectory.unprotectedRenameTo: " + "failed to rename " + src + " to " + dst); throw new IOException("rename from " + src + " to " + dst + " failed."); } + /* + * Calculate QuotaCounts based on parent directory and storage policy + * 1. If the storage policy of src and dst are different, + * calculate the QuotaCounts of src and dst respectively. + * 2. If all parent nodes of src and dst are not set with Quota, + * there is no need to calculate QuotaCount. + * 3. if parent nodes of src and dst have Quota configured, + * the QuotaCount is calculated once using the storage policy of src. + * */ + private static void computeQuotaCounts( + QuotaCounts srcStoragePolicyCounts, + QuotaCounts dstStoragePolicyCounts, + INodesInPath srcIIP, + INodesInPath dstIIP, + BlockStoragePolicySuite bsps, + RenameOperation tx) { +INode dstParent = dstIIP.getINode(-2); +