[ https://issues.apache.org/jira/browse/HDFS-17568?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17868923#comment-17868923 ]
ASF GitHub Bot commented on HDFS-17568: --------------------------------------- hadoop-yetus commented on PR #6921: URL: https://github.com/apache/hadoop/pull/6921#issuecomment-2252531489 :broken_heart: **-1 overall** | Vote | Subsystem | Runtime | Logfile | Comment | |:----:|----------:|--------:|:--------:|:-------:| | +0 :ok: | reexec | 0m 54s | | Docker mode activated. | |||| _ Prechecks _ | | +1 :green_heart: | dupname | 0m 0s | | No case conflicting files found. | | +0 :ok: | codespell | 0m 0s | | codespell was not available. | | +0 :ok: | detsecrets | 0m 0s | | detect-secrets was not available. | | +1 :green_heart: | @author | 0m 0s | | The patch does not contain any @author tags. | | +1 :green_heart: | test4tests | 0m 0s | | The patch appears to include 1 new or modified test files. | |||| _ trunk Compile Tests _ | | +1 :green_heart: | mvninstall | 49m 3s | | trunk passed | | +1 :green_heart: | compile | 1m 27s | | trunk passed with JDK Ubuntu-11.0.23+9-post-Ubuntu-1ubuntu120.04.2 | | +1 :green_heart: | compile | 1m 10s | | trunk passed with JDK Private Build-1.8.0_412-8u412-ga-1~20.04.1-b08 | | +1 :green_heart: | checkstyle | 1m 12s | | trunk passed | | +1 :green_heart: | mvnsite | 1m 21s | | trunk passed | | +1 :green_heart: | javadoc | 1m 8s | | trunk passed with JDK Ubuntu-11.0.23+9-post-Ubuntu-1ubuntu120.04.2 | | +1 :green_heart: | javadoc | 1m 40s | | trunk passed with JDK Private Build-1.8.0_412-8u412-ga-1~20.04.1-b08 | | +1 :green_heart: | spotbugs | 3m 30s | | trunk passed | | +1 :green_heart: | shadedclient | 40m 27s | | branch has no errors when building and testing our client artifacts. | | -0 :warning: | patch | 40m 48s | | Used diff version of patch file. Binary files and potentially other changes not applied. Please rebase and squash commits if necessary. | |||| _ Patch Compile Tests _ | | +1 :green_heart: | mvninstall | 1m 10s | | the patch passed | | +1 :green_heart: | compile | 1m 19s | | the patch passed with JDK Ubuntu-11.0.23+9-post-Ubuntu-1ubuntu120.04.2 | | +1 :green_heart: | javac | 1m 19s | | the patch passed | | +1 :green_heart: | compile | 1m 11s | | the patch passed with JDK Private Build-1.8.0_412-8u412-ga-1~20.04.1-b08 | | +1 :green_heart: | javac | 1m 11s | | the patch passed | | +1 :green_heart: | blanks | 0m 0s | | The patch has no blanks issues. | | -0 :warning: | checkstyle | 1m 5s | [/results-checkstyle-hadoop-hdfs-project_hadoop-hdfs.txt](https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-6921/9/artifact/out/results-checkstyle-hadoop-hdfs-project_hadoop-hdfs.txt) | hadoop-hdfs-project/hadoop-hdfs: The patch generated 78 new + 115 unchanged - 0 fixed = 193 total (was 115) | | +1 :green_heart: | mvnsite | 1m 17s | | the patch passed | | +1 :green_heart: | javadoc | 0m 59s | | the patch passed with JDK Ubuntu-11.0.23+9-post-Ubuntu-1ubuntu120.04.2 | | +1 :green_heart: | javadoc | 1m 41s | | the patch passed with JDK Private Build-1.8.0_412-8u412-ga-1~20.04.1-b08 | | +1 :green_heart: | spotbugs | 3m 32s | | the patch passed | | +1 :green_heart: | shadedclient | 41m 20s | | patch has no errors when building and testing our client artifacts. | |||| _ Other Tests _ | | -1 :x: | unit | 264m 30s | [/patch-unit-hadoop-hdfs-project_hadoop-hdfs.txt](https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-6921/9/artifact/out/patch-unit-hadoop-hdfs-project_hadoop-hdfs.txt) | hadoop-hdfs in the patch passed. | | +1 :green_heart: | asflicense | 0m 54s | | The patch does not generate ASF License warnings. | | | | 418m 19s | | | | Reason | Tests | |-------:|:------| | Failed junit tests | hadoop.hdfs.server.datanode.TestDataNodeMetrics | | Subsystem | Report/Notes | |----------:|:-------------| | Docker | ClientAPI=1.44 ServerAPI=1.44 base: https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-6921/9/artifact/out/Dockerfile | | GITHUB PR | https://github.com/apache/hadoop/pull/6921 | | Optional Tests | dupname asflicense compile javac javadoc mvninstall mvnsite unit shadedclient spotbugs checkstyle codespell detsecrets | | uname | Linux 360ba56ffa1a 5.15.0-94-generic #104-Ubuntu SMP Tue Jan 9 15:25:40 UTC 2024 x86_64 x86_64 x86_64 GNU/Linux | | Build tool | maven | | Personality | dev-support/bin/hadoop.sh | | git revision | trunk / 9a63bfa0ee6c97100f531d03cddb4f8b27dc5707 | | Default Java | Private Build-1.8.0_412-8u412-ga-1~20.04.1-b08 | | Multi-JDK versions | /usr/lib/jvm/java-11-openjdk-amd64:Ubuntu-11.0.23+9-post-Ubuntu-1ubuntu120.04.2 /usr/lib/jvm/java-8-openjdk-amd64:Private Build-1.8.0_412-8u412-ga-1~20.04.1-b08 | | Test Results | https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-6921/9/testReport/ | | Max. process+thread count | 3040 (vs. ulimit of 5500) | | modules | C: hadoop-hdfs-project/hadoop-hdfs U: hadoop-hdfs-project/hadoop-hdfs | | Console output | https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-6921/9/console | | versions | git=2.25.1 maven=3.6.3 spotbugs=4.2.2 | | Powered by | Apache Yetus 0.14.0 https://yetus.apache.org | This message was automatically generated. > [Decommission]Show Aggregated Reason for Why Low-Redundancy Block is Skipped > for Reconstruction > ----------------------------------------------------------------------------------------------- > > Key: HDFS-17568 > URL: https://issues.apache.org/jira/browse/HDFS-17568 > Project: Hadoop HDFS > Issue Type: Improvement > Reporter: wuchang > Priority: Major > Labels: pull-request-available > > The troubleshooting for decommission is very hard even in DEBUG mode. > In somehow cases when decommission has taken a lot of time but we are not > sure whether it is still in-progress, we run {{-refreshNodes}} to try to > re-trigger the decommission(In fact it is useless). > Then we check the NameNode logs, but it is a pity that we cannot find any > useful log about whether or not our {{refreshNodes}} subcommand has taken any > effect or not. > So, my change is: > * I changed this critical log 's level from TRACE to INFO since this log is > not a repeatedly log and it gives administrator critical information for > about what happened. > {code:java} > } else { > LOG.info("startDecommission: Node {} in {}, nothing to do.", > node, node.getAdminState()); > }{code} > * When the Reconstruction which is trigged by node decommission is skipped, > we want to know the reason. The reason is divided to 3 catagories > ## No source node is available > ## No Target node is available > ## ReconstructionWork is built but validation failed > I put these reasons in a single enum {{{}ReconstructionSkipReason{}}}. > In DEBUG mode, the reason will be aggregated and showed to users. > The log aggregated by {{ReconstructionSkipReason}} is as below: > {code:java} > 2024-07-10 02:59:09,707 [Thread-990] DEBUG blockmanagement.BlockManager: > Block blk_3_0 is not scheduled for reconstruction since: [ source node or > storage unavailable on node [DISK]storageID_0_3:NORMAL:127.0.0.1:9866. Detail > : [stored replica state is corrupt or excess] source node or storage > unavailable on node [DISK]storageID_1_3:NORMAL:127.0.0.1:9866. Detail : > [replica is already decommissioned] ]{code} -- This message was sent by Atlassian Jira (v8.20.10#820010) --------------------------------------------------------------------- To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org