[jira] [Updated] (HDFS-16614) Improve balancer operation strategy and performance

2022-06-01 Thread JiangHua Zhu (Jira)


 [ 
https://issues.apache.org/jira/browse/HDFS-16614?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

JiangHua Zhu updated HDFS-16614:

Description: 
When the Balancer program is run, it does some work in the following order:
1. Obtain available datanode information from NameNode.
2. Classify and calculate the average utilization according to StorageType. 
Here, some sets will be obtained in combination with the set thresholds: 
overUtilized, aboveAvgUtilized, belowAvgUtilized, and underUtilized.
3. According to some calculations, the source and target related to the 
transfer data are obtained. The source is used for the source end, and the 
target is used for the data receiving end.
4. Start the data transfer work in parallel.
In this process, run iteratively. In this process, the threshold is unified and 
applied to all StorageTypes, which seems to be a bit rough, because one of the 
StorageTypes cannot be distinguished, which is based on the currently supported 
heterogeneous storage.

There is an online cluster with more than 2000 nodes, and there is an imbalance 
in node storage. E.g:
 !image-2022-06-02-13-18-33-213.png! 

Here, the average utilization of the cluster is 78%, but the utilization of 
most nodes is between 85% and 90%. When the balancer is turned on, we find that 
85% of the nodes are working as sources. In this case, we think it is not 
reasonable, because it will occupy more network resources in the cluster, and 
it will be beneficial to the normal work of the cluster to do some effective 
restrictions.
So here are some changes to make:
1. When the balancer is running, we should actively prompt the suggested value 
of the threshold related to StorageType. For example: [[DISK, 10%], [SSD, 
8%]...]
2. Support to set threshold according to StorageType and work.
3. Add an option to prohibit nodes below the threshold from joining the Source 
set. This is to allow nodes with high utilization to transfer data as soon as 
possible, which is good for balance.
4. Add new support. If there are a lot of datanode usage in the cluster, it 
should remain unchanged. For example, the utilization rate of 40% of the nodes 
in the cluster is 75% to 80%, and these nodes should not join the Source set. 
Of course this support needs to be specified by the user at runtime.

  was:
When the Balancer program is run, it does some work in the following order:
1. Obtain available datanode information from NameNode.
2. Classify and calculate the average utilization according to StorageType. 
Here, some sets will be obtained in combination with the set thresholds: 
overUtilized, aboveAvgUtilized, belowAvgUtilized, and underUtilized.
3. According to some calculations, the source and target related to the 
transfer data are obtained. The source is used for the source end, and the 
target is used for the data receiving end.
4. Start the data transfer work in parallel.
In this process, run iteratively. In this process, the threshold is unified and 
applied to all StorageTypes, which seems to be a bit rough, because one of the 
StorageTypes cannot be distinguished, which is based on the currently supported 
heterogeneous storage.

There is an online cluster with more than 2000 nodes, and there is an imbalance 
in node storage. E.g:
 !image-2022-06-02-13-18-33-213.png! 

Here, the average utilization of the cluster is 78%, but the utilization of 
most nodes is between 85% and 90%. When the balancer is turned on, we find that 
85% of the nodes are working as sources. In this case, we think it is not 
reasonable, because it will occupy more network resources in the cluster, and 
it will be beneficial to the normal work of the cluster to do some effective 
restrictions.
So here are some changes to make:
1. When the balancer is running, it should try to prompt the threshold related 
to StorageType. For example [[DISK, 10%], [SSD, 8%]...]
2. Support to set threshold according to StorageType and work.
3. Add an option to prohibit nodes below the threshold from joining the Source 
set. This is to allow nodes with high utilization to transfer data as soon as 
possible, which is good for balance.
4. Add new support. If there are a lot of datanode usage in the cluster, it 
should remain unchanged. For example, the utilization rate of 40% of the nodes 
in the cluster is 75% to 80%, and these nodes should not join the Source set. 
Of course this support needs to be specified by the user at runtime.


> Improve balancer operation strategy and performance
> ---
>
> Key: HDFS-16614
> URL: https://issues.apache.org/jira/browse/HDFS-16614
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: balancer & mover, namenode
>Affects Versions: 2.9.2
>Reporter: JiangHua Zhu
>Assignee: JiangHua Zhu
>Priority: Major
> Att

[jira] [Updated] (HDFS-16614) Improve balancer operation strategy and performance

2022-06-01 Thread JiangHua Zhu (Jira)


 [ 
https://issues.apache.org/jira/browse/HDFS-16614?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

JiangHua Zhu updated HDFS-16614:

Affects Version/s: 2.9.2
   (was: 3.3.0)

> Improve balancer operation strategy and performance
> ---
>
> Key: HDFS-16614
> URL: https://issues.apache.org/jira/browse/HDFS-16614
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: balancer & mover, namenode
>Affects Versions: 2.9.2
>Reporter: JiangHua Zhu
>Assignee: JiangHua Zhu
>Priority: Major
> Attachments: image-2022-06-02-13-18-33-213.png
>
>
> When the Balancer program is run, it does some work in the following order:
> 1. Obtain available datanode information from NameNode.
> 2. Classify and calculate the average utilization according to StorageType. 
> Here, some sets will be obtained in combination with the set thresholds: 
> overUtilized, aboveAvgUtilized, belowAvgUtilized, and underUtilized.
> 3. According to some calculations, the source and target related to the 
> transfer data are obtained. The source is used for the source end, and the 
> target is used for the data receiving end.
> 4. Start the data transfer work in parallel.
> In this process, run iteratively. In this process, the threshold is unified 
> and applied to all StorageTypes, which seems to be a bit rough, because one 
> of the StorageTypes cannot be distinguished, which is based on the currently 
> supported heterogeneous storage.
> There is an online cluster with more than 2000 nodes, and there is an 
> imbalance in node storage. E.g:
>  !image-2022-06-02-13-18-33-213.png! 
> Here, the average utilization of the cluster is 78%, but the utilization of 
> most nodes is between 85% and 90%. When the balancer is turned on, we find 
> that 85% of the nodes are working as sources. In this case, we think it is 
> not reasonable, because it will occupy more network resources in the cluster, 
> and it will be beneficial to the normal work of the cluster to do some 
> effective restrictions.
> So here are some changes to make:
> 1. When the balancer is running, it should try to prompt the threshold 
> related to StorageType. For example [[DISK, 10%], [SSD, 8%]...]
> 2. Support to set threshold according to StorageType and work.
> 3. Add an option to prohibit nodes below the threshold from joining the 
> Source set. This is to allow nodes with high utilization to transfer data as 
> soon as possible, which is good for balance.
> 4. Add new support. If there are a lot of datanode usage in the cluster, it 
> should remain unchanged. For example, the utilization rate of 40% of the 
> nodes in the cluster is 75% to 80%, and these nodes should not join the 
> Source set. Of course this support needs to be specified by the user at 
> runtime.



--
This message was sent by Atlassian Jira
(v8.20.7#820007)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Created] (HDFS-16614) Improve balancer operation strategy and performance

2022-06-01 Thread JiangHua Zhu (Jira)
JiangHua Zhu created HDFS-16614:
---

 Summary: Improve balancer operation strategy and performance
 Key: HDFS-16614
 URL: https://issues.apache.org/jira/browse/HDFS-16614
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: balancer & mover, namenode
Affects Versions: 3.3.0
Reporter: JiangHua Zhu
 Attachments: image-2022-06-02-13-18-33-213.png

When the Balancer program is run, it does some work in the following order:
1. Obtain available datanode information from NameNode.
2. Classify and calculate the average utilization according to StorageType. 
Here, some sets will be obtained in combination with the set thresholds: 
overUtilized, aboveAvgUtilized, belowAvgUtilized, and underUtilized.
3. According to some calculations, the source and target related to the 
transfer data are obtained. The source is used for the source end, and the 
target is used for the data receiving end.
4. Start the data transfer work in parallel.
In this process, run iteratively. In this process, the threshold is unified and 
applied to all StorageTypes, which seems to be a bit rough, because one of the 
StorageTypes cannot be distinguished, which is based on the currently supported 
heterogeneous storage.

There is an online cluster with more than 2000 nodes, and there is an imbalance 
in node storage. E.g:
 !image-2022-06-02-13-18-33-213.png! 

Here, the average utilization of the cluster is 78%, but the utilization of 
most nodes is between 85% and 90%. When the balancer is turned on, we find that 
85% of the nodes are working as sources. In this case, we think it is not 
reasonable, because it will occupy more network resources in the cluster, and 
it will be beneficial to the normal work of the cluster to do some effective 
restrictions.
So here are some changes to make:
1. When the balancer is running, it should try to prompt the threshold related 
to StorageType. For example [[DISK, 10%], [SSD, 8%]...]
2. Support to set threshold according to StorageType and work.
3. Add an option to prohibit nodes below the threshold from joining the Source 
set. This is to allow nodes with high utilization to transfer data as soon as 
possible, which is good for balance.
4. Add new support. If there are a lot of datanode usage in the cluster, it 
should remain unchanged. For example, the utilization rate of 40% of the nodes 
in the cluster is 75% to 80%, and these nodes should not join the Source set. 
Of course this support needs to be specified by the user at runtime.



--
This message was sent by Atlassian Jira
(v8.20.7#820007)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Assigned] (HDFS-16614) Improve balancer operation strategy and performance

2022-06-01 Thread JiangHua Zhu (Jira)


 [ 
https://issues.apache.org/jira/browse/HDFS-16614?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

JiangHua Zhu reassigned HDFS-16614:
---

Assignee: JiangHua Zhu

> Improve balancer operation strategy and performance
> ---
>
> Key: HDFS-16614
> URL: https://issues.apache.org/jira/browse/HDFS-16614
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: balancer & mover, namenode
>Affects Versions: 3.3.0
>Reporter: JiangHua Zhu
>Assignee: JiangHua Zhu
>Priority: Major
> Attachments: image-2022-06-02-13-18-33-213.png
>
>
> When the Balancer program is run, it does some work in the following order:
> 1. Obtain available datanode information from NameNode.
> 2. Classify and calculate the average utilization according to StorageType. 
> Here, some sets will be obtained in combination with the set thresholds: 
> overUtilized, aboveAvgUtilized, belowAvgUtilized, and underUtilized.
> 3. According to some calculations, the source and target related to the 
> transfer data are obtained. The source is used for the source end, and the 
> target is used for the data receiving end.
> 4. Start the data transfer work in parallel.
> In this process, run iteratively. In this process, the threshold is unified 
> and applied to all StorageTypes, which seems to be a bit rough, because one 
> of the StorageTypes cannot be distinguished, which is based on the currently 
> supported heterogeneous storage.
> There is an online cluster with more than 2000 nodes, and there is an 
> imbalance in node storage. E.g:
>  !image-2022-06-02-13-18-33-213.png! 
> Here, the average utilization of the cluster is 78%, but the utilization of 
> most nodes is between 85% and 90%. When the balancer is turned on, we find 
> that 85% of the nodes are working as sources. In this case, we think it is 
> not reasonable, because it will occupy more network resources in the cluster, 
> and it will be beneficial to the normal work of the cluster to do some 
> effective restrictions.
> So here are some changes to make:
> 1. When the balancer is running, it should try to prompt the threshold 
> related to StorageType. For example [[DISK, 10%], [SSD, 8%]...]
> 2. Support to set threshold according to StorageType and work.
> 3. Add an option to prohibit nodes below the threshold from joining the 
> Source set. This is to allow nodes with high utilization to transfer data as 
> soon as possible, which is good for balance.
> 4. Add new support. If there are a lot of datanode usage in the cluster, it 
> should remain unchanged. For example, the utilization rate of 40% of the 
> nodes in the cluster is 75% to 80%, and these nodes should not join the 
> Source set. Of course this support needs to be specified by the user at 
> runtime.



--
This message was sent by Atlassian Jira
(v8.20.7#820007)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Work logged] (HDFS-16595) Slow peer metrics - add median, mad and upper latency limits

2022-06-01 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HDFS-16595?focusedWorklogId=777232&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-777232
 ]

ASF GitHub Bot logged work on HDFS-16595:
-

Author: ASF GitHub Bot
Created on: 02/Jun/22 01:20
Start Date: 02/Jun/22 01:20
Worklog Time Spent: 10m 
  Work Description: virajjasani commented on code in PR #4357:
URL: https://github.com/apache/hadoop/pull/4357#discussion_r887419980


##
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/protocolPB/PBHelper.java:
##
@@ -853,11 +854,14 @@ public static List 
convertSlowPeerInfo(
 
 List slowPeerInfoProtos =
 new ArrayList<>(slowPeers.getSlowPeers().size());
-for (Map.Entry entry :
-slowPeers.getSlowPeers().entrySet()) {
-  slowPeerInfoProtos.add(SlowPeerReportProto.newBuilder()
+for (Map.Entry entry : 
slowPeers.getSlowPeers().entrySet()) {
+  OutlierMetrics outlierMetrics = entry.getValue();
+  slowPeerInfoProtos.add(
+  SlowPeerReportProto.newBuilder()
   .setDataNodeId(entry.getKey())
-  .setAggregateLatency(entry.getValue())
+  .setAggregateLatency(outlierMetrics.getActualLatency())
+  
.setMedian(outlierMetrics.getMedian()).setMad(outlierMetrics.getMad())

Review Comment:
   Done, thanks for the suggestion.





Issue Time Tracking
---

Worklog Id: (was: 777232)
Time Spent: 2.5h  (was: 2h 20m)

> Slow peer metrics - add median, mad and upper latency limits
> 
>
> Key: HDFS-16595
> URL: https://issues.apache.org/jira/browse/HDFS-16595
> Project: Hadoop HDFS
>  Issue Type: New Feature
>Reporter: Viraj Jasani
>Assignee: Viraj Jasani
>Priority: Major
>  Labels: pull-request-available
>  Time Spent: 2.5h
>  Remaining Estimate: 0h
>
> Slow datanode metrics include slow node and it's reporting node details. With 
> HDFS-16582, we added the aggregate latency that is perceived by the reporting 
> nodes.
> In order to get more insights into how the outlier slownode's latencies 
> differ from the rest of the nodes, we should also expose median, median 
> absolute deviation and the calculated upper latency limit details.



--
This message was sent by Atlassian Jira
(v8.20.7#820007)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Work logged] (HDFS-16595) Slow peer metrics - add median, mad and upper latency limits

2022-06-01 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HDFS-16595?focusedWorklogId=777224&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-777224
 ]

ASF GitHub Bot logged work on HDFS-16595:
-

Author: ASF GitHub Bot
Created on: 02/Jun/22 00:33
Start Date: 02/Jun/22 00:33
Worklog Time Spent: 10m 
  Work Description: jojochuang commented on code in PR #4357:
URL: https://github.com/apache/hadoop/pull/4357#discussion_r887401881


##
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/protocolPB/PBHelper.java:
##
@@ -853,11 +854,14 @@ public static List 
convertSlowPeerInfo(
 
 List slowPeerInfoProtos =
 new ArrayList<>(slowPeers.getSlowPeers().size());
-for (Map.Entry entry :
-slowPeers.getSlowPeers().entrySet()) {
-  slowPeerInfoProtos.add(SlowPeerReportProto.newBuilder()
+for (Map.Entry entry : 
slowPeers.getSlowPeers().entrySet()) {
+  OutlierMetrics outlierMetrics = entry.getValue();
+  slowPeerInfoProtos.add(
+  SlowPeerReportProto.newBuilder()
   .setDataNodeId(entry.getKey())
-  .setAggregateLatency(entry.getValue())
+  .setAggregateLatency(outlierMetrics.getActualLatency())
+  
.setMedian(outlierMetrics.getMedian()).setMad(outlierMetrics.getMad())

Review Comment:
   move .setMad to its own line?
   ```suggestion
 .setMedian(outlierMetrics.getMedian())
 .setMad(outlierMetrics.getMad())
   ```





Issue Time Tracking
---

Worklog Id: (was: 777224)
Time Spent: 2h 20m  (was: 2h 10m)

> Slow peer metrics - add median, mad and upper latency limits
> 
>
> Key: HDFS-16595
> URL: https://issues.apache.org/jira/browse/HDFS-16595
> Project: Hadoop HDFS
>  Issue Type: New Feature
>Reporter: Viraj Jasani
>Assignee: Viraj Jasani
>Priority: Major
>  Labels: pull-request-available
>  Time Spent: 2h 20m
>  Remaining Estimate: 0h
>
> Slow datanode metrics include slow node and it's reporting node details. With 
> HDFS-16582, we added the aggregate latency that is perceived by the reporting 
> nodes.
> In order to get more insights into how the outlier slownode's latencies 
> differ from the rest of the nodes, we should also expose median, median 
> absolute deviation and the calculated upper latency limit details.



--
This message was sent by Atlassian Jira
(v8.20.7#820007)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Work logged] (HDFS-16610) Make fsck read timeout configurable

2022-06-01 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HDFS-16610?focusedWorklogId=777124&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-777124
 ]

ASF GitHub Bot logged work on HDFS-16610:
-

Author: ASF GitHub Bot
Created on: 01/Jun/22 19:36
Start Date: 01/Jun/22 19:36
Worklog Time Spent: 10m 
  Work Description: sodonnel merged PR #4384:
URL: https://github.com/apache/hadoop/pull/4384




Issue Time Tracking
---

Worklog Id: (was: 777124)
Time Spent: 1h 50m  (was: 1h 40m)

> Make fsck read timeout configurable
> ---
>
> Key: HDFS-16610
> URL: https://issues.apache.org/jira/browse/HDFS-16610
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: hdfs-client
>Reporter: Stephen O'Donnell
>Assignee: Stephen O'Donnell
>Priority: Major
>  Labels: pull-request-available
>  Time Spent: 1h 50m
>  Remaining Estimate: 0h
>
> In a cluster with a lot of small files, we encountered a case where fsck was 
> very slow. I believe it is due to contention with many other threads reading 
> / writing data on the cluster.
> Sometimes fsck does not report any progress for more than 60 seconds and the 
> client times out. Currently the connect and read timeout are hardcoded to 60 
> seconds. This change is to make them configurable.



--
This message was sent by Atlassian Jira
(v8.20.7#820007)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Work logged] (HDFS-16610) Make fsck read timeout configurable

2022-06-01 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HDFS-16610?focusedWorklogId=777123&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-777123
 ]

ASF GitHub Bot logged work on HDFS-16610:
-

Author: ASF GitHub Bot
Created on: 01/Jun/22 19:35
Start Date: 01/Jun/22 19:35
Worklog Time Spent: 10m 
  Work Description: sodonnel commented on PR #4384:
URL: https://github.com/apache/hadoop/pull/4384#issuecomment-1144054446

   Good build this time, so I will commit. Thanks all for the reviews.




Issue Time Tracking
---

Worklog Id: (was: 777123)
Time Spent: 1h 40m  (was: 1.5h)

> Make fsck read timeout configurable
> ---
>
> Key: HDFS-16610
> URL: https://issues.apache.org/jira/browse/HDFS-16610
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: hdfs-client
>Reporter: Stephen O'Donnell
>Assignee: Stephen O'Donnell
>Priority: Major
>  Labels: pull-request-available
>  Time Spent: 1h 40m
>  Remaining Estimate: 0h
>
> In a cluster with a lot of small files, we encountered a case where fsck was 
> very slow. I believe it is due to contention with many other threads reading 
> / writing data on the cluster.
> Sometimes fsck does not report any progress for more than 60 seconds and the 
> client times out. Currently the connect and read timeout are hardcoded to 60 
> seconds. This change is to make them configurable.



--
This message was sent by Atlassian Jira
(v8.20.7#820007)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Work logged] (HDFS-16610) Make fsck read timeout configurable

2022-06-01 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HDFS-16610?focusedWorklogId=777018&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-777018
 ]

ASF GitHub Bot logged work on HDFS-16610:
-

Author: ASF GitHub Bot
Created on: 01/Jun/22 17:06
Start Date: 01/Jun/22 17:06
Worklog Time Spent: 10m 
  Work Description: hadoop-yetus commented on PR #4384:
URL: https://github.com/apache/hadoop/pull/4384#issuecomment-1143886604

   :broken_heart: **-1 overall**
   
   
   
   
   
   
   | Vote | Subsystem | Runtime |  Logfile | Comment |
   |::|--:|:|::|:---:|
   | +0 :ok: |  reexec  |   1m  9s |  |  Docker mode activated.  |
    _ Prechecks _ |
   | +1 :green_heart: |  dupname  |   0m  0s |  |  No case conflicting files 
found.  |
   | +0 :ok: |  codespell  |   0m  1s |  |  codespell was not available.  |
   | +0 :ok: |  detsecrets  |   0m  1s |  |  detect-secrets was not available.  
|
   | +0 :ok: |  xmllint  |   0m  1s |  |  xmllint was not available.  |
   | +1 :green_heart: |  @author  |   0m  0s |  |  The patch does not contain 
any @author tags.  |
   | -1 :x: |  test4tests  |   0m  0s |  |  The patch doesn't appear to include 
any new or modified tests. Please justify why no new tests are needed for this 
patch. Also please list what manual steps were performed to verify this patch.  
|
    _ trunk Compile Tests _ |
   | +0 :ok: |  mvndep  |  14m 16s |  |  Maven dependency ordering for branch  |
   | +1 :green_heart: |  mvninstall  |  27m 52s |  |  trunk passed  |
   | +1 :green_heart: |  compile  |   7m  2s |  |  trunk passed with JDK 
Private Build-11.0.15+10-Ubuntu-0ubuntu0.20.04.1  |
   | +1 :green_heart: |  compile  |   6m 22s |  |  trunk passed with JDK 
Private Build-1.8.0_312-8u312-b07-0ubuntu1~20.04-b07  |
   | +1 :green_heart: |  checkstyle  |   1m 34s |  |  trunk passed  |
   | +1 :green_heart: |  mvnsite  |   2m 49s |  |  trunk passed  |
   | +1 :green_heart: |  javadoc  |   2m 12s |  |  trunk passed with JDK 
Private Build-11.0.15+10-Ubuntu-0ubuntu0.20.04.1  |
   | +1 :green_heart: |  javadoc  |   2m 28s |  |  trunk passed with JDK 
Private Build-1.8.0_312-8u312-b07-0ubuntu1~20.04-b07  |
   | +1 :green_heart: |  spotbugs  |   6m 33s |  |  trunk passed  |
   | +1 :green_heart: |  shadedclient  |  26m 15s |  |  branch has no errors 
when building and testing our client artifacts.  |
    _ Patch Compile Tests _ |
   | +0 :ok: |  mvndep  |   0m 27s |  |  Maven dependency ordering for patch  |
   | +1 :green_heart: |  mvninstall  |   2m 35s |  |  the patch passed  |
   | +1 :green_heart: |  compile  |   7m 39s |  |  the patch passed with JDK 
Private Build-11.0.15+10-Ubuntu-0ubuntu0.20.04.1  |
   | +1 :green_heart: |  javac  |   7m 39s |  |  the patch passed  |
   | +1 :green_heart: |  compile  |   7m 27s |  |  the patch passed with JDK 
Private Build-1.8.0_312-8u312-b07-0ubuntu1~20.04-b07  |
   | +1 :green_heart: |  javac  |   7m 27s |  |  the patch passed  |
   | +1 :green_heart: |  blanks  |   0m  0s |  |  The patch has no blanks 
issues.  |
   | +1 :green_heart: |  checkstyle  |   1m 26s |  |  the patch passed  |
   | +1 :green_heart: |  mvnsite  |   2m 26s |  |  the patch passed  |
   | +1 :green_heart: |  javadoc  |   1m 50s |  |  the patch passed with JDK 
Private Build-11.0.15+10-Ubuntu-0ubuntu0.20.04.1  |
   | +1 :green_heart: |  javadoc  |   2m  9s |  |  the patch passed with JDK 
Private Build-1.8.0_312-8u312-b07-0ubuntu1~20.04-b07  |
   | +1 :green_heart: |  spotbugs  |   6m 38s |  |  the patch passed  |
   | +1 :green_heart: |  shadedclient  |  26m  1s |  |  patch has no errors 
when building and testing our client artifacts.  |
    _ Other Tests _ |
   | +1 :green_heart: |  unit  |   2m 28s |  |  hadoop-hdfs-client in the patch 
passed.  |
   | +1 :green_heart: |  unit  | 331m 58s |  |  hadoop-hdfs in the patch 
passed.  |
   | +1 :green_heart: |  asflicense  |   1m  0s |  |  The patch does not 
generate ASF License warnings.  |
   |  |   | 490m 17s |  |  |
   
   
   | Subsystem | Report/Notes |
   |--:|:-|
   | Docker | ClientAPI=1.41 ServerAPI=1.41 base: 
https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-4384/3/artifact/out/Dockerfile
 |
   | GITHUB PR | https://github.com/apache/hadoop/pull/4384 |
   | Optional Tests | dupname asflicense compile javac javadoc mvninstall 
mvnsite unit shadedclient spotbugs checkstyle codespell detsecrets xmllint |
   | uname | Linux bd88f84c1b08 4.15.0-175-generic #184-Ubuntu SMP Thu Mar 24 
17:48:36 UTC 2022 x86_64 x86_64 x86_64 GNU/Linux |
   | Build tool | maven |
   | Personality | dev-support/bin/hadoop.sh |
   | git revision | trunk / 0ecf65945d28b2c3243495f826057f390eb1fad6 |
   | Default Java | Private Build-1.8.0_312-8u312-b07-0ubuntu1~20.04-b07 |
   | Multi-JDK versions | /usr/lib/jvm/java-11-openjdk-amd64:Private 
Build-11.0.15+10-Ubuntu-0ubuntu0.20

[jira] [Resolved] (HDFS-16612) impove import * In HDFS Project

2022-06-01 Thread fanshilun (Jira)


 [ 
https://issues.apache.org/jira/browse/HDFS-16612?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

fanshilun resolved HDFS-16612.
--
Resolution: Not A Problem

> impove import * In HDFS Project
> ---
>
> Key: HDFS-16612
> URL: https://issues.apache.org/jira/browse/HDFS-16612
> Project: Hadoop HDFS
>  Issue Type: Improvement
>Affects Versions: 3.4.0
>Reporter: fanshilun
>Assignee: fanshilun
>Priority: Minor
>




--
This message was sent by Atlassian Jira
(v8.20.7#820007)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Work logged] (HDFS-16602) Use "defined" directive along with #if

2022-06-01 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HDFS-16602?focusedWorklogId=776795&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-776795
 ]

ASF GitHub Bot logged work on HDFS-16602:
-

Author: ASF GitHub Bot
Created on: 01/Jun/22 11:04
Start Date: 01/Jun/22 11:04
Worklog Time Spent: 10m 
  Work Description: hadoop-yetus commented on PR #4371:
URL: https://github.com/apache/hadoop/pull/4371#issuecomment-1143459438

   :broken_heart: **-1 overall**
   
   
   
   
   
   
   | Vote | Subsystem | Runtime |  Logfile | Comment |
   |::|--:|:|::|:---:|
   | +0 :ok: |  reexec  |  12m  1s |  |  Docker mode activated.  |
    _ Prechecks _ |
   | +1 :green_heart: |  dupname  |   0m  0s |  |  No case conflicting files 
found.  |
   | +0 :ok: |  codespell  |   0m  1s |  |  codespell was not available.  |
   | +0 :ok: |  detsecrets  |   0m  1s |  |  detect-secrets was not available.  
|
   | +1 :green_heart: |  @author  |   0m  0s |  |  The patch does not contain 
any @author tags.  |
   | +1 :green_heart: |  test4tests  |   0m  0s |  |  The patch appears to 
include 1 new or modified test files.  |
    _ trunk Compile Tests _ |
   | +1 :green_heart: |  mvninstall  |  24m 44s |  |  trunk passed  |
   | -1 :x: |  compile  |   0m 45s | 
[/branch-compile-hadoop-hdfs-project_hadoop-hdfs-native-client.txt](https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-4371/2/artifact/out/branch-compile-hadoop-hdfs-project_hadoop-hdfs-native-client.txt)
 |  hadoop-hdfs-native-client in trunk failed.  |
   | +1 :green_heart: |  mvnsite  |   0m 41s |  |  trunk passed  |
   | +1 :green_heart: |  shadedclient  |  52m  7s |  |  branch has no errors 
when building and testing our client artifacts.  |
    _ Patch Compile Tests _ |
   | +1 :green_heart: |  mvninstall  |   0m 23s |  |  the patch passed  |
   | -1 :x: |  compile  |   0m 28s | 
[/patch-compile-hadoop-hdfs-project_hadoop-hdfs-native-client.txt](https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-4371/2/artifact/out/patch-compile-hadoop-hdfs-project_hadoop-hdfs-native-client.txt)
 |  hadoop-hdfs-native-client in the patch failed.  |
   | -1 :x: |  cc  |   0m 28s | 
[/patch-compile-hadoop-hdfs-project_hadoop-hdfs-native-client.txt](https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-4371/2/artifact/out/patch-compile-hadoop-hdfs-project_hadoop-hdfs-native-client.txt)
 |  hadoop-hdfs-native-client in the patch failed.  |
   | -1 :x: |  golang  |   0m 28s | 
[/patch-compile-hadoop-hdfs-project_hadoop-hdfs-native-client.txt](https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-4371/2/artifact/out/patch-compile-hadoop-hdfs-project_hadoop-hdfs-native-client.txt)
 |  hadoop-hdfs-native-client in the patch failed.  |
   | -1 :x: |  javac  |   0m 28s | 
[/patch-compile-hadoop-hdfs-project_hadoop-hdfs-native-client.txt](https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-4371/2/artifact/out/patch-compile-hadoop-hdfs-project_hadoop-hdfs-native-client.txt)
 |  hadoop-hdfs-native-client in the patch failed.  |
   | +1 :green_heart: |  blanks  |   0m  0s |  |  The patch has no blanks 
issues.  |
   | +1 :green_heart: |  mvnsite  |   0m 24s |  |  the patch passed  |
   | +1 :green_heart: |  shadedclient  |  26m 18s |  |  patch has no errors 
when building and testing our client artifacts.  |
    _ Other Tests _ |
   | -1 :x: |  unit  |   0m 37s | 
[/patch-unit-hadoop-hdfs-project_hadoop-hdfs-native-client.txt](https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-4371/2/artifact/out/patch-unit-hadoop-hdfs-project_hadoop-hdfs-native-client.txt)
 |  hadoop-hdfs-native-client in the patch failed.  |
   | +1 :green_heart: |  asflicense  |   0m 45s |  |  The patch does not 
generate ASF License warnings.  |
   |  |   |  95m 33s |  |  |
   
   
   | Subsystem | Report/Notes |
   |--:|:-|
   | Docker | ClientAPI=1.41 ServerAPI=1.41 base: 
https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-4371/2/artifact/out/Dockerfile
 |
   | GITHUB PR | https://github.com/apache/hadoop/pull/4371 |
   | Optional Tests | dupname asflicense compile cc mvnsite javac unit 
codespell detsecrets golang |
   | uname | Linux 24833ef1266f 4.15.0-58-generic #64-Ubuntu SMP Tue Aug 6 
11:12:41 UTC 2019 x86_64 GNU/Linux |
   | Build tool | maven |
   | Personality | dev-support/bin/hadoop.sh |
   | git revision | trunk / b472514254aeaae500853979626918c80c8387b4 |
   | Default Java | Debian-11.0.15+10-post-Debian-1deb10u1 |
   |  Test Results | 
https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-4371/2/testReport/ |
   | Max. process+thread count | 698 (vs. ulimit of 5500) |
   | modules | C: hadoop-hdfs-project/hadoop-hdfs-native-client U: 
hadoop-hdfs-project/hadoop-hdfs-native-client |
   | Console output | 
https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-4371/2/console |
   | vers

[jira] [Work logged] (HDFS-16602) Use "defined" directive along with #if

2022-06-01 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HDFS-16602?focusedWorklogId=776737&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-776737
 ]

ASF GitHub Bot logged work on HDFS-16602:
-

Author: ASF GitHub Bot
Created on: 01/Jun/22 09:28
Start Date: 01/Jun/22 09:28
Worklog Time Spent: 10m 
  Work Description: hadoop-yetus commented on PR #4371:
URL: https://github.com/apache/hadoop/pull/4371#issuecomment-1143358246

   :confetti_ball: **+1 overall**
   
   
   
   
   
   
   | Vote | Subsystem | Runtime |  Logfile | Comment |
   |::|--:|:|::|:---:|
   | +0 :ok: |  reexec  |  22m 58s |  |  Docker mode activated.  |
    _ Prechecks _ |
   | +1 :green_heart: |  dupname  |   0m  0s |  |  No case conflicting files 
found.  |
   | +0 :ok: |  codespell  |   0m  0s |  |  codespell was not available.  |
   | +0 :ok: |  detsecrets  |   0m  0s |  |  detect-secrets was not available.  
|
   | +1 :green_heart: |  @author  |   0m  0s |  |  The patch does not contain 
any @author tags.  |
   | +1 :green_heart: |  test4tests  |   0m  0s |  |  The patch appears to 
include 1 new or modified test files.  |
    _ trunk Compile Tests _ |
   | +1 :green_heart: |  mvninstall  |  23m 14s |  |  trunk passed  |
   | +1 :green_heart: |  compile  |   4m 12s |  |  trunk passed  |
   | +1 :green_heart: |  mvnsite  |   1m  4s |  |  trunk passed  |
   | +1 :green_heart: |  shadedclient  |  47m 55s |  |  branch has no errors 
when building and testing our client artifacts.  |
    _ Patch Compile Tests _ |
   | +1 :green_heart: |  mvninstall  |   0m 32s |  |  the patch passed  |
   | +1 :green_heart: |  compile  |   3m 46s |  |  the patch passed  |
   | +1 :green_heart: |  cc  |   3m 46s |  |  the patch passed  |
   | +1 :green_heart: |  golang  |   3m 46s |  |  the patch passed  |
   | +1 :green_heart: |  javac  |   3m 46s |  |  the patch passed  |
   | +1 :green_heart: |  blanks  |   0m  0s |  |  The patch has no blanks 
issues.  |
   | +1 :green_heart: |  mvnsite  |   0m 35s |  |  the patch passed  |
   | +1 :green_heart: |  shadedclient  |  19m 18s |  |  patch has no errors 
when building and testing our client artifacts.  |
    _ Other Tests _ |
   | +1 :green_heart: |  unit  |  33m 12s |  |  hadoop-hdfs-native-client in 
the patch passed.  |
   | +1 :green_heart: |  asflicense  |   1m  2s |  |  The patch does not 
generate ASF License warnings.  |
   |  |   | 131m 46s |  |  |
   
   
   | Subsystem | Report/Notes |
   |--:|:-|
   | Docker | ClientAPI=1.41 ServerAPI=1.41 base: 
https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-4371/2/artifact/out/Dockerfile
 |
   | GITHUB PR | https://github.com/apache/hadoop/pull/4371 |
   | Optional Tests | dupname asflicense compile cc mvnsite javac unit 
codespell detsecrets golang |
   | uname | Linux 0700fbc3045c 4.15.0-58-generic #64-Ubuntu SMP Tue Aug 6 
11:12:41 UTC 2019 x86_64 x86_64 x86_64 GNU/Linux |
   | Build tool | maven |
   | Personality | dev-support/bin/hadoop.sh |
   | git revision | trunk / b472514254aeaae500853979626918c80c8387b4 |
   | Default Java | Red Hat, Inc.-1.8.0_312-b07 |
   |  Test Results | 
https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-4371/2/testReport/ |
   | Max. process+thread count | 698 (vs. ulimit of 5500) |
   | modules | C: hadoop-hdfs-project/hadoop-hdfs-native-client U: 
hadoop-hdfs-project/hadoop-hdfs-native-client |
   | Console output | 
https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-4371/2/console |
   | versions | git=2.27.0 maven=3.6.3 |
   | Powered by | Apache Yetus 0.14.0 https://yetus.apache.org |
   
   
   This message was automatically generated.
   
   




Issue Time Tracking
---

Worklog Id: (was: 776737)
Time Spent: 40m  (was: 0.5h)

> Use "defined" directive along with #if
> --
>
> Key: HDFS-16602
> URL: https://issues.apache.org/jira/browse/HDFS-16602
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: libhdfs++
>Affects Versions: 3.4.0
>Reporter: Gautham Banasandra
>Assignee: Gautham Banasandra
>Priority: Major
>  Labels: libhdfscpp, pull-request-available
>  Time Spent: 40m
>  Remaining Estimate: 0h
>
> The #if preprocessor directive expects a boolean expression. Thus, we need to 
> use the "defined" directive as well to check if the macro has been defined.



--
This message was sent by Atlassian Jira
(v8.20.7#820007)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Work logged] (HDFS-16610) Make fsck read timeout configurable

2022-06-01 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HDFS-16610?focusedWorklogId=776723&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-776723
 ]

ASF GitHub Bot logged work on HDFS-16610:
-

Author: ASF GitHub Bot
Created on: 01/Jun/22 08:55
Start Date: 01/Jun/22 08:55
Worklog Time Spent: 10m 
  Work Description: sodonnel commented on PR #4384:
URL: https://github.com/apache/hadoop/pull/4384#issuecomment-1143322132

   I rebased against the latest trunk and force pushed it. Lets see if we get a 
better build this time.




Issue Time Tracking
---

Worklog Id: (was: 776723)
Time Spent: 1h 20m  (was: 1h 10m)

> Make fsck read timeout configurable
> ---
>
> Key: HDFS-16610
> URL: https://issues.apache.org/jira/browse/HDFS-16610
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: hdfs-client
>Reporter: Stephen O'Donnell
>Assignee: Stephen O'Donnell
>Priority: Major
>  Labels: pull-request-available
>  Time Spent: 1h 20m
>  Remaining Estimate: 0h
>
> In a cluster with a lot of small files, we encountered a case where fsck was 
> very slow. I believe it is due to contention with many other threads reading 
> / writing data on the cluster.
> Sometimes fsck does not report any progress for more than 60 seconds and the 
> client times out. Currently the connect and read timeout are hardcoded to 60 
> seconds. This change is to make them configurable.



--
This message was sent by Atlassian Jira
(v8.20.7#820007)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Work logged] (HDFS-16463) Make dirent cross platform compatible

2022-06-01 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HDFS-16463?focusedWorklogId=776700&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-776700
 ]

ASF GitHub Bot logged work on HDFS-16463:
-

Author: ASF GitHub Bot
Created on: 01/Jun/22 07:41
Start Date: 01/Jun/22 07:41
Worklog Time Spent: 10m 
  Work Description: hadoop-yetus commented on PR #4370:
URL: https://github.com/apache/hadoop/pull/4370#issuecomment-1143228728

   :broken_heart: **-1 overall**
   
   
   
   
   
   
   | Vote | Subsystem | Runtime |  Logfile | Comment |
   |::|--:|:|::|:---:|
   | +0 :ok: |  reexec  |  46m  8s |  |  Docker mode activated.  |
    _ Prechecks _ |
   | +1 :green_heart: |  dupname  |   0m  0s |  |  No case conflicting files 
found.  |
   | +0 :ok: |  codespell  |   0m  0s |  |  codespell was not available.  |
   | +0 :ok: |  detsecrets  |   0m  0s |  |  detect-secrets was not available.  
|
   | +1 :green_heart: |  @author  |   0m  0s |  |  The patch does not contain 
any @author tags.  |
   | +1 :green_heart: |  test4tests  |   0m  0s |  |  The patch appears to 
include 5 new or modified test files.  |
    _ trunk Compile Tests _ |
   | +1 :green_heart: |  mvninstall  |  40m 55s |  |  trunk passed  |
   | +1 :green_heart: |  compile  |   3m 55s |  |  trunk passed  |
   | +1 :green_heart: |  mvnsite  |   0m 38s |  |  trunk passed  |
   | +1 :green_heart: |  shadedclient  |  67m 37s |  |  branch has no errors 
when building and testing our client artifacts.  |
    _ Patch Compile Tests _ |
   | +1 :green_heart: |  mvninstall  |   0m 20s |  |  the patch passed  |
   | +1 :green_heart: |  compile  |   3m 47s |  |  the patch passed  |
   | -1 :x: |  cc  |   3m 47s | 
[/results-compile-cc-hadoop-hdfs-project_hadoop-hdfs-native-client.txt](https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-4370/7/artifact/out/results-compile-cc-hadoop-hdfs-project_hadoop-hdfs-native-client.txt)
 |  hadoop-hdfs-project_hadoop-hdfs-native-client generated 2 new + 8 unchanged 
- 0 fixed = 10 total (was 8)  |
   | +1 :green_heart: |  golang  |   3m 47s |  |  the patch passed  |
   | +1 :green_heart: |  javac  |   3m 47s |  |  the patch passed  |
   | -1 :x: |  blanks  |   0m  0s | 
[/blanks-eol.txt](https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-4370/7/artifact/out/blanks-eol.txt)
 |  The patch has 1 line(s) that end in blanks. Use git apply --whitespace=fix 
<>. Refer https://git-scm.com/docs/git-apply  |
   | +1 :green_heart: |  mvnsite  |   0m 22s |  |  the patch passed  |
   | +1 :green_heart: |  shadedclient  |  22m  9s |  |  patch has no errors 
when building and testing our client artifacts.  |
    _ Other Tests _ |
   | -1 :x: |  unit  |   5m  2s | 
[/patch-unit-hadoop-hdfs-project_hadoop-hdfs-native-client.txt](https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-4370/7/artifact/out/patch-unit-hadoop-hdfs-project_hadoop-hdfs-native-client.txt)
 |  hadoop-hdfs-native-client in the patch failed.  |
   | +1 :green_heart: |  asflicense  |   0m 46s |  |  The patch does not 
generate ASF License warnings.  |
   |  |   | 148m 51s |  |  |
   
   
   | Reason | Tests |
   |---:|:--|
   | Failed CTEST tests | test_test_libhdfs_ops_hdfs_static |
   |   | test_test_libhdfs_threaded_hdfs_static |
   |   | test_test_libhdfs_zerocopy_hdfs_static |
   |   | test_test_native_mini_dfs |
   |   | test_libhdfs_threaded_hdfspp_test_shim_static |
   |   | test_hdfspp_mini_dfs_smoke_hdfspp_test_shim_static |
   |   | libhdfs_mini_stress_valgrind_hdfspp_test_static |
   |   | memcheck_libhdfs_mini_stress_valgrind_hdfspp_test_static |
   |   | test_libhdfs_mini_stress_hdfspp_test_shim_static |
   |   | test_hdfs_ext_hdfspp_test_shim_static |
   |   | x_platform_dirent_c_test |
   
   
   | Subsystem | Report/Notes |
   |--:|:-|
   | Docker | ClientAPI=1.41 ServerAPI=1.41 base: 
https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-4370/7/artifact/out/Dockerfile
 |
   | GITHUB PR | https://github.com/apache/hadoop/pull/4370 |
   | Optional Tests | dupname asflicense compile cc mvnsite javac unit 
codespell detsecrets golang |
   | uname | Linux 628acf824e94 4.15.0-175-generic #184-Ubuntu SMP Thu Mar 24 
17:48:36 UTC 2022 x86_64 x86_64 x86_64 GNU/Linux |
   | Build tool | maven |
   | Personality | dev-support/bin/hadoop.sh |
   | git revision | trunk / 27b3dc8c60092c35879010ad8f0a55eef265625a |
   | Default Java | Red Hat, Inc.-1.8.0_332-b09 |
   | CTEST | 
https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-4370/7/artifact/out/patch-hadoop-hdfs-project_hadoop-hdfs-native-client-ctest.txt
 |
   |  Test Results | 
https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-4370/7/testReport/ |
   | Max. process+thread count | 600 (vs. ulimit of 5500) |
   | modules | C: hadoop-hdfs-project/hadoop-hdfs-native-client U: 
hadoop-hdfs-project

[jira] [Work logged] (HDFS-16598) All datanodes [DatanodeInfoWithStorage[127.0.0.1:57448,DS-1b5f7e33-a2bf-4edc-9122-a74c995a99f5,DISK]] are bad. Aborting...

2022-06-01 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HDFS-16598?focusedWorklogId=776694&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-776694
 ]

ASF GitHub Bot logged work on HDFS-16598:
-

Author: ASF GitHub Bot
Created on: 01/Jun/22 07:20
Start Date: 01/Jun/22 07:20
Worklog Time Spent: 10m 
  Work Description: ZanderXu commented on PR #4366:
URL: https://github.com/apache/hadoop/pull/4366#issuecomment-1143208824

   getReplicaInfo(ExtendedBlock b) will check gs, and getReplicaInfo(String 
bpid, long blkid) will not check the gs.




Issue Time Tracking
---

Worklog Id: (was: 776694)
Time Spent: 1h 10m  (was: 1h)

> All datanodes 
> [DatanodeInfoWithStorage[127.0.0.1:57448,DS-1b5f7e33-a2bf-4edc-9122-a74c995a99f5,DISK]]
>  are bad. Aborting...
> --
>
> Key: HDFS-16598
> URL: https://issues.apache.org/jira/browse/HDFS-16598
> Project: Hadoop HDFS
>  Issue Type: Bug
>Reporter: ZanderXu
>Assignee: ZanderXu
>Priority: Major
>  Labels: pull-request-available
>  Time Spent: 1h 10m
>  Remaining Estimate: 0h
>
> org.apache.hadoop.hdfs.testPipelineRecoveryOnRestartFailure failed with the 
> stack like:
> {code:java}
> java.io.IOException: All datanodes 
> [DatanodeInfoWithStorage[127.0.0.1:57448,DS-1b5f7e33-a2bf-4edc-9122-a74c995a99f5,DISK]]
>  are bad. Aborting...
>   at 
> org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1667)
>   at 
> org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1601)
>   at 
> org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1587)
>   at 
> org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1371)
>   at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:674)
> {code}
> After tracing the root cause, this bug was introduced by 
> [HDFS-16534|https://issues.apache.org/jira/browse/HDFS-16534]. Because the 
> block GS of client may be smaller than DN when pipeline recovery failed.



--
This message was sent by Atlassian Jira
(v8.20.7#820007)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Work logged] (HDFS-16602) Use "defined" directive along with #if

2022-06-01 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HDFS-16602?focusedWorklogId=776693&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-776693
 ]

ASF GitHub Bot logged work on HDFS-16602:
-

Author: ASF GitHub Bot
Created on: 01/Jun/22 07:16
Start Date: 01/Jun/22 07:16
Worklog Time Spent: 10m 
  Work Description: hadoop-yetus commented on PR #4371:
URL: https://github.com/apache/hadoop/pull/4371#issuecomment-1143205258

   :confetti_ball: **+1 overall**
   
   
   
   
   
   
   | Vote | Subsystem | Runtime |  Logfile | Comment |
   |::|--:|:|::|:---:|
   | +0 :ok: |  reexec  |   0m 39s |  |  Docker mode activated.  |
    _ Prechecks _ |
   | +1 :green_heart: |  dupname  |   0m  0s |  |  No case conflicting files 
found.  |
   | +0 :ok: |  codespell  |   0m  0s |  |  codespell was not available.  |
   | +0 :ok: |  detsecrets  |   0m  0s |  |  detect-secrets was not available.  
|
   | +1 :green_heart: |  @author  |   0m  0s |  |  The patch does not contain 
any @author tags.  |
   | +1 :green_heart: |  test4tests  |   0m  0s |  |  The patch appears to 
include 1 new or modified test files.  |
    _ trunk Compile Tests _ |
   | +1 :green_heart: |  mvninstall  |  37m 42s |  |  trunk passed  |
   | +1 :green_heart: |  compile  |   4m  0s |  |  trunk passed  |
   | +1 :green_heart: |  mvnsite  |   0m 49s |  |  trunk passed  |
   | +1 :green_heart: |  shadedclient  |  61m 35s |  |  branch has no errors 
when building and testing our client artifacts.  |
    _ Patch Compile Tests _ |
   | +1 :green_heart: |  mvninstall  |   0m 24s |  |  the patch passed  |
   | +1 :green_heart: |  compile  |   3m 43s |  |  the patch passed  |
   | +1 :green_heart: |  cc  |   3m 43s |  |  the patch passed  |
   | +1 :green_heart: |  golang  |   3m 43s |  |  the patch passed  |
   | +1 :green_heart: |  javac  |   3m 43s |  |  the patch passed  |
   | +1 :green_heart: |  blanks  |   0m  0s |  |  The patch has no blanks 
issues.  |
   | +1 :green_heart: |  mvnsite  |   0m 27s |  |  the patch passed  |
   | +1 :green_heart: |  shadedclient  |  19m  1s |  |  patch has no errors 
when building and testing our client artifacts.  |
    _ Other Tests _ |
   | +1 :green_heart: |  unit  |  32m 47s |  |  hadoop-hdfs-native-client in 
the patch passed.  |
   | +1 :green_heart: |  asflicense  |   0m 49s |  |  The patch does not 
generate ASF License warnings.  |
   |  |   | 122m  3s |  |  |
   
   
   | Subsystem | Report/Notes |
   |--:|:-|
   | Docker | ClientAPI=1.41 ServerAPI=1.41 base: 
https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-4371/2/artifact/out/Dockerfile
 |
   | GITHUB PR | https://github.com/apache/hadoop/pull/4371 |
   | Optional Tests | dupname asflicense compile cc mvnsite javac unit 
codespell detsecrets golang |
   | uname | Linux 8f37b16f5a4a 4.15.0-58-generic #64-Ubuntu SMP Tue Aug 6 
11:12:41 UTC 2019 x86_64 x86_64 x86_64 GNU/Linux |
   | Build tool | maven |
   | Personality | dev-support/bin/hadoop.sh |
   | git revision | trunk / b472514254aeaae500853979626918c80c8387b4 |
   | Default Java | Red Hat, Inc.-1.8.0_332-b09 |
   |  Test Results | 
https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-4371/2/testReport/ |
   | Max. process+thread count | 666 (vs. ulimit of 5500) |
   | modules | C: hadoop-hdfs-project/hadoop-hdfs-native-client U: 
hadoop-hdfs-project/hadoop-hdfs-native-client |
   | Console output | 
https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-4371/2/console |
   | versions | git=2.9.5 maven=3.6.3 |
   | Powered by | Apache Yetus 0.14.0 https://yetus.apache.org |
   
   
   This message was automatically generated.
   
   




Issue Time Tracking
---

Worklog Id: (was: 776693)
Time Spent: 0.5h  (was: 20m)

> Use "defined" directive along with #if
> --
>
> Key: HDFS-16602
> URL: https://issues.apache.org/jira/browse/HDFS-16602
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: libhdfs++
>Affects Versions: 3.4.0
>Reporter: Gautham Banasandra
>Assignee: Gautham Banasandra
>Priority: Major
>  Labels: libhdfscpp, pull-request-available
>  Time Spent: 0.5h
>  Remaining Estimate: 0h
>
> The #if preprocessor directive expects a boolean expression. Thus, we need to 
> use the "defined" directive as well to check if the macro has been defined.



--
This message was sent by Atlassian Jira
(v8.20.7#820007)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Work logged] (HDFS-16598) All datanodes [DatanodeInfoWithStorage[127.0.0.1:57448,DS-1b5f7e33-a2bf-4edc-9122-a74c995a99f5,DISK]] are bad. Aborting...

2022-06-01 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HDFS-16598?focusedWorklogId=776691&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-776691
 ]

ASF GitHub Bot logged work on HDFS-16598:
-

Author: ASF GitHub Bot
Created on: 01/Jun/22 07:14
Start Date: 01/Jun/22 07:14
Worklog Time Spent: 10m 
  Work Description: MingXiangLi commented on PR #4366:
URL: https://github.com/apache/hadoop/pull/4366#issuecomment-1143203294

   Thanks for your feedback.But I don't quite understand why this PR can solves 
the problem.It seems that you only change getReplicaInfo(x) to 
getReplicaInfo(x,x), the logic in two method was almost same.




Issue Time Tracking
---

Worklog Id: (was: 776691)
Time Spent: 1h  (was: 50m)

> All datanodes 
> [DatanodeInfoWithStorage[127.0.0.1:57448,DS-1b5f7e33-a2bf-4edc-9122-a74c995a99f5,DISK]]
>  are bad. Aborting...
> --
>
> Key: HDFS-16598
> URL: https://issues.apache.org/jira/browse/HDFS-16598
> Project: Hadoop HDFS
>  Issue Type: Bug
>Reporter: ZanderXu
>Assignee: ZanderXu
>Priority: Major
>  Labels: pull-request-available
>  Time Spent: 1h
>  Remaining Estimate: 0h
>
> org.apache.hadoop.hdfs.testPipelineRecoveryOnRestartFailure failed with the 
> stack like:
> {code:java}
> java.io.IOException: All datanodes 
> [DatanodeInfoWithStorage[127.0.0.1:57448,DS-1b5f7e33-a2bf-4edc-9122-a74c995a99f5,DISK]]
>  are bad. Aborting...
>   at 
> org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1667)
>   at 
> org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1601)
>   at 
> org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1587)
>   at 
> org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1371)
>   at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:674)
> {code}
> After tracing the root cause, this bug was introduced by 
> [HDFS-16534|https://issues.apache.org/jira/browse/HDFS-16534]. Because the 
> block GS of client may be smaller than DN when pipeline recovery failed.



--
This message was sent by Atlassian Jira
(v8.20.7#820007)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org