This is an automated email from the ASF dual-hosted git repository.
ayushsaxena pushed a commit to branch trunk
in repository https://gitbox.apache.org/repos/asf/hadoop.git
The following commit(s) were added to refs/heads/trunk by this push:
new cb64e99 HDFS-15347. Replace the deprecated method shaHex. Contributed
by bianqi.
cb64e99 is described below
commit cb64e993c270a68cb0c056ff41ad522e9306c88b
Author: Ayush Saxena <[email protected]>
AuthorDate: Sun May 10 00:11:46 2020 +0530
HDFS-15347. Replace the deprecated method shaHex. Contributed by bianqi.
---
.../java/org/apache/hadoop/hdfs/server/datanode/DiskBalancer.java | 2 +-
.../hadoop/hdfs/server/diskbalancer/command/CancelCommand.java | 2 +-
.../hadoop/hdfs/server/diskbalancer/command/ExecuteCommand.java | 2 +-
.../apache/hadoop/hdfs/server/diskbalancer/TestDiskBalancer.java | 4 ++--
.../hadoop/hdfs/server/diskbalancer/TestDiskBalancerRPC.java | 2 +-
.../hdfs/server/diskbalancer/TestDiskBalancerWithMockMover.java | 8 ++++----
6 files changed, 10 insertions(+), 10 deletions(-)
diff --git
a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DiskBalancer.java
b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DiskBalancer.java
index ffa8524..958c0cfee 100644
---
a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DiskBalancer.java
+++
b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DiskBalancer.java
@@ -399,7 +399,7 @@ public class DiskBalancer {
if ((planID == null) ||
(planID.length() != sha1Length) ||
- !DigestUtils.shaHex(plan.getBytes(Charset.forName("UTF-8")))
+ !DigestUtils.sha1Hex(plan.getBytes(Charset.forName("UTF-8")))
.equalsIgnoreCase(planID)) {
LOG.error("Disk Balancer - Invalid plan hash.");
throw new DiskBalancerException("Invalid or mis-matched hash.",
diff --git
a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/diskbalancer/command/CancelCommand.java
b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/diskbalancer/command/CancelCommand.java
index 007272e..8cd4e14 100644
---
a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/diskbalancer/command/CancelCommand.java
+++
b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/diskbalancer/command/CancelCommand.java
@@ -94,7 +94,7 @@ public class CancelCommand extends Command {
String dataNodeAddress = plan.getNodeName() + ":" + plan.getPort();
Preconditions.checkNotNull(dataNodeAddress);
ClientDatanodeProtocol dataNode = getDataNodeProxy(dataNodeAddress);
- String planHash = DigestUtils.shaHex(planData);
+ String planHash = DigestUtils.sha1Hex(planData);
try {
dataNode.cancelDiskBalancePlan(planHash);
} catch (DiskBalancerException ex) {
diff --git
a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/diskbalancer/command/ExecuteCommand.java
b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/diskbalancer/command/ExecuteCommand.java
index c7cb089..be4a3e3 100644
---
a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/diskbalancer/command/ExecuteCommand.java
+++
b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/diskbalancer/command/ExecuteCommand.java
@@ -99,7 +99,7 @@ public class ExecuteCommand extends Command {
String dataNodeAddress = plan.getNodeName() + ":" + plan.getPort();
Preconditions.checkNotNull(dataNodeAddress);
ClientDatanodeProtocol dataNode = getDataNodeProxy(dataNodeAddress);
- String planHash = DigestUtils.shaHex(planData);
+ String planHash = DigestUtils.sha1Hex(planData);
try {
dataNode.submitDiskBalancerPlan(planHash, DiskBalancerCLI.PLAN_VERSION,
planFile, planData, skipDateCheck);
diff --git
a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/diskbalancer/TestDiskBalancer.java
b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/diskbalancer/TestDiskBalancer.java
index 931bdb4..0b4df1d 100644
---
a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/diskbalancer/TestDiskBalancer.java
+++
b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/diskbalancer/TestDiskBalancer.java
@@ -642,7 +642,7 @@ public class TestDiskBalancer {
node = cluster.getDataNodes().get(dataNodeIndex);
String planJson = plan.toJson();
- String planID = DigestUtils.shaHex(planJson);
+ String planID = DigestUtils.sha1Hex(planJson);
// Submit the plan and wait till the execution is done.
node.submitDiskBalancerPlan(planID, 1, PLAN_FILE, planJson,
@@ -740,7 +740,7 @@ public class TestDiskBalancer {
reconfigThread.start();
String planJson = plan.toJson();
- String planID = DigestUtils.shaHex(planJson);
+ String planID = DigestUtils.sha1Hex(planJson);
diskBalancer.submitPlan(planID, 1, PLAN_FILE, planJson, false);
GenericTestUtils.waitFor(new Supplier<Boolean>() {
diff --git
a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/diskbalancer/TestDiskBalancerRPC.java
b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/diskbalancer/TestDiskBalancerRPC.java
index 9c68f8d..21c9a59 100644
---
a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/diskbalancer/TestDiskBalancerRPC.java
+++
b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/diskbalancer/TestDiskBalancerRPC.java
@@ -317,7 +317,7 @@ public class TestDiskBalancerRPC {
plan = new NodePlan(node.getDataNodeName(), node.getDataNodePort());
planner.balanceVolumeSet(node, node.getVolumeSets().get("DISK"), plan);
planVersion = 1;
- planHash = DigestUtils.shaHex(plan.toJson());
+ planHash = DigestUtils.sha1Hex(plan.toJson());
return this;
}
}
diff --git
a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/diskbalancer/TestDiskBalancerWithMockMover.java
b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/diskbalancer/TestDiskBalancerWithMockMover.java
index f6a5d9c..4ca0af4 100644
---
a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/diskbalancer/TestDiskBalancerWithMockMover.java
+++
b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/diskbalancer/TestDiskBalancerWithMockMover.java
@@ -126,7 +126,7 @@ public class TestDiskBalancerWithMockMover {
private void executeSubmitPlan(NodePlan plan, DiskBalancer balancer,
int version) throws IOException {
String planJson = plan.toJson();
- String planID = DigestUtils.shaHex(planJson);
+ String planID = DigestUtils.sha1Hex(planJson);
balancer.submitPlan(planID, version, PLAN_FILE, planJson, false);
}
@@ -215,7 +215,7 @@ public class TestDiskBalancerWithMockMover {
NodePlan plan = mockMoverHelper.getPlan();
DiskBalancer balancer = mockMoverHelper.getBalancer();
String planJson = plan.toJson();
- String planID = DigestUtils.shaHex(planJson);
+ String planID = DigestUtils.sha1Hex(planJson);
thrown.expect(DiskBalancerException.class);
thrown.expect(new DiskBalancerResultVerifier(DiskBalancerException
@@ -232,7 +232,7 @@ public class TestDiskBalancerWithMockMover {
String planJson = plan.toJson();
- String planID = DigestUtils.shaHex(planJson);
+ String planID = DigestUtils.sha1Hex(planJson);
char repChar = planID.charAt(0);
repChar++;
@@ -262,7 +262,7 @@ public class TestDiskBalancerWithMockMover {
String planJson = plan.toJson();
- String planID = DigestUtils.shaHex(planJson);
+ String planID = DigestUtils.sha1Hex(planJson);
balancer.cancelPlan(planID);
DiskBalancerWorkStatus status = balancer.queryWorkStatus();
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]