This is an automated email from the ASF dual-hosted git repository.

elek pushed a commit to branch trunk
in repository https://gitbox.apache.org/repos/asf/hadoop.git


The following commit(s) were added to refs/heads/trunk by this push:
     new fb7c1ca  Revert "HDDS-1384. TestBlockOutputStreamWithFailures is 
failing"
fb7c1ca is described below

commit fb7c1cad0ea93406a7272872c888d06e4e56620a
Author: Márton Elek <e...@apache.org>
AuthorDate: Mon May 6 20:17:00 2019 +0200

    Revert "HDDS-1384. TestBlockOutputStreamWithFailures is failing"
    
    This reverts commit dead9b4049484c31e0608956e53a9ef07a45819d.
---
 .../apache/hadoop/ozone/MiniOzoneClusterImpl.java  | 62 +++++-----------------
 1 file changed, 14 insertions(+), 48 deletions(-)

diff --git 
a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/MiniOzoneClusterImpl.java
 
b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/MiniOzoneClusterImpl.java
index c306f22..9fbdad7 100644
--- 
a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/MiniOzoneClusterImpl.java
+++ 
b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/MiniOzoneClusterImpl.java
@@ -18,7 +18,6 @@
 package org.apache.hadoop.ozone;
 
 import java.io.File;
-import java.net.ServerSocket;
 import java.util.ArrayList;
 import java.util.List;
 import java.util.Optional;
@@ -64,7 +63,6 @@ import java.nio.file.Paths;
 import java.util.UUID;
 import java.util.concurrent.TimeUnit;
 import java.util.concurrent.TimeoutException;
-import java.util.concurrent.atomic.AtomicInteger;
 
 import static org.apache.hadoop.hdds.HddsConfigKeys.HDDS_HEARTBEAT_INTERVAL;
 import static org.apache.hadoop.hdds.protocol.proto.HddsProtos.NodeState
@@ -390,9 +388,6 @@ public class MiniOzoneClusterImpl implements 
MiniOzoneCluster {
    */
   public static class Builder extends MiniOzoneCluster.Builder {
 
-    private static AtomicInteger lastUsedPort =
-        new AtomicInteger(1000);
-
     /**
      * Creates a new Builder.
      *
@@ -534,16 +529,14 @@ public class MiniOzoneClusterImpl implements 
MiniOzoneCluster {
      */
     List<HddsDatanodeService> createHddsDatanodes(
         StorageContainerManager scm) throws IOException {
-
-      String scmAddress = scm.getDatanodeRpcAddress().getHostString() +
+      configureHddsDatanodes();
+      String scmAddress =  scm.getDatanodeRpcAddress().getHostString() +
           ":" + scm.getDatanodeRpcAddress().getPort();
       String[] args = new String[] {};
       conf.setStrings(ScmConfigKeys.OZONE_SCM_NAMES, scmAddress);
-
       List<HddsDatanodeService> hddsDatanodes = new ArrayList<>();
       for (int i = 0; i < numOfDatanodes; i++) {
         OzoneConfiguration dnConf = new OzoneConfiguration(conf);
-        configureHddsDatanodes(dnConf);
         String datanodeBaseDir = path + "/datanode-" + Integer.toString(i);
         Path metaDir = Paths.get(datanodeBaseDir, "meta");
         Path dataDir = Paths.get(datanodeBaseDir, "data", "containers");
@@ -570,14 +563,10 @@ public class MiniOzoneClusterImpl implements 
MiniOzoneCluster {
     }
 
     private void configureSCM() {
-      conf.set(ScmConfigKeys.OZONE_SCM_CLIENT_ADDRESS_KEY,
-          "127.0.0.1:" + findPort());
-      conf.set(ScmConfigKeys.OZONE_SCM_BLOCK_CLIENT_ADDRESS_KEY,
-          "127.0.0.1:" + findPort());
-      conf.set(ScmConfigKeys.OZONE_SCM_DATANODE_ADDRESS_KEY,
-          "127.0.0.1:" + findPort());
-      conf.set(ScmConfigKeys.OZONE_SCM_HTTP_ADDRESS_KEY,
-          "127.0.0.1:" + findPort());
+      conf.set(ScmConfigKeys.OZONE_SCM_CLIENT_ADDRESS_KEY, "127.0.0.1:0");
+      conf.set(ScmConfigKeys.OZONE_SCM_BLOCK_CLIENT_ADDRESS_KEY, 
"127.0.0.1:0");
+      conf.set(ScmConfigKeys.OZONE_SCM_DATANODE_ADDRESS_KEY, "127.0.0.1:0");
+      conf.set(ScmConfigKeys.OZONE_SCM_HTTP_ADDRESS_KEY, "127.0.0.1:0");
       conf.setInt(ScmConfigKeys.OZONE_SCM_HANDLER_COUNT_KEY, numOfScmHandlers);
       configureSCMheartbeat();
     }
@@ -608,42 +597,19 @@ public class MiniOzoneClusterImpl implements 
MiniOzoneCluster {
 
 
     private void configureOM() {
-      conf.set(OMConfigKeys.OZONE_OM_ADDRESS_KEY, "127.0.0.1:" + findPort());
-      conf.set(OMConfigKeys.OZONE_OM_HTTP_ADDRESS_KEY,
-          "127.0.0.1:" + findPort());
+      conf.set(OMConfigKeys.OZONE_OM_ADDRESS_KEY, "127.0.0.1:0");
+      conf.set(OMConfigKeys.OZONE_OM_HTTP_ADDRESS_KEY, "127.0.0.1:0");
       conf.setInt(OMConfigKeys.OZONE_OM_HANDLER_COUNT_KEY, numOfOmHandlers);
     }
 
-    /**
-     * Return an available TCP port if available.
-     * <p>
-     * As we have a static counter the port should be unique inside the JVM..
-     */
-    private int findPort() {
-      while (lastUsedPort.get() < 65536) {
-        try {
-          int nextPort = lastUsedPort.incrementAndGet();
-          ServerSocket socket = new ServerSocket(nextPort);
-          socket.close();
-          return nextPort;
-        } catch (IOException ex) {
-          //port is not available, let's try the next one.
-          continue;
-        }
-      }
-      throw new RuntimeException("No available port");
-    }
-
-    private void configureHddsDatanodes(OzoneConfiguration dnConf) {
-      dnConf.set(ScmConfigKeys.HDDS_REST_HTTP_ADDRESS_KEY,
-          "0.0.0.0:" + findPort());
-      dnConf.set(HddsConfigKeys.HDDS_DATANODE_HTTP_ADDRESS_KEY,
-          "0.0.0.0:" + findPort());
-      dnConf.set(HDDS_DATANODE_PLUGINS_KEY,
+    private void configureHddsDatanodes() {
+      conf.set(ScmConfigKeys.HDDS_REST_HTTP_ADDRESS_KEY, "0.0.0.0:0");
+      conf.set(HddsConfigKeys.HDDS_DATANODE_HTTP_ADDRESS_KEY, "0.0.0.0:0");
+      conf.set(HDDS_DATANODE_PLUGINS_KEY,
           "org.apache.hadoop.ozone.web.OzoneHddsDatanodeService");
-      dnConf.setBoolean(OzoneConfigKeys.DFS_CONTAINER_IPC_RANDOM_PORT,
+      conf.setBoolean(OzoneConfigKeys.DFS_CONTAINER_IPC_RANDOM_PORT,
           randomContainerPort);
-      dnConf.setBoolean(OzoneConfigKeys.DFS_CONTAINER_RATIS_IPC_RANDOM_PORT,
+      conf.setBoolean(OzoneConfigKeys.DFS_CONTAINER_RATIS_IPC_RANDOM_PORT,
           randomContainerPort);
     }
 


---------------------------------------------------------------------
To unsubscribe, e-mail: common-commits-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-commits-h...@hadoop.apache.org

Reply via email to