[GitHub] [hadoop-ozone] nandakumar131 commented on issue #313: HDDS-2242. Avoid unnecessary rpc needed to discover the pipeline leader.

2019-12-12 Thread GitBox
nandakumar131 commented on issue #313: HDDS-2242. Avoid unnecessary rpc needed 
to discover the pipeline leader.
URL: https://github.com/apache/hadoop-ozone/pull/313#issuecomment-565333775
 
 
   > > The code will become a lot simpler if the leader node in `Pipeline` is 
represented as `DatanodeDetails` instead of just `UUID`.
   > 
   > Hi, @nandakumar131 this was changed based on your suggestion in HDDS-1868 
:-)
   > Although it isn't affecting performance, can we refactor later, this patch 
was meant as downstream work and I was hoping not to change proto structure for 
it.
   
   @swagle Even in HDDS-1868 the suggestion was to use `DatanodeDetails` 
instead of just `UUID`.
   We can do it in a follow-up jira.
   
   I will merge this shortly.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[GitHub] [hadoop-ozone] xiaoyuyao commented on a change in pull request #291: HDDS-1574 Average out pipeline allocation on datanodes and add metrcs/test

2019-12-12 Thread GitBox
xiaoyuyao commented on a change in pull request #291: HDDS-1574 Average out 
pipeline allocation on datanodes and add metrcs/test
URL: https://github.com/apache/hadoop-ozone/pull/291#discussion_r357511301
 
 

 ##
 File path: 
hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/pipeline/SCMPipelineManager.java
 ##
 @@ -163,6 +177,18 @@ public synchronized Pipeline 
createPipeline(ReplicationType type,
 metrics.incNumPipelineCreated();
 metrics.createPerPipelineMetrics(pipeline);
   }
+  Pipeline overlapPipeline = RatisPipelineUtils
+  .checkPipelineContainSameDatanodes(stateManager, pipeline);
+  if (overlapPipeline != null) {
+metrics.incNumPipelineContainSameDatanodes();
+//TODO remove until pipeline allocation is proved equally distributed.
+LOG.info("Pipeline: " + pipeline.getId().toString() +
 
 Review comment:
   Talked offline. OK to have this at INFO level for now. 


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[GitHub] [hadoop-ozone] xiaoyuyao commented on a change in pull request #291: HDDS-1574 Average out pipeline allocation on datanodes and add metrcs/test

2019-12-12 Thread GitBox
xiaoyuyao commented on a change in pull request #291: HDDS-1574 Average out 
pipeline allocation on datanodes and add metrcs/test
URL: https://github.com/apache/hadoop-ozone/pull/291#discussion_r357509565
 
 

 ##
 File path: 
hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/pipeline/RatisPipelineUtils.java
 ##
 @@ -100,4 +105,45 @@ static void destroyPipeline(DatanodeDetails dn, 
PipelineID pipelineID,
   true, p.getId());
 }
   }
+
+  static int encodeNodeIdsOfFactorThreePipeline(List nodes) {
+if (nodes.size() != HddsProtos.ReplicationFactor.THREE.getNumber()) {
+  return 0;
+}
+List nodeIds = nodes.stream()
+.map(DatanodeDetails::getUuid).distinct()
+.collect(Collectors.toList());
+nodeIds.sort(new ComparableComparator());
+// Only for Factor THREE pipeline.
+return new HashCodeBuilder()
+.append(nodeIds.get(0).toString())
+.append(nodeIds.get(1).toString())
+.append(nodeIds.get(2).toString())
+.toHashCode();
+  }
+
+  /**
+   * Return first existed pipeline which share the same set of datanodes
+   * with the input pipeline.
+   * @param stateManager PipelineStateManager
+   * @param pipeline input pipeline
+   * @return first matched pipeline
+   */
+  static Pipeline checkPipelineContainSameDatanodes(
 
 Review comment:
   Both should work just non-static seems cleaner. But if there are too much 
additional work to fix the unit tests. I'm OK with current approach. 


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[GitHub] [hadoop-ozone] adoroszlai commented on issue #341: HDDS-2414. Simplify robot tests with removing output greps

2019-12-12 Thread GitBox
adoroszlai commented on issue #341: HDDS-2414. Simplify robot tests with 
removing output greps
URL: https://github.com/apache/hadoop-ozone/pull/341#issuecomment-565327920
 
 
   @elek please review


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[GitHub] [hadoop-ozone] adoroszlai commented on issue #342: HDDS-2709. Maven property skipShade should not skip ozonefs compilation

2019-12-12 Thread GitBox
adoroszlai commented on issue #342: HDDS-2709. Maven property skipShade should 
not skip ozonefs compilation
URL: https://github.com/apache/hadoop-ozone/pull/342#issuecomment-565327979
 
 
   @elek please review


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[GitHub] [hadoop-ozone] adoroszlai commented on a change in pull request #348: HDDS-2035 Implement datanode level CLI to reveal pipeline relation.

2019-12-12 Thread GitBox
adoroszlai commented on a change in pull request #348: HDDS-2035 Implement 
datanode level CLI to reveal pipeline relation.
URL: https://github.com/apache/hadoop-ozone/pull/348#discussion_r357503054
 
 

 ##
 File path: 
hadoop-hdds/tools/src/main/java/org/apache/hadoop/hdds/scm/cli/datanode/ListInfoSubcommand.java
 ##
 @@ -0,0 +1,117 @@
+/**
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ * 
+ * http://www.apache.org/licenses/LICENSE-2.0
+ * 
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+package org.apache.hadoop.hdds.scm.cli.datanode;
+
+import org.apache.hadoop.hdds.cli.HddsVersionProvider;
+import org.apache.hadoop.hdds.protocol.DatanodeDetails;
+import org.apache.hadoop.hdds.protocol.proto.HddsProtos;
+import org.apache.hadoop.hdds.scm.client.ScmClient;
+import org.apache.hadoop.hdds.scm.pipeline.Pipeline;
+import picocli.CommandLine;
+
+import java.io.IOException;
+import java.util.ArrayList;
+import java.util.List;
+import java.util.concurrent.Callable;
+import java.util.stream.Collectors;
+
+/**
+ * Handler of list datanodes info command.
+ */
+@CommandLine.Command(
+name = "list",
+description = "List info of datanodes",
+mixinStandardHelpOptions = true,
+versionProvider = HddsVersionProvider.class)
+public class ListInfoSubcommand implements Callable {
+
+  @CommandLine.ParentCommand
+  private DatanodeCommands parent;
+
+  @CommandLine.Option(names = {"-ip", "--byIp"},
+  description = "Show info by ip address.",
+  defaultValue = "",
+  required = false)
+  private String ipaddress;
+
+  @CommandLine.Option(names = {"-id", "--byUuid"},
+  description = "Show info by datanode UUID.",
+  defaultValue = "",
+  required = false)
+  private String uuid;
+
+  private List pipelines;
+
+
+  @Override
+  public Void call() throws Exception {
+try (ScmClient scmClient = parent.getParent().createScmClient()) {
+  pipelines = scmClient.listPipelines();
+  if (isNullOrEmpty(ipaddress) && isNullOrEmpty(uuid)) {
+getAllNodes(scmClient).stream().forEach(p -> printDatanodeInfo(p));
+  } else {
+getAllNodes(scmClient).stream().filter(
+p -> ((isNullOrEmpty(ipaddress) ||
+(p.getIpAddress().compareToIgnoreCase(ipaddress) == 0))
+&& (isNullOrEmpty(uuid) ||
+(p.getUuid().equals(uuid)
+.forEach(p -> printDatanodeInfo(p));
+  }
 
 Review comment:
   Structuring this as below would reduce duplication, and make it easier to 
add further filters if needed:
   
   ```
   Stream<...> stream = getAllNodes(scmClient).stream();
   if (!isNullOrEmpty(ipaddress)) {
 stream = stream.filter(...);
   }
   if (!isNullOrEmpty(uuid)) {
 stream = stream.filter(...);
   }
   stream.forEach(...);
   ```


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[GitHub] [hadoop-ozone] adoroszlai commented on a change in pull request #348: HDDS-2035 Implement datanode level CLI to reveal pipeline relation.

2019-12-12 Thread GitBox
adoroszlai commented on a change in pull request #348: HDDS-2035 Implement 
datanode level CLI to reveal pipeline relation.
URL: https://github.com/apache/hadoop-ozone/pull/348#discussion_r357502346
 
 

 ##
 File path: 
hadoop-hdds/tools/src/main/java/org/apache/hadoop/hdds/scm/cli/datanode/ListInfoSubcommand.java
 ##
 @@ -0,0 +1,117 @@
+/**
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ * 
+ * http://www.apache.org/licenses/LICENSE-2.0
+ * 
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+package org.apache.hadoop.hdds.scm.cli.datanode;
+
+import org.apache.hadoop.hdds.cli.HddsVersionProvider;
+import org.apache.hadoop.hdds.protocol.DatanodeDetails;
+import org.apache.hadoop.hdds.protocol.proto.HddsProtos;
+import org.apache.hadoop.hdds.scm.client.ScmClient;
+import org.apache.hadoop.hdds.scm.pipeline.Pipeline;
+import picocli.CommandLine;
+
+import java.io.IOException;
+import java.util.ArrayList;
+import java.util.List;
+import java.util.concurrent.Callable;
+import java.util.stream.Collectors;
+
+/**
+ * Handler of list datanodes info command.
+ */
+@CommandLine.Command(
+name = "list",
+description = "List info of datanodes",
+mixinStandardHelpOptions = true,
+versionProvider = HddsVersionProvider.class)
+public class ListInfoSubcommand implements Callable {
+
+  @CommandLine.ParentCommand
+  private DatanodeCommands parent;
+
+  @CommandLine.Option(names = {"-ip", "--byIp"},
+  description = "Show info by ip address.",
+  defaultValue = "",
+  required = false)
+  private String ipaddress;
+
+  @CommandLine.Option(names = {"-id", "--byUuid"},
 
 Review comment:
   Similarly, suggesting only `--uuid`.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[GitHub] [hadoop-ozone] adoroszlai commented on a change in pull request #348: HDDS-2035 Implement datanode level CLI to reveal pipeline relation.

2019-12-12 Thread GitBox
adoroszlai commented on a change in pull request #348: HDDS-2035 Implement 
datanode level CLI to reveal pipeline relation.
URL: https://github.com/apache/hadoop-ozone/pull/348#discussion_r357507729
 
 

 ##
 File path: 
hadoop-hdds/tools/src/main/java/org/apache/hadoop/hdds/scm/cli/datanode/ListInfoSubcommand.java
 ##
 @@ -0,0 +1,117 @@
+/**
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ * 
+ * http://www.apache.org/licenses/LICENSE-2.0
+ * 
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+package org.apache.hadoop.hdds.scm.cli.datanode;
+
+import org.apache.hadoop.hdds.cli.HddsVersionProvider;
+import org.apache.hadoop.hdds.protocol.DatanodeDetails;
+import org.apache.hadoop.hdds.protocol.proto.HddsProtos;
+import org.apache.hadoop.hdds.scm.client.ScmClient;
+import org.apache.hadoop.hdds.scm.pipeline.Pipeline;
+import picocli.CommandLine;
+
+import java.io.IOException;
+import java.util.ArrayList;
+import java.util.List;
+import java.util.concurrent.Callable;
+import java.util.stream.Collectors;
+
+/**
+ * Handler of list datanodes info command.
+ */
+@CommandLine.Command(
+name = "list",
+description = "List info of datanodes",
+mixinStandardHelpOptions = true,
+versionProvider = HddsVersionProvider.class)
+public class ListInfoSubcommand implements Callable {
+
+  @CommandLine.ParentCommand
+  private DatanodeCommands parent;
+
+  @CommandLine.Option(names = {"-ip", "--byIp"},
+  description = "Show info by ip address.",
+  defaultValue = "",
+  required = false)
+  private String ipaddress;
+
+  @CommandLine.Option(names = {"-id", "--byUuid"},
+  description = "Show info by datanode UUID.",
+  defaultValue = "",
+  required = false)
+  private String uuid;
+
+  private List pipelines;
+
+
+  @Override
+  public Void call() throws Exception {
+try (ScmClient scmClient = parent.getParent().createScmClient()) {
+  pipelines = scmClient.listPipelines();
+  if (isNullOrEmpty(ipaddress) && isNullOrEmpty(uuid)) {
+getAllNodes(scmClient).stream().forEach(p -> printDatanodeInfo(p));
+  } else {
+getAllNodes(scmClient).stream().filter(
+p -> ((isNullOrEmpty(ipaddress) ||
+(p.getIpAddress().compareToIgnoreCase(ipaddress) == 0))
+&& (isNullOrEmpty(uuid) ||
+(p.getUuid().equals(uuid)
+.forEach(p -> printDatanodeInfo(p));
+  }
+  return null;
+}
+  }
+
+  private List getAllNodes(ScmClient scmClient)
+  throws IOException {
+List nodes = scmClient.queryNode(
+HddsProtos.NodeState.HEALTHY, HddsProtos.QueryScope.CLUSTER, "");
+List datanodes = new ArrayList<>(nodes.size());
+nodes.stream().forEach(
+p -> datanodes.add(DatanodeDetails.getFromProtoBuf(p.getNodeID(;
+return datanodes;
+  }
+
+  private void printDatanodeInfo(DatanodeDetails datanode) {
+if (pipelines.isEmpty()) {
+  System.out.println("Datanode: " + datanode.getUuid().toString() +
+  " (" + datanode.getIpAddress() + "/"
+  + datanode.getHostName() + "). \n No pipeline created.");
+  return;
+}
+List relatedPipelines = pipelines.stream().filter(
+p -> p.getNodes().contains(datanode)).collect(Collectors.toList());
+StringBuilder pipelineList = new StringBuilder()
+.append("Related pipelines:\n");
+relatedPipelines.stream().forEach(
+p -> pipelineList.append(p.getId().getId().toString())
+.append("/").append(p.getFactor().toString()).append("/")
+.append(p.getType().toString()).append("/")
+.append(p.getPipelineState().toString()).append("/")
+.append(datanode.getUuid().equals(p.getLeaderId()) ?
+"Leader" : "Follower")
+.append(System.getProperty("line.separator")));
+System.out.println("Datanode: " + datanode.getUuid().toString() +
+" (" + datanode.getIpAddress() + "/"
++ datanode.getHostName() + "/" + relatedPipelines.size() +
+" pipelines). \n" + pipelineList);
 
 Review comment:
   Even if there are pipelines, the datanode may happen to be not part of any.  
Handling "no related pipelines" here would also allow removing the "No pipeline 
created" special case above.


[GitHub] [hadoop-ozone] adoroszlai commented on a change in pull request #348: HDDS-2035 Implement datanode level CLI to reveal pipeline relation.

2019-12-12 Thread GitBox
adoroszlai commented on a change in pull request #348: HDDS-2035 Implement 
datanode level CLI to reveal pipeline relation.
URL: https://github.com/apache/hadoop-ozone/pull/348#discussion_r357500392
 
 

 ##
 File path: 
hadoop-hdds/tools/src/main/java/org/apache/hadoop/hdds/scm/cli/datanode/ListInfoSubcommand.java
 ##
 @@ -0,0 +1,117 @@
+/**
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ * 
+ * http://www.apache.org/licenses/LICENSE-2.0
+ * 
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+package org.apache.hadoop.hdds.scm.cli.datanode;
+
+import org.apache.hadoop.hdds.cli.HddsVersionProvider;
+import org.apache.hadoop.hdds.protocol.DatanodeDetails;
+import org.apache.hadoop.hdds.protocol.proto.HddsProtos;
+import org.apache.hadoop.hdds.scm.client.ScmClient;
+import org.apache.hadoop.hdds.scm.pipeline.Pipeline;
+import picocli.CommandLine;
+
+import java.io.IOException;
+import java.util.ArrayList;
+import java.util.List;
+import java.util.concurrent.Callable;
+import java.util.stream.Collectors;
+
+/**
+ * Handler of list datanodes info command.
+ */
+@CommandLine.Command(
+name = "list",
+description = "List info of datanodes",
+mixinStandardHelpOptions = true,
+versionProvider = HddsVersionProvider.class)
+public class ListInfoSubcommand implements Callable {
 
 Review comment:
   `ListDatanodesSubcommand` might better reflect what this subcommand is about.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[GitHub] [hadoop-ozone] adoroszlai commented on a change in pull request #348: HDDS-2035 Implement datanode level CLI to reveal pipeline relation.

2019-12-12 Thread GitBox
adoroszlai commented on a change in pull request #348: HDDS-2035 Implement 
datanode level CLI to reveal pipeline relation.
URL: https://github.com/apache/hadoop-ozone/pull/348#discussion_r357502263
 
 

 ##
 File path: 
hadoop-hdds/tools/src/main/java/org/apache/hadoop/hdds/scm/cli/datanode/ListInfoSubcommand.java
 ##
 @@ -0,0 +1,117 @@
+/**
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ * 
+ * http://www.apache.org/licenses/LICENSE-2.0
+ * 
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+package org.apache.hadoop.hdds.scm.cli.datanode;
+
+import org.apache.hadoop.hdds.cli.HddsVersionProvider;
+import org.apache.hadoop.hdds.protocol.DatanodeDetails;
+import org.apache.hadoop.hdds.protocol.proto.HddsProtos;
+import org.apache.hadoop.hdds.scm.client.ScmClient;
+import org.apache.hadoop.hdds.scm.pipeline.Pipeline;
+import picocli.CommandLine;
+
+import java.io.IOException;
+import java.util.ArrayList;
+import java.util.List;
+import java.util.concurrent.Callable;
+import java.util.stream.Collectors;
+
+/**
+ * Handler of list datanodes info command.
+ */
+@CommandLine.Command(
+name = "list",
+description = "List info of datanodes",
+mixinStandardHelpOptions = true,
+versionProvider = HddsVersionProvider.class)
+public class ListInfoSubcommand implements Callable {
+
+  @CommandLine.ParentCommand
+  private DatanodeCommands parent;
+
+  @CommandLine.Option(names = {"-ip", "--byIp"},
 
 Review comment:
   I think `-` should be reserved for single-char options.  I suggest adding 
only `--ip` for simplicity.  It's only slightly longer than `-ip`, but let's us 
avoid confusion with combined single-char options.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[GitHub] [hadoop-ozone] adoroszlai commented on a change in pull request #348: HDDS-2035 Implement datanode level CLI to reveal pipeline relation.

2019-12-12 Thread GitBox
adoroszlai commented on a change in pull request #348: HDDS-2035 Implement 
datanode level CLI to reveal pipeline relation.
URL: https://github.com/apache/hadoop-ozone/pull/348#discussion_r357505137
 
 

 ##
 File path: 
hadoop-hdds/tools/src/main/java/org/apache/hadoop/hdds/scm/cli/datanode/ListInfoSubcommand.java
 ##
 @@ -0,0 +1,117 @@
+/**
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ * 
+ * http://www.apache.org/licenses/LICENSE-2.0
+ * 
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+package org.apache.hadoop.hdds.scm.cli.datanode;
+
+import org.apache.hadoop.hdds.cli.HddsVersionProvider;
+import org.apache.hadoop.hdds.protocol.DatanodeDetails;
+import org.apache.hadoop.hdds.protocol.proto.HddsProtos;
+import org.apache.hadoop.hdds.scm.client.ScmClient;
+import org.apache.hadoop.hdds.scm.pipeline.Pipeline;
+import picocli.CommandLine;
+
+import java.io.IOException;
+import java.util.ArrayList;
+import java.util.List;
+import java.util.concurrent.Callable;
+import java.util.stream.Collectors;
+
+/**
+ * Handler of list datanodes info command.
+ */
+@CommandLine.Command(
+name = "list",
+description = "List info of datanodes",
+mixinStandardHelpOptions = true,
+versionProvider = HddsVersionProvider.class)
+public class ListInfoSubcommand implements Callable {
+
+  @CommandLine.ParentCommand
+  private DatanodeCommands parent;
+
+  @CommandLine.Option(names = {"-ip", "--byIp"},
+  description = "Show info by ip address.",
+  defaultValue = "",
+  required = false)
+  private String ipaddress;
+
+  @CommandLine.Option(names = {"-id", "--byUuid"},
+  description = "Show info by datanode UUID.",
+  defaultValue = "",
+  required = false)
+  private String uuid;
+
+  private List pipelines;
+
+
+  @Override
+  public Void call() throws Exception {
+try (ScmClient scmClient = parent.getParent().createScmClient()) {
+  pipelines = scmClient.listPipelines();
+  if (isNullOrEmpty(ipaddress) && isNullOrEmpty(uuid)) {
+getAllNodes(scmClient).stream().forEach(p -> printDatanodeInfo(p));
+  } else {
+getAllNodes(scmClient).stream().filter(
+p -> ((isNullOrEmpty(ipaddress) ||
+(p.getIpAddress().compareToIgnoreCase(ipaddress) == 0))
+&& (isNullOrEmpty(uuid) ||
+(p.getUuid().equals(uuid)
+.forEach(p -> printDatanodeInfo(p));
+  }
+  return null;
+}
+  }
+
+  private List getAllNodes(ScmClient scmClient)
+  throws IOException {
+List nodes = scmClient.queryNode(
+HddsProtos.NodeState.HEALTHY, HddsProtos.QueryScope.CLUSTER, "");
+List datanodes = new ArrayList<>(nodes.size());
+nodes.stream().forEach(
+p -> datanodes.add(DatanodeDetails.getFromProtoBuf(p.getNodeID(;
+return datanodes;
 
 Review comment:
   Please try to avoid using `forEach` to fill an existing list, prefer 
collectors instead, eg. `stream().map(...).collect(toList())`.  It gives the 
JDK more chance to improve.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[jira] [Updated] (HDDS-2722) Let ChunkManager read/write ChunkBuffer instead of ByteBuffer

2019-12-12 Thread Attila Doroszlai (Jira)


 [ 
https://issues.apache.org/jira/browse/HDDS-2722?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Attila Doroszlai updated HDDS-2722:
---
Status: Patch Available  (was: In Progress)

> Let ChunkManager read/write ChunkBuffer instead of ByteBuffer
> -
>
> Key: HDDS-2722
> URL: https://issues.apache.org/jira/browse/HDDS-2722
> Project: Hadoop Distributed Data Store
>  Issue Type: Improvement
>  Components: Ozone Datanode
>Reporter: Attila Doroszlai
>Assignee: Attila Doroszlai
>Priority: Major
>  Labels: pull-request-available
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> Change {{ChunkManager}} read/write methods to accept/return {{ChunkBuffer}} 
> instead of {{ByteBuffer}}.  This allows seamlessly passing multiple buffers 
> without further interface change.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[jira] [Updated] (HDDS-2722) Let ChunkManager read/write ChunkBuffer instead of ByteBuffer

2019-12-12 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HDDS-2722?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated HDDS-2722:
-
Labels: pull-request-available  (was: )

> Let ChunkManager read/write ChunkBuffer instead of ByteBuffer
> -
>
> Key: HDDS-2722
> URL: https://issues.apache.org/jira/browse/HDDS-2722
> Project: Hadoop Distributed Data Store
>  Issue Type: Improvement
>  Components: Ozone Datanode
>Reporter: Attila Doroszlai
>Assignee: Attila Doroszlai
>Priority: Major
>  Labels: pull-request-available
>
> Change {{ChunkManager}} read/write methods to accept/return {{ChunkBuffer}} 
> instead of {{ByteBuffer}}.  This allows seamlessly passing multiple buffers 
> without further interface change.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[GitHub] [hadoop-ozone] adoroszlai opened a new pull request #352: HDDS-2722. Let ChunkManager read/write ChunkBuffer instead of ByteBuffer

2019-12-12 Thread GitBox
adoroszlai opened a new pull request #352: HDDS-2722. Let ChunkManager 
read/write ChunkBuffer instead of ByteBuffer
URL: https://github.com/apache/hadoop-ozone/pull/352
 
 
   ## What changes were proposed in this pull request?
   
   Change `ChunkManager` read/write methods to accept/return `ChunkBuffer` 
instead of `ByteBuffer`.  This allows seamlessly passing multiple buffers 
without further interface change.
   
   https://issues.apache.org/jira/browse/HDDS-2722
   
   ## How was this patch tested?
   
   Added unit test for new functionality in `ChunkBuffer`.  The rest is covered 
by existing tests.
   
   https://github.com/adoroszlai/hadoop-ozone/runs/346822935


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[jira] [Assigned] (HDDS-1590) Add per pipeline metrics

2019-12-12 Thread Xiaoyu Yao (Jira)


 [ 
https://issues.apache.org/jira/browse/HDDS-1590?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xiaoyu Yao reassigned HDDS-1590:


Assignee: Li Cheng

> Add per pipeline metrics
> 
>
> Key: HDDS-1590
> URL: https://issues.apache.org/jira/browse/HDDS-1590
> Project: Hadoop Distributed Data Store
>  Issue Type: Sub-task
>  Components: SCM
>Reporter: Siddharth Wagle
>Assignee: Li Cheng
>Priority: Major
>
> Add metrics on per pipeline basis:
> - bytes read/written
> - container metrics (state changes)



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[jira] [Updated] (HDDS-1575) Add support for storage policies to Pipelines

2019-12-12 Thread Xiaoyu Yao (Jira)


 [ 
https://issues.apache.org/jira/browse/HDDS-1575?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xiaoyu Yao updated HDDS-1575:
-
Parent: (was: HDDS-1564)
Issue Type: Improvement  (was: Sub-task)

> Add support for storage policies to Pipelines
> -
>
> Key: HDDS-1575
> URL: https://issues.apache.org/jira/browse/HDDS-1575
> Project: Hadoop Distributed Data Store
>  Issue Type: Improvement
>  Components: SCM
>Reporter: Siddharth Wagle
>Priority: Major
>
> The Pipeline level storage policy can be thought of segregating write 
> bandwidth for high throughput writes or random reads use cases when SSDs, 
> NVMe or RAM disks are involved by not having to figure out write pipelines on 
> the fly. The Datanode would need to read the pipeline policy and route write 
> to appropriate storage volumes. An Ozone mover can be provided for the 
> archival data use case. 
> Additionally, it makes sense to permeate the policy information to 
> Containers, this will help when allocateBlock call for a key with storage 
> policy set to ALL_SSD, to be routed to an open container on a Pipeline with 
> the same policy setting. The caveat is that at least one open pipeline needs 
> to be maintained at all times in order to support storage the policy setting. 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[jira] [Commented] (HDDS-1575) Add support for storage policies to Pipelines

2019-12-12 Thread Xiaoyu Yao (Jira)


[ 
https://issues.apache.org/jira/browse/HDDS-1575?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16995366#comment-16995366
 ] 

Xiaoyu Yao commented on HDDS-1575:
--

Move this out of multi-raft support. This can be implemented without 
multi-raft. 

> Add support for storage policies to Pipelines
> -
>
> Key: HDDS-1575
> URL: https://issues.apache.org/jira/browse/HDDS-1575
> Project: Hadoop Distributed Data Store
>  Issue Type: Sub-task
>  Components: SCM
>Reporter: Siddharth Wagle
>Priority: Major
>
> The Pipeline level storage policy can be thought of segregating write 
> bandwidth for high throughput writes or random reads use cases when SSDs, 
> NVMe or RAM disks are involved by not having to figure out write pipelines on 
> the fly. The Datanode would need to read the pipeline policy and route write 
> to appropriate storage volumes. An Ozone mover can be provided for the 
> archival data use case. 
> Additionally, it makes sense to permeate the policy information to 
> Containers, this will help when allocateBlock call for a key with storage 
> policy set to ALL_SSD, to be routed to an open container on a Pipeline with 
> the same policy setting. The caveat is that at least one open pipeline needs 
> to be maintained at all times in order to support storage the policy setting. 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[GitHub] [hadoop-ozone] avijayanhwx commented on issue #331: HDDS-2188. Implement LocatedFileStatus & getFileBlockLocations to pro…

2019-12-12 Thread GitBox
avijayanhwx commented on issue #331: HDDS-2188. Implement LocatedFileStatus & 
getFileBlockLocations to pro…
URL: https://github.com/apache/hadoop-ozone/pull/331#issuecomment-565305120
 
 
   The acceptance test failures are due to this patch. I am looking into it. 


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[jira] [Updated] (HDDS-2176) Add new pipeline state “CLOSING” and new CLOSE_PIPELINE_STATUS command

2019-12-12 Thread Xiaoyu Yao (Jira)


 [ 
https://issues.apache.org/jira/browse/HDDS-2176?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xiaoyu Yao updated HDDS-2176:
-
Parent: (was: HDDS-1564)
Issue Type: Improvement  (was: Sub-task)

> Add new pipeline state “CLOSING” and new CLOSE_PIPELINE_STATUS command
> --
>
> Key: HDDS-2176
> URL: https://issues.apache.org/jira/browse/HDDS-2176
> Project: Hadoop Distributed Data Store
>  Issue Type: Improvement
>Reporter: Sammi Chen
>Assignee: Sammi Chen
>Priority: Major
>
> Currrent pipeline has 3 states,  ALLOCATED, OPEN and CLOSED. 
> When create pipeline command is sent out to datanodes from SCM, this pipeline 
> is marked as ALLOCATED in SCM.  Once SCM received all 3 datanodes 
> confirmation of pipelinee creation, SCM will change pipeline's state from 
> ALLOCATED to OPEN.  
> Close pipeline process is similar.  Add a new CLOSING state to pipeline. When 
> close pipeline command is sent out to datanodes, pipeline will be marked as 
> CLOSING.  When all 3 datanode confirmed,  change pipeline state from CLOSING 
> to CLOSED.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[jira] [Updated] (HDDS-2034) Async RATIS pipeline creation and destroy through heartbeat commands

2019-12-12 Thread Xiaoyu Yao (Jira)


 [ 
https://issues.apache.org/jira/browse/HDDS-2034?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xiaoyu Yao updated HDDS-2034:
-
Parent: (was: HDDS-1564)
Issue Type: New Feature  (was: Sub-task)

> Async RATIS pipeline creation and destroy through heartbeat commands
> 
>
> Key: HDDS-2034
> URL: https://issues.apache.org/jira/browse/HDDS-2034
> Project: Hadoop Distributed Data Store
>  Issue Type: New Feature
>Reporter: Sammi Chen
>Assignee: Sammi Chen
>Priority: Major
>  Labels: pull-request-available
> Fix For: 0.4.2
>
>  Time Spent: 14h
>  Remaining Estimate: 0h
>
> Currently, pipeline creation and destroy are synchronous operations. SCM 
> directly connect to each datanode of the pipeline through gRPC channel to 
> create the pipeline to destroy the pipeline.  
> This task is to remove the gRPC channel, send pipeline creation and destroy 
> action through heartbeat command to each datanode.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[jira] [Comment Edited] (HDDS-2036) Multi-raft support on single datanode integration test

2019-12-12 Thread Xiaoyu Yao (Jira)


[ 
https://issues.apache.org/jira/browse/HDDS-2036?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16995363#comment-16995363
 ] 

Xiaoyu Yao edited comment on HDDS-2036 at 12/13/19 5:16 AM:


The test change is covered by HDDS-1569.


was (Author: xyao):
The test change is covered by HDDS-1576.

> Multi-raft support on single datanode integration test
> --
>
> Key: HDDS-2036
> URL: https://issues.apache.org/jira/browse/HDDS-2036
> Project: Hadoop Distributed Data Store
>  Issue Type: Sub-task
>Reporter: Sammi Chen
>Priority: Major
>
> Improve MiniOzoneCluster to support multi-raft group



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[jira] [Resolved] (HDDS-2036) Multi-raft support on single datanode integration test

2019-12-12 Thread Xiaoyu Yao (Jira)


 [ 
https://issues.apache.org/jira/browse/HDDS-2036?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xiaoyu Yao resolved HDDS-2036.
--
Resolution: Fixed

The test change is covered by HDDS-1576.

> Multi-raft support on single datanode integration test
> --
>
> Key: HDDS-2036
> URL: https://issues.apache.org/jira/browse/HDDS-2036
> Project: Hadoop Distributed Data Store
>  Issue Type: Sub-task
>Reporter: Sammi Chen
>Priority: Major
>
> Improve MiniOzoneCluster to support multi-raft group



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[jira] [Resolved] (HDDS-1573) Add scrubber metrics and pipeline metrics

2019-12-12 Thread Xiaoyu Yao (Jira)


 [ 
https://issues.apache.org/jira/browse/HDDS-1573?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xiaoyu Yao resolved HDDS-1573.
--
Resolution: Implemented

There metrics have been implemented as part of the feature. 

> Add scrubber metrics and pipeline metrics
> -
>
> Key: HDDS-1573
> URL: https://issues.apache.org/jira/browse/HDDS-1573
> Project: Hadoop Distributed Data Store
>  Issue Type: Sub-task
>  Components: SCM
>Reporter: Siddharth Wagle
>Assignee: Li Cheng
>Priority: Major
>
> - Add metrics for how many pipelines per datanode
> - Add metrics for pipelines that are chosen by the scrubber
> - Add metrics for pipelines that are in violation



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[jira] [Commented] (HDDS-1590) Add per pipeline metrics

2019-12-12 Thread Xiaoyu Yao (Jira)


[ 
https://issues.apache.org/jira/browse/HDDS-1590?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16995357#comment-16995357
 ] 

Xiaoyu Yao commented on HDDS-1590:
--

TODO: add SCM per pipeline bytes write metrics.

> Add per pipeline metrics
> 
>
> Key: HDDS-1590
> URL: https://issues.apache.org/jira/browse/HDDS-1590
> Project: Hadoop Distributed Data Store
>  Issue Type: Sub-task
>  Components: SCM
>Reporter: Siddharth Wagle
>Priority: Major
>
> Add metrics on per pipeline basis:
> - bytes read/written
> - container metrics (state changes)



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[GitHub] [hadoop-ozone] adoroszlai commented on issue #322: HDDS-2684. Refactor common test utilities to hadoop-hdds/common

2019-12-12 Thread GitBox
adoroszlai commented on issue #322: HDDS-2684. Refactor common test utilities 
to hadoop-hdds/common
URL: https://github.com/apache/hadoop-ozone/pull/322#issuecomment-565287537
 
 
   Thanks @arp7 for reviewing and committing it.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[jira] [Updated] (HDDS-2684) Refactor common test utilities to hadoop-hdds/common

2019-12-12 Thread Arpit Agarwal (Jira)


 [ 
https://issues.apache.org/jira/browse/HDDS-2684?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arpit Agarwal updated HDDS-2684:

   Fix Version/s: 0.5.0
Target Version/s:   (was: 0.5.0)
  Resolution: Fixed
  Status: Resolved  (was: Patch Available)

+1

Committed via GitHub. Thanks for the contribution [~adoroszlai] and thanks for 
the review [~xyao].

> Refactor common test utilities to hadoop-hdds/common
> 
>
> Key: HDDS-2684
> URL: https://issues.apache.org/jira/browse/HDDS-2684
> Project: Hadoop Distributed Data Store
>  Issue Type: Improvement
>  Components: test
>Reporter: Attila Doroszlai
>Assignee: Attila Doroszlai
>Priority: Major
>  Labels: pull-request-available
> Fix For: 0.5.0
>
>  Time Spent: 20m
>  Remaining Estimate: 0h
>
> Expose test code from {{hadoop-hdds/common}} to other modules.  Move some 
> "common" test utilities.  Example: random {{DatanodeDetails}} creation.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[GitHub] [hadoop-ozone] arp7 merged pull request #322: HDDS-2684. Refactor common test utilities to hadoop-hdds/common

2019-12-12 Thread GitBox
arp7 merged pull request #322: HDDS-2684. Refactor common test utilities to 
hadoop-hdds/common
URL: https://github.com/apache/hadoop-ozone/pull/322
 
 
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[GitHub] [hadoop-ozone] arp7 commented on issue #322: HDDS-2684. Refactor common test utilities to hadoop-hdds/common

2019-12-12 Thread GitBox
arp7 commented on issue #322: HDDS-2684. Refactor common test utilities to 
hadoop-hdds/common
URL: https://github.com/apache/hadoop-ozone/pull/322#issuecomment-565259614
 
 
   +1
   
   Looks like Xiaoyu's comments are addressed. Xiaoyu - if not we can address 
in a followup Jira.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[GitHub] [hadoop-ozone] bharatviswa504 opened a new pull request #351: HDDS-1989. Fix ApplyTransaction error handling in OzoneManagerStateMachine.

2019-12-12 Thread GitBox
bharatviswa504 opened a new pull request #351: HDDS-1989. Fix ApplyTransaction 
error handling in OzoneManagerStateMachine.
URL: https://github.com/apache/hadoop-ozone/pull/351
 
 
   
   
   ## What changes were proposed in this pull request?
   
   Terminate OM, after a critical error in apply transaction. Right now 
terminate for INTERNAL_ERROR and METADATA_ERROR.
   
   ## What is the link to the Apache JIRA
   
   https://issues.apache.org/jira/browse/HDDS-1989
   
   Please replace this section with the link to the Apache JIRA)
   
   ## How was this patch tested?
   
   Need to see if any test for this failure scenario can be added. Mostly will 
try to come up with UT once HDDS-2682 merges in.
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[jira] [Updated] (HDDS-1989) Fix ApplyTransaction error handling in OzoneManagerStateMachine

2019-12-12 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HDDS-1989?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated HDDS-1989:
-
Labels: pull-request-available  (was: )

> Fix ApplyTransaction error handling in OzoneManagerStateMachine
> ---
>
> Key: HDDS-1989
> URL: https://issues.apache.org/jira/browse/HDDS-1989
> Project: Hadoop Distributed Data Store
>  Issue Type: Sub-task
>Reporter: Bharat Viswanadham
>Priority: Major
>  Labels: pull-request-available
>
> Right now, applyTransaction calls validateAndUpdateCache. This does not 
> return any error. We should check the OmResponse Status code, and see if it 
> is critical error we should completeExceptionally and stop ratis server.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[GitHub] [hadoop-ozone] bharatviswa504 closed pull request #350: WIP: HDDS-1989. Fix ApplyTransaction error handling in OzoneManagerStateMachine

2019-12-12 Thread GitBox
bharatviswa504 closed pull request #350: WIP: HDDS-1989. Fix ApplyTransaction 
error handling in OzoneManagerStateMachine
URL: https://github.com/apache/hadoop-ozone/pull/350
 
 
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[GitHub] [hadoop-ozone] bharatviswa504 opened a new pull request #350: WIP: HDDS-1989. Fix ApplyTransaction error handling in OzoneManagerStateMachine

2019-12-12 Thread GitBox
bharatviswa504 opened a new pull request #350: WIP: HDDS-1989. Fix 
ApplyTransaction error handling in OzoneManagerStateMachine
URL: https://github.com/apache/hadoop-ozone/pull/350
 
 
   ## What changes were proposed in this pull request?
   
   Right now, applyTransaction calls validateAndUpdateCache. This does not 
return any error. We should check the OmResponse Status code, and see if it is 
critical error we should stop OM.
   
   
   https://issues.apache.org/jira/browse/HDDS-1989
   
   ## How was this patch tested?
   
   Existing HA tests ran, need to add any UT for this scenario.
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[jira] [Updated] (HDDS-2188) Implement LocatedFileStatus & getFileBlockLocations to provide node/localization information to Yarn/Mapreduce

2019-12-12 Thread Aravindan Vijayan (Jira)


 [ 
https://issues.apache.org/jira/browse/HDDS-2188?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Aravindan Vijayan updated HDDS-2188:

Status: Patch Available  (was: In Progress)

> Implement LocatedFileStatus & getFileBlockLocations to provide 
> node/localization information to Yarn/Mapreduce
> --
>
> Key: HDDS-2188
> URL: https://issues.apache.org/jira/browse/HDDS-2188
> Project: Hadoop Distributed Data Store
>  Issue Type: Bug
>  Components: Ozone Filesystem
>Affects Versions: 0.5.0
>Reporter: Mukul Kumar Singh
>Assignee: Aravindan Vijayan
>Priority: Major
>  Labels: pull-request-available
>  Time Spent: 1h 40m
>  Remaining Estimate: 0h
>
> For applications like Hive/MapReduce to take advantage of the data locality 
> in Ozone, Ozone should return the location of the Ozone blocks. This is 
> needed for better read performance for Hadoop Applications.
> {code}
> if (file instanceof LocatedFileStatus) {
>   blkLocations = ((LocatedFileStatus) file).getBlockLocations();
> } else {
>   blkLocations = fs.getFileBlockLocations(file, 0, length);
> }
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[jira] [Assigned] (HDDS-2714) Add Ratis leader info for scm pipeline info command

2019-12-12 Thread Siddharth Wagle (Jira)


 [ 
https://issues.apache.org/jira/browse/HDDS-2714?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Siddharth Wagle reassigned HDDS-2714:
-

Assignee: Siddharth Wagle

> Add Ratis leader info for scm pipeline info command
> ---
>
> Key: HDDS-2714
> URL: https://issues.apache.org/jira/browse/HDDS-2714
> Project: Hadoop Distributed Data Store
>  Issue Type: Bug
>  Components: SCM
>Affects Versions: 0.5.0
>Reporter: Shashikant Banerjee
>Assignee: Siddharth Wagle
>Priority: Major
> Fix For: 0.5.0
>
>
> SCMCli prints the datanode info with the pipeline ID. it would be of great 
> help in debugging if it also displays the ratis leader info along with it as 
> this info is already populated by datanode to scm.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[jira] [Assigned] (HDDS-2716) ReadStateMachine failure should close down the ratis pipeline

2019-12-12 Thread Siddharth Wagle (Jira)


 [ 
https://issues.apache.org/jira/browse/HDDS-2716?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Siddharth Wagle reassigned HDDS-2716:
-

Assignee: Siddharth Wagle

> ReadStateMachine failure should close down the ratis pipeline 
> --
>
> Key: HDDS-2716
> URL: https://issues.apache.org/jira/browse/HDDS-2716
> Project: Hadoop Distributed Data Store
>  Issue Type: Bug
>  Components: Ozone Datanode
>Affects Versions: 0.5.0
>Reporter: Shashikant Banerjee
>Assignee: Siddharth Wagle
>Priority: Major
> Fix For: 0.5.0
>
>
> Currently, in case of readStateMachineFailures, the containerStateMachine is 
> marked unhealthy which stalls all the pipeline operations. But the pipeline 
> will get destroyed in the next snapshot execution. The idea here is to just 
> destroy the pipeline as soon as readStateMachine call fails.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[jira] [Commented] (HDDS-2703) OzoneFSInputStream to support ByteBufferReadable

2019-12-12 Thread YiSheng Lien (Jira)


[ 
https://issues.apache.org/jira/browse/HDDS-2703?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16995169#comment-16995169
 ] 

YiSheng Lien commented on HDDS-2703:


Note:
Cause the current dependency version of Hadoop in Ozone is 3.2.0, 

I don't implement {{ByteBufferPositionedReadable}}  for now. (Supported in 
Hadoop 3.3.0)

The same is {{StreamCapabilities}}, the {{StreamCapabilities}} support 
{{StreamCapabilities.READBYTEBUFFER}} for Hadoop 3.3.0 now.

Should we create a JIRA for these ?

> OzoneFSInputStream to support ByteBufferReadable
> 
>
> Key: HDDS-2703
> URL: https://issues.apache.org/jira/browse/HDDS-2703
> Project: Hadoop Distributed Data Store
>  Issue Type: Improvement
>Reporter: Wei-Chiu Chuang
>Assignee: YiSheng Lien
>Priority: Major
>  Labels: pull-request-available
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> This was found by [~cxorm] via HDDS-2443.
> ByteBufferReadable could help certain application performance, such as 
> Impala. (See HDFS-14111)
> Additionally, if we support ByteBufferPositionedReadable, it would benefit 
> HBase. (see HDFS-3246)
> Finally, we should add StreamCapabilities to let client probe for these 
> abilities. (See HDFS-11644)



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[jira] [Comment Edited] (HDDS-2703) OzoneFSInputStream to support ByteBufferReadable

2019-12-12 Thread YiSheng Lien (Jira)


[ 
https://issues.apache.org/jira/browse/HDDS-2703?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16995169#comment-16995169
 ] 

YiSheng Lien edited comment on HDDS-2703 at 12/12/19 10:41 PM:
---

Note:
Cause the current dependency version of Hadoop in Ozone is 3.2.0, 

I don't implement {{ByteBufferPositionedReadable}}  for now. (Supported in 
Hadoop 3.3.0)

The same is {{StreamCapabilities}}, the {{StreamCapabilities}} support 
{{StreamCapabilities.READBYTEBUFFER}} only for Hadoop 3.3.0 now.

Should we create a JIRA for these ?


was (Author: cxorm):
Note:
Cause the current dependency version of Hadoop in Ozone is 3.2.0, 

I don't implement {{ByteBufferPositionedReadable}}  for now. (Supported in 
Hadoop 3.3.0)

The same is {{StreamCapabilities}}, the {{StreamCapabilities}} support 
{{StreamCapabilities.READBYTEBUFFER}} for Hadoop 3.3.0 now.

Should we create a JIRA for these ?

> OzoneFSInputStream to support ByteBufferReadable
> 
>
> Key: HDDS-2703
> URL: https://issues.apache.org/jira/browse/HDDS-2703
> Project: Hadoop Distributed Data Store
>  Issue Type: Improvement
>Reporter: Wei-Chiu Chuang
>Assignee: YiSheng Lien
>Priority: Major
>  Labels: pull-request-available
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> This was found by [~cxorm] via HDDS-2443.
> ByteBufferReadable could help certain application performance, such as 
> Impala. (See HDFS-14111)
> Additionally, if we support ByteBufferPositionedReadable, it would benefit 
> HBase. (see HDFS-3246)
> Finally, we should add StreamCapabilities to let client probe for these 
> abilities. (See HDFS-11644)



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[GitHub] [hadoop-ozone] sodonnel commented on issue #309: HDDS-2593. DatanodeAdminMonitor should track under replicated containers and complete the admin workflow accordingly

2019-12-12 Thread GitBox
sodonnel commented on issue #309: HDDS-2593. DatanodeAdminMonitor should track 
under replicated containers and complete the admin workflow accordingly
URL: https://github.com/apache/hadoop-ozone/pull/309#issuecomment-565213144
 
 
   That test is still failing event after merging master to the branch. I 
though these commits would have fixed it, but I was confusing them with 
something else:
   
   ```
   Addendum HDDS-2646. Start acceptance tests only if at least one THREE 
pipeline is available (#282)
   HDDS-2646. Start acceptance tests only if at least one THREE pipeline is 
available (#282)
   ```
   I think this is a problem with the MockNodeManager handing the new 
operational states on this branch. I will look at it further.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[GitHub] [hadoop-ozone] cxorm commented on a change in pull request #345: HDDS-2703. OzoneFSInputStream to support ByteBufferReadable

2019-12-12 Thread GitBox
cxorm commented on a change in pull request #345: HDDS-2703. OzoneFSInputStream 
to support ByteBufferReadable
URL: https://github.com/apache/hadoop-ozone/pull/345#discussion_r357394244
 
 

 ##
 File path: 
hadoop-ozone/ozonefs/src/main/java/org/apache/hadoop/fs/ozone/OzoneFSInputStream.java
 ##
 @@ -76,4 +79,21 @@ public boolean seekToNewSource(long targetPos) throws 
IOException {
   public int available() throws IOException {
 return inputStream.available();
   }
+
+  /**
+   * @param buf the ByteBuffer to receive the results of the read operation.
+   * @return the number of bytes read, possibly zero.
 
 Review comment:
   Thanks @jojochuang for the review.
   Fixed the description, and updated the implementation. 


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[GitHub] [hadoop-ozone] cxorm commented on a change in pull request #345: HDDS-2703. OzoneFSInputStream to support ByteBufferReadable

2019-12-12 Thread GitBox
cxorm commented on a change in pull request #345: HDDS-2703. OzoneFSInputStream 
to support ByteBufferReadable
URL: https://github.com/apache/hadoop-ozone/pull/345#discussion_r357393478
 
 

 ##
 File path: 
hadoop-ozone/ozonefs/src/main/java/org/apache/hadoop/fs/ozone/OzoneFSInputStream.java
 ##
 @@ -76,4 +79,21 @@ public boolean seekToNewSource(long targetPos) throws 
IOException {
   public int available() throws IOException {
 return inputStream.available();
   }
+
+  /**
+   * @param buf the ByteBuffer to receive the results of the read operation.
+   * @return the number of bytes read, possibly zero.
+   * @throws IOException if there is some error performing the read
+   */
+  @Override
+  public int read(ByteBuffer buf) throws IOException {
+
+int bufInitPos = buf.position();
+while (buf.hasRemaining() && (inputStream.available() > 0)) {
 
 Review comment:
   Thanks @bharatviswa504 for the review.
   Fixed, and work well.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[GitHub] [hadoop-ozone] cxorm commented on a change in pull request #345: HDDS-2703. OzoneFSInputStream to support ByteBufferReadable

2019-12-12 Thread GitBox
cxorm commented on a change in pull request #345: HDDS-2703. OzoneFSInputStream 
to support ByteBufferReadable
URL: https://github.com/apache/hadoop-ozone/pull/345#discussion_r357393104
 
 

 ##
 File path: 
hadoop-ozone/ozonefs/src/main/java/org/apache/hadoop/fs/ozone/OzoneFSInputStream.java
 ##
 @@ -76,4 +81,39 @@ public boolean seekToNewSource(long targetPos) throws 
IOException {
   public int available() throws IOException {
 return inputStream.available();
   }
+
+  /**
+   * @param buf the ByteBuffer to receive the results of the read operation.
+   * @return the number of bytes read, possibly zero, or -1 if
+   * reach end-of-stream
+   * @throws IOException if there is some error performing the read
+   */
+  @Override
+  public int read(ByteBuffer buf) throws IOException {
+
+int bufInitPos = buf.position();
+int readLen = Math.min(buf.remaining(), inputStream.available());
+
+byte[] readData = new byte[readLen];
+int bytesRead = inputStream.read(readData, bufInitPos, readLen);
+buf.put(readData);
+
+return bytesRead;
+  }
+
+  /**
+   * Query the stream for a specific capability.
+   *
+   * @param capability string to query the stream support for.
+   * @return True if the stream supports capability.
+   */
+  @Override
+  public boolean hasCapability(String capability) {
+switch (StringUtils.toLowerCase(capability)) {
+case "in:readbytebuffer":
 
 Review comment:
   This case should be ```StreamCapabilities.READBYTEBUFFER``` that is 
supported in Hadoop 3.3.0 for now.
   I use ```in:readbytebuffer``` here cause the dependency version of hadoop in 
Ozone is 3.2.0.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[GitHub] [hadoop-ozone] jojochuang commented on a change in pull request #345: HDDS-2703. OzoneFSInputStream to support ByteBufferReadable

2019-12-12 Thread GitBox
jojochuang commented on a change in pull request #345: HDDS-2703. 
OzoneFSInputStream to support ByteBufferReadable
URL: https://github.com/apache/hadoop-ozone/pull/345#discussion_r357386990
 
 

 ##
 File path: 
hadoop-ozone/ozonefs/src/main/java/org/apache/hadoop/fs/ozone/OzoneFSInputStream.java
 ##
 @@ -76,4 +79,21 @@ public boolean seekToNewSource(long targetPos) throws 
IOException {
   public int available() throws IOException {
 return inputStream.available();
   }
+
+  /**
+   * @param buf the ByteBuffer to receive the results of the read operation.
+   * @return the number of bytes read, possibly zero.
 
 Review comment:
   According to the ByteBufferReadable javadoc, it should return -1 if end of 
file.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[jira] [Commented] (HDDS-2443) Python client/interface for Ozone

2019-12-12 Thread Wei-Chiu Chuang (Jira)


[ 
https://issues.apache.org/jira/browse/HDDS-2443?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16995138#comment-16995138
 ] 

Wei-Chiu Chuang commented on HDDS-2443:
---

Thanks for the doc. It looks so easy to support Python clients!
Could you make the instruction under Ozone Recipes? 
https://hadoop.apache.org/ozone/docs/0.4.1-alpha/recipe.html

> Python client/interface for Ozone
> -
>
> Key: HDDS-2443
> URL: https://issues.apache.org/jira/browse/HDDS-2443
> Project: Hadoop Distributed Data Store
>  Issue Type: New Feature
>  Components: Ozone Client
>Reporter: Li Cheng
>Priority: Major
> Attachments: Ozone with pyarrow.html, Ozone with pyarrow.odt, 
> OzoneS3.py
>
>
> Original ideas: item#25 in 
> [https://cwiki.apache.org/confluence/display/HADOOP/Ozone+project+ideas+for+new+contributors]
> Ozone Client(Python) for Data Science Notebook such as Jupyter.
>  # Size: Large
>  # PyArrow: [https://pypi.org/project/pyarrow/]
>  # Python -> libhdfs HDFS JNI library (HDFS, S3,...) -> Java client API 
> Impala uses  libhdfs
>  
> Path to try:
> # s3 interface: Ozone s3 gateway(already supported) + AWS python client 
> (boto3)
> # python native RPC
> # pyarrow + libhdfs, which use the Java client under the hood.
> # python + C interface of go / rust ozone library. I created POC go / rust 
> clients earlier which can be improved if the libhdfs interface is not good 
> enough. [By [~elek]]



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[jira] [Updated] (HDDS-2723) Wrap filesystem.open() with try-catch in TestOzoneFSInputStream

2019-12-12 Thread YiSheng Lien (Jira)


 [ 
https://issues.apache.org/jira/browse/HDDS-2723?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

YiSheng Lien updated HDDS-2723:
---
Description: 
Thanks [~weichiu] for the proposal.

To ensure resource is closed properly, we could wrap the fs.open() in a try.
The detail is on https://github.com/apache/hadoop-ozone/pull/345.



  was:
Thanks [~weichiu] for the proposal.

The detail is on https://github.com/apache/hadoop-ozone/pull/345.




> Wrap filesystem.open() with try-catch in TestOzoneFSInputStream
> ---
>
> Key: HDDS-2723
> URL: https://issues.apache.org/jira/browse/HDDS-2723
> Project: Hadoop Distributed Data Store
>  Issue Type: Improvement
>  Components: Ozone Filesystem
>Reporter: YiSheng Lien
>Assignee: YiSheng Lien
>Priority: Major
>
> Thanks [~weichiu] for the proposal.
> To ensure resource is closed properly, we could wrap the fs.open() in a try.
> The detail is on https://github.com/apache/hadoop-ozone/pull/345.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[jira] [Updated] (HDDS-2723) Wrap filesystem.open() with try-catch in TestOzoneFSInputStream

2019-12-12 Thread YiSheng Lien (Jira)


 [ 
https://issues.apache.org/jira/browse/HDDS-2723?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

YiSheng Lien updated HDDS-2723:
---
Summary: Wrap filesystem.open() with try-catch in TestOzoneFSInputStream  
(was: Wrap filesystem.open() with try-catch to ensure resource is closed 
properly)

> Wrap filesystem.open() with try-catch in TestOzoneFSInputStream
> ---
>
> Key: HDDS-2723
> URL: https://issues.apache.org/jira/browse/HDDS-2723
> Project: Hadoop Distributed Data Store
>  Issue Type: Improvement
>  Components: Ozone Filesystem
>Reporter: YiSheng Lien
>Assignee: YiSheng Lien
>Priority: Major
>
> Thanks [~weichiu] for the proposal.
> The detail is on https://github.com/apache/hadoop-ozone/pull/345.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[GitHub] [hadoop-ozone] cxorm commented on a change in pull request #345: HDDS-2703. OzoneFSInputStream to support ByteBufferReadable

2019-12-12 Thread GitBox
cxorm commented on a change in pull request #345: HDDS-2703. OzoneFSInputStream 
to support ByteBufferReadable
URL: https://github.com/apache/hadoop-ozone/pull/345#discussion_r357364575
 
 

 ##
 File path: 
hadoop-ozone/ozonefs/src/test/java/org/apache/hadoop/fs/ozone/TestOzoneFSInputStream.java
 ##
 @@ -134,4 +135,20 @@ public void testO3FSMultiByteRead() throws IOException {
 Assert.assertTrue(Arrays.equals(value, data));
 inputStream.close();
   }
+
+  @Test
+  public void testO3FSByteBufferRead() throws IOException {
+FSDataInputStream inputStream = fs.open(filePath);
 
 Review comment:
   Thanks @jojochuang for the suggest.
   JIRA created for this issue. 
([HDDS-2723](https://issues.apache.org/jira/browse/HDDS-2723))


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[jira] [Updated] (HDDS-2723) Wrap filesystem.open() with try-catch to ensure resource is closed properly

2019-12-12 Thread YiSheng Lien (Jira)


 [ 
https://issues.apache.org/jira/browse/HDDS-2723?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

YiSheng Lien updated HDDS-2723:
---
External issue ID:   (was: HDDS-2703)

> Wrap filesystem.open() with try-catch to ensure resource is closed properly
> ---
>
> Key: HDDS-2723
> URL: https://issues.apache.org/jira/browse/HDDS-2723
> Project: Hadoop Distributed Data Store
>  Issue Type: Improvement
>  Components: Ozone Filesystem
> Environment: [PR of HDDS-2703|http://example.com]
>Reporter: YiSheng Lien
>Assignee: YiSheng Lien
>Priority: Major
>
> Thanks [~weichiu] for the proposal.
> The detail is on https://github.com/apache/hadoop-ozone/pull/345.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[jira] [Updated] (HDDS-2723) Wrap filesystem.open() with try-catch to ensure resource is closed properly

2019-12-12 Thread YiSheng Lien (Jira)


 [ 
https://issues.apache.org/jira/browse/HDDS-2723?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

YiSheng Lien updated HDDS-2723:
---
Environment: (was: [PR of HDDS-2703|http://example.com])

> Wrap filesystem.open() with try-catch to ensure resource is closed properly
> ---
>
> Key: HDDS-2723
> URL: https://issues.apache.org/jira/browse/HDDS-2723
> Project: Hadoop Distributed Data Store
>  Issue Type: Improvement
>  Components: Ozone Filesystem
>Reporter: YiSheng Lien
>Assignee: YiSheng Lien
>Priority: Major
>
> Thanks [~weichiu] for the proposal.
> The detail is on https://github.com/apache/hadoop-ozone/pull/345.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[jira] [Created] (HDDS-2723) Wrap filesystem.open() with try-catch to ensure resource is closed properly

2019-12-12 Thread YiSheng Lien (Jira)
YiSheng Lien created HDDS-2723:
--

 Summary: Wrap filesystem.open() with try-catch to ensure resource 
is closed properly
 Key: HDDS-2723
 URL: https://issues.apache.org/jira/browse/HDDS-2723
 Project: Hadoop Distributed Data Store
  Issue Type: Improvement
  Components: Ozone Filesystem
 Environment: [PR of HDDS-2703|http://example.com]
Reporter: YiSheng Lien
Assignee: YiSheng Lien


Thanks [~weichiu] for the proposal.

The detail is on https://github.com/apache/hadoop-ozone/pull/345.





--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[GitHub] [hadoop-ozone] cxorm commented on a change in pull request #329: HDDS-2423. Add the recover-trash command client side code

2019-12-12 Thread GitBox
cxorm commented on a change in pull request #329: HDDS-2423. Add the 
recover-trash command client side code
URL: https://github.com/apache/hadoop-ozone/pull/329#discussion_r357352410
 
 

 ##
 File path: 
hadoop-ozone/client/src/main/java/org/apache/hadoop/ozone/client/rpc/RpcClient.java
 ##
 @@ -720,6 +720,14 @@ public void renameKey(String volumeName, String 
bucketName,
 keyPrefix, maxKeys);
   }
 
+  @Override
+  public boolean recoverTrash(String volumeName, String bucketName,
+  String keyName, String destinationName) throws IOException {
 
 Review comment:
   Thanks @xiaoyuyao for this NIT.
   Fixed.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[GitHub] [hadoop-ozone] cxorm commented on a change in pull request #329: HDDS-2423. Add the recover-trash command client side code

2019-12-12 Thread GitBox
cxorm commented on a change in pull request #329: HDDS-2423. Add the 
recover-trash command client side code
URL: https://github.com/apache/hadoop-ozone/pull/329#discussion_r357350576
 
 

 ##
 File path: 
hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/OzoneManager.java
 ##
 @@ -2238,6 +2238,15 @@ public void deleteKey(OmKeyArgs args) throws 
IOException {
 }
   }
 
+  // TODO: HDDS-2424. recover-trash command server side handling.
+  @Override
+  public boolean recoverTrash(String volumeName, String bucketName,
+  String keyName, String destinationBucket) throws IOException {
+
+boolean recoverOperation = true;
+return recoverOperation;
 
 Review comment:
   Thanks @xiaoyuyao for the review.
   
   I updated the UT that is prototype for now,
   and it can be extended for the following server-side development.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[GitHub] [hadoop-ozone] jojochuang commented on a change in pull request #345: HDDS-2703. OzoneFSInputStream to support ByteBufferReadable

2019-12-12 Thread GitBox
jojochuang commented on a change in pull request #345: HDDS-2703. 
OzoneFSInputStream to support ByteBufferReadable
URL: https://github.com/apache/hadoop-ozone/pull/345#discussion_r357315867
 
 

 ##
 File path: 
hadoop-ozone/ozonefs/src/test/java/org/apache/hadoop/fs/ozone/TestOzoneFSInputStream.java
 ##
 @@ -134,4 +135,20 @@ public void testO3FSMultiByteRead() throws IOException {
 Assert.assertTrue(Arrays.equals(value, data));
 inputStream.close();
   }
+
+  @Test
+  public void testO3FSByteBufferRead() throws IOException {
+FSDataInputStream inputStream = fs.open(filePath);
 
 Review comment:
   i would really love to see this wrapped in a try with block to ensure 
resource is closed properly. But I see that all test cases in the file follows 
the same pattern so maybe we can have a separate jira to correct that.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[jira] [Created] (HDDS-2722) Let ChunkManager read/write ChunkBuffer instead of ByteBuffer

2019-12-12 Thread Attila Doroszlai (Jira)
Attila Doroszlai created HDDS-2722:
--

 Summary: Let ChunkManager read/write ChunkBuffer instead of 
ByteBuffer
 Key: HDDS-2722
 URL: https://issues.apache.org/jira/browse/HDDS-2722
 Project: Hadoop Distributed Data Store
  Issue Type: Improvement
  Components: Ozone Datanode
Reporter: Attila Doroszlai
Assignee: Attila Doroszlai


Change {{ChunkManager}} read/write methods to accept/return {{ChunkBuffer}} 
instead of {{ByteBuffer}}.  This allows seamlessly passing multiple buffers 
without further interface change.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[jira] [Created] (HDDS-2721) Allow wrapping list of ByteBuffers with ChunkBuffer

2019-12-12 Thread Attila Doroszlai (Jira)
Attila Doroszlai created HDDS-2721:
--

 Summary: Allow wrapping list of ByteBuffers with ChunkBuffer
 Key: HDDS-2721
 URL: https://issues.apache.org/jira/browse/HDDS-2721
 Project: Hadoop Distributed Data Store
  Issue Type: Improvement
Reporter: Attila Doroszlai
Assignee: Attila Doroszlai


{{ChunkBuffer}} is a useful abstraction over {{ByteBuffer}} to hide whether 
it's backed by a single {{ByteBuffer}} or multiple ones 
({{IncrementalChunkBuffer}}).  However, {{IncrementalChunkBuffer}} allocates 
its own {{ByteBuffer}} instances and only works with uniform buffer sizes.  The 
goal of this task is to allow wrapping an existing  {{List}} in 
{{ChunkBuffer}}.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[jira] [Created] (HDDS-2720) Ozone Failure injection Service

2019-12-12 Thread Prashant Pogde (Jira)
Prashant Pogde created HDDS-2720:


 Summary: Ozone Failure injection Service
 Key: HDDS-2720
 URL: https://issues.apache.org/jira/browse/HDDS-2720
 Project: Hadoop Distributed Data Store
  Issue Type: New Feature
  Components: Ozone Datanode, Ozone Filesystem, Ozone Manager, Ozone 
Recon, SCM
Affects Versions: 0.5.0
Reporter: Prashant Pogde
Assignee: Prashant Pogde
 Fix For: 0.5.0
 Attachments: OzoneNoiseInjection.pdf

This will be used to track development for failure injection service. This can 
be used to inject various failures/delays in an ozone cluster and validate 
ozone in presence of these failures or extreme conditions.

Attached document provides a brief overview for this failure injection service 
and how it could be leveraged to validate ozone in stressful environments.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[jira] [Updated] (HDDS-2719) Rename S3Token to S3AuthInfo

2019-12-12 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HDDS-2719?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated HDDS-2719:
-
Labels: pull-request-available  (was: )

> Rename S3Token to S3AuthInfo
> 
>
> Key: HDDS-2719
> URL: https://issues.apache.org/jira/browse/HDDS-2719
> Project: Hadoop Distributed Data Store
>  Issue Type: Improvement
>  Components: S3
>Reporter: Mohammad
>Assignee: Mohammad
>Priority: Minor
>  Labels: pull-request-available
> Fix For: 0.5.0
>
>
> Rename S3Token to something like S3AuthInfo
> Item #24 in the list of ideas for new contributors to work on
> https://cwiki.apache.org/confluence/display/HADOOP/Ozone+project+ideas+for+new+contributors



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[GitHub] [hadoop-ozone] mohammadjkhan opened a new pull request #349: HDDS-2719. Rename S3Token to S3AuthInfo

2019-12-12 Thread GitBox
mohammadjkhan opened a new pull request #349: HDDS-2719. Rename S3Token to 
S3AuthInfo
URL: https://github.com/apache/hadoop-ozone/pull/349
 
 
   ## What changes were proposed in this pull request?
   Renamed references of S3TOKEN to S3AUTHINFO
   
   ## What is the link to the Apache JIRA
   https://issues.apache.org/jira/browse/HDDS-2719
   
   ## How was this patch tested?
   Ran project unit tests
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[GitHub] [hadoop-ozone] cxorm commented on issue #296: HDDS-2659. KeyValueHandler#handleCreateContainer should log the exception on container creation failure

2019-12-12 Thread GitBox
cxorm commented on issue #296: HDDS-2659. KeyValueHandler#handleCreateContainer 
should log the exception on container creation failure
URL: https://github.com/apache/hadoop-ozone/pull/296#issuecomment-565110871
 
 
   Thanks @xiaoyuyao, @anuengineer and @bharatviswa504 for the reviews.
   And thanks @mukul1987 for the merge.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[jira] [Updated] (HDDS-2035) Implement datanode level CLI to reveal pipeline info

2019-12-12 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HDDS-2035?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated HDDS-2035:
-
Labels: pull-request-available  (was: )

> Implement datanode level CLI to reveal pipeline info
> 
>
> Key: HDDS-2035
> URL: https://issues.apache.org/jira/browse/HDDS-2035
> Project: Hadoop Distributed Data Store
>  Issue Type: Sub-task
>Reporter: Sammi Chen
>Assignee: Li Cheng
>Priority: Major
>  Labels: pull-request-available
>
> 1. filter pipeline by datanode



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[GitHub] [hadoop-ozone] timmylicheng opened a new pull request #348: HDDS-2035 Implement datanode level CLI to reveal pipeline relation.

2019-12-12 Thread GitBox
timmylicheng opened a new pull request #348: HDDS-2035 Implement datanode level 
CLI to reveal pipeline relation.
URL: https://github.com/apache/hadoop-ozone/pull/348
 
 
   ## What changes were proposed in this pull request?
   
   #HDDS-2035 Implement datanode level CLI to reveal pipeline info
   
   ## What is the link to the Apache JIRA
   
   https://issues.apache.org/jira/browse/HDDS-2035
   
   (Please create an issue in ASF JIRA before opening a pull request,
   and you need to set the title of the pull request which starts with
   the corresponding JIRA issue number. (e.g. HDDS-. Fix a typo in YYY.)
   
   Please replace this section with the link to the Apache JIRA)
   
   ## How was this patch tested?
   
   bash-4.2$ ozone scmcli datanode list
   Datanode: 7a68a668-3dbe-4366-9fed-271f221c1db0 
(172.20.0.2/ozone_datanode_1.ozone_default).
   Related pipelines:
   cd060313-b724-46cf-a83a-8764ba066ae2/ONE/RATIS/OPEN
   
   (Please explain how this patch was tested. Ex: unit tests, manual tests)
   (If this patch involves UI changes, please attach a screen-shot; otherwise, 
remove this)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[jira] [Updated] (HDDS-2675) Sonar : fix issues in PipelineStateManager

2019-12-12 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HDDS-2675?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated HDDS-2675:
-
Labels: pull-request-available  (was: )

> Sonar : fix issues in PipelineStateManager
> --
>
> Key: HDDS-2675
> URL: https://issues.apache.org/jira/browse/HDDS-2675
> Project: Hadoop Distributed Data Store
>  Issue Type: Improvement
>  Components: SCM
>Reporter: Supratim Deka
>Assignee: Supratim Deka
>Priority: Major
>  Labels: pull-request-available
>
> Sonar "Code smell" category issues:
> https://sonarcloud.io/project/issues?id=hadoop-ozone=AW5md-uCKcVY8lQ4ZsFf=AW5md-uCKcVY8lQ4ZsFf
> https://sonarcloud.io/project/issues?id=hadoop-ozone=AW5md-uCKcVY8lQ4ZsFd=AW5md-uCKcVY8lQ4ZsFd
> https://sonarcloud.io/project/issues?id=hadoop-ozone=AW5md-uCKcVY8lQ4ZsFg=AW5md-uCKcVY8lQ4ZsFg
> https://sonarcloud.io/project/issues?id=hadoop-ozone=AW5md-uCKcVY8lQ4ZsFe=AW5md-uCKcVY8lQ4ZsFe
> https://sonarcloud.io/project/issues?id=hadoop-ozone=AW5md-uCKcVY8lQ4ZsFh=AW5md-uCKcVY8lQ4ZsFh



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[GitHub] [hadoop-ozone] supratimdeka opened a new pull request #347: HDDS-2675. Sonar : fix issues in PipelineStateManager

2019-12-12 Thread GitBox
supratimdeka opened a new pull request #347: HDDS-2675. Sonar : fix issues in 
PipelineStateManager
URL: https://github.com/apache/hadoop-ozone/pull/347
 
 
   ## What changes were proposed in this pull request?
   Resolved all Sonar "code smell" category issues in 
PipelineStateManager.java, including removal of unused parameter 'conf' from 
the constructor.
   
   ## What is the link to the Apache JIRA
   
   https://issues.apache.org/jira/browse/HDDS-2675
   
   ## How was this patch tested?
   existing junits:
   TestRatisPipelineProvider.java
   TestPipelineStateManager.java
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[GitHub] [hadoop-ozone] sodonnel edited a comment on issue #309: HDDS-2593. DatanodeAdminMonitor should track under replicated containers and complete the admin workflow accordingly

2019-12-12 Thread GitBox
sodonnel edited a comment on issue #309: HDDS-2593. DatanodeAdminMonitor should 
track under replicated containers and complete the admin workflow accordingly
URL: https://github.com/apache/hadoop-ozone/pull/309#issuecomment-565064730
 
 
   The unit test failure is still this one, unrelated to the changes I believe:
   
   ```[ERROR] Tests run: 9, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 
4.211 s <<< FAILURE! - in org.apache.hadoop.hdds.scm.block.TestBlockManager
   [ERROR] 
testMultipleBlockAllocationWithClosedContainer(org.apache.hadoop.hdds.scm.block.TestBlockManager)
  Time elapsed: 2.248 s  <<< ERROR!
   org.apache.hadoop.hdds.scm.pipeline.InsufficientDatanodesException: Cannot 
create pipeline of factor 3 using 2 nodes.
at 
org.apache.hadoop.hdds.scm.pipeline.RatisPipelineProvider.create(RatisPipelineProvider.java:153)
at 
org.apache.hadoop.hdds.scm.pipeline.PipelineFactory.create(PipelineFactory.java:58)
at 
org.apache.hadoop.hdds.scm.pipeline.SCMPipelineManager.createPipeline(SCMPipelineManager.java:155)
at 
org.apache.hadoop.hdds.scm.block.TestBlockManager.testMultipleBlockAllocationWithClosedContainer(TestBlockManager.java:306)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at 
org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:47)
at 
org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at 
org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:44)
   ```
   
   It should be fixed when we next merge master into this branch, as I believe 
that test has been fixed on master now.
   
   I agree there is a chance we could move the calculation into replication 
manager and track the metrics of under-replicated blocks in the NodeManager, 
but for now I think we should move forward and commit this and once we have a 
working decommission solution we can look to optimise and refactor it further.
   
   If you are happy can you please go ahead and commit this change?


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[GitHub] [hadoop-ozone] sodonnel edited a comment on issue #309: HDDS-2593. DatanodeAdminMonitor should track under replicated containers and complete the admin workflow accordingly

2019-12-12 Thread GitBox
sodonnel edited a comment on issue #309: HDDS-2593. DatanodeAdminMonitor should 
track under replicated containers and complete the admin workflow accordingly
URL: https://github.com/apache/hadoop-ozone/pull/309#issuecomment-565064730
 
 
   The unit test failure is still this one, unrelated to the changes I believe:
   
   ```[ERROR] Tests run: 9, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 
4.211 s <<< FAILURE! - in org.apache.hadoop.hdds.scm.block.TestBlockManager
   [ERROR] 
testMultipleBlockAllocationWithClosedContainer(org.apache.hadoop.hdds.scm.block.TestBlockManager)
  Time elapsed: 2.248 s  <<< ERROR!
   org.apache.hadoop.hdds.scm.pipeline.InsufficientDatanodesException: Cannot 
create pipeline of factor 3 using 2 nodes.
at 
org.apache.hadoop.hdds.scm.pipeline.RatisPipelineProvider.create(RatisPipelineProvider.java:153)
at 
org.apache.hadoop.hdds.scm.pipeline.PipelineFactory.create(PipelineFactory.java:58)
at 
org.apache.hadoop.hdds.scm.pipeline.SCMPipelineManager.createPipeline(SCMPipelineManager.java:155)
at 
org.apache.hadoop.hdds.scm.block.TestBlockManager.testMultipleBlockAllocationWithClosedContainer(TestBlockManager.java:306)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at 
org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:47)
at 
org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at 
org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:44)
   ```
   
   It should be fixed when we next merge master into this branch, as I believe 
that test has been fixed on master now.
   
   I agree there is a chance we could move the calculation into replication 
manager and track the metrics of under-replicated blocks in the NodeManager, 
but for now I think we should move forward and commit this and once we have a 
working decommission solution we can look to optimised and refactor it further.
   
   If you are happy can you please go ahead and commit this change?


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[GitHub] [hadoop-ozone] sodonnel commented on issue #309: HDDS-2593. DatanodeAdminMonitor should track under replicated containers and complete the admin workflow accordingly

2019-12-12 Thread GitBox
sodonnel commented on issue #309: HDDS-2593. DatanodeAdminMonitor should track 
under replicated containers and complete the admin workflow accordingly
URL: https://github.com/apache/hadoop-ozone/pull/309#issuecomment-565064730
 
 
   The unit test failure is still this one, unrelated to the changes I believe:
   
   ```[ERROR] Tests run: 9, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 
4.211 s <<< FAILURE! - in org.apache.hadoop.hdds.scm.block.TestBlockManager
   [ERROR] 
testMultipleBlockAllocationWithClosedContainer(org.apache.hadoop.hdds.scm.block.TestBlockManager)
  Time elapsed: 2.248 s  <<< ERROR!
   org.apache.hadoop.hdds.scm.pipeline.InsufficientDatanodesException: Cannot 
create pipeline of factor 3 using 2 nodes.
at 
org.apache.hadoop.hdds.scm.pipeline.RatisPipelineProvider.create(RatisPipelineProvider.java:153)
at 
org.apache.hadoop.hdds.scm.pipeline.PipelineFactory.create(PipelineFactory.java:58)
at 
org.apache.hadoop.hdds.scm.pipeline.SCMPipelineManager.createPipeline(SCMPipelineManager.java:155)
at 
org.apache.hadoop.hdds.scm.block.TestBlockManager.testMultipleBlockAllocationWithClosedContainer(TestBlockManager.java:306)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at 
org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:47)
at 
org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at 
org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:44)```
   
   It should be fixed when we next merge master into this branch, as I believe 
that test has been fixed on master now.
   
   I agree there is a chance we could move the calculation into replication 
manager and track the metrics of under-replicated blocks in the NodeManager, 
but for now I think we should move forward and commit this and once we have a 
working decommission solution we can look to optimised and refactor it further.
   
   If you are happy can you please go ahead and commit this change?


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[GitHub] [hadoop-ozone] elek commented on issue #337: HDDS-2706. Add class CRLCodec - used for certificate revocation list.

2019-12-12 Thread GitBox
elek commented on issue #337: HDDS-2706. Add class CRLCodec - used for 
certificate revocation list.
URL: https://github.com/apache/hadoop-ozone/pull/337#issuecomment-565018418
 
 
   Thank you very much the patch @abhishekaypurohit Is there any related design 
doc which can help to understand the goal and review the change? I didn't find 
it in the Jira but I might miss it...


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[jira] [Resolved] (HDDS-1792) Use ConcurrentHashSet in place of ConcurrentHashMap in ContainerStteMachine.

2019-12-12 Thread Mukul Kumar Singh (Jira)


 [ 
https://issues.apache.org/jira/browse/HDDS-1792?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mukul Kumar Singh resolved HDDS-1792.
-
Resolution: Invalid

> Use ConcurrentHashSet in place of ConcurrentHashMap in ContainerStteMachine.
> 
>
> Key: HDDS-1792
> URL: https://issues.apache.org/jira/browse/HDDS-1792
> Project: Hadoop Distributed Data Store
>  Issue Type: Bug
>  Components: Ozone Datanode
>Affects Versions: 0.4.0
>Reporter: Mukul Kumar Singh
>Assignee: Mukul Kumar Singh
>Priority: Major
>
> Use ConcurrentHashSet in place of ConcurrentHashMap in ContainerStteMachine.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[jira] [Resolved] (HDDS-1679) TestBCSID failing because of dangling db references

2019-12-12 Thread Mukul Kumar Singh (Jira)


 [ 
https://issues.apache.org/jira/browse/HDDS-1679?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mukul Kumar Singh resolved HDDS-1679.
-
Resolution: Cannot Reproduce

Haven't seen this in a while. resolving this.

> TestBCSID failing because of dangling db references
> ---
>
> Key: HDDS-1679
> URL: https://issues.apache.org/jira/browse/HDDS-1679
> Project: Hadoop Distributed Data Store
>  Issue Type: Bug
>  Components: Ozone Datanode
>Reporter: Mukul Kumar Singh
>Assignee: Mukul Kumar Singh
>Priority: Major
>  Time Spent: 1h 10m
>  Remaining Estimate: 0h
>
> TestBCSID failing because of dangling db references.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[jira] [Resolved] (HDDS-2663) Add OzoneFileSystem to MiniOzoneChaosCluster

2019-12-12 Thread Mukul Kumar Singh (Jira)


 [ 
https://issues.apache.org/jira/browse/HDDS-2663?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mukul Kumar Singh resolved HDDS-2663.
-
Resolution: Duplicate

> Add OzoneFileSystem to MiniOzoneChaosCluster
> 
>
> Key: HDDS-2663
> URL: https://issues.apache.org/jira/browse/HDDS-2663
> Project: Hadoop Distributed Data Store
>  Issue Type: Bug
>  Components: test
>Reporter: Mukul Kumar Singh
>Assignee: Mukul Kumar Singh
>Priority: Major
>
> This jira adds ozonefilesystem functionality to MiniOzoneChaosCluster.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[jira] [Updated] (HDDS-2719) Rename S3Token to S3AuthInfo

2019-12-12 Thread Mohammad (Jira)


 [ 
https://issues.apache.org/jira/browse/HDDS-2719?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mohammad updated HDDS-2719:
---
Fix Version/s: (was: 0.4.2)
   0.5.0

> Rename S3Token to S3AuthInfo
> 
>
> Key: HDDS-2719
> URL: https://issues.apache.org/jira/browse/HDDS-2719
> Project: Hadoop Distributed Data Store
>  Issue Type: Improvement
>Reporter: Mohammad
>Assignee: Mohammad
>Priority: Minor
> Fix For: 0.5.0
>
>
> Rename S3Token to something like S3AuthInfo
> Item #24 in the list of ideas for new contributors to work on
> https://cwiki.apache.org/confluence/display/HADOOP/Ozone+project+ideas+for+new+contributors



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[jira] [Updated] (HDDS-2719) Rename S3Token to S3AuthInfo

2019-12-12 Thread Mohammad (Jira)


 [ 
https://issues.apache.org/jira/browse/HDDS-2719?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mohammad updated HDDS-2719:
---
Component/s: S3

> Rename S3Token to S3AuthInfo
> 
>
> Key: HDDS-2719
> URL: https://issues.apache.org/jira/browse/HDDS-2719
> Project: Hadoop Distributed Data Store
>  Issue Type: Improvement
>  Components: S3
>Reporter: Mohammad
>Assignee: Mohammad
>Priority: Minor
> Fix For: 0.5.0
>
>
> Rename S3Token to something like S3AuthInfo
> Item #24 in the list of ideas for new contributors to work on
> https://cwiki.apache.org/confluence/display/HADOOP/Ozone+project+ideas+for+new+contributors



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[jira] [Updated] (HDDS-2719) Rename S3Token to S3AuthInfo

2019-12-12 Thread Mohammad (Jira)


 [ 
https://issues.apache.org/jira/browse/HDDS-2719?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mohammad updated HDDS-2719:
---
Component/s: (was: SCM)

> Rename S3Token to S3AuthInfo
> 
>
> Key: HDDS-2719
> URL: https://issues.apache.org/jira/browse/HDDS-2719
> Project: Hadoop Distributed Data Store
>  Issue Type: Improvement
>Reporter: Mohammad
>Assignee: Mohammad
>Priority: Minor
> Fix For: 0.4.2
>
>
> Rename S3Token to something like S3AuthInfo
> Item #24 in the list of ideas for new contributors to work on
> https://cwiki.apache.org/confluence/display/HADOOP/Ozone+project+ideas+for+new+contributors



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[jira] [Updated] (HDDS-2719) Rename S3Token to S3AuthInfo

2019-12-12 Thread Mohammad (Jira)


 [ 
https://issues.apache.org/jira/browse/HDDS-2719?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mohammad updated HDDS-2719:
---
Description: 
Rename S3Token to something like S3AuthInfo

Item #24 in the list of ideas for new contributors to work on

https://cwiki.apache.org/confluence/display/HADOOP/Ozone+project+ideas+for+new+contributors

  was:
Sonar report :
[https://sonarcloud.io/project/issues?id=hadoop-ozone=AW5md-_2KcVY8lQ4ZsVd=false=BUG|https://sonarcloud.io/project/issues?id=hadoop-ozone=AW5md-4jKcVY8lQ4ZsPQ=AW5md-4jKcVY8lQ4ZsPQ]


> Rename S3Token to S3AuthInfo
> 
>
> Key: HDDS-2719
> URL: https://issues.apache.org/jira/browse/HDDS-2719
> Project: Hadoop Distributed Data Store
>  Issue Type: Improvement
>  Components: SCM
>Reporter: Mohammad
>Assignee: Mohammad
>Priority: Minor
> Fix For: 0.4.2
>
>
> Rename S3Token to something like S3AuthInfo
> Item #24 in the list of ideas for new contributors to work on
> https://cwiki.apache.org/confluence/display/HADOOP/Ozone+project+ideas+for+new+contributors



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[jira] [Created] (HDDS-2719) Rename S3Token to S3AuthInfo

2019-12-12 Thread Mohammad (Jira)
Mohammad created HDDS-2719:
--

 Summary: Rename S3Token to S3AuthInfo
 Key: HDDS-2719
 URL: https://issues.apache.org/jira/browse/HDDS-2719
 Project: Hadoop Distributed Data Store
  Issue Type: Improvement
  Components: SCM
Reporter: Mohammad
Assignee: Mohammad
 Fix For: 0.4.2


Sonar report :
[https://sonarcloud.io/project/issues?id=hadoop-ozone=AW5md-_2KcVY8lQ4ZsVd=false=BUG|https://sonarcloud.io/project/issues?id=hadoop-ozone=AW5md-4jKcVY8lQ4ZsPQ=AW5md-4jKcVY8lQ4ZsPQ]



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[jira] [Updated] (HDDS-2588) Consolidate compose environments

2019-12-12 Thread Attila Doroszlai (Jira)


 [ 
https://issues.apache.org/jira/browse/HDDS-2588?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Attila Doroszlai updated HDDS-2588:
---
Issue Type: Improvement  (was: Bug)

> Consolidate compose environments
> 
>
> Key: HDDS-2588
> URL: https://issues.apache.org/jira/browse/HDDS-2588
> Project: Hadoop Distributed Data Store
>  Issue Type: Improvement
>  Components: docker
>Reporter: Attila Doroszlai
>Assignee: Attila Doroszlai
>Priority: Major
>  Labels: pull-request-available
> Fix For: 0.5.0
>
>  Time Spent: 20m
>  Remaining Estimate: 0h
>
> There are a few slightly different sample docker compose environments: ozone, 
> ozoneperf, ozones3, ozone-recon. This issue proposes to merge these 4 by 
> minor additions to ozoneperf:
>  # add {{recon}} service from {{ozone-recon}}
>  # run GDPR and S3 tests
>  # expose datanode web port (eg. for profiling)
> Plus: also run ozone-shell test (from basic suite), which is currently run 
> only in ozonesecure
> I also propose to rename {{ozoneperf}} to {{ozone}} for simplicity.
> Consolidating these 4 environments would slightly reduce both code 
> duplication and the time needed for acceptance tests.
> CC [~elek]



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[jira] [Updated] (HDDS-2588) Consolidate compose environments

2019-12-12 Thread Attila Doroszlai (Jira)


 [ 
https://issues.apache.org/jira/browse/HDDS-2588?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Attila Doroszlai updated HDDS-2588:
---
Fix Version/s: 0.5.0

> Consolidate compose environments
> 
>
> Key: HDDS-2588
> URL: https://issues.apache.org/jira/browse/HDDS-2588
> Project: Hadoop Distributed Data Store
>  Issue Type: Bug
>  Components: docker
>Reporter: Attila Doroszlai
>Assignee: Attila Doroszlai
>Priority: Major
>  Labels: pull-request-available
> Fix For: 0.5.0
>
>  Time Spent: 20m
>  Remaining Estimate: 0h
>
> There are a few slightly different sample docker compose environments: ozone, 
> ozoneperf, ozones3, ozone-recon. This issue proposes to merge these 4 by 
> minor additions to ozoneperf:
>  # add {{recon}} service from {{ozone-recon}}
>  # run GDPR and S3 tests
>  # expose datanode web port (eg. for profiling)
> Plus: also run ozone-shell test (from basic suite), which is currently run 
> only in ozonesecure
> I also propose to rename {{ozoneperf}} to {{ozone}} for simplicity.
> Consolidating these 4 environments would slightly reduce both code 
> duplication and the time needed for acceptance tests.
> CC [~elek]



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[GitHub] [hadoop-ozone] adoroszlai commented on issue #238: HDDS-2588. Consolidate compose environments

2019-12-12 Thread GitBox
adoroszlai commented on issue #238: HDDS-2588. Consolidate compose environments
URL: https://github.com/apache/hadoop-ozone/pull/238#issuecomment-564959372
 
 
   Thanks @anuengineer, @avijayanhwx, @bharatviswa504 for review, and @elek for 
review and commit.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[GitHub] [hadoop-ozone] adoroszlai commented on issue #324: HDDS-2688. Refactor container response builders to hadoop-hdds/common

2019-12-12 Thread GitBox
adoroszlai commented on issue #324: HDDS-2688. Refactor container response 
builders to hadoop-hdds/common
URL: https://github.com/apache/hadoop-ozone/pull/324#issuecomment-564958466
 
 
   Thanks @bshashikant for reviewing and committing this.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[jira] [Resolved] (HDDS-2659) KeyValueHandler#handleCreateContainer should log the exception on container creation failure

2019-12-12 Thread Mukul Kumar Singh (Jira)


 [ 
https://issues.apache.org/jira/browse/HDDS-2659?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mukul Kumar Singh resolved HDDS-2659.
-
Fix Version/s: 0.5.0
   Resolution: Fixed

Thanks for the contribution [~cxorm]. I have committed this.

> KeyValueHandler#handleCreateContainer should log the exception on container 
> creation failure
> 
>
> Key: HDDS-2659
> URL: https://issues.apache.org/jira/browse/HDDS-2659
> Project: Hadoop Distributed Data Store
>  Issue Type: Bug
>Reporter: Nilotpal Nandi
>Assignee: YiSheng Lien
>Priority: Major
>  Labels: newbie, pull-request-available
> Fix For: 0.5.0
>
>  Time Spent: 20m
>  Remaining Estimate: 0h
>
> KeyValueHandler#handleCreateContainer does not log the stack trace of the 
> StorageContainerException cause. This leads to difficulty in debugging the 
> reason for failure.
>  
> {code:java}
> 2019-11-04 00:40:56,719 INFO 
> org.apache.hadoop.ozone.container.keyvalue.KeyValueHandler: Operation: 
> CreateContainer : Trace ID: 
> 1f7530ed73cf2e47:74588d4de9f1d702:1f7530ed73cf2e47:0 : Message: Container 
> creation failed. : Result: CONTAINER_INTERNAL_ERROR
> 2019-11-04 00:40:56,720 INFO 
> org.apache.hadoop.ozone.container.common.impl.HddsDispatcher: Operation: 
> WriteChunk : Trace ID: 1f7530ed73cf2e47:74588d4de9f1d702:1f7530ed73cf2e47:0 : 
> Message: ContainerID 11 creation failed : Result: CONTAINER_INTERNAL_ERROR
> {code}
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[GitHub] [hadoop-ozone] elek closed pull request #238: HDDS-2588. Consolidate compose environments

2019-12-12 Thread GitBox
elek closed pull request #238: HDDS-2588. Consolidate compose environments
URL: https://github.com/apache/hadoop-ozone/pull/238
 
 
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[GitHub] [hadoop-ozone] mukul1987 merged pull request #296: HDDS-2659. KeyValueHandler#handleCreateContainer should log the exception on container creation failure

2019-12-12 Thread GitBox
mukul1987 merged pull request #296: HDDS-2659. 
KeyValueHandler#handleCreateContainer should log the exception on container 
creation failure
URL: https://github.com/apache/hadoop-ozone/pull/296
 
 
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[jira] [Updated] (HDDS-2688) Refactor container response builders to hdds-common

2019-12-12 Thread Shashikant Banerjee (Jira)


 [ 
https://issues.apache.org/jira/browse/HDDS-2688?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Shashikant Banerjee updated HDDS-2688:
--
Fix Version/s: 0.5.0
   Resolution: Fixed
   Status: Resolved  (was: Patch Available)

Thanks [~adoroszlai] for working on this. I have committed this.

> Refactor container response builders to hdds-common
> ---
>
> Key: HDDS-2688
> URL: https://issues.apache.org/jira/browse/HDDS-2688
> Project: Hadoop Distributed Data Store
>  Issue Type: Improvement
>  Components: Ozone Datanode
>Reporter: Attila Doroszlai
>Assignee: Attila Doroszlai
>Priority: Minor
>  Labels: pull-request-available
> Fix For: 0.5.0
>
>  Time Spent: 20m
>  Remaining Estimate: 0h
>
> {{ContainerUtils}} and {{BlockUtils}} have some helper functions to build 
> responses to container commands.  These would be useful for client-side unit 
> tests, but {{client}} does not depend on {{container-service}} since the 
> interfaces and messages it needs are defined in {{common}}.  This issue 
> proposes to move these helpers to {{common}} to avoid duplicating the 
> functionality for tests.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[GitHub] [hadoop-ozone] bshashikant merged pull request #324: HDDS-2688. Refactor container response builders to hadoop-hdds/common

2019-12-12 Thread GitBox
bshashikant merged pull request #324: HDDS-2688. Refactor container response 
builders to hadoop-hdds/common
URL: https://github.com/apache/hadoop-ozone/pull/324
 
 
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[jira] [Created] (HDDS-2718) Send chunks in increments to datanode

2019-12-12 Thread Attila Doroszlai (Jira)
Attila Doroszlai created HDDS-2718:
--

 Summary: Send chunks in increments to datanode
 Key: HDDS-2718
 URL: https://issues.apache.org/jira/browse/HDDS-2718
 Project: Hadoop Distributed Data Store
  Issue Type: Improvement
  Components: SCM Client
Reporter: Attila Doroszlai
Assignee: Attila Doroszlai


Change client to send chunks in multiple parts based on stream buffer increment 
size config.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[jira] [Updated] (HDDS-2717) Handle chunk increments in datanode

2019-12-12 Thread Attila Doroszlai (Jira)


 [ 
https://issues.apache.org/jira/browse/HDDS-2717?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Attila Doroszlai updated HDDS-2717:
---
Issue Type: Improvement  (was: Task)

> Handle chunk increments in datanode
> ---
>
> Key: HDDS-2717
> URL: https://issues.apache.org/jira/browse/HDDS-2717
> Project: Hadoop Distributed Data Store
>  Issue Type: Improvement
>  Components: Ozone Datanode
>Reporter: Attila Doroszlai
>Assignee: Attila Doroszlai
>Priority: Major
>
> Let datanode handle incremental additions to chunks (data with non-zero 
> offset).



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[jira] [Created] (HDDS-2717) Handle chunk increments in datanode

2019-12-12 Thread Attila Doroszlai (Jira)
Attila Doroszlai created HDDS-2717:
--

 Summary: Handle chunk increments in datanode
 Key: HDDS-2717
 URL: https://issues.apache.org/jira/browse/HDDS-2717
 Project: Hadoop Distributed Data Store
  Issue Type: Task
  Components: Ozone Datanode
Reporter: Attila Doroszlai
Assignee: Attila Doroszlai


Let datanode handle incremental additions to chunks (data with non-zero offset).



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[jira] [Created] (HDDS-2716) ReadStateMachine failure should close down the ratis pipeline

2019-12-12 Thread Shashikant Banerjee (Jira)
Shashikant Banerjee created HDDS-2716:
-

 Summary: ReadStateMachine failure should close down the ratis 
pipeline 
 Key: HDDS-2716
 URL: https://issues.apache.org/jira/browse/HDDS-2716
 Project: Hadoop Distributed Data Store
  Issue Type: Bug
  Components: Ozone Datanode
Affects Versions: 0.5.0
Reporter: Shashikant Banerjee
 Fix For: 0.5.0


Currently, in case of readStateMachineFailures, the containerStateMachine is 
marked unhealthy which stalls all the pipeline operations. But the pipeline 
will get destroyed in the next snapshot execution. The idea here is to just 
destroy the pipeline as soon as readStateMachine call fails.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[GitHub] [hadoop-ozone] elek commented on issue #323: HDDS-2686. Use protobuf 3 instead of protobuf 2

2019-12-12 Thread GitBox
elek commented on issue #323: HDDS-2686. Use protobuf 3 instead of protobuf 2
URL: https://github.com/apache/hadoop-ozone/pull/323#issuecomment-564928421
 
 
   > but there are acceptance test failures.
   
   Yes, the problem is in the mr test with hadoop-32. With hadoop 2.7 - 3.1 we 
have no problems as we use properly isolated classpath environments ("legacy" 
ozonesf jar) but until now we were compatible with hadoop 3.2. 
   
   With this change the `current` jar will be compatible with the (unreleased) 
hadoop 3.3 and we need to use the classpath separation for 3.2.
   
   I updated the tests and created a new mr test for 3.3 (based on a custom 
docker image which can be replaced after the release with the official apache 
image).
   
   Note: shading is not working for the protobuf only the classpath isolation. 
With the `current` jar we are not shading the hadoop API which includes 
references to the protobuf GenericMessage. Without shading the Hadoop API we 
can't shade the protobuf files, therefore the only option is to use the 
classpath isolation. 


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[jira] [Created] (HDDS-2715) Add config for client buffer increment size

2019-12-12 Thread Attila Doroszlai (Jira)
Attila Doroszlai created HDDS-2715:
--

 Summary: Add config for client buffer increment size
 Key: HDDS-2715
 URL: https://issues.apache.org/jira/browse/HDDS-2715
 Project: Hadoop Distributed Data Store
  Issue Type: Task
  Components: Ozone Client
Reporter: Attila Doroszlai
Assignee: Attila Doroszlai


Introduce a new config property for client-side buffer increment size along the 
lines of {{ozone.client.stream.buffer.flush.size}} and 
{{ozone.client.stream.buffer.max.size}}.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[jira] [Created] (HDDS-2714) Add Ratis leader info for scm pipeline info command

2019-12-12 Thread Shashikant Banerjee (Jira)
Shashikant Banerjee created HDDS-2714:
-

 Summary: Add Ratis leader info for scm pipeline info command
 Key: HDDS-2714
 URL: https://issues.apache.org/jira/browse/HDDS-2714
 Project: Hadoop Distributed Data Store
  Issue Type: Bug
  Components: SCM
Affects Versions: 0.5.0
Reporter: Shashikant Banerjee
 Fix For: 0.5.0


SCMCli prints the datanode info with the pipeline ID. it would be of great help 
in debugging if it also displays the ratis leader info along with it as this 
info is already populated by datanode to scm.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org



[jira] [Created] (HDDS-2713) Decouple client write size and datanode chunk size

2019-12-12 Thread Attila Doroszlai (Jira)
Attila Doroszlai created HDDS-2713:
--

 Summary: Decouple client write size and datanode chunk size
 Key: HDDS-2713
 URL: https://issues.apache.org/jira/browse/HDDS-2713
 Project: Hadoop Distributed Data Store
  Issue Type: Improvement
  Components: Ozone Datanode, SCM Client
Reporter: Attila Doroszlai
Assignee: Attila Doroszlai


Currently Datanode creates chunk files as they are received from HDDS client.  
This creates a conflict between requirements: client would like to use less 
memory for buffering, but datanode needs to avoid small files.

The goal of this task is to decouple client and server write sizes to allow the 
client to send data in smaller increments without affecting datanode storage.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org