[jira] [Updated] (HDFS-14935) Refactor DFSNetworkTopology#isNodeInScope
[ https://issues.apache.org/jira/browse/HDFS-14935?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Lisheng Sun updated HDFS-14935: --- Attachment: HDFS-14935.003.patch > Refactor DFSNetworkTopology#isNodeInScope > - > > Key: HDFS-14935 > URL: https://issues.apache.org/jira/browse/HDFS-14935 > Project: Hadoop HDFS > Issue Type: Improvement >Reporter: Lisheng Sun >Assignee: Lisheng Sun >Priority: Major > Attachments: HDFS-14935.001.patch, HDFS-14935.002.patch, > HDFS-14935.003.patch > > > {code:java} > private boolean isNodeInScope(Node node, String scope) { > if (!scope.endsWith("/")) { > scope += "/"; > } > String nodeLocation = node.getNetworkLocation() + "/"; > return nodeLocation.startsWith(scope); > } > {code} > NodeBase#normalize() is used to normalize scope. > so i refator DFSNetworkTopology#isNodeInScope. > > -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-14935) Refactor DFSNetworkTopology#isNodeInScope
[ https://issues.apache.org/jira/browse/HDFS-14935?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16960293#comment-16960293 ] Lisheng Sun commented on HDFS-14935: [~ayushtkn] i think if there is illegalArgumentException, it should throw this Exception and stop. As NetworkTopology#countNumOfAvailableNodes() is also done. {code:java} @VisibleForTesting public int countNumOfAvailableNodes(String scope, Collection excludedNodes) { boolean isExcluded=false; if (scope.startsWith("~")) { isExcluded=true; scope=scope.substring(1); } scope = NodeBase.normalize(scope); . }{code} Please correct me if i was wrong. Thank you [~ayushtkn] > Refactor DFSNetworkTopology#isNodeInScope > - > > Key: HDFS-14935 > URL: https://issues.apache.org/jira/browse/HDFS-14935 > Project: Hadoop HDFS > Issue Type: Improvement >Reporter: Lisheng Sun >Assignee: Lisheng Sun >Priority: Major > Attachments: HDFS-14935.001.patch, HDFS-14935.002.patch > > > {code:java} > private boolean isNodeInScope(Node node, String scope) { > if (!scope.endsWith("/")) { > scope += "/"; > } > String nodeLocation = node.getNetworkLocation() + "/"; > return nodeLocation.startsWith(scope); > } > {code} > NodeBase#normalize() is used to normalize scope. > so i refator DFSNetworkTopology#isNodeInScope. > > -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Comment Edited] (HDFS-14935) Refactor DFSNetworkTopology#isNodeInScope
[ https://issues.apache.org/jira/browse/HDFS-14935?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16959623#comment-16959623 ] Lisheng Sun edited comment on HDFS-14935 at 10/25/19 3:27 PM: -- hi [~ayushtkn] this Jira is used to refactor code and no need to write repetitive code. use existing code as follow: {code:java} scope = NodeBase.normalize(scope); public static String normalize(String path) { if (path == null) { throw new IllegalArgumentException( "Network Location is null "); } if (path.length() == 0) { return ROOT; } if (path.charAt(0) != PATH_SEPARATOR) { throw new IllegalArgumentException( "Network Location path does not start with " +PATH_SEPARATOR_STR+ ": "+path); } int len = path.length(); if (path.charAt(len-1) == PATH_SEPARATOR) { return path.substring(0, len-1); } return path; } {code} was (Author: leosun08): hi [~ayushtkn] this Jira is used to optimize code and no need to write repetitive code. use existing code as follow: {code:java} scope = NodeBase.normalize(scope); public static String normalize(String path) { if (path == null) { throw new IllegalArgumentException( "Network Location is null "); } if (path.length() == 0) { return ROOT; } if (path.charAt(0) != PATH_SEPARATOR) { throw new IllegalArgumentException( "Network Location path does not start with " +PATH_SEPARATOR_STR+ ": "+path); } int len = path.length(); if (path.charAt(len-1) == PATH_SEPARATOR) { return path.substring(0, len-1); } return path; } {code} > Refactor DFSNetworkTopology#isNodeInScope > - > > Key: HDFS-14935 > URL: https://issues.apache.org/jira/browse/HDFS-14935 > Project: Hadoop HDFS > Issue Type: Improvement >Reporter: Lisheng Sun >Assignee: Lisheng Sun >Priority: Major > Attachments: HDFS-14935.001.patch, HDFS-14935.002.patch > > > {code:java} > private boolean isNodeInScope(Node node, String scope) { > if (!scope.endsWith("/")) { > scope += "/"; > } > String nodeLocation = node.getNetworkLocation() + "/"; > return nodeLocation.startsWith(scope); > } > {code} > NodeBase#normalize() is used to normalize scope. > so i refator DFSNetworkTopology#isNodeInScope. > > -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-14935) Refactor DFSNetworkTopology#isNodeInScope
[ https://issues.apache.org/jira/browse/HDFS-14935?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Lisheng Sun updated HDFS-14935: --- Summary: Refactor DFSNetworkTopology#isNodeInScope (was: Optimize DFSNetworkTopology#isNodeInScope) > Refactor DFSNetworkTopology#isNodeInScope > - > > Key: HDFS-14935 > URL: https://issues.apache.org/jira/browse/HDFS-14935 > Project: Hadoop HDFS > Issue Type: Improvement >Reporter: Lisheng Sun >Assignee: Lisheng Sun >Priority: Major > Attachments: HDFS-14935.001.patch, HDFS-14935.002.patch > > > {code:java} > private boolean isNodeInScope(Node node, String scope) { > if (!scope.endsWith("/")) { > scope += "/"; > } > String nodeLocation = node.getNetworkLocation() + "/"; > return nodeLocation.startsWith(scope); > } > {code} > NodeBase#normalize() is used to normalize scope. > so i refator > > -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-14935) Refactor DFSNetworkTopology#isNodeInScope
[ https://issues.apache.org/jira/browse/HDFS-14935?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Lisheng Sun updated HDFS-14935: --- Description: {code:java} private boolean isNodeInScope(Node node, String scope) { if (!scope.endsWith("/")) { scope += "/"; } String nodeLocation = node.getNetworkLocation() + "/"; return nodeLocation.startsWith(scope); } {code} NodeBase#normalize() is used to normalize scope. so i refator DFSNetworkTopology#isNodeInScope. was: {code:java} private boolean isNodeInScope(Node node, String scope) { if (!scope.endsWith("/")) { scope += "/"; } String nodeLocation = node.getNetworkLocation() + "/"; return nodeLocation.startsWith(scope); } {code} NodeBase#normalize() is used to normalize scope. so i refator isNodeInScope > Refactor DFSNetworkTopology#isNodeInScope > - > > Key: HDFS-14935 > URL: https://issues.apache.org/jira/browse/HDFS-14935 > Project: Hadoop HDFS > Issue Type: Improvement >Reporter: Lisheng Sun >Assignee: Lisheng Sun >Priority: Major > Attachments: HDFS-14935.001.patch, HDFS-14935.002.patch > > > {code:java} > private boolean isNodeInScope(Node node, String scope) { > if (!scope.endsWith("/")) { > scope += "/"; > } > String nodeLocation = node.getNetworkLocation() + "/"; > return nodeLocation.startsWith(scope); > } > {code} > NodeBase#normalize() is used to normalize scope. > so i refator DFSNetworkTopology#isNodeInScope. > > -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-14935) Refactor DFSNetworkTopology#isNodeInScope
[ https://issues.apache.org/jira/browse/HDFS-14935?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Lisheng Sun updated HDFS-14935: --- Description: {code:java} private boolean isNodeInScope(Node node, String scope) { if (!scope.endsWith("/")) { scope += "/"; } String nodeLocation = node.getNetworkLocation() + "/"; return nodeLocation.startsWith(scope); } {code} NodeBase#normalize() is used to normalize scope. so i refator isNodeInScope was: {code:java} private boolean isNodeInScope(Node node, String scope) { if (!scope.endsWith("/")) { scope += "/"; } String nodeLocation = node.getNetworkLocation() + "/"; return nodeLocation.startsWith(scope); } {code} NodeBase#normalize() is used to normalize scope. so i refator > Refactor DFSNetworkTopology#isNodeInScope > - > > Key: HDFS-14935 > URL: https://issues.apache.org/jira/browse/HDFS-14935 > Project: Hadoop HDFS > Issue Type: Improvement >Reporter: Lisheng Sun >Assignee: Lisheng Sun >Priority: Major > Attachments: HDFS-14935.001.patch, HDFS-14935.002.patch > > > {code:java} > private boolean isNodeInScope(Node node, String scope) { > if (!scope.endsWith("/")) { > scope += "/"; > } > String nodeLocation = node.getNetworkLocation() + "/"; > return nodeLocation.startsWith(scope); > } > {code} > NodeBase#normalize() is used to normalize scope. > so i refator isNodeInScope > > -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-14935) Optimize DFSNetworkTopology#isNodeInScope
[ https://issues.apache.org/jira/browse/HDFS-14935?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Lisheng Sun updated HDFS-14935: --- Description: {code:java} private boolean isNodeInScope(Node node, String scope) { if (!scope.endsWith("/")) { scope += "/"; } String nodeLocation = node.getNetworkLocation() + "/"; return nodeLocation.startsWith(scope); } {code} NodeBase#normalize() is used to normalize scope. so i refator > Optimize DFSNetworkTopology#isNodeInScope > - > > Key: HDFS-14935 > URL: https://issues.apache.org/jira/browse/HDFS-14935 > Project: Hadoop HDFS > Issue Type: Improvement >Reporter: Lisheng Sun >Assignee: Lisheng Sun >Priority: Major > Attachments: HDFS-14935.001.patch, HDFS-14935.002.patch > > > {code:java} > private boolean isNodeInScope(Node node, String scope) { > if (!scope.endsWith("/")) { > scope += "/"; > } > String nodeLocation = node.getNetworkLocation() + "/"; > return nodeLocation.startsWith(scope); > } > {code} > NodeBase#normalize() is used to normalize scope. > so i refator > > -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Comment Edited] (HDFS-14935) Optimize DFSNetworkTopology#isNodeInScope
[ https://issues.apache.org/jira/browse/HDFS-14935?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16959623#comment-16959623 ] Lisheng Sun edited comment on HDFS-14935 at 10/25/19 9:59 AM: -- hi [~ayushtkn] this Jira is used to optimize code and no need to write repetitive code. use existing code as follow: {code:java} scope = NodeBase.normalize(scope); public static String normalize(String path) { if (path == null) { throw new IllegalArgumentException( "Network Location is null "); } if (path.length() == 0) { return ROOT; } if (path.charAt(0) != PATH_SEPARATOR) { throw new IllegalArgumentException( "Network Location path does not start with " +PATH_SEPARATOR_STR+ ": "+path); } int len = path.length(); if (path.charAt(len-1) == PATH_SEPARATOR) { return path.substring(0, len-1); } return path; } {code} was (Author: leosun08): hi [~ayushtkn] this Jira is used to optimize code and no need to write repetitive code. use existing code as follow: {code:java} scope = NodeBase.normalize(scope); public static String normalize(String path) { if (path == null) { throw new IllegalArgumentException( "Network Location is null "); } if (path.length() == 0) { return ROOT; } if (path.charAt(0) != PATH_SEPARATOR) { throw new IllegalArgumentException( "Network Location path does not start with " +PATH_SEPARATOR_STR+ ": "+path); } int len = path.length(); if (path.charAt(len-1) == PATH_SEPARATOR) { return path.substring(0, len-1); } return path; } {code} https://issues.apache.org/jira/secure/attachment/12984010/HDFS-14935.002.patch > Optimize DFSNetworkTopology#isNodeInScope > - > > Key: HDFS-14935 > URL: https://issues.apache.org/jira/browse/HDFS-14935 > Project: Hadoop HDFS > Issue Type: Improvement >Reporter: Lisheng Sun >Assignee: Lisheng Sun >Priority: Major > Attachments: HDFS-14935.001.patch, HDFS-14935.002.patch > > -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Comment Edited] (HDFS-14935) Optimize DFSNetworkTopology#isNodeInScope
[ https://issues.apache.org/jira/browse/HDFS-14935?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16959623#comment-16959623 ] Lisheng Sun edited comment on HDFS-14935 at 10/25/19 9:58 AM: -- hi [~ayushtkn] this Jira is used to optimize code and no need to write repetitive code. use existing code as follow: {code:java} scope = NodeBase.normalize(scope); public static String normalize(String path) { if (path == null) { throw new IllegalArgumentException( "Network Location is null "); } if (path.length() == 0) { return ROOT; } if (path.charAt(0) != PATH_SEPARATOR) { throw new IllegalArgumentException( "Network Location path does not start with " +PATH_SEPARATOR_STR+ ": "+path); } int len = path.length(); if (path.charAt(len-1) == PATH_SEPARATOR) { return path.substring(0, len-1); } return path; } {code} was (Author: leosun08): hi [~ayushtkn] this Jira is used to optimize code and no need to write repetitive code.https://issues.apache.org/jira/secure/attachment/12984010/HDFS-14935.002.patch use existing code as follow: {code:java} scope = NodeBase.normalize(scope); public static String normalize(String path) { if (path == null) { throw new IllegalArgumentException( "Network Location is null "); } if (path.length() == 0) { return ROOT; } if (path.charAt(0) != PATH_SEPARATOR) { throw new IllegalArgumentException( "Network Location path does not start with " +PATH_SEPARATOR_STR+ ": "+path); } int len = path.length(); if (path.charAt(len-1) == PATH_SEPARATOR) { return path.substring(0, len-1); } return path; } {code} > Optimize DFSNetworkTopology#isNodeInScope > - > > Key: HDFS-14935 > URL: https://issues.apache.org/jira/browse/HDFS-14935 > Project: Hadoop HDFS > Issue Type: Improvement >Reporter: Lisheng Sun >Assignee: Lisheng Sun >Priority: Major > Attachments: HDFS-14935.001.patch, HDFS-14935.002.patch > > -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Comment Edited] (HDFS-14935) Optimize DFSNetworkTopology#isNodeInScope
[ https://issues.apache.org/jira/browse/HDFS-14935?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16959623#comment-16959623 ] Lisheng Sun edited comment on HDFS-14935 at 10/25/19 9:58 AM: -- hi [~ayushtkn] this Jira is used to optimize code and no need to write repetitive code. use existing code as follow: {code:java} scope = NodeBase.normalize(scope); public static String normalize(String path) { if (path == null) { throw new IllegalArgumentException( "Network Location is null "); } if (path.length() == 0) { return ROOT; } if (path.charAt(0) != PATH_SEPARATOR) { throw new IllegalArgumentException( "Network Location path does not start with " +PATH_SEPARATOR_STR+ ": "+path); } int len = path.length(); if (path.charAt(len-1) == PATH_SEPARATOR) { return path.substring(0, len-1); } return path; } {code} https://issues.apache.org/jira/secure/attachment/12984010/HDFS-14935.002.patch was (Author: leosun08): hi [~ayushtkn] this Jira is used to optimize code and no need to write repetitive code. use existing code as follow: {code:java} scope = NodeBase.normalize(scope); public static String normalize(String path) { if (path == null) { throw new IllegalArgumentException( "Network Location is null "); } if (path.length() == 0) { return ROOT; } if (path.charAt(0) != PATH_SEPARATOR) { throw new IllegalArgumentException( "Network Location path does not start with " +PATH_SEPARATOR_STR+ ": "+path); } int len = path.length(); if (path.charAt(len-1) == PATH_SEPARATOR) { return path.substring(0, len-1); } return path; } {code} > Optimize DFSNetworkTopology#isNodeInScope > - > > Key: HDFS-14935 > URL: https://issues.apache.org/jira/browse/HDFS-14935 > Project: Hadoop HDFS > Issue Type: Improvement >Reporter: Lisheng Sun >Assignee: Lisheng Sun >Priority: Major > Attachments: HDFS-14935.001.patch, HDFS-14935.002.patch > > -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Comment Edited] (HDFS-14935) Optimize DFSNetworkTopology#isNodeInScope
[ https://issues.apache.org/jira/browse/HDFS-14935?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16959623#comment-16959623 ] Lisheng Sun edited comment on HDFS-14935 at 10/25/19 9:57 AM: -- hi [~ayushtkn] this Jira is used to optimize code and no need to write repetitive code.https://issues.apache.org/jira/secure/attachment/12984010/HDFS-14935.002.patch use existing code as follow: {code:java} scope = NodeBase.normalize(scope); public static String normalize(String path) { if (path == null) { throw new IllegalArgumentException( "Network Location is null "); } if (path.length() == 0) { return ROOT; } if (path.charAt(0) != PATH_SEPARATOR) { throw new IllegalArgumentException( "Network Location path does not start with " +PATH_SEPARATOR_STR+ ": "+path); } int len = path.length(); if (path.charAt(len-1) == PATH_SEPARATOR) { return path.substring(0, len-1); } return path; } {code} was (Author: leosun08): hi [~ayushtkn] this Jira is used to optimize code and no need to write repetitive code. use existing code as follow: {code:java} scope = NodeBase.normalize(scope); public static String normalize(String path) { if (path == null) { throw new IllegalArgumentException( "Network Location is null "); } if (path.length() == 0) { return ROOT; } if (path.charAt(0) != PATH_SEPARATOR) { throw new IllegalArgumentException( "Network Location path does not start with " +PATH_SEPARATOR_STR+ ": "+path); } int len = path.length(); if (path.charAt(len-1) == PATH_SEPARATOR) { return path.substring(0, len-1); } return path; } {code} > Optimize DFSNetworkTopology#isNodeInScope > - > > Key: HDFS-14935 > URL: https://issues.apache.org/jira/browse/HDFS-14935 > Project: Hadoop HDFS > Issue Type: Improvement >Reporter: Lisheng Sun >Assignee: Lisheng Sun >Priority: Major > Attachments: HDFS-14935.001.patch, HDFS-14935.002.patch > > -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Comment Edited] (HDFS-14935) Optimize DFSNetworkTopology#isNodeInScope
[ https://issues.apache.org/jira/browse/HDFS-14935?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16959623#comment-16959623 ] Lisheng Sun edited comment on HDFS-14935 at 10/25/19 9:55 AM: -- hi [~ayushtkn] this Jira is used to optimize code and no need to write repetitive code. use existing code as follow: {code:java} scope = NodeBase.normalize(scope); public static String normalize(String path) { if (path == null) { throw new IllegalArgumentException( "Network Location is null "); } if (path.length() == 0) { return ROOT; } if (path.charAt(0) != PATH_SEPARATOR) { throw new IllegalArgumentException( "Network Location path does not start with " +PATH_SEPARATOR_STR+ ": "+path); } int len = path.length(); if (path.charAt(len-1) == PATH_SEPARATOR) { return path.substring(0, len-1); } return path; } {code} was (Author: leosun08): this Jira is used to optimize code and no need to write repetitive code. use existing code as follow: {code:java} scope = NodeBase.normalize(scope); public static String normalize(String path) { if (path == null) { throw new IllegalArgumentException( "Network Location is null "); } if (path.length() == 0) { return ROOT; } if (path.charAt(0) != PATH_SEPARATOR) { throw new IllegalArgumentException( "Network Location path does not start with " +PATH_SEPARATOR_STR+ ": "+path); } int len = path.length(); if (path.charAt(len-1) == PATH_SEPARATOR) { return path.substring(0, len-1); } return path; } {code} > Optimize DFSNetworkTopology#isNodeInScope > - > > Key: HDFS-14935 > URL: https://issues.apache.org/jira/browse/HDFS-14935 > Project: Hadoop HDFS > Issue Type: Improvement >Reporter: Lisheng Sun >Assignee: Lisheng Sun >Priority: Major > Attachments: HDFS-14935.001.patch, HDFS-14935.002.patch > > -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-14935) Optimize DFSNetworkTopology#isNodeInScope
[ https://issues.apache.org/jira/browse/HDFS-14935?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16959623#comment-16959623 ] Lisheng Sun commented on HDFS-14935: this Jira is used to optimize code and no need to write repetitive code. use existing code as follow: {code:java} scope = NodeBase.normalize(scope); public static String normalize(String path) { if (path == null) { throw new IllegalArgumentException( "Network Location is null "); } if (path.length() == 0) { return ROOT; } if (path.charAt(0) != PATH_SEPARATOR) { throw new IllegalArgumentException( "Network Location path does not start with " +PATH_SEPARATOR_STR+ ": "+path); } int len = path.length(); if (path.charAt(len-1) == PATH_SEPARATOR) { return path.substring(0, len-1); } return path; } {code} > Optimize DFSNetworkTopology#isNodeInScope > - > > Key: HDFS-14935 > URL: https://issues.apache.org/jira/browse/HDFS-14935 > Project: Hadoop HDFS > Issue Type: Improvement >Reporter: Lisheng Sun >Assignee: Lisheng Sun >Priority: Major > Attachments: HDFS-14935.001.patch, HDFS-14935.002.patch > > -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-14935) Optimize DFSNetworkTopology#isNodeInScope
[ https://issues.apache.org/jira/browse/HDFS-14935?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Lisheng Sun updated HDFS-14935: --- Attachment: HDFS-14935.002.patch > Optimize DFSNetworkTopology#isNodeInScope > - > > Key: HDFS-14935 > URL: https://issues.apache.org/jira/browse/HDFS-14935 > Project: Hadoop HDFS > Issue Type: Improvement >Reporter: Lisheng Sun >Assignee: Lisheng Sun >Priority: Major > Attachments: HDFS-14935.001.patch, HDFS-14935.002.patch > > -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-14935) Optimize DFSNetworkTopology#isNodeInScope
[ https://issues.apache.org/jira/browse/HDFS-14935?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Lisheng Sun updated HDFS-14935: --- Summary: Optimize DFSNetworkTopology#isNodeInScope (was: Unified constant in DFSNetworkTopology#isNodeInScope) > Optimize DFSNetworkTopology#isNodeInScope > - > > Key: HDFS-14935 > URL: https://issues.apache.org/jira/browse/HDFS-14935 > Project: Hadoop HDFS > Issue Type: Improvement >Reporter: Lisheng Sun >Assignee: Lisheng Sun >Priority: Major > Attachments: HDFS-14935.001.patch > > -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-14935) Unified constant in DFSNetworkTopology#isNodeInScope
[ https://issues.apache.org/jira/browse/HDFS-14935?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Lisheng Sun updated HDFS-14935: --- Attachment: (was: HDFS-14935.001.patch) > Unified constant in DFSNetworkTopology#isNodeInScope > > > Key: HDFS-14935 > URL: https://issues.apache.org/jira/browse/HDFS-14935 > Project: Hadoop HDFS > Issue Type: Improvement >Reporter: Lisheng Sun >Assignee: Lisheng Sun >Priority: Major > Attachments: HDFS-14935.001.patch > > -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-14935) Unified constant in DFSNetworkTopology#isNodeInScope
[ https://issues.apache.org/jira/browse/HDFS-14935?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Lisheng Sun updated HDFS-14935: --- Attachment: HDFS-14935.001.patch > Unified constant in DFSNetworkTopology#isNodeInScope > > > Key: HDFS-14935 > URL: https://issues.apache.org/jira/browse/HDFS-14935 > Project: Hadoop HDFS > Issue Type: Improvement >Reporter: Lisheng Sun >Assignee: Lisheng Sun >Priority: Major > Attachments: HDFS-14935.001.patch > > -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Assigned] (HDFS-14935) Unified constant in DFSNetworkTopology#isNodeInScope
[ https://issues.apache.org/jira/browse/HDFS-14935?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Lisheng Sun reassigned HDFS-14935: -- Assignee: Lisheng Sun > Unified constant in DFSNetworkTopology#isNodeInScope > > > Key: HDFS-14935 > URL: https://issues.apache.org/jira/browse/HDFS-14935 > Project: Hadoop HDFS > Issue Type: Improvement >Reporter: Lisheng Sun >Assignee: Lisheng Sun >Priority: Major > Attachments: HDFS-14935.001.patch > > -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-14935) Unified constant in DFSNetworkTopology#isNodeInScope
[ https://issues.apache.org/jira/browse/HDFS-14935?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Lisheng Sun updated HDFS-14935: --- Attachment: HDFS-14935.001.patch Status: Patch Available (was: Open) > Unified constant in DFSNetworkTopology#isNodeInScope > > > Key: HDFS-14935 > URL: https://issues.apache.org/jira/browse/HDFS-14935 > Project: Hadoop HDFS > Issue Type: Improvement >Reporter: Lisheng Sun >Priority: Major > Attachments: HDFS-14935.001.patch > > -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Created] (HDFS-14935) Unified constant in DFSNetworkTopology#isNodeInScope
Lisheng Sun created HDFS-14935: -- Summary: Unified constant in DFSNetworkTopology#isNodeInScope Key: HDFS-14935 URL: https://issues.apache.org/jira/browse/HDFS-14935 Project: Hadoop HDFS Issue Type: Improvement Reporter: Lisheng Sun -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-11430) Separate class InnerNode from class NetworkTopology and make it extendable
[ https://issues.apache.org/jira/browse/HDFS-11430?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16958493#comment-16958493 ] Lisheng Sun commented on HDFS-11430: {code:java} @Override public Node getLeaf(int leafIndex, Node excludedNode) { int count=0; // check if the excluded node a leaf boolean isLeaf = !(excludedNode instanceof InnerNode); // calculate the total number of excluded leaf nodes int numOfExcludedLeaves = isLeaf ? 1 : ((InnerNode)excludedNode).getNumOfLeaves(); if (isLeafParent()) { // children are leaves if (isLeaf) { // excluded node is a leaf node if (excludedNode != null && childrenMap.containsKey(excludedNode.getName())) { int excludedIndex = children.indexOf(excludedNode); if (excludedIndex != -1 && leafIndex >= 0) { // excluded node is one of the children so adjust the leaf index leafIndex = leafIndex>=excludedIndex ? leafIndex+1 : leafIndex; } } } // range check if (leafIndex<0 || leafIndex>=this.getNumOfChildren()) { return null; } return children.get(leafIndex); } ... }{code} [~szetszwo] [~vagarychen] [~ayushtkn] [~elgoiri] the code InnerNodeImpl#getLeaf() as above i think it has two problems: 1.if childrenMap.containsKey(excludedNode.getName()) return true, children.indexOf(excludedNode) must return > -1, so if (excludedIndex != -1) is it necessary? 2. if excludedindex = children.size() -1 as current code: leafIndex = leafIndex>=excludedIndex ? leafIndex+1 : leafIndex; leafIndex will be out of index and return null. Actually there are nodes that can be returned. i think it should add the judgement excludedIndex == children.size() -1 Please correct me if i was wrong. Thank you. > Separate class InnerNode from class NetworkTopology and make it extendable > -- > > Key: HDFS-11430 > URL: https://issues.apache.org/jira/browse/HDFS-11430 > Project: Hadoop HDFS > Issue Type: Sub-task > Components: namenode >Reporter: Chen Liang >Assignee: Tsz-wo Sze >Priority: Major > Fix For: 2.9.0, 3.0.0-alpha4 > > Attachments: HDFS-11430-branch-2.001.patch, h11430_20170217.patch, > h11430_20170218.patch > > > The approach we will take in HDFS-11419 is to annotate topology's inner node > with more information, such that it chooses a subtree that meets storage type > requirement. However, {{InnerNode}} is not specific to HDFS, so our change > should affect other components using this class. > This JIRA separates {{InnerNode}} out of {{NetworkTopology}} and makes it > extendable. Therefore HDFS can have it's own customized inner node class, > while other services can still have inner node as what it is right now. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-14919) Provide Non DFS Used per DataNode in DataNode UI
[ https://issues.apache.org/jira/browse/HDFS-14919?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Lisheng Sun updated HDFS-14919: --- Attachment: HDFS-14919.001.patch Status: Patch Available (was: Open) > Provide Non DFS Used per DataNode in DataNode UI > > > Key: HDFS-14919 > URL: https://issues.apache.org/jira/browse/HDFS-14919 > Project: Hadoop HDFS > Issue Type: Bug >Reporter: Lisheng Sun >Assignee: Lisheng Sun >Priority: Major > Attachments: HDFS-14919.001.patch, hadoop2.6_datanode_ui.png, > hadoop3.1_datanode_ui.png > > -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-14283) DFSInputStream to prefer cached replica
[ https://issues.apache.org/jira/browse/HDFS-14283?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16957217#comment-16957217 ] Lisheng Sun commented on HDFS-14283: Thanx [~ayushtkn] [~hexiaoqiao] for your good suggestions. I added the switch for prefering cached replica and turned it off by default. Could you mind review the v004 patch? Thank you. > DFSInputStream to prefer cached replica > --- > > Key: HDFS-14283 > URL: https://issues.apache.org/jira/browse/HDFS-14283 > Project: Hadoop HDFS > Issue Type: Improvement >Affects Versions: 2.6.0 > Environment: HDFS Caching >Reporter: Wei-Chiu Chuang >Assignee: Lisheng Sun >Priority: Major > Attachments: HDFS-14283.001.patch, HDFS-14283.002.patch, > HDFS-14283.003.patch, HDFS-14283.004.patch > > > HDFS Caching offers performance benefits. However, currently NameNode does > not treat cached replica with higher priority, so HDFS caching is only useful > when cache replication = 3, that is to say, all replicas are cached in > memory, so that a client doesn't randomly pick an uncached replica. > HDFS-6846 proposed to let NameNode give higher priority to cached replica. > Changing a logic in NameNode is always tricky so that didn't get much > traction. Here I propose a different approach: let client (DFSInputStream) > prefer cached replica. > A {{LocatedBlock}} object already contains cached replica location so a > client has the needed information. I think we can change > {{DFSInputStream#getBestNodeDNAddrPair()}} for this purpose. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-14283) DFSInputStream to prefer cached replica
[ https://issues.apache.org/jira/browse/HDFS-14283?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Lisheng Sun updated HDFS-14283: --- Attachment: HDFS-14283.004.patch > DFSInputStream to prefer cached replica > --- > > Key: HDFS-14283 > URL: https://issues.apache.org/jira/browse/HDFS-14283 > Project: Hadoop HDFS > Issue Type: Improvement >Affects Versions: 2.6.0 > Environment: HDFS Caching >Reporter: Wei-Chiu Chuang >Assignee: Lisheng Sun >Priority: Major > Attachments: HDFS-14283.001.patch, HDFS-14283.002.patch, > HDFS-14283.003.patch, HDFS-14283.004.patch > > > HDFS Caching offers performance benefits. However, currently NameNode does > not treat cached replica with higher priority, so HDFS caching is only useful > when cache replication = 3, that is to say, all replicas are cached in > memory, so that a client doesn't randomly pick an uncached replica. > HDFS-6846 proposed to let NameNode give higher priority to cached replica. > Changing a logic in NameNode is always tricky so that didn't get much > traction. Here I propose a different approach: let client (DFSInputStream) > prefer cached replica. > A {{LocatedBlock}} object already contains cached replica location so a > client has the needed information. I think we can change > {{DFSInputStream#getBestNodeDNAddrPair()}} for this purpose. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-14919) Provide Non DFS Used per DataNode in DataNode UI
[ https://issues.apache.org/jira/browse/HDFS-14919?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Lisheng Sun updated HDFS-14919: --- Summary: Provide Non DFS Used per DataNode in DataNode UI (was: Provide Non DFS Used per disk in DataNode UI) > Provide Non DFS Used per DataNode in DataNode UI > > > Key: HDFS-14919 > URL: https://issues.apache.org/jira/browse/HDFS-14919 > Project: Hadoop HDFS > Issue Type: Bug >Reporter: Lisheng Sun >Assignee: Lisheng Sun >Priority: Major > Attachments: hadoop2.6_datanode_ui.png, hadoop3.1_datanode_ui.png > > -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-14919) Provide Non DFS Used per disk in DataNode UI
[ https://issues.apache.org/jira/browse/HDFS-14919?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16956605#comment-16956605 ] Lisheng Sun commented on HDFS-14919: [~elgoiri] [~hemanthboyina] I think we need to look at the no dfs used of each datanode from a cluster perspective. From a popup of the capacity bar, we can not see no dfs used of all datanode. It is necessary to add it back. Please correct me if was wrong. Thank you. > Provide Non DFS Used per disk in DataNode UI > > > Key: HDFS-14919 > URL: https://issues.apache.org/jira/browse/HDFS-14919 > Project: Hadoop HDFS > Issue Type: Bug >Reporter: Lisheng Sun >Assignee: Lisheng Sun >Priority: Major > Attachments: hadoop2.6_datanode_ui.png, hadoop3.1_datanode_ui.png > > -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Assigned] (HDFS-14919) Provide Non DFS Used per disk in DataNode UI
[ https://issues.apache.org/jira/browse/HDFS-14919?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Lisheng Sun reassigned HDFS-14919: -- Assignee: Lisheng Sun > Provide Non DFS Used per disk in DataNode UI > > > Key: HDFS-14919 > URL: https://issues.apache.org/jira/browse/HDFS-14919 > Project: Hadoop HDFS > Issue Type: Bug >Reporter: Lisheng Sun >Assignee: Lisheng Sun >Priority: Major > Attachments: hadoop2.6_datanode_ui.png, hadoop3.1_datanode_ui.png > > -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-14919) Provide Non DFS Used per disk in DataNode UI
[ https://issues.apache.org/jira/browse/HDFS-14919?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Lisheng Sun updated HDFS-14919: --- Attachment: hadoop3.1_datanode_ui.png hadoop2.6_datanode_ui.png > Provide Non DFS Used per disk in DataNode UI > > > Key: HDFS-14919 > URL: https://issues.apache.org/jira/browse/HDFS-14919 > Project: Hadoop HDFS > Issue Type: Bug >Reporter: Lisheng Sun >Priority: Major > Attachments: hadoop2.6_datanode_ui.png, hadoop3.1_datanode_ui.png > > -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Created] (HDFS-14919) Provide Non DFS Used per disk in DataNode UI
Lisheng Sun created HDFS-14919: -- Summary: Provide Non DFS Used per disk in DataNode UI Key: HDFS-14919 URL: https://issues.apache.org/jira/browse/HDFS-14919 Project: Hadoop HDFS Issue Type: Bug Reporter: Lisheng Sun -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-14283) DFSInputStream to prefer cached replica
[ https://issues.apache.org/jira/browse/HDFS-14283?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16955388#comment-16955388 ] Lisheng Sun commented on HDFS-14283: hi [~smeng] [~weichiu] [~ayushtkn] [~elgoiri] Could you have time to take a reivew for this patch? Thank you. > DFSInputStream to prefer cached replica > --- > > Key: HDFS-14283 > URL: https://issues.apache.org/jira/browse/HDFS-14283 > Project: Hadoop HDFS > Issue Type: Improvement >Affects Versions: 2.6.0 > Environment: HDFS Caching >Reporter: Wei-Chiu Chuang >Assignee: Lisheng Sun >Priority: Major > Attachments: HDFS-14283.001.patch, HDFS-14283.002.patch, > HDFS-14283.003.patch > > > HDFS Caching offers performance benefits. However, currently NameNode does > not treat cached replica with higher priority, so HDFS caching is only useful > when cache replication = 3, that is to say, all replicas are cached in > memory, so that a client doesn't randomly pick an uncached replica. > HDFS-6846 proposed to let NameNode give higher priority to cached replica. > Changing a logic in NameNode is always tricky so that didn't get much > traction. Here I propose a different approach: let client (DFSInputStream) > prefer cached replica. > A {{LocatedBlock}} object already contains cached replica location so a > client has the needed information. I think we can change > {{DFSInputStream#getBestNodeDNAddrPair()}} for this purpose. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-14913) Correct the value of available count in DFSNetworkTopology#chooseRandomWithStorageType()
[ https://issues.apache.org/jira/browse/HDFS-14913?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16955088#comment-16955088 ] Lisheng Sun commented on HDFS-14913: [^HDFS-14913-02.patch] LGTM +1. > Correct the value of available count in > DFSNetworkTopology#chooseRandomWithStorageType() > - > > Key: HDFS-14913 > URL: https://issues.apache.org/jira/browse/HDFS-14913 > Project: Hadoop HDFS > Issue Type: Bug >Reporter: Ayush Saxena >Assignee: Ayush Saxena >Priority: Major > Attachments: HDFS-14913-01.patch, HDFS-14913-02.patch > > > Presently if excluded scope is /default/rack1 and excluded node is > /default/rack10/node. Then the available count is not deducted. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-14283) DFSInputStream to prefer cached replica
[ https://issues.apache.org/jira/browse/HDFS-14283?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16950371#comment-16950371 ] Lisheng Sun commented on HDFS-14283: hi [~smeng] [~weichiu] Could you help review this patch? 1.the UT of TestDatanodeManager# testGetBlockLocationsWithCache for when the block is cached on one or more DataNodes, it should return the location of the nearest DataNode that has the cached block 2. the UT of TestDatanodeManager# testGetBlockLocationsWithNoCache for the block *isn't cached* on any DataNodes, fall back to the strategy without block cache (i.e. return the location of the nearest DataNode). Thank you. > DFSInputStream to prefer cached replica > --- > > Key: HDFS-14283 > URL: https://issues.apache.org/jira/browse/HDFS-14283 > Project: Hadoop HDFS > Issue Type: Improvement >Affects Versions: 2.6.0 > Environment: HDFS Caching >Reporter: Wei-Chiu Chuang >Assignee: Lisheng Sun >Priority: Major > Attachments: HDFS-14283.001.patch, HDFS-14283.002.patch, > HDFS-14283.003.patch > > > HDFS Caching offers performance benefits. However, currently NameNode does > not treat cached replica with higher priority, so HDFS caching is only useful > when cache replication = 3, that is to say, all replicas are cached in > memory, so that a client doesn't randomly pick an uncached replica. > HDFS-6846 proposed to let NameNode give higher priority to cached replica. > Changing a logic in NameNode is always tricky so that didn't get much > traction. Here I propose a different approach: let client (DFSInputStream) > prefer cached replica. > A {{LocatedBlock}} object already contains cached replica location so a > client has the needed information. I think we can change > {{DFSInputStream#getBestNodeDNAddrPair()}} for this purpose. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-6524) Choosing datanode retries times considering with block replica number
[ https://issues.apache.org/jira/browse/HDFS-6524?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16950286#comment-16950286 ] Lisheng Sun commented on HDFS-6524: --- [~ayushtkn] sorry, I did not understand the meaning of this failure before. According to this defination, it is not related to numbers of replicas. I apologize for this. > Choosing datanode retries times considering with block replica number > -- > > Key: HDFS-6524 > URL: https://issues.apache.org/jira/browse/HDFS-6524 > Project: Hadoop HDFS > Issue Type: Improvement > Components: hdfs-client >Affects Versions: 3.0.0-alpha1 >Reporter: Liang Xie >Assignee: Lisheng Sun >Priority: Minor > Labels: BB2015-05-TBR > Attachments: HDFS-6524.001.patch, HDFS-6524.002.patch, > HDFS-6524.003.patch, HDFS-6524.004.patch, HDFS-6524.005(2).patch, > HDFS-6524.005.patch, HDFS-6524.006.patch, HDFS-6524.007.patch, HDFS-6524.txt > > > Currently the chooseDataNode() does retry with the setting: > dfsClientConf.maxBlockAcquireFailures, which by default is 3 > (DFS_CLIENT_MAX_BLOCK_ACQUIRE_FAILURES_DEFAULT = 3), it would be better > having another option, block replication factor. One cluster with only two > block replica setting, or using Reed-solomon encoding solution with one > replica factor. It helps to reduce the long tail latency. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-14283) DFSInputStream to prefer cached replica
[ https://issues.apache.org/jira/browse/HDFS-14283?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Lisheng Sun updated HDFS-14283: --- Attachment: HDFS-14283.003.patch > DFSInputStream to prefer cached replica > --- > > Key: HDFS-14283 > URL: https://issues.apache.org/jira/browse/HDFS-14283 > Project: Hadoop HDFS > Issue Type: Improvement >Affects Versions: 2.6.0 > Environment: HDFS Caching >Reporter: Wei-Chiu Chuang >Assignee: Lisheng Sun >Priority: Major > Attachments: HDFS-14283.001.patch, HDFS-14283.002.patch, > HDFS-14283.003.patch > > > HDFS Caching offers performance benefits. However, currently NameNode does > not treat cached replica with higher priority, so HDFS caching is only useful > when cache replication = 3, that is to say, all replicas are cached in > memory, so that a client doesn't randomly pick an uncached replica. > HDFS-6846 proposed to let NameNode give higher priority to cached replica. > Changing a logic in NameNode is always tricky so that didn't get much > traction. Here I propose a different approach: let client (DFSInputStream) > prefer cached replica. > A {{LocatedBlock}} object already contains cached replica location so a > client has the needed information. I think we can change > {{DFSInputStream#getBestNodeDNAddrPair()}} for this purpose. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-6524) Choosing datanode retries times considering with block replica number
[ https://issues.apache.org/jira/browse/HDFS-6524?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16950239#comment-16950239 ] Lisheng Sun commented on HDFS-6524: --- Thanks [~ayushtkn] for your good suggestions. I updated the patch and uploaded the v007 patch. Could you help continue to review it? Thank you. > Choosing datanode retries times considering with block replica number > -- > > Key: HDFS-6524 > URL: https://issues.apache.org/jira/browse/HDFS-6524 > Project: Hadoop HDFS > Issue Type: Improvement > Components: hdfs-client >Affects Versions: 3.0.0-alpha1 >Reporter: Liang Xie >Assignee: Lisheng Sun >Priority: Minor > Labels: BB2015-05-TBR > Attachments: HDFS-6524.001.patch, HDFS-6524.002.patch, > HDFS-6524.003.patch, HDFS-6524.004.patch, HDFS-6524.005(2).patch, > HDFS-6524.005.patch, HDFS-6524.006.patch, HDFS-6524.007.patch, HDFS-6524.txt > > > Currently the chooseDataNode() does retry with the setting: > dfsClientConf.maxBlockAcquireFailures, which by default is 3 > (DFS_CLIENT_MAX_BLOCK_ACQUIRE_FAILURES_DEFAULT = 3), it would be better > having another option, block replication factor. One cluster with only two > block replica setting, or using Reed-solomon encoding solution with one > replica factor. It helps to reduce the long tail latency. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-6524) Choosing datanode retries times considering with block replica number
[ https://issues.apache.org/jira/browse/HDFS-6524?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Lisheng Sun updated HDFS-6524: -- Attachment: HDFS-6524.007.patch > Choosing datanode retries times considering with block replica number > -- > > Key: HDFS-6524 > URL: https://issues.apache.org/jira/browse/HDFS-6524 > Project: Hadoop HDFS > Issue Type: Improvement > Components: hdfs-client >Affects Versions: 3.0.0-alpha1 >Reporter: Liang Xie >Assignee: Lisheng Sun >Priority: Minor > Labels: BB2015-05-TBR > Attachments: HDFS-6524.001.patch, HDFS-6524.002.patch, > HDFS-6524.003.patch, HDFS-6524.004.patch, HDFS-6524.005(2).patch, > HDFS-6524.005.patch, HDFS-6524.006.patch, HDFS-6524.007.patch, HDFS-6524.txt > > > Currently the chooseDataNode() does retry with the setting: > dfsClientConf.maxBlockAcquireFailures, which by default is 3 > (DFS_CLIENT_MAX_BLOCK_ACQUIRE_FAILURES_DEFAULT = 3), it would be better > having another option, block replication factor. One cluster with only two > block replica setting, or using Reed-solomon encoding solution with one > replica factor. It helps to reduce the long tail latency. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-6524) Choosing datanode retries times considering with block replica number
[ https://issues.apache.org/jira/browse/HDFS-6524?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16950228#comment-16950228 ] Lisheng Sun commented on HDFS-6524: --- [~surendrasingh] [~weichiu] [~ayushtkn] Could you help review this patch? Thank you. > Choosing datanode retries times considering with block replica number > -- > > Key: HDFS-6524 > URL: https://issues.apache.org/jira/browse/HDFS-6524 > Project: Hadoop HDFS > Issue Type: Improvement > Components: hdfs-client >Affects Versions: 3.0.0-alpha1 >Reporter: Liang Xie >Assignee: Lisheng Sun >Priority: Minor > Labels: BB2015-05-TBR > Attachments: HDFS-6524.001.patch, HDFS-6524.002.patch, > HDFS-6524.003.patch, HDFS-6524.004.patch, HDFS-6524.005(2).patch, > HDFS-6524.005.patch, HDFS-6524.006.patch, HDFS-6524.txt > > > Currently the chooseDataNode() does retry with the setting: > dfsClientConf.maxBlockAcquireFailures, which by default is 3 > (DFS_CLIENT_MAX_BLOCK_ACQUIRE_FAILURES_DEFAULT = 3), it would be better > having another option, block replication factor. One cluster with only two > block replica setting, or using Reed-solomon encoding solution with one > replica factor. It helps to reduce the long tail latency. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Comment Edited] (HDFS-6524) Choosing datanode retries times considering with block replica number
[ https://issues.apache.org/jira/browse/HDFS-6524?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16950228#comment-16950228 ] Lisheng Sun edited comment on HDFS-6524 at 10/13/19 6:26 AM: - [~surendrasingh] [~weichiu] [~ayushtkn] [~smeng] Could you help review this patch? Thank you. was (Author: leosun08): [~surendrasingh] [~weichiu] [~ayushtkn] Could you help review this patch? Thank you. > Choosing datanode retries times considering with block replica number > -- > > Key: HDFS-6524 > URL: https://issues.apache.org/jira/browse/HDFS-6524 > Project: Hadoop HDFS > Issue Type: Improvement > Components: hdfs-client >Affects Versions: 3.0.0-alpha1 >Reporter: Liang Xie >Assignee: Lisheng Sun >Priority: Minor > Labels: BB2015-05-TBR > Attachments: HDFS-6524.001.patch, HDFS-6524.002.patch, > HDFS-6524.003.patch, HDFS-6524.004.patch, HDFS-6524.005(2).patch, > HDFS-6524.005.patch, HDFS-6524.006.patch, HDFS-6524.txt > > > Currently the chooseDataNode() does retry with the setting: > dfsClientConf.maxBlockAcquireFailures, which by default is 3 > (DFS_CLIENT_MAX_BLOCK_ACQUIRE_FAILURES_DEFAULT = 3), it would be better > having another option, block replication factor. One cluster with only two > block replica setting, or using Reed-solomon encoding solution with one > replica factor. It helps to reduce the long tail latency. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-14895) Define LOG instead of BlockPlacementPolicy.LOG in DatanodeDescriptor#chooseStorage4Block
[ https://issues.apache.org/jira/browse/HDFS-14895?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16950226#comment-16950226 ] Lisheng Sun commented on HDFS-14895: [~ayushtkn] Could you have time to continue to take a review for this patch? Thank you. > Define LOG instead of BlockPlacementPolicy.LOG in > DatanodeDescriptor#chooseStorage4Block > > > Key: HDFS-14895 > URL: https://issues.apache.org/jira/browse/HDFS-14895 > Project: Hadoop HDFS > Issue Type: Bug >Reporter: Lisheng Sun >Assignee: Lisheng Sun >Priority: Major > Attachments: HDFS-14895.001.patch > > > There is a noisy log with BlockPlacementPolicy.LOG, it's too hard to debug > problem. Define LOG instead of it in DatanodeDescriptor#chooseStorage4Block. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-6524) Choosing datanode retries times considering with block replica number
[ https://issues.apache.org/jira/browse/HDFS-6524?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16949052#comment-16949052 ] Lisheng Sun commented on HDFS-6524: --- ping [~elgoiri] Should we commit this patch to trunk and all other branches? Thank you. > Choosing datanode retries times considering with block replica number > -- > > Key: HDFS-6524 > URL: https://issues.apache.org/jira/browse/HDFS-6524 > Project: Hadoop HDFS > Issue Type: Improvement > Components: hdfs-client >Affects Versions: 3.0.0-alpha1 >Reporter: Liang Xie >Assignee: Lisheng Sun >Priority: Minor > Labels: BB2015-05-TBR > Attachments: HDFS-6524.001.patch, HDFS-6524.002.patch, > HDFS-6524.003.patch, HDFS-6524.004.patch, HDFS-6524.005(2).patch, > HDFS-6524.005.patch, HDFS-6524.006.patch, HDFS-6524.txt > > > Currently the chooseDataNode() does retry with the setting: > dfsClientConf.maxBlockAcquireFailures, which by default is 3 > (DFS_CLIENT_MAX_BLOCK_ACQUIRE_FAILURES_DEFAULT = 3), it would be better > having another option, block replication factor. One cluster with only two > block replica setting, or using Reed-solomon encoding solution with one > replica factor. It helps to reduce the long tail latency. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-14896) Make MIN_BLOCKS_FOR_WRITE configurable
[ https://issues.apache.org/jira/browse/HDFS-14896?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Lisheng Sun updated HDFS-14896: --- Attachment: HDFS-14896.005.patch > Make MIN_BLOCKS_FOR_WRITE configurable > -- > > Key: HDFS-14896 > URL: https://issues.apache.org/jira/browse/HDFS-14896 > Project: Hadoop HDFS > Issue Type: Improvement >Reporter: Lisheng Sun >Assignee: Lisheng Sun >Priority: Minor > Attachments: HDFS-14896.001.patch, HDFS-14896.002.patch, > HDFS-14896.003(2).patch, HDFS-14896.003.patch, HDFS-14896.004.patch, > HDFS-14896.005.patch > > -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-14896) Make MIN_BLOCKS_FOR_WRITE configurable
[ https://issues.apache.org/jira/browse/HDFS-14896?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16947792#comment-16947792 ] Lisheng Sun commented on HDFS-14896: [~ayushtkn] Could you help continue to take a review for the v005 patch? Thank you. > Make MIN_BLOCKS_FOR_WRITE configurable > -- > > Key: HDFS-14896 > URL: https://issues.apache.org/jira/browse/HDFS-14896 > Project: Hadoop HDFS > Issue Type: Improvement >Reporter: Lisheng Sun >Assignee: Lisheng Sun >Priority: Minor > Attachments: HDFS-14896.001.patch, HDFS-14896.002.patch, > HDFS-14896.003(2).patch, HDFS-14896.003.patch, HDFS-14896.004.patch, > HDFS-14896.005.patch > > -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-14896) Make MIN_BLOCKS_FOR_WRITE configurable
[ https://issues.apache.org/jira/browse/HDFS-14896?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Lisheng Sun updated HDFS-14896: --- Attachment: HDFS-14896.004.patch > Make MIN_BLOCKS_FOR_WRITE configurable > -- > > Key: HDFS-14896 > URL: https://issues.apache.org/jira/browse/HDFS-14896 > Project: Hadoop HDFS > Issue Type: Improvement >Reporter: Lisheng Sun >Assignee: Lisheng Sun >Priority: Minor > Attachments: HDFS-14896.001.patch, HDFS-14896.002.patch, > HDFS-14896.003(2).patch, HDFS-14896.003.patch, HDFS-14896.004.patch > > -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-14896) Make MIN_BLOCKS_FOR_WRITE configurable
[ https://issues.apache.org/jira/browse/HDFS-14896?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Lisheng Sun updated HDFS-14896: --- Attachment: HDFS-14896.003(2).patch > Make MIN_BLOCKS_FOR_WRITE configurable > -- > > Key: HDFS-14896 > URL: https://issues.apache.org/jira/browse/HDFS-14896 > Project: Hadoop HDFS > Issue Type: Improvement >Reporter: Lisheng Sun >Assignee: Lisheng Sun >Priority: Minor > Attachments: HDFS-14896.001.patch, HDFS-14896.002.patch, > HDFS-14896.003(2).patch, HDFS-14896.003.patch > > -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-14896) Make MIN_BLOCKS_FOR_WRITE configurable
[ https://issues.apache.org/jira/browse/HDFS-14896?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Lisheng Sun updated HDFS-14896: --- Attachment: HDFS-14896.003.patch > Make MIN_BLOCKS_FOR_WRITE configurable > -- > > Key: HDFS-14896 > URL: https://issues.apache.org/jira/browse/HDFS-14896 > Project: Hadoop HDFS > Issue Type: Improvement >Reporter: Lisheng Sun >Assignee: Lisheng Sun >Priority: Minor > Attachments: HDFS-14896.001.patch, HDFS-14896.002.patch, > HDFS-14896.003.patch > > -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-14896) Make MIN_BLOCKS_FOR_WRITE configurable
[ https://issues.apache.org/jira/browse/HDFS-14896?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Lisheng Sun updated HDFS-14896: --- Attachment: HDFS-14896.003.patch > Make MIN_BLOCKS_FOR_WRITE configurable > -- > > Key: HDFS-14896 > URL: https://issues.apache.org/jira/browse/HDFS-14896 > Project: Hadoop HDFS > Issue Type: Improvement >Reporter: Lisheng Sun >Assignee: Lisheng Sun >Priority: Minor > Attachments: HDFS-14896.001.patch, HDFS-14896.002.patch > > -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-14896) Make MIN_BLOCKS_FOR_WRITE configurable
[ https://issues.apache.org/jira/browse/HDFS-14896?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Lisheng Sun updated HDFS-14896: --- Attachment: (was: HDFS-14896.003.patch) > Make MIN_BLOCKS_FOR_WRITE configurable > -- > > Key: HDFS-14896 > URL: https://issues.apache.org/jira/browse/HDFS-14896 > Project: Hadoop HDFS > Issue Type: Improvement >Reporter: Lisheng Sun >Assignee: Lisheng Sun >Priority: Minor > Attachments: HDFS-14896.001.patch, HDFS-14896.002.patch > > -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-6524) Choosing datanode retries times considering with block replica number
[ https://issues.apache.org/jira/browse/HDFS-6524?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16946665#comment-16946665 ] Lisheng Sun commented on HDFS-6524: --- [~elgoiri] Could you help commit this patch to trunk? Thank you a alot. > Choosing datanode retries times considering with block replica number > -- > > Key: HDFS-6524 > URL: https://issues.apache.org/jira/browse/HDFS-6524 > Project: Hadoop HDFS > Issue Type: Improvement > Components: hdfs-client >Affects Versions: 3.0.0-alpha1 >Reporter: Liang Xie >Assignee: Lisheng Sun >Priority: Minor > Labels: BB2015-05-TBR > Attachments: HDFS-6524.001.patch, HDFS-6524.002.patch, > HDFS-6524.003.patch, HDFS-6524.004.patch, HDFS-6524.005(2).patch, > HDFS-6524.005.patch, HDFS-6524.006.patch, HDFS-6524.txt > > > Currently the chooseDataNode() does retry with the setting: > dfsClientConf.maxBlockAcquireFailures, which by default is 3 > (DFS_CLIENT_MAX_BLOCK_ACQUIRE_FAILURES_DEFAULT = 3), it would be better > having another option, block replication factor. One cluster with only two > block replica setting, or using Reed-solomon encoding solution with one > replica factor. It helps to reduce the long tail latency. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-14895) Define LOG instead of BlockPlacementPolicy.LOG in DatanodeDescriptor#chooseStorage4Block
[ https://issues.apache.org/jira/browse/HDFS-14895?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16946401#comment-16946401 ] Lisheng Sun commented on HDFS-14895: [~ayushtkn] -HDFS-9023 defines BlockPlacementPolicy.LOG.debug instead of LOG.debug. i think LOG.debug is more reasonable. Unified definition of log mode. Please correct me if was wrong. Thank you a lot [~ayushtkn] > Define LOG instead of BlockPlacementPolicy.LOG in > DatanodeDescriptor#chooseStorage4Block > > > Key: HDFS-14895 > URL: https://issues.apache.org/jira/browse/HDFS-14895 > Project: Hadoop HDFS > Issue Type: Bug >Reporter: Lisheng Sun >Assignee: Lisheng Sun >Priority: Major > Attachments: HDFS-14895.001.patch > > > There is a noisy log with BlockPlacementPolicy.LOG, it's too hard to debug > problem. Define LOG instead of it in DatanodeDescriptor#chooseStorage4Block. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-14895) Define LOG instead of BlockPlacementPolicy.LOG in DatanodeDescriptor#chooseStorage4Block
[ https://issues.apache.org/jira/browse/HDFS-14895?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16945561#comment-16945561 ] Lisheng Sun commented on HDFS-14895: hi [~ayushtkn], Could you have time to take a reivew for this patch? Thank you. > Define LOG instead of BlockPlacementPolicy.LOG in > DatanodeDescriptor#chooseStorage4Block > > > Key: HDFS-14895 > URL: https://issues.apache.org/jira/browse/HDFS-14895 > Project: Hadoop HDFS > Issue Type: Bug >Reporter: Lisheng Sun >Assignee: Lisheng Sun >Priority: Major > Attachments: HDFS-14895.001.patch > > > There is a noisy log with BlockPlacementPolicy.LOG, it's too hard to debug > problem. Define LOG instead of it in DatanodeDescriptor#chooseStorage4Block. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-14896) Make MIN_BLOCKS_FOR_WRITE configurable
[ https://issues.apache.org/jira/browse/HDFS-14896?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Lisheng Sun updated HDFS-14896: --- Attachment: HDFS-14896.002.patch > Make MIN_BLOCKS_FOR_WRITE configurable > -- > > Key: HDFS-14896 > URL: https://issues.apache.org/jira/browse/HDFS-14896 > Project: Hadoop HDFS > Issue Type: Improvement >Reporter: Lisheng Sun >Assignee: Lisheng Sun >Priority: Minor > Attachments: HDFS-14896.001.patch, HDFS-14896.002.patch > > -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-14896) Make MIN_BLOCKS_FOR_WRITE configurable
[ https://issues.apache.org/jira/browse/HDFS-14896?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Lisheng Sun updated HDFS-14896: --- Attachment: HDFS-14896.001.patch Status: Patch Available (was: Open) > Make MIN_BLOCKS_FOR_WRITE configurable > -- > > Key: HDFS-14896 > URL: https://issues.apache.org/jira/browse/HDFS-14896 > Project: Hadoop HDFS > Issue Type: Improvement >Reporter: Lisheng Sun >Assignee: Lisheng Sun >Priority: Minor > Attachments: HDFS-14896.001.patch > > -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-14896) Make MIN_BLOCKS_FOR_WRITE configurable
[ https://issues.apache.org/jira/browse/HDFS-14896?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Lisheng Sun updated HDFS-14896: --- Summary: Make MIN_BLOCKS_FOR_WRITE configurable (was: Make MIN_BLOCKS_FOR_WRITE configuable) > Make MIN_BLOCKS_FOR_WRITE configurable > -- > > Key: HDFS-14896 > URL: https://issues.apache.org/jira/browse/HDFS-14896 > Project: Hadoop HDFS > Issue Type: Improvement >Reporter: Lisheng Sun >Assignee: Lisheng Sun >Priority: Minor > -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Created] (HDFS-14896) Make MIN_BLOCKS_FOR_WRITE configuable
Lisheng Sun created HDFS-14896: -- Summary: Make MIN_BLOCKS_FOR_WRITE configuable Key: HDFS-14896 URL: https://issues.apache.org/jira/browse/HDFS-14896 Project: Hadoop HDFS Issue Type: Improvement Reporter: Lisheng Sun Assignee: Lisheng Sun -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-14895) Define LOG instead of BlockPlacementPolicy.LOG in DatanodeDescriptor#chooseStorage4Block
[ https://issues.apache.org/jira/browse/HDFS-14895?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Lisheng Sun updated HDFS-14895: --- Attachment: HDFS-14895.001.patch Status: Patch Available (was: Open) > Define LOG instead of BlockPlacementPolicy.LOG in > DatanodeDescriptor#chooseStorage4Block > > > Key: HDFS-14895 > URL: https://issues.apache.org/jira/browse/HDFS-14895 > Project: Hadoop HDFS > Issue Type: Bug >Reporter: Lisheng Sun >Assignee: Lisheng Sun >Priority: Major > Attachments: HDFS-14895.001.patch > > > There is a noisy log with BlockPlacementPolicy.LOG, it's too hard to debug > problem. Define LOG instead of it in DatanodeDescriptor#chooseStorage4Block. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Created] (HDFS-14895) Define LOG instead of BlockPlacementPolicy.LOG in DatanodeDescriptor#chooseStorage4Block
Lisheng Sun created HDFS-14895: -- Summary: Define LOG instead of BlockPlacementPolicy.LOG in DatanodeDescriptor#chooseStorage4Block Key: HDFS-14895 URL: https://issues.apache.org/jira/browse/HDFS-14895 Project: Hadoop HDFS Issue Type: Bug Reporter: Lisheng Sun Assignee: Lisheng Sun There is a noisy log with BlockPlacementPolicy.LOG, it's too hard to debug problem. Define LOG instead of it in DatanodeDescriptor#chooseStorage4Block. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-14877) Remove unused imports from TestClose.java
[ https://issues.apache.org/jira/browse/HDFS-14877?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Lisheng Sun updated HDFS-14877: --- Resolution: Implemented Status: Resolved (was: Patch Available) > Remove unused imports from TestClose.java > - > > Key: HDFS-14877 > URL: https://issues.apache.org/jira/browse/HDFS-14877 > Project: Hadoop HDFS > Issue Type: Bug > Components: test >Reporter: Lisheng Sun >Assignee: Lisheng Sun >Priority: Minor > Attachments: HDFS-14877.000.patch > > > There 1 unused import in TestClose.java. Let's clean them up. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-6524) Choosing datanode retries times considering with block replica number
[ https://issues.apache.org/jira/browse/HDFS-6524?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16944208#comment-16944208 ] Lisheng Sun commented on HDFS-6524: --- [~elgoiri] In current code TestDFSClientRetries#testFailuresArePerOperation and TestDFSClientRetries#testDFSClientRetriesOnBusyBlocks have a replication factor of 1. After updated patch TestDFSClientRetries#testFailuresArePerOperation have a replication of 1 and TestDFSClientRetries#testDFSClientRetriesOnBusyBlocks have replication factor of 3. Both of these two replication factor are tested. Please correct me if I was wrong. Thanks a lot [~elgoiri]. > Choosing datanode retries times considering with block replica number > -- > > Key: HDFS-6524 > URL: https://issues.apache.org/jira/browse/HDFS-6524 > Project: Hadoop HDFS > Issue Type: Improvement > Components: hdfs-client >Affects Versions: 3.0.0-alpha1 >Reporter: Liang Xie >Assignee: Lisheng Sun >Priority: Minor > Labels: BB2015-05-TBR > Attachments: HDFS-6524.001.patch, HDFS-6524.002.patch, > HDFS-6524.003.patch, HDFS-6524.004.patch, HDFS-6524.005(2).patch, > HDFS-6524.005.patch, HDFS-6524.006.patch, HDFS-6524.txt > > > Currently the chooseDataNode() does retry with the setting: > dfsClientConf.maxBlockAcquireFailures, which by default is 3 > (DFS_CLIENT_MAX_BLOCK_ACQUIRE_FAILURES_DEFAULT = 3), it would be better > having another option, block replication factor. One cluster with only two > block replica setting, or using Reed-solomon encoding solution with one > replica factor. It helps to reduce the long tail latency. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-14283) DFSInputStream to prefer cached replica
[ https://issues.apache.org/jira/browse/HDFS-14283?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16942552#comment-16942552 ] Lisheng Sun commented on HDFS-14283: Thanks [~smeng] for good comment and review. Current *block.getCachedLocations()* which doesn't get DNs in priority order. I think your option that some DN could become a hotspot. So i add priority order for cache locations. But i have a problem that current *block.getLocations()* which gets a list of DataNodes in priority order does not consider choosed DN LOAD, bandwidth etc. I think it is necessary to add this logic later. Please correct me if I was wrong. > DFSInputStream to prefer cached replica > --- > > Key: HDFS-14283 > URL: https://issues.apache.org/jira/browse/HDFS-14283 > Project: Hadoop HDFS > Issue Type: Improvement >Affects Versions: 2.6.0 > Environment: HDFS Caching >Reporter: Wei-Chiu Chuang >Assignee: Lisheng Sun >Priority: Major > Attachments: HDFS-14283.001.patch, HDFS-14283.002.patch > > > HDFS Caching offers performance benefits. However, currently NameNode does > not treat cached replica with higher priority, so HDFS caching is only useful > when cache replication = 3, that is to say, all replicas are cached in > memory, so that a client doesn't randomly pick an uncached replica. > HDFS-6846 proposed to let NameNode give higher priority to cached replica. > Changing a logic in NameNode is always tricky so that didn't get much > traction. Here I propose a different approach: let client (DFSInputStream) > prefer cached replica. > A {{LocatedBlock}} object already contains cached replica location so a > client has the needed information. I think we can change > {{DFSInputStream#getBestNodeDNAddrPair()}} for this purpose. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-14283) DFSInputStream to prefer cached replica
[ https://issues.apache.org/jira/browse/HDFS-14283?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Lisheng Sun updated HDFS-14283: --- Attachment: HDFS-14283.002.patch > DFSInputStream to prefer cached replica > --- > > Key: HDFS-14283 > URL: https://issues.apache.org/jira/browse/HDFS-14283 > Project: Hadoop HDFS > Issue Type: Improvement >Affects Versions: 2.6.0 > Environment: HDFS Caching >Reporter: Wei-Chiu Chuang >Assignee: Lisheng Sun >Priority: Major > Attachments: HDFS-14283.001.patch, HDFS-14283.002.patch > > > HDFS Caching offers performance benefits. However, currently NameNode does > not treat cached replica with higher priority, so HDFS caching is only useful > when cache replication = 3, that is to say, all replicas are cached in > memory, so that a client doesn't randomly pick an uncached replica. > HDFS-6846 proposed to let NameNode give higher priority to cached replica. > Changing a logic in NameNode is always tricky so that didn't get much > traction. Here I propose a different approach: let client (DFSInputStream) > prefer cached replica. > A {{LocatedBlock}} object already contains cached replica location so a > client has the needed information. I think we can change > {{DFSInputStream#getBestNodeDNAddrPair()}} for this purpose. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-14876) Remove unused imports from TestBlockMissingException.java and TestClose.java
[ https://issues.apache.org/jira/browse/HDFS-14876?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16940073#comment-16940073 ] Lisheng Sun commented on HDFS-14876: Hi [~ayushtkn] HDFS-14849 remove unused imports in TestBlockMissingException.java. So i updated this patch and uploaded the v002 patch. Would you mind taking a review for patch? Thank you. > Remove unused imports from TestBlockMissingException.java and TestClose.java > > > Key: HDFS-14876 > URL: https://issues.apache.org/jira/browse/HDFS-14876 > Project: Hadoop HDFS > Issue Type: Bug > Components: test >Reporter: Lisheng Sun >Assignee: Lisheng Sun >Priority: Minor > Attachments: HDFS-14876.000.patch, HDFS-14876.001.patch, > HDFS-14876.002.patch > > > There 3 unused imports in TestBlockMissingException.java and TestClose.java. > Let's clean them up. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-14876) Remove unused imports from TestBlockMissingException.java and TestClose.java
[ https://issues.apache.org/jira/browse/HDFS-14876?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Lisheng Sun updated HDFS-14876: --- Attachment: HDFS-14876.002.patch > Remove unused imports from TestBlockMissingException.java and TestClose.java > > > Key: HDFS-14876 > URL: https://issues.apache.org/jira/browse/HDFS-14876 > Project: Hadoop HDFS > Issue Type: Bug > Components: test >Reporter: Lisheng Sun >Assignee: Lisheng Sun >Priority: Minor > Attachments: HDFS-14876.000.patch, HDFS-14876.001.patch, > HDFS-14876.002.patch > > > There 3 unused imports in TestBlockMissingException.java and TestClose.java. > Let's clean them up. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-14850) Optimize FileSystemAccessService#getFileSystemConfiguration
[ https://issues.apache.org/jira/browse/HDFS-14850?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16939823#comment-16939823 ] Lisheng Sun commented on HDFS-14850: hi [~elgoiri] Should we commit this patch to trunk? Thank you. > Optimize FileSystemAccessService#getFileSystemConfiguration > --- > > Key: HDFS-14850 > URL: https://issues.apache.org/jira/browse/HDFS-14850 > Project: Hadoop HDFS > Issue Type: Improvement > Components: httpfs, performance >Reporter: Lisheng Sun >Assignee: Lisheng Sun >Priority: Major > Attachments: HDFS-14850.001.patch, HDFS-14850.002.patch, > HDFS-14850.003.patch, HDFS-14850.004(2).patch, HDFS-14850.004.patch, > HDFS-14850.005.patch > > > {code:java} > @Override > public Configuration getFileSystemConfiguration() { > Configuration conf = new Configuration(true); > ConfigurationUtils.copy(serviceHadoopConf, conf); > conf.setBoolean(FILE_SYSTEM_SERVICE_CREATED, true); > // Force-clear server-side umask to make HttpFS match WebHDFS behavior > conf.set(FsPermission.UMASK_LABEL, "000"); > return conf; > } > {code} > As above code,when call > FileSystemAccessService#getFileSystemConfiguration,current code new > Configuration every time. > It is not necessary and affects performance. I think it only need to new > Configuration in FileSystemAccessService#init once and > FileSystemAccessService#getFileSystemConfiguration get it. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-6524) Choosing datanode retries times considering with block replica number
[ https://issues.apache.org/jira/browse/HDFS-6524?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16939822#comment-16939822 ] Lisheng Sun commented on HDFS-6524: --- [~elgoiri] Could you have time to continue to reveiew this patch? Thank you. > Choosing datanode retries times considering with block replica number > -- > > Key: HDFS-6524 > URL: https://issues.apache.org/jira/browse/HDFS-6524 > Project: Hadoop HDFS > Issue Type: Improvement > Components: hdfs-client >Affects Versions: 3.0.0-alpha1 >Reporter: Liang Xie >Assignee: Lisheng Sun >Priority: Minor > Labels: BB2015-05-TBR > Attachments: HDFS-6524.001.patch, HDFS-6524.002.patch, > HDFS-6524.003.patch, HDFS-6524.004.patch, HDFS-6524.005(2).patch, > HDFS-6524.005.patch, HDFS-6524.006.patch, HDFS-6524.txt > > > Currently the chooseDataNode() does retry with the setting: > dfsClientConf.maxBlockAcquireFailures, which by default is 3 > (DFS_CLIENT_MAX_BLOCK_ACQUIRE_FAILURES_DEFAULT = 3), it would be better > having another option, block replication factor. One cluster with only two > block replica setting, or using Reed-solomon encoding solution with one > replica factor. It helps to reduce the long tail latency. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-6524) Choosing datanode retries times considering with block replica number
[ https://issues.apache.org/jira/browse/HDFS-6524?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16939482#comment-16939482 ] Lisheng Sun commented on HDFS-6524: --- all failed unit tests are not related to this patch. > Choosing datanode retries times considering with block replica number > -- > > Key: HDFS-6524 > URL: https://issues.apache.org/jira/browse/HDFS-6524 > Project: Hadoop HDFS > Issue Type: Improvement > Components: hdfs-client >Affects Versions: 3.0.0-alpha1 >Reporter: Liang Xie >Assignee: Lisheng Sun >Priority: Minor > Labels: BB2015-05-TBR > Attachments: HDFS-6524.001.patch, HDFS-6524.002.patch, > HDFS-6524.003.patch, HDFS-6524.004.patch, HDFS-6524.005(2).patch, > HDFS-6524.005.patch, HDFS-6524.006.patch, HDFS-6524.txt > > > Currently the chooseDataNode() does retry with the setting: > dfsClientConf.maxBlockAcquireFailures, which by default is 3 > (DFS_CLIENT_MAX_BLOCK_ACQUIRE_FAILURES_DEFAULT = 3), it would be better > having another option, block replication factor. One cluster with only two > block replica setting, or using Reed-solomon encoding solution with one > replica factor. It helps to reduce the long tail latency. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-14876) Remove unused imports from TestBlockMissingException.java and TestClose.java
[ https://issues.apache.org/jira/browse/HDFS-14876?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16939476#comment-16939476 ] Lisheng Sun commented on HDFS-14876: Thanks [~ayushtkn] for good suggestions. I updated this patch and uploaded the v001 patch. Thank you. > Remove unused imports from TestBlockMissingException.java and TestClose.java > > > Key: HDFS-14876 > URL: https://issues.apache.org/jira/browse/HDFS-14876 > Project: Hadoop HDFS > Issue Type: Bug > Components: test >Reporter: Lisheng Sun >Assignee: Lisheng Sun >Priority: Minor > Attachments: HDFS-14876.000.patch, HDFS-14876.001.patch > > > There 3 unused imports in TestBlockMissingException.java and TestClose.java. > Let's clean them up. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-14876) Remove unused imports from TestBlockMissingException.java and TestClose.java
[ https://issues.apache.org/jira/browse/HDFS-14876?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Lisheng Sun updated HDFS-14876: --- Description: There 3 unused imports in TestBlockMissingException.java and TestClose.java. Let's clean them up. (was: There 2 unused imports in TestBlockMissingException.java. Let's clean them up.) > Remove unused imports from TestBlockMissingException.java and TestClose.java > > > Key: HDFS-14876 > URL: https://issues.apache.org/jira/browse/HDFS-14876 > Project: Hadoop HDFS > Issue Type: Bug > Components: test >Reporter: Lisheng Sun >Assignee: Lisheng Sun >Priority: Minor > Attachments: HDFS-14876.000.patch, HDFS-14876.001.patch > > > There 3 unused imports in TestBlockMissingException.java and TestClose.java. > Let's clean them up. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-14876) Remove unused imports from TestBlockMissingException.java and TestClose.java
[ https://issues.apache.org/jira/browse/HDFS-14876?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Lisheng Sun updated HDFS-14876: --- Summary: Remove unused imports from TestBlockMissingException.java and TestClose.java (was: Remove unused imports from TestBlockMissingException.java) > Remove unused imports from TestBlockMissingException.java and TestClose.java > > > Key: HDFS-14876 > URL: https://issues.apache.org/jira/browse/HDFS-14876 > Project: Hadoop HDFS > Issue Type: Bug > Components: test >Reporter: Lisheng Sun >Assignee: Lisheng Sun >Priority: Minor > Attachments: HDFS-14876.000.patch, HDFS-14876.001.patch > > > There 2 unused imports in TestBlockMissingException.java. Let's clean them up. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-14876) Remove unused imports from TestBlockMissingException.java
[ https://issues.apache.org/jira/browse/HDFS-14876?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Lisheng Sun updated HDFS-14876: --- Attachment: HDFS-14876.001.patch > Remove unused imports from TestBlockMissingException.java > - > > Key: HDFS-14876 > URL: https://issues.apache.org/jira/browse/HDFS-14876 > Project: Hadoop HDFS > Issue Type: Bug > Components: test >Reporter: Lisheng Sun >Assignee: Lisheng Sun >Priority: Minor > Attachments: HDFS-14876.000.patch, HDFS-14876.001.patch > > > There 2 unused imports in TestBlockMissingException.java. Let's clean them up. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-14877) Remove unused imports from TestClose.java
[ https://issues.apache.org/jira/browse/HDFS-14877?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Lisheng Sun updated HDFS-14877: --- Attachment: HDFS-14877.000.patch Status: Patch Available (was: Open) > Remove unused imports from TestClose.java > - > > Key: HDFS-14877 > URL: https://issues.apache.org/jira/browse/HDFS-14877 > Project: Hadoop HDFS > Issue Type: Bug > Components: test >Reporter: Lisheng Sun >Assignee: Lisheng Sun >Priority: Minor > Attachments: HDFS-14877.000.patch > > > There 1 unused import in TestClose.java. Let's clean them up. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Created] (HDFS-14877) Remove unused imports from TestClose.java
Lisheng Sun created HDFS-14877: -- Summary: Remove unused imports from TestClose.java Key: HDFS-14877 URL: https://issues.apache.org/jira/browse/HDFS-14877 Project: Hadoop HDFS Issue Type: Bug Components: test Reporter: Lisheng Sun Assignee: Lisheng Sun There 1 unused import in TestClose.java. Let's clean them up. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-14876) Remove unused imports from TestBlockMissingException.java
[ https://issues.apache.org/jira/browse/HDFS-14876?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Lisheng Sun updated HDFS-14876: --- Attachment: HDFS-14876.000.patch Status: Patch Available (was: Open) > Remove unused imports from TestBlockMissingException.java > - > > Key: HDFS-14876 > URL: https://issues.apache.org/jira/browse/HDFS-14876 > Project: Hadoop HDFS > Issue Type: Bug > Components: test >Reporter: Lisheng Sun >Assignee: Lisheng Sun >Priority: Minor > Attachments: HDFS-14876.000.patch > > > There 2 unused imports in TestBlockMissingException.java. Let's clean them up. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Created] (HDFS-14876) Remove unused imports from TestBlockMissingException.java
Lisheng Sun created HDFS-14876: -- Summary: Remove unused imports from TestBlockMissingException.java Key: HDFS-14876 URL: https://issues.apache.org/jira/browse/HDFS-14876 Project: Hadoop HDFS Issue Type: Bug Components: test Reporter: Lisheng Sun Assignee: Lisheng Sun There 2 unused imports in TestBlockMissingException.java. Let's clean them up. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-6524) Choosing datanode retries times considering with block replica number
[ https://issues.apache.org/jira/browse/HDFS-6524?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16938796#comment-16938796 ] Lisheng Sun commented on HDFS-6524: --- Thanks [~elgoiri] for your good suggestions. Updated UT. TestDFSClientRetries#testFailuresArePerOperation for 1 DN, and TestDFSClientRetries#testDFSClientRetriesOnBusyBlocks for 3 DN. Uploaded the v006 patch. Thank you a lot. > Choosing datanode retries times considering with block replica number > -- > > Key: HDFS-6524 > URL: https://issues.apache.org/jira/browse/HDFS-6524 > Project: Hadoop HDFS > Issue Type: Improvement > Components: hdfs-client >Affects Versions: 3.0.0-alpha1 >Reporter: Liang Xie >Assignee: Lisheng Sun >Priority: Minor > Labels: BB2015-05-TBR > Attachments: HDFS-6524.001.patch, HDFS-6524.002.patch, > HDFS-6524.003.patch, HDFS-6524.004.patch, HDFS-6524.005(2).patch, > HDFS-6524.005.patch, HDFS-6524.006.patch, HDFS-6524.txt > > > Currently the chooseDataNode() does retry with the setting: > dfsClientConf.maxBlockAcquireFailures, which by default is 3 > (DFS_CLIENT_MAX_BLOCK_ACQUIRE_FAILURES_DEFAULT = 3), it would be better > having another option, block replication factor. One cluster with only two > block replica setting, or using Reed-solomon encoding solution with one > replica factor. It helps to reduce the long tail latency. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-6524) Choosing datanode retries times considering with block replica number
[ https://issues.apache.org/jira/browse/HDFS-6524?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Lisheng Sun updated HDFS-6524: -- Attachment: HDFS-6524.006.patch > Choosing datanode retries times considering with block replica number > -- > > Key: HDFS-6524 > URL: https://issues.apache.org/jira/browse/HDFS-6524 > Project: Hadoop HDFS > Issue Type: Improvement > Components: hdfs-client >Affects Versions: 3.0.0-alpha1 >Reporter: Liang Xie >Assignee: Lisheng Sun >Priority: Minor > Labels: BB2015-05-TBR > Attachments: HDFS-6524.001.patch, HDFS-6524.002.patch, > HDFS-6524.003.patch, HDFS-6524.004.patch, HDFS-6524.005(2).patch, > HDFS-6524.005.patch, HDFS-6524.006.patch, HDFS-6524.txt > > > Currently the chooseDataNode() does retry with the setting: > dfsClientConf.maxBlockAcquireFailures, which by default is 3 > (DFS_CLIENT_MAX_BLOCK_ACQUIRE_FAILURES_DEFAULT = 3), it would be better > having another option, block replication factor. One cluster with only two > block replica setting, or using Reed-solomon encoding solution with one > replica factor. It helps to reduce the long tail latency. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-6524) Choosing datanode retries times considering with block replica number
[ https://issues.apache.org/jira/browse/HDFS-6524?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Lisheng Sun updated HDFS-6524: -- Attachment: (was: HDFS-6524.006.patch) > Choosing datanode retries times considering with block replica number > -- > > Key: HDFS-6524 > URL: https://issues.apache.org/jira/browse/HDFS-6524 > Project: Hadoop HDFS > Issue Type: Improvement > Components: hdfs-client >Affects Versions: 3.0.0-alpha1 >Reporter: Liang Xie >Assignee: Lisheng Sun >Priority: Minor > Labels: BB2015-05-TBR > Attachments: HDFS-6524.001.patch, HDFS-6524.002.patch, > HDFS-6524.003.patch, HDFS-6524.004.patch, HDFS-6524.005(2).patch, > HDFS-6524.005.patch, HDFS-6524.txt > > > Currently the chooseDataNode() does retry with the setting: > dfsClientConf.maxBlockAcquireFailures, which by default is 3 > (DFS_CLIENT_MAX_BLOCK_ACQUIRE_FAILURES_DEFAULT = 3), it would be better > having another option, block replication factor. One cluster with only two > block replica setting, or using Reed-solomon encoding solution with one > replica factor. It helps to reduce the long tail latency. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-6524) Choosing datanode retries times considering with block replica number
[ https://issues.apache.org/jira/browse/HDFS-6524?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Lisheng Sun updated HDFS-6524: -- Attachment: HDFS-6524.006.patch > Choosing datanode retries times considering with block replica number > -- > > Key: HDFS-6524 > URL: https://issues.apache.org/jira/browse/HDFS-6524 > Project: Hadoop HDFS > Issue Type: Improvement > Components: hdfs-client >Affects Versions: 3.0.0-alpha1 >Reporter: Liang Xie >Assignee: Lisheng Sun >Priority: Minor > Labels: BB2015-05-TBR > Attachments: HDFS-6524.001.patch, HDFS-6524.002.patch, > HDFS-6524.003.patch, HDFS-6524.004.patch, HDFS-6524.005(2).patch, > HDFS-6524.005.patch, HDFS-6524.006.patch, HDFS-6524.txt > > > Currently the chooseDataNode() does retry with the setting: > dfsClientConf.maxBlockAcquireFailures, which by default is 3 > (DFS_CLIENT_MAX_BLOCK_ACQUIRE_FAILURES_DEFAULT = 3), it would be better > having another option, block replication factor. One cluster with only two > block replica setting, or using Reed-solomon encoding solution with one > replica factor. It helps to reduce the long tail latency. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Comment Edited] (HDFS-14850) Optimize FileSystemAccessService#getFileSystemConfiguration
[ https://issues.apache.org/jira/browse/HDFS-14850?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16937864#comment-16937864 ] Lisheng Sun edited comment on HDFS-14850 at 9/25/19 3:54 PM: - [~elgoiri] If the new method is static, this static method does not get FileSystemAccessService# serviceHadoopConf. And the newest patch is [^HDFS-14850.005.patch] ^Thank you a lot.^ was (Author: leosun08): [~elgoiri] If the new method is static, this static method does not get FileSystemAccessService# serviceHadoopConf. And the newest patch is [^HDFS-14850.005.patch] > Optimize FileSystemAccessService#getFileSystemConfiguration > --- > > Key: HDFS-14850 > URL: https://issues.apache.org/jira/browse/HDFS-14850 > Project: Hadoop HDFS > Issue Type: Improvement > Components: httpfs, performance >Reporter: Lisheng Sun >Assignee: Lisheng Sun >Priority: Major > Attachments: HDFS-14850.001.patch, HDFS-14850.002.patch, > HDFS-14850.003.patch, HDFS-14850.004(2).patch, HDFS-14850.004.patch, > HDFS-14850.005.patch > > > {code:java} > @Override > public Configuration getFileSystemConfiguration() { > Configuration conf = new Configuration(true); > ConfigurationUtils.copy(serviceHadoopConf, conf); > conf.setBoolean(FILE_SYSTEM_SERVICE_CREATED, true); > // Force-clear server-side umask to make HttpFS match WebHDFS behavior > conf.set(FsPermission.UMASK_LABEL, "000"); > return conf; > } > {code} > As above code,when call > FileSystemAccessService#getFileSystemConfiguration,current code new > Configuration every time. > It is not necessary and affects performance. I think it only need to new > Configuration in FileSystemAccessService#init once and > FileSystemAccessService#getFileSystemConfiguration get it. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Comment Edited] (HDFS-14850) Optimize FileSystemAccessService#getFileSystemConfiguration
[ https://issues.apache.org/jira/browse/HDFS-14850?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16937864#comment-16937864 ] Lisheng Sun edited comment on HDFS-14850 at 9/25/19 3:53 PM: - [~elgoiri] If the new method is static, this static method does not get FileSystemAccessService# serviceHadoopConf. And the newest patch is [^HDFS-14850.005.patch] was (Author: leosun08): [~elgoiri] If the new method is static, this static method does not get FileSystemAccessService# serviceHadoopConf. And the newest patch is HDFS-14850.005.patch > Optimize FileSystemAccessService#getFileSystemConfiguration > --- > > Key: HDFS-14850 > URL: https://issues.apache.org/jira/browse/HDFS-14850 > Project: Hadoop HDFS > Issue Type: Improvement > Components: httpfs, performance >Reporter: Lisheng Sun >Assignee: Lisheng Sun >Priority: Major > Attachments: HDFS-14850.001.patch, HDFS-14850.002.patch, > HDFS-14850.003.patch, HDFS-14850.004(2).patch, HDFS-14850.004.patch, > HDFS-14850.005.patch > > > {code:java} > @Override > public Configuration getFileSystemConfiguration() { > Configuration conf = new Configuration(true); > ConfigurationUtils.copy(serviceHadoopConf, conf); > conf.setBoolean(FILE_SYSTEM_SERVICE_CREATED, true); > // Force-clear server-side umask to make HttpFS match WebHDFS behavior > conf.set(FsPermission.UMASK_LABEL, "000"); > return conf; > } > {code} > As above code,when call > FileSystemAccessService#getFileSystemConfiguration,current code new > Configuration every time. > It is not necessary and affects performance. I think it only need to new > Configuration in FileSystemAccessService#init once and > FileSystemAccessService#getFileSystemConfiguration get it. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-14850) Optimize FileSystemAccessService#getFileSystemConfiguration
[ https://issues.apache.org/jira/browse/HDFS-14850?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16937864#comment-16937864 ] Lisheng Sun commented on HDFS-14850: [~elgoiri] If the new method is static, this static method does not get FileSystemAccessService# serviceHadoopConf. And the newest patch is HDFS-14850.005.patch > Optimize FileSystemAccessService#getFileSystemConfiguration > --- > > Key: HDFS-14850 > URL: https://issues.apache.org/jira/browse/HDFS-14850 > Project: Hadoop HDFS > Issue Type: Improvement > Components: httpfs, performance >Reporter: Lisheng Sun >Assignee: Lisheng Sun >Priority: Major > Attachments: HDFS-14850.001.patch, HDFS-14850.002.patch, > HDFS-14850.003.patch, HDFS-14850.004(2).patch, HDFS-14850.004.patch, > HDFS-14850.005.patch > > > {code:java} > @Override > public Configuration getFileSystemConfiguration() { > Configuration conf = new Configuration(true); > ConfigurationUtils.copy(serviceHadoopConf, conf); > conf.setBoolean(FILE_SYSTEM_SERVICE_CREATED, true); > // Force-clear server-side umask to make HttpFS match WebHDFS behavior > conf.set(FsPermission.UMASK_LABEL, "000"); > return conf; > } > {code} > As above code,when call > FileSystemAccessService#getFileSystemConfiguration,current code new > Configuration every time. > It is not necessary and affects performance. I think it only need to new > Configuration in FileSystemAccessService#init once and > FileSystemAccessService#getFileSystemConfiguration get it. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Resolved] (HDFS-14871) StagingTestBase uses methods not available in Mockito 1.8.5 in branch-3.1
[ https://issues.apache.org/jira/browse/HDFS-14871?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Lisheng Sun resolved HDFS-14871. Resolution: Duplicate > StagingTestBase uses methods not available in Mockito 1.8.5 in branch-3.1 > - > > Key: HDFS-14871 > URL: https://issues.apache.org/jira/browse/HDFS-14871 > Project: Hadoop HDFS > Issue Type: Bug >Affects Versions: 3.1.0 >Reporter: Lisheng Sun >Priority: Major > > details see HADOOP-15398 > Problem: hadoop trunk compilation is failing > Root Cause: > compilation error is coming from > org.apache.hadoop.fs.s3a.commit.staging.StagingTestBase. Compilation error is > "The method getArgumentAt(int, Class) is undefined for the > type InvocationOnMock". > StagingTestBase is using getArgumentAt(int, Class) method > which is not available in mockito-all 1.8.5 version. getArgumentAt(int, > Class) method is available only from version 2.0.0-beta -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-14871) StagingTestBase uses methods not available in Mockito 1.8.5 in branch-3.1
[ https://issues.apache.org/jira/browse/HDFS-14871?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Lisheng Sun updated HDFS-14871: --- Summary: StagingTestBase uses methods not available in Mockito 1.8.5 in branch-3.1 (was: StagingTestBase uses methods not available in Mockito 1.8.5) > StagingTestBase uses methods not available in Mockito 1.8.5 in branch-3.1 > - > > Key: HDFS-14871 > URL: https://issues.apache.org/jira/browse/HDFS-14871 > Project: Hadoop HDFS > Issue Type: Bug >Affects Versions: 3.1.0 >Reporter: Lisheng Sun >Priority: Major > > details see HADOOP-15398 > Problem: hadoop trunk compilation is failing > Root Cause: > compilation error is coming from > org.apache.hadoop.fs.s3a.commit.staging.StagingTestBase. Compilation error is > "The method getArgumentAt(int, Class) is undefined for the > type InvocationOnMock". > StagingTestBase is using getArgumentAt(int, Class) method > which is not available in mockito-all 1.8.5 version. getArgumentAt(int, > Class) method is available only from version 2.0.0-beta -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Created] (HDFS-14871) StagingTestBase uses methods not available in Mockito 1.8.5
Lisheng Sun created HDFS-14871: -- Summary: StagingTestBase uses methods not available in Mockito 1.8.5 Key: HDFS-14871 URL: https://issues.apache.org/jira/browse/HDFS-14871 Project: Hadoop HDFS Issue Type: Bug Affects Versions: 3.1.0 Reporter: Lisheng Sun details see HADOOP-15398 Problem: hadoop trunk compilation is failing Root Cause: compilation error is coming from org.apache.hadoop.fs.s3a.commit.staging.StagingTestBase. Compilation error is "The method getArgumentAt(int, Class) is undefined for the type InvocationOnMock". StagingTestBase is using getArgumentAt(int, Class) method which is not available in mockito-all 1.8.5 version. getArgumentAt(int, Class) method is available only from version 2.0.0-beta -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Comment Edited] (HDFS-14850) Optimize FileSystemAccessService#getFileSystemConfiguration
[ https://issues.apache.org/jira/browse/HDFS-14850?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16936513#comment-16936513 ] Lisheng Sun edited comment on HDFS-14850 at 9/24/19 7:33 AM: - fixed the failures and uploaded the v005 patch. Could you have time to continue to review this patch? Thank you. [~elgoiri] was (Author: leosun08): fixed the failures and uploaded the v005 patch. Could you have time to continue to review this patch? Thank you. > Optimize FileSystemAccessService#getFileSystemConfiguration > --- > > Key: HDFS-14850 > URL: https://issues.apache.org/jira/browse/HDFS-14850 > Project: Hadoop HDFS > Issue Type: Improvement > Components: httpfs, performance >Reporter: Lisheng Sun >Assignee: Lisheng Sun >Priority: Major > Attachments: HDFS-14850.001.patch, HDFS-14850.002.patch, > HDFS-14850.003.patch, HDFS-14850.004(2).patch, HDFS-14850.004.patch, > HDFS-14850.005.patch > > > {code:java} > @Override > public Configuration getFileSystemConfiguration() { > Configuration conf = new Configuration(true); > ConfigurationUtils.copy(serviceHadoopConf, conf); > conf.setBoolean(FILE_SYSTEM_SERVICE_CREATED, true); > // Force-clear server-side umask to make HttpFS match WebHDFS behavior > conf.set(FsPermission.UMASK_LABEL, "000"); > return conf; > } > {code} > As above code,when call > FileSystemAccessService#getFileSystemConfiguration,current code new > Configuration every time. > It is not necessary and affects performance. I think it only need to new > Configuration in FileSystemAccessService#init once and > FileSystemAccessService#getFileSystemConfiguration get it. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-14850) Optimize FileSystemAccessService#getFileSystemConfiguration
[ https://issues.apache.org/jira/browse/HDFS-14850?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16936513#comment-16936513 ] Lisheng Sun commented on HDFS-14850: fixed the failures and uploaded the v005 patch. Could you have time to continue to review this patch? Thank you. > Optimize FileSystemAccessService#getFileSystemConfiguration > --- > > Key: HDFS-14850 > URL: https://issues.apache.org/jira/browse/HDFS-14850 > Project: Hadoop HDFS > Issue Type: Improvement > Components: httpfs, performance >Reporter: Lisheng Sun >Assignee: Lisheng Sun >Priority: Major > Attachments: HDFS-14850.001.patch, HDFS-14850.002.patch, > HDFS-14850.003.patch, HDFS-14850.004(2).patch, HDFS-14850.004.patch, > HDFS-14850.005.patch > > > {code:java} > @Override > public Configuration getFileSystemConfiguration() { > Configuration conf = new Configuration(true); > ConfigurationUtils.copy(serviceHadoopConf, conf); > conf.setBoolean(FILE_SYSTEM_SERVICE_CREATED, true); > // Force-clear server-side umask to make HttpFS match WebHDFS behavior > conf.set(FsPermission.UMASK_LABEL, "000"); > return conf; > } > {code} > As above code,when call > FileSystemAccessService#getFileSystemConfiguration,current code new > Configuration every time. > It is not necessary and affects performance. I think it only need to new > Configuration in FileSystemAccessService#init once and > FileSystemAccessService#getFileSystemConfiguration get it. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-14850) Optimize FileSystemAccessService#getFileSystemConfiguration
[ https://issues.apache.org/jira/browse/HDFS-14850?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Lisheng Sun updated HDFS-14850: --- Attachment: HDFS-14850.005.patch > Optimize FileSystemAccessService#getFileSystemConfiguration > --- > > Key: HDFS-14850 > URL: https://issues.apache.org/jira/browse/HDFS-14850 > Project: Hadoop HDFS > Issue Type: Improvement > Components: httpfs, performance >Reporter: Lisheng Sun >Assignee: Lisheng Sun >Priority: Major > Attachments: HDFS-14850.001.patch, HDFS-14850.002.patch, > HDFS-14850.003.patch, HDFS-14850.004(2).patch, HDFS-14850.004.patch, > HDFS-14850.005.patch > > > {code:java} > @Override > public Configuration getFileSystemConfiguration() { > Configuration conf = new Configuration(true); > ConfigurationUtils.copy(serviceHadoopConf, conf); > conf.setBoolean(FILE_SYSTEM_SERVICE_CREATED, true); > // Force-clear server-side umask to make HttpFS match WebHDFS behavior > conf.set(FsPermission.UMASK_LABEL, "000"); > return conf; > } > {code} > As above code,when call > FileSystemAccessService#getFileSystemConfiguration,current code new > Configuration every time. > It is not necessary and affects performance. I think it only need to new > Configuration in FileSystemAccessService#init once and > FileSystemAccessService#getFileSystemConfiguration get it. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-14850) Optimize FileSystemAccessService#getFileSystemConfiguration
[ https://issues.apache.org/jira/browse/HDFS-14850?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Lisheng Sun updated HDFS-14850: --- Attachment: HDFS-14850.004(2).patch > Optimize FileSystemAccessService#getFileSystemConfiguration > --- > > Key: HDFS-14850 > URL: https://issues.apache.org/jira/browse/HDFS-14850 > Project: Hadoop HDFS > Issue Type: Improvement > Components: httpfs, performance >Reporter: Lisheng Sun >Assignee: Lisheng Sun >Priority: Major > Attachments: HDFS-14850.001.patch, HDFS-14850.002.patch, > HDFS-14850.003.patch, HDFS-14850.004(2).patch, HDFS-14850.004.patch > > > {code:java} > @Override > public Configuration getFileSystemConfiguration() { > Configuration conf = new Configuration(true); > ConfigurationUtils.copy(serviceHadoopConf, conf); > conf.setBoolean(FILE_SYSTEM_SERVICE_CREATED, true); > // Force-clear server-side umask to make HttpFS match WebHDFS behavior > conf.set(FsPermission.UMASK_LABEL, "000"); > return conf; > } > {code} > As above code,when call > FileSystemAccessService#getFileSystemConfiguration,current code new > Configuration every time. > It is not necessary and affects performance. I think it only need to new > Configuration in FileSystemAccessService#init once and > FileSystemAccessService#getFileSystemConfiguration get it. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-14283) DFSInputStream to prefer cached replica
[ https://issues.apache.org/jira/browse/HDFS-14283?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16935758#comment-16935758 ] Lisheng Sun commented on HDFS-14283: hi [~smeng] Could you have time to take a review for this patch? Thank you. > DFSInputStream to prefer cached replica > --- > > Key: HDFS-14283 > URL: https://issues.apache.org/jira/browse/HDFS-14283 > Project: Hadoop HDFS > Issue Type: Improvement >Affects Versions: 2.6.0 > Environment: HDFS Caching >Reporter: Wei-Chiu Chuang >Assignee: Lisheng Sun >Priority: Major > Attachments: HDFS-14283.001.patch > > > HDFS Caching offers performance benefits. However, currently NameNode does > not treat cached replica with higher priority, so HDFS caching is only useful > when cache replication = 3, that is to say, all replicas are cached in > memory, so that a client doesn't randomly pick an uncached replica. > HDFS-6846 proposed to let NameNode give higher priority to cached replica. > Changing a logic in NameNode is always tricky so that didn't get much > traction. Here I propose a different approach: let client (DFSInputStream) > prefer cached replica. > A {{LocatedBlock}} object already contains cached replica location so a > client has the needed information. I think we can change > {{DFSInputStream#getBestNodeDNAddrPair()}} for this purpose. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-6524) Choosing datanode retries times considering with block replica number
[ https://issues.apache.org/jira/browse/HDFS-6524?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16935755#comment-16935755 ] Lisheng Sun commented on HDFS-6524: --- hi [~elgoiri] {quote} Can we have some more coverage for this? We should cover the old behavior and the new one {quote} Updated unit tests for this patch. do you mean that i add comments for this patch? Thank you. > Choosing datanode retries times considering with block replica number > -- > > Key: HDFS-6524 > URL: https://issues.apache.org/jira/browse/HDFS-6524 > Project: Hadoop HDFS > Issue Type: Improvement > Components: hdfs-client >Affects Versions: 3.0.0-alpha1 >Reporter: Liang Xie >Assignee: Lisheng Sun >Priority: Minor > Labels: BB2015-05-TBR > Attachments: HDFS-6524.001.patch, HDFS-6524.002.patch, > HDFS-6524.003.patch, HDFS-6524.004.patch, HDFS-6524.005(2).patch, > HDFS-6524.005.patch, HDFS-6524.txt > > > Currently the chooseDataNode() does retry with the setting: > dfsClientConf.maxBlockAcquireFailures, which by default is 3 > (DFS_CLIENT_MAX_BLOCK_ACQUIRE_FAILURES_DEFAULT = 3), it would be better > having another option, block replication factor. One cluster with only two > block replica setting, or using Reed-solomon encoding solution with one > replica factor. It helps to reduce the long tail latency. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-14850) Optimize FileSystemAccessService#getFileSystemConfiguration
[ https://issues.apache.org/jira/browse/HDFS-14850?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Lisheng Sun updated HDFS-14850: --- Attachment: HDFS-14850.004.patch > Optimize FileSystemAccessService#getFileSystemConfiguration > --- > > Key: HDFS-14850 > URL: https://issues.apache.org/jira/browse/HDFS-14850 > Project: Hadoop HDFS > Issue Type: Improvement > Components: httpfs, performance >Reporter: Lisheng Sun >Assignee: Lisheng Sun >Priority: Major > Attachments: HDFS-14850.001.patch, HDFS-14850.002.patch, > HDFS-14850.003.patch, HDFS-14850.004.patch > > > {code:java} > @Override > public Configuration getFileSystemConfiguration() { > Configuration conf = new Configuration(true); > ConfigurationUtils.copy(serviceHadoopConf, conf); > conf.setBoolean(FILE_SYSTEM_SERVICE_CREATED, true); > // Force-clear server-side umask to make HttpFS match WebHDFS behavior > conf.set(FsPermission.UMASK_LABEL, "000"); > return conf; > } > {code} > As above code,when call > FileSystemAccessService#getFileSystemConfiguration,current code new > Configuration every time. > It is not necessary and affects performance. I think it only need to new > Configuration in FileSystemAccessService#init once and > FileSystemAccessService#getFileSystemConfiguration get it. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-14844) Make buffer of BlockReaderRemote#newBlockReader#BufferedOutputStream configurable
[ https://issues.apache.org/jira/browse/HDFS-14844?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16932987#comment-16932987 ] Lisheng Sun commented on HDFS-14844: all failed UTs are ok in local. So they are unrelated to this patch. [~elgoiri] Can we commit this patch to trunk? Thank you a lot. > Make buffer of BlockReaderRemote#newBlockReader#BufferedOutputStream > configurable > -- > > Key: HDFS-14844 > URL: https://issues.apache.org/jira/browse/HDFS-14844 > Project: Hadoop HDFS > Issue Type: Improvement >Reporter: Lisheng Sun >Assignee: Lisheng Sun >Priority: Minor > Attachments: HDFS-14844.001.patch, HDFS-14844.002.patch, > HDFS-14844.003.patch, HDFS-14844.004.patch, HDFS-14844.005.patch > > > details for HDFS-14820 -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-14850) Optimize FileSystemAccessService#getFileSystemConfiguration
[ https://issues.apache.org/jira/browse/HDFS-14850?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16932986#comment-16932986 ] Lisheng Sun commented on HDFS-14850: hi [~jojochuang] [~elgoiri] Could you help take a review for this patch? Thank you. > Optimize FileSystemAccessService#getFileSystemConfiguration > --- > > Key: HDFS-14850 > URL: https://issues.apache.org/jira/browse/HDFS-14850 > Project: Hadoop HDFS > Issue Type: Improvement > Components: httpfs, performance >Reporter: Lisheng Sun >Assignee: Lisheng Sun >Priority: Major > Attachments: HDFS-14850.001.patch, HDFS-14850.002.patch, > HDFS-14850.003.patch > > > {code:java} > @Override > public Configuration getFileSystemConfiguration() { > Configuration conf = new Configuration(true); > ConfigurationUtils.copy(serviceHadoopConf, conf); > conf.setBoolean(FILE_SYSTEM_SERVICE_CREATED, true); > // Force-clear server-side umask to make HttpFS match WebHDFS behavior > conf.set(FsPermission.UMASK_LABEL, "000"); > return conf; > } > {code} > As above code,when call > FileSystemAccessService#getFileSystemConfiguration,current code new > Configuration every time. > It is not necessary and affects performance. I think it only need to new > Configuration in FileSystemAccessService#init once and > FileSystemAccessService#getFileSystemConfiguration get it. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-14850) Optimize FileSystemAccessService#getFileSystemConfiguration
[ https://issues.apache.org/jira/browse/HDFS-14850?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Lisheng Sun updated HDFS-14850: --- Attachment: HDFS-14850.003.patch > Optimize FileSystemAccessService#getFileSystemConfiguration > --- > > Key: HDFS-14850 > URL: https://issues.apache.org/jira/browse/HDFS-14850 > Project: Hadoop HDFS > Issue Type: Improvement > Components: httpfs, performance >Reporter: Lisheng Sun >Assignee: Lisheng Sun >Priority: Major > Attachments: HDFS-14850.001.patch, HDFS-14850.002.patch, > HDFS-14850.003.patch > > > {code:java} > @Override > public Configuration getFileSystemConfiguration() { > Configuration conf = new Configuration(true); > ConfigurationUtils.copy(serviceHadoopConf, conf); > conf.setBoolean(FILE_SYSTEM_SERVICE_CREATED, true); > // Force-clear server-side umask to make HttpFS match WebHDFS behavior > conf.set(FsPermission.UMASK_LABEL, "000"); > return conf; > } > {code} > As above code,when call > FileSystemAccessService#getFileSystemConfiguration,current code new > Configuration every time. > It is not necessary and affects performance. I think it only need to new > Configuration in FileSystemAccessService#init once and > FileSystemAccessService#getFileSystemConfiguration get it. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-14844) Make buffer of BlockReaderRemote#newBlockReader#BufferedOutputStream configurable
[ https://issues.apache.org/jira/browse/HDFS-14844?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Lisheng Sun updated HDFS-14844: --- Attachment: HDFS-14844.005.patch > Make buffer of BlockReaderRemote#newBlockReader#BufferedOutputStream > configurable > -- > > Key: HDFS-14844 > URL: https://issues.apache.org/jira/browse/HDFS-14844 > Project: Hadoop HDFS > Issue Type: Improvement >Reporter: Lisheng Sun >Assignee: Lisheng Sun >Priority: Minor > Attachments: HDFS-14844.001.patch, HDFS-14844.002.patch, > HDFS-14844.003.patch, HDFS-14844.004.patch, HDFS-14844.005.patch > > > details for HDFS-14820 -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-14844) Make buffer of BlockReaderRemote#newBlockReader#BufferedOutputStream configurable
[ https://issues.apache.org/jira/browse/HDFS-14844?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16932013#comment-16932013 ] Lisheng Sun commented on HDFS-14844: [~elgoiri] I agree your option. And I confirmed all failed UTs are ok in local. So they are unrelated to this patch. Fixed the the indentation (too many spaces) in BlockReaderRemote#401-402. Uploaded the v005 patch. Thank you a lot. [~elgoiri] > Make buffer of BlockReaderRemote#newBlockReader#BufferedOutputStream > configurable > -- > > Key: HDFS-14844 > URL: https://issues.apache.org/jira/browse/HDFS-14844 > Project: Hadoop HDFS > Issue Type: Improvement >Reporter: Lisheng Sun >Assignee: Lisheng Sun >Priority: Minor > Attachments: HDFS-14844.001.patch, HDFS-14844.002.patch, > HDFS-14844.003.patch, HDFS-14844.004.patch, HDFS-14844.005.patch > > > details for HDFS-14820 -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-14850) Optimize FileSystemAccessService#getFileSystemConfiguration
[ https://issues.apache.org/jira/browse/HDFS-14850?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Lisheng Sun updated HDFS-14850: --- Component/s: performance > Optimize FileSystemAccessService#getFileSystemConfiguration > --- > > Key: HDFS-14850 > URL: https://issues.apache.org/jira/browse/HDFS-14850 > Project: Hadoop HDFS > Issue Type: Improvement > Components: httpfs, performance >Reporter: Lisheng Sun >Assignee: Lisheng Sun >Priority: Major > Attachments: HDFS-14850.001.patch, HDFS-14850.002.patch > > > {code:java} > @Override > public Configuration getFileSystemConfiguration() { > Configuration conf = new Configuration(true); > ConfigurationUtils.copy(serviceHadoopConf, conf); > conf.setBoolean(FILE_SYSTEM_SERVICE_CREATED, true); > // Force-clear server-side umask to make HttpFS match WebHDFS behavior > conf.set(FsPermission.UMASK_LABEL, "000"); > return conf; > } > {code} > As above code,when call > FileSystemAccessService#getFileSystemConfiguration,current code new > Configuration every time. > It is not necessary and affects performance. I think it only need to new > Configuration in FileSystemAccessService#init once and > FileSystemAccessService#getFileSystemConfiguration get it. -- This message was sent by Atlassian Jira (v8.3.2#803003) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-14850) Optimize FileSystemAccessService#getFileSystemConfiguration
[ https://issues.apache.org/jira/browse/HDFS-14850?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Lisheng Sun updated HDFS-14850: --- Component/s: httpfs > Optimize FileSystemAccessService#getFileSystemConfiguration > --- > > Key: HDFS-14850 > URL: https://issues.apache.org/jira/browse/HDFS-14850 > Project: Hadoop HDFS > Issue Type: Improvement > Components: httpfs >Reporter: Lisheng Sun >Assignee: Lisheng Sun >Priority: Major > Attachments: HDFS-14850.001.patch, HDFS-14850.002.patch > > > {code:java} > @Override > public Configuration getFileSystemConfiguration() { > Configuration conf = new Configuration(true); > ConfigurationUtils.copy(serviceHadoopConf, conf); > conf.setBoolean(FILE_SYSTEM_SERVICE_CREATED, true); > // Force-clear server-side umask to make HttpFS match WebHDFS behavior > conf.set(FsPermission.UMASK_LABEL, "000"); > return conf; > } > {code} > As above code,when call > FileSystemAccessService#getFileSystemConfiguration,current code new > Configuration every time. > It is not necessary and affects performance. I think it only need to new > Configuration in FileSystemAccessService#init once and > FileSystemAccessService#getFileSystemConfiguration get it. -- This message was sent by Atlassian Jira (v8.3.2#803003) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-14850) Optimize FileSystemAccessService#getFileSystemConfiguration
[ https://issues.apache.org/jira/browse/HDFS-14850?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Lisheng Sun updated HDFS-14850: --- Attachment: HDFS-14850.002.patch > Optimize FileSystemAccessService#getFileSystemConfiguration > --- > > Key: HDFS-14850 > URL: https://issues.apache.org/jira/browse/HDFS-14850 > Project: Hadoop HDFS > Issue Type: Improvement >Reporter: Lisheng Sun >Assignee: Lisheng Sun >Priority: Major > Attachments: HDFS-14850.001.patch, HDFS-14850.002.patch > > > {code:java} > @Override > public Configuration getFileSystemConfiguration() { > Configuration conf = new Configuration(true); > ConfigurationUtils.copy(serviceHadoopConf, conf); > conf.setBoolean(FILE_SYSTEM_SERVICE_CREATED, true); > // Force-clear server-side umask to make HttpFS match WebHDFS behavior > conf.set(FsPermission.UMASK_LABEL, "000"); > return conf; > } > {code} > As above code,when call > FileSystemAccessService#getFileSystemConfiguration,current code new > Configuration every time. > It is not necessary and affects performance. I think it only need to new > Configuration in FileSystemAccessService#init once and > FileSystemAccessService#getFileSystemConfiguration get it. -- This message was sent by Atlassian Jira (v8.3.2#803003) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-6524) Choosing datanode retries times considering with block replica number
[ https://issues.apache.org/jira/browse/HDFS-6524?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Lisheng Sun updated HDFS-6524: -- Attachment: HDFS-6524.005(2).patch > Choosing datanode retries times considering with block replica number > -- > > Key: HDFS-6524 > URL: https://issues.apache.org/jira/browse/HDFS-6524 > Project: Hadoop HDFS > Issue Type: Improvement > Components: hdfs-client >Affects Versions: 3.0.0-alpha1 >Reporter: Liang Xie >Assignee: Lisheng Sun >Priority: Minor > Labels: BB2015-05-TBR > Attachments: HDFS-6524.001.patch, HDFS-6524.002.patch, > HDFS-6524.003.patch, HDFS-6524.004.patch, HDFS-6524.005(2).patch, > HDFS-6524.005.patch, HDFS-6524.txt > > > Currently the chooseDataNode() does retry with the setting: > dfsClientConf.maxBlockAcquireFailures, which by default is 3 > (DFS_CLIENT_MAX_BLOCK_ACQUIRE_FAILURES_DEFAULT = 3), it would be better > having another option, block replication factor. One cluster with only two > block replica setting, or using Reed-solomon encoding solution with one > replica factor. It helps to reduce the long tail latency. -- This message was sent by Atlassian Jira (v8.3.2#803003) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-6524) Choosing datanode retries times considering with block replica number
[ https://issues.apache.org/jira/browse/HDFS-6524?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Lisheng Sun updated HDFS-6524: -- Attachment: (was: HDFS-6524.005.patch) > Choosing datanode retries times considering with block replica number > -- > > Key: HDFS-6524 > URL: https://issues.apache.org/jira/browse/HDFS-6524 > Project: Hadoop HDFS > Issue Type: Improvement > Components: hdfs-client >Affects Versions: 3.0.0-alpha1 >Reporter: Liang Xie >Assignee: Lisheng Sun >Priority: Minor > Labels: BB2015-05-TBR > Attachments: HDFS-6524.001.patch, HDFS-6524.002.patch, > HDFS-6524.003.patch, HDFS-6524.004.patch, HDFS-6524.005.patch, HDFS-6524.txt > > > Currently the chooseDataNode() does retry with the setting: > dfsClientConf.maxBlockAcquireFailures, which by default is 3 > (DFS_CLIENT_MAX_BLOCK_ACQUIRE_FAILURES_DEFAULT = 3), it would be better > having another option, block replication factor. One cluster with only two > block replica setting, or using Reed-solomon encoding solution with one > replica factor. It helps to reduce the long tail latency. -- This message was sent by Atlassian Jira (v8.3.2#803003) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org