[jira] [Commented] (HADOOP-17374) AliyunOSS: support ListObjectsV2
[ https://issues.apache.org/jira/browse/HADOOP-17374?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17438502#comment-17438502 ] wujinhu commented on HADOOP-17374: -- Thanks [~cheersyang], please merge this change to branch-3.1, branch-3.2, branch-3.3, too. > AliyunOSS: support ListObjectsV2 > > > Key: HADOOP-17374 > URL: https://issues.apache.org/jira/browse/HADOOP-17374 > Project: Hadoop Common > Issue Type: Sub-task > Components: fs/oss >Affects Versions: 2.9.2, 3.0.3, 3.3.0, 3.2.1, 3.1.4, 2.10.1 >Reporter: wujinhu >Assignee: wujinhu >Priority: Major > Labels: pull-request-available > Time Spent: 10m > Remaining Estimate: 0h > > OSS supports > ListObjectsV2([https://help.aliyun.com/document_detail/187544.html?spm=a2c4g.11186623.6.1589.e0623d9fE1b64S)] > to optimize versioning bucket list. We should support this feature in > AliyunOSS module. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Updated] (HADOOP-17374) AliyunOSS: support ListObjectsV2
[ https://issues.apache.org/jira/browse/HADOOP-17374?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] wujinhu updated HADOOP-17374: - Attachment: (was: HADOOP-17374.001.patch) > AliyunOSS: support ListObjectsV2 > > > Key: HADOOP-17374 > URL: https://issues.apache.org/jira/browse/HADOOP-17374 > Project: Hadoop Common > Issue Type: Sub-task > Components: fs/oss >Affects Versions: 2.9.2, 3.0.3, 3.3.0, 3.2.1, 3.1.4, 2.10.1 >Reporter: wujinhu >Assignee: wujinhu >Priority: Major > Labels: pull-request-available > Time Spent: 10m > Remaining Estimate: 0h > > OSS supports > ListObjectsV2([https://help.aliyun.com/document_detail/187544.html?spm=a2c4g.11186623.6.1589.e0623d9fE1b64S)] > to optimize versioning bucket list. We should support this feature in > AliyunOSS module. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[GitHub] [hadoop] hadoop-yetus commented on pull request #3613: HDFS-16296. RouterRpcFairnessPolicyController add rejected permits for each nameservice
hadoop-yetus commented on pull request #3613: URL: https://github.com/apache/hadoop/pull/3613#issuecomment-960483188 :broken_heart: **-1 overall** | Vote | Subsystem | Runtime | Logfile | Comment | |::|--:|:|::|:---:| | +0 :ok: | reexec | 1m 15s | | Docker mode activated. | _ Prechecks _ | | +1 :green_heart: | dupname | 0m 0s | | No case conflicting files found. | | +0 :ok: | codespell | 0m 0s | | codespell was not available. | | +1 :green_heart: | @author | 0m 0s | | The patch does not contain any @author tags. | | +1 :green_heart: | test4tests | 0m 0s | | The patch appears to include 2 new or modified test files. | _ trunk Compile Tests _ | | +1 :green_heart: | mvninstall | 33m 0s | | trunk passed | | +1 :green_heart: | compile | 0m 47s | | trunk passed with JDK Ubuntu-11.0.11+9-Ubuntu-0ubuntu2.20.04 | | +1 :green_heart: | compile | 0m 41s | | trunk passed with JDK Private Build-1.8.0_292-8u292-b10-0ubuntu1~20.04-b10 | | +1 :green_heart: | checkstyle | 0m 28s | | trunk passed | | +1 :green_heart: | mvnsite | 0m 43s | | trunk passed | | +1 :green_heart: | javadoc | 0m 45s | | trunk passed with JDK Ubuntu-11.0.11+9-Ubuntu-0ubuntu2.20.04 | | +1 :green_heart: | javadoc | 0m 52s | | trunk passed with JDK Private Build-1.8.0_292-8u292-b10-0ubuntu1~20.04-b10 | | +1 :green_heart: | spotbugs | 1m 20s | | trunk passed | | +1 :green_heart: | shadedclient | 20m 21s | | branch has no errors when building and testing our client artifacts. | _ Patch Compile Tests _ | | +1 :green_heart: | mvninstall | 0m 33s | | the patch passed | | +1 :green_heart: | compile | 0m 33s | | the patch passed with JDK Ubuntu-11.0.11+9-Ubuntu-0ubuntu2.20.04 | | +1 :green_heart: | javac | 0m 33s | | the patch passed | | +1 :green_heart: | compile | 0m 31s | | the patch passed with JDK Private Build-1.8.0_292-8u292-b10-0ubuntu1~20.04-b10 | | +1 :green_heart: | javac | 0m 31s | | the patch passed | | +1 :green_heart: | blanks | 0m 0s | | The patch has no blanks issues. | | +1 :green_heart: | checkstyle | 0m 17s | | the patch passed | | +1 :green_heart: | mvnsite | 0m 33s | | the patch passed | | +1 :green_heart: | javadoc | 0m 33s | | the patch passed with JDK Ubuntu-11.0.11+9-Ubuntu-0ubuntu2.20.04 | | +1 :green_heart: | javadoc | 0m 49s | | the patch passed with JDK Private Build-1.8.0_292-8u292-b10-0ubuntu1~20.04-b10 | | +1 :green_heart: | spotbugs | 1m 19s | | the patch passed | | +1 :green_heart: | shadedclient | 20m 5s | | patch has no errors when building and testing our client artifacts. | _ Other Tests _ | | -1 :x: | unit | 34m 16s | [/patch-unit-hadoop-hdfs-project_hadoop-hdfs-rbf.txt](https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-3613/4/artifact/out/patch-unit-hadoop-hdfs-project_hadoop-hdfs-rbf.txt) | hadoop-hdfs-rbf in the patch passed. | | +1 :green_heart: | asflicense | 0m 38s | | The patch does not generate ASF License warnings. | | | | 121m 55s | | | | Reason | Tests | |---:|:--| | Failed junit tests | hadoop.hdfs.rbfbalance.TestRouterDistCpProcedure | | Subsystem | Report/Notes | |--:|:-| | Docker | ClientAPI=1.41 ServerAPI=1.41 base: https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-3613/4/artifact/out/Dockerfile | | GITHUB PR | https://github.com/apache/hadoop/pull/3613 | | Optional Tests | dupname asflicense compile javac javadoc mvninstall mvnsite unit shadedclient spotbugs checkstyle codespell | | uname | Linux 6b97e564ae90 4.15.0-112-generic #113-Ubuntu SMP Thu Jul 9 23:41:39 UTC 2020 x86_64 x86_64 x86_64 GNU/Linux | | Build tool | maven | | Personality | dev-support/bin/hadoop.sh | | git revision | trunk / 9df5178b099d7039f1848b9e526e6c66b6c2a8bd | | Default Java | Private Build-1.8.0_292-8u292-b10-0ubuntu1~20.04-b10 | | Multi-JDK versions | /usr/lib/jvm/java-11-openjdk-amd64:Ubuntu-11.0.11+9-Ubuntu-0ubuntu2.20.04 /usr/lib/jvm/java-8-openjdk-amd64:Private Build-1.8.0_292-8u292-b10-0ubuntu1~20.04-b10 | | Test Results | https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-3613/4/testReport/ | | Max. process+thread count | 2706 (vs. ulimit of 5500) | | modules | C: hadoop-hdfs-project/hadoop-hdfs-rbf U: hadoop-hdfs-project/hadoop-hdfs-rbf | | Console output | https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-3613/4/console | | versions | git=2.25.1 maven=3.6.3 spotbugs=4.2.2 | | Powered by | Apache Yetus 0.14.0-SNAPSHOT https://yetus.apache.org | This message was automatically generated. -- This is an automated message from the Apache Git Service. To r
[jira] [Commented] (HADOOP-17374) AliyunOSS: support ListObjectsV2
[ https://issues.apache.org/jira/browse/HADOOP-17374?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17438479#comment-17438479 ] Weiwei Yang commented on HADOOP-17374: -- Looks good, I have merged the PR to the trunk. What else branches needed for this change? > AliyunOSS: support ListObjectsV2 > > > Key: HADOOP-17374 > URL: https://issues.apache.org/jira/browse/HADOOP-17374 > Project: Hadoop Common > Issue Type: Sub-task > Components: fs/oss >Affects Versions: 2.9.2, 3.0.3, 3.3.0, 3.2.1, 3.1.4, 2.10.1 >Reporter: wujinhu >Assignee: wujinhu >Priority: Major > Labels: pull-request-available > Attachments: HADOOP-17374.001.patch > > Time Spent: 10m > Remaining Estimate: 0h > > OSS supports > ListObjectsV2([https://help.aliyun.com/document_detail/187544.html?spm=a2c4g.11186623.6.1589.e0623d9fE1b64S)] > to optimize versioning bucket list. We should support this feature in > AliyunOSS module. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Work logged] (HADOOP-17374) AliyunOSS: support ListObjectsV2
[ https://issues.apache.org/jira/browse/HADOOP-17374?focusedWorklogId=676204&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-676204 ] ASF GitHub Bot logged work on HADOOP-17374: --- Author: ASF GitHub Bot Created on: 04/Nov/21 04:48 Start Date: 04/Nov/21 04:48 Worklog Time Spent: 10m Work Description: yangwwei merged pull request #3587: URL: https://github.com/apache/hadoop/pull/3587 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org Issue Time Tracking --- Worklog Id: (was: 676204) Remaining Estimate: 0h Time Spent: 10m > AliyunOSS: support ListObjectsV2 > > > Key: HADOOP-17374 > URL: https://issues.apache.org/jira/browse/HADOOP-17374 > Project: Hadoop Common > Issue Type: Sub-task > Components: fs/oss >Affects Versions: 2.9.2, 3.0.3, 3.3.0, 3.2.1, 3.1.4, 2.10.1 >Reporter: wujinhu >Assignee: wujinhu >Priority: Major > Labels: pull-request-available > Attachments: HADOOP-17374.001.patch > > Time Spent: 10m > Remaining Estimate: 0h > > OSS supports > ListObjectsV2([https://help.aliyun.com/document_detail/187544.html?spm=a2c4g.11186623.6.1589.e0623d9fE1b64S)] > to optimize versioning bucket list. We should support this feature in > AliyunOSS module. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[GitHub] [hadoop] yangwwei merged pull request #3587: HADOOP-17374. support listObjectV2
yangwwei merged pull request #3587: URL: https://github.com/apache/hadoop/pull/3587 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[GitHub] [hadoop] ferhui commented on pull request #3613: HDFS-16296. RouterRpcFairnessPolicyController add rejected permits for each nameservice
ferhui commented on pull request #3613: URL: https://github.com/apache/hadoop/pull/3613#issuecomment-960449761 @symious Thanks for contribution. @goiri Thanks for review. Will merge tomorrow if no other comments. -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[GitHub] [hadoop] ferhui commented on pull request #3602: HDFS-16291.Make the comment of INode#ReclaimContext more standardized.
ferhui commented on pull request #3602: URL: https://github.com/apache/hadoop/pull/3602#issuecomment-960369289 @jianghuazhu Thanks for contribution. @virajjasani @tomscut Thanks for review! Merged -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[GitHub] [hadoop] tomscut commented on pull request #3538: HDFS-16266. Add remote port information to HDFS audit log
tomscut commented on pull request #3538: URL: https://github.com/apache/hadoop/pull/3538#issuecomment-960359159 Thanks @tasanuma for the merge. Thanks all for your reviews and suggestions. -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[GitHub] [hadoop] ferhui merged pull request #3602: HDFS-16291.Make the comment of INode#ReclaimContext more standardized.
ferhui merged pull request #3602: URL: https://github.com/apache/hadoop/pull/3602 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Work logged] (HADOOP-17981) Support etag-assisted renames in FileOutputCommitter
[ https://issues.apache.org/jira/browse/HADOOP-17981?focusedWorklogId=676132&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-676132 ] ASF GitHub Bot logged work on HADOOP-17981: --- Author: ASF GitHub Bot Created on: 04/Nov/21 02:01 Start Date: 04/Nov/21 02:01 Worklog Time Spent: 10m Work Description: hadoop-yetus removed a comment on pull request #3597: URL: https://github.com/apache/hadoop/pull/3597#issuecomment-956975430 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org Issue Time Tracking --- Worklog Id: (was: 676132) Time Spent: 8h 20m (was: 8h 10m) > Support etag-assisted renames in FileOutputCommitter > > > Key: HADOOP-17981 > URL: https://issues.apache.org/jira/browse/HADOOP-17981 > Project: Hadoop Common > Issue Type: New Feature > Components: fs, fs/azure >Affects Versions: 3.4.0 >Reporter: Steve Loughran >Assignee: Steve Loughran >Priority: Major > Labels: pull-request-available > Time Spent: 8h 20m > Remaining Estimate: 0h > > To deal with some throttling/retry issues in object stores, > pass the FileStatus entries retrieved during listing > into a private interface ResilientCommitByRename which filesystems > may implement to use extra attributes in the listing (etag, version) > to constrain and validate the operation. > Although targeting azure, GCS and others could use. no point in S3A as they > shouldn't use this committer. > # And we are not going to do any changes to FileSystem as there are explicit > guarantees of public use and stability. > I am not going to make that change as the hive thing that will suddenly start > expecting it to work forever. > # I'm not planning to merge this in, as the manifest committer is going to > include this and more (MAPREDUCE-7341) > However, I do need to get this in on a branch, so am doing this work on trunk > for dev & test and for others to review -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[GitHub] [hadoop] hadoop-yetus removed a comment on pull request #3597: HADOOP-17981 Support etag-assisted renames in FileOutputCommitter
hadoop-yetus removed a comment on pull request #3597: URL: https://github.com/apache/hadoop/pull/3597#issuecomment-956975430 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[GitHub] [hadoop] sodonnel merged pull request #3593: HDFS-16286. Add a debug tool to verify the correctness of erasure coding on file
sodonnel merged pull request #3593: URL: https://github.com/apache/hadoop/pull/3593 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[GitHub] [hadoop] tasanuma commented on pull request #3538: HDFS-16266. Add remote port information to HDFS audit log
tasanuma commented on pull request #3538: URL: https://github.com/apache/hadoop/pull/3538#issuecomment-960354847 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[GitHub] [hadoop] ferhui commented on pull request #3613: HDFS-16296. RouterRpcFairnessPolicyController add rejected permits for each nameservice
ferhui commented on pull request #3613: URL: https://github.com/apache/hadoop/pull/3613#issuecomment-958680152 @symious Thanks for contribution, it looks good. Let's wait for the CI reports. -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Work logged] (HADOOP-17981) Support etag-assisted renames in FileOutputCommitter
[ https://issues.apache.org/jira/browse/HADOOP-17981?focusedWorklogId=676063&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-676063 ] ASF GitHub Bot logged work on HADOOP-17981: --- Author: ASF GitHub Bot Created on: 04/Nov/21 01:55 Start Date: 04/Nov/21 01:55 Worklog Time Spent: 10m Work Description: sidseth commented on pull request #3597: URL: https://github.com/apache/hadoop/pull/3597#issuecomment-958658416 > > This mechanism becomes very FileSystem specific. Implemented by Azure right now. > > I agree, which is why the API is restricted for its uses to mr-client-core only. as abfs is the only one which needs it for correctness under load, And I'm not worried about that specifity. Can I point to how much of the hadoop fs api are hdfs-only -and they are public. > > > Other users of rename will not see the benefits without changing interfaces, which in turn requires shimming etc. > > Please don't try and use this particular interface in Hive. > Was referring to any potential usage - including Hive. > > Would it be better for AzureFileSystem rename itself to add a config parameter which can lookup the src etag (at the cost of a performance hit for consistency), so that downstream components / any users of the rename operation can benefit from this change without having to change interfaces. > > We are going straight from a listing (1 request/500 entries) to that rename. doing a HEAD first cuts the throughtput in half. so no. > In the scenario where this is encountered. Would not be the default behaviour, and limits the change to Abfs. Could also have the less consistent version which is not etag based, and responds only on failures. Again - limited to Abfs. > > Also, if the performance penalty is a big problem - Abfs could create very short-lived caches for FileStatus objects, and handle errors on discrepancies with the cached copy. > > Possible but convoluted. > Agree. Quite convoluted. Tossing in potential options - to avoid a new public API. > > Essentially - don't force usage of the new interface to get the benefits. > > I understand the interests of the hive team, but this fix is not the place to do a better API. > > Briefly cacheing the source FS entries is something to consider though. Not this week. > > What I could do with is some help getting #2735 in, then we can start on a public rename() builder API which will take a file status, as openFile does. > This particular change would be FSImpl agnostic, and potentially remove the need for the new interface here? > > Side note: The fs.getStatus within ResilientCommitByRenameHelper for FileSystems where this new functionality is not supported will lead to a performance penalty for the other FileSystems (performing a getFileStatus on src). > > There is an option to say "i know it is not there"; this skips the check. the committer passes this option down because it issues a delete call first. > EOD - this ends up being a new API (almost on the FileSystem), which is used by the committer first; then someone discovers it and decides to make use of it. > FWIW the manifest committer will make that pre-rename commit optional, saving that IO request. I am curious as to how well that will work I went executed on well formed tables. -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org Issue Time Tracking --- Worklog Id: (was: 676063) Time Spent: 8h 10m (was: 8h) > Support etag-assisted renames in FileOutputCommitter > > > Key: HADOOP-17981 > URL: https://issues.apache.org/jira/browse/HADOOP-17981 > Project: Hadoop Common > Issue Type: New Feature > Components: fs, fs/azure >Affects Versions: 3.4.0 >Reporter: Steve Loughran >Assignee: Steve Loughran >Priority: Major > Labels: pull-request-available > Time Spent: 8h 10m > Remaining Estimate: 0h > > To deal with some throttling/retry issues in object stores, > pass the FileStatus entries retrieved during listing > into a private interface ResilientCommitByRename which filesystems > may implement to use extra attributes in the listing (etag, version) > to constrain and validate the operation. > Although targeting azure, GCS and others could use. no point in S3A as they > shouldn't use this committer. > # And we are not going to do any changes to FileSystem as
[jira] [Work logged] (HADOOP-17981) Support etag-assisted renames in FileOutputCommitter
[ https://issues.apache.org/jira/browse/HADOOP-17981?focusedWorklogId=676062&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-676062 ] ASF GitHub Bot logged work on HADOOP-17981: --- Author: ASF GitHub Bot Created on: 04/Nov/21 01:55 Start Date: 04/Nov/21 01:55 Worklog Time Spent: 10m Work Description: steveloughran closed pull request #3597: URL: https://github.com/apache/hadoop/pull/3597 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org Issue Time Tracking --- Worklog Id: (was: 676062) Time Spent: 8h (was: 7h 50m) > Support etag-assisted renames in FileOutputCommitter > > > Key: HADOOP-17981 > URL: https://issues.apache.org/jira/browse/HADOOP-17981 > Project: Hadoop Common > Issue Type: New Feature > Components: fs, fs/azure >Affects Versions: 3.4.0 >Reporter: Steve Loughran >Assignee: Steve Loughran >Priority: Major > Labels: pull-request-available > Time Spent: 8h > Remaining Estimate: 0h > > To deal with some throttling/retry issues in object stores, > pass the FileStatus entries retrieved during listing > into a private interface ResilientCommitByRename which filesystems > may implement to use extra attributes in the listing (etag, version) > to constrain and validate the operation. > Although targeting azure, GCS and others could use. no point in S3A as they > shouldn't use this committer. > # And we are not going to do any changes to FileSystem as there are explicit > guarantees of public use and stability. > I am not going to make that change as the hive thing that will suddenly start > expecting it to work forever. > # I'm not planning to merge this in, as the manifest committer is going to > include this and more (MAPREDUCE-7341) > However, I do need to get this in on a branch, so am doing this work on trunk > for dev & test and for others to review -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[GitHub] [hadoop] haiyang1987 commented on pull request #3596: HDFS-16287. Support to make dfs.namenode.avoid.read.slow.datanode reconfigurable
haiyang1987 commented on pull request #3596: URL: https://github.com/apache/hadoop/pull/3596#issuecomment-958731868 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[GitHub] [hadoop] jianghuazhu commented on a change in pull request #3602: HDFS-16291.Make the comment of INode#ReclaimContext more standardized.
jianghuazhu commented on a change in pull request #3602: URL: https://github.com/apache/hadoop/pull/3602#discussion_r741662220 ## File path: hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/INode.java ## @@ -993,15 +993,13 @@ public long getNsDelta() { private final QuotaDelta quotaDelta; /** - * @param bsps - * block storage policy suite to calculate intended storage type Review comment: Thanks @ferhui for the comment and review. I will update it later. The new style will look like this: ![image](https://user-images.githubusercontent.com/6416939/140021341-f0016022-613a-4de2-a658-a63c336b2b14.png) -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[GitHub] [hadoop] sidseth commented on pull request #3597: HADOOP-17981 Support etag-assisted renames in FileOutputCommitter
sidseth commented on pull request #3597: URL: https://github.com/apache/hadoop/pull/3597#issuecomment-958658416 > > This mechanism becomes very FileSystem specific. Implemented by Azure right now. > > I agree, which is why the API is restricted for its uses to mr-client-core only. as abfs is the only one which needs it for correctness under load, And I'm not worried about that specifity. Can I point to how much of the hadoop fs api are hdfs-only -and they are public. > > > Other users of rename will not see the benefits without changing interfaces, which in turn requires shimming etc. > > Please don't try and use this particular interface in Hive. > Was referring to any potential usage - including Hive. > > Would it be better for AzureFileSystem rename itself to add a config parameter which can lookup the src etag (at the cost of a performance hit for consistency), so that downstream components / any users of the rename operation can benefit from this change without having to change interfaces. > > We are going straight from a listing (1 request/500 entries) to that rename. doing a HEAD first cuts the throughtput in half. so no. > In the scenario where this is encountered. Would not be the default behaviour, and limits the change to Abfs. Could also have the less consistent version which is not etag based, and responds only on failures. Again - limited to Abfs. > > Also, if the performance penalty is a big problem - Abfs could create very short-lived caches for FileStatus objects, and handle errors on discrepancies with the cached copy. > > Possible but convoluted. > Agree. Quite convoluted. Tossing in potential options - to avoid a new public API. > > Essentially - don't force usage of the new interface to get the benefits. > > I understand the interests of the hive team, but this fix is not the place to do a better API. > > Briefly cacheing the source FS entries is something to consider though. Not this week. > > What I could do with is some help getting #2735 in, then we can start on a public rename() builder API which will take a file status, as openFile does. > This particular change would be FSImpl agnostic, and potentially remove the need for the new interface here? > > Side note: The fs.getStatus within ResilientCommitByRenameHelper for FileSystems where this new functionality is not supported will lead to a performance penalty for the other FileSystems (performing a getFileStatus on src). > > There is an option to say "i know it is not there"; this skips the check. the committer passes this option down because it issues a delete call first. > EOD - this ends up being a new API (almost on the FileSystem), which is used by the committer first; then someone discovers it and decides to make use of it. > FWIW the manifest committer will make that pre-rename commit optional, saving that IO request. I am curious as to how well that will work I went executed on well formed tables. -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[GitHub] [hadoop] steveloughran closed pull request #3597: HADOOP-17981 Support etag-assisted renames in FileOutputCommitter
steveloughran closed pull request #3597: URL: https://github.com/apache/hadoop/pull/3597 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[GitHub] [hadoop] tomscut commented on a change in pull request #3596: HDFS-16287. Support to make dfs.namenode.avoid.read.slow.datanode reconfigurable
tomscut commented on a change in pull request #3596: URL: https://github.com/apache/hadoop/pull/3596#discussion_r742015162 ## File path: hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/DatanodeManager.java ## @@ -511,7 +505,16 @@ private boolean isInactive(DatanodeInfo datanode) { private boolean isSlowNode(String dnUuid) { return avoidSlowDataNodesForRead && slowNodesUuidSet.contains(dnUuid); } - + + public void setAvoidSlowDataNodesForReadEnabled(boolean enable) { Review comment: Thanks @haiyang1987 for your comment. I think the logic right now is no problem. I mean that when ```excludeSlowNodesEnabled``` is set to true we ```startSlowPeerCollector```, and ```stopSlowPeerCollector``` when ```excludeSlowNodesEnabled``` is set to false. There is no extra overhead. What do you think? ## File path: hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/DatanodeManager.java ## @@ -511,7 +505,16 @@ private boolean isInactive(DatanodeInfo datanode) { private boolean isSlowNode(String dnUuid) { return avoidSlowDataNodesForRead && slowNodesUuidSet.contains(dnUuid); } - + + public void setAvoidSlowDataNodesForReadEnabled(boolean enable) { Review comment: Thanks @haiyang1987 for your comment. I think the logic right now is no problem. I mean that only when ```excludeSlowNodesEnabled``` is set to true we ```startSlowPeerCollector```, and ```stopSlowPeerCollector``` when ```excludeSlowNodesEnabled``` is set to false. There is no extra overhead. What do you think? ## File path: hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/DatanodeManager.java ## @@ -511,7 +505,16 @@ private boolean isInactive(DatanodeInfo datanode) { private boolean isSlowNode(String dnUuid) { return avoidSlowDataNodesForRead && slowNodesUuidSet.contains(dnUuid); } - + + public void setAvoidSlowDataNodesForReadEnabled(boolean enable) { Review comment: Thanks @haiyang1987 for your comment. I think the logic right now is no problem. I mean that only when `excludeSlowNodesEnabled` is set to `true` we `startSlowPeerCollector`, and `stopSlowPeerCollector` when `excludeSlowNodesEnabled` is set to `false`. There is no extra overhead. What do you think? -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[GitHub] [hadoop] hadoop-yetus commented on pull request #3602: HDFS-16291.Make the comment of INode#ReclaimContext more standardized.
hadoop-yetus commented on pull request #3602: URL: https://github.com/apache/hadoop/pull/3602#issuecomment-959335456 :broken_heart: **-1 overall** | Vote | Subsystem | Runtime | Logfile | Comment | |::|--:|:|::|:---:| | +0 :ok: | reexec | 0m 56s | | Docker mode activated. | _ Prechecks _ | | +1 :green_heart: | dupname | 0m 0s | | No case conflicting files found. | | +0 :ok: | codespell | 0m 2s | | codespell was not available. | | +1 :green_heart: | @author | 0m 0s | | The patch does not contain any @author tags. | | -1 :x: | test4tests | 0m 0s | | The patch doesn't appear to include any new or modified tests. Please justify why no new tests are needed for this patch. Also please list what manual steps were performed to verify this patch. | _ trunk Compile Tests _ | | +1 :green_heart: | mvninstall | 35m 13s | | trunk passed | | +1 :green_heart: | compile | 1m 22s | | trunk passed with JDK Ubuntu-11.0.11+9-Ubuntu-0ubuntu2.20.04 | | +1 :green_heart: | compile | 1m 15s | | trunk passed with JDK Private Build-1.8.0_292-8u292-b10-0ubuntu1~20.04-b10 | | +1 :green_heart: | checkstyle | 1m 2s | | trunk passed | | +1 :green_heart: | mvnsite | 1m 22s | | trunk passed | | +1 :green_heart: | javadoc | 0m 57s | | trunk passed with JDK Ubuntu-11.0.11+9-Ubuntu-0ubuntu2.20.04 | | +1 :green_heart: | javadoc | 1m 26s | | trunk passed with JDK Private Build-1.8.0_292-8u292-b10-0ubuntu1~20.04-b10 | | +1 :green_heart: | spotbugs | 3m 15s | | trunk passed | | +1 :green_heart: | shadedclient | 25m 3s | | branch has no errors when building and testing our client artifacts. | _ Patch Compile Tests _ | | +1 :green_heart: | mvninstall | 1m 15s | | the patch passed | | +1 :green_heart: | compile | 1m 17s | | the patch passed with JDK Ubuntu-11.0.11+9-Ubuntu-0ubuntu2.20.04 | | +1 :green_heart: | javac | 1m 17s | | the patch passed | | +1 :green_heart: | compile | 1m 10s | | the patch passed with JDK Private Build-1.8.0_292-8u292-b10-0ubuntu1~20.04-b10 | | +1 :green_heart: | javac | 1m 10s | | the patch passed | | +1 :green_heart: | blanks | 0m 0s | | The patch has no blanks issues. | | +1 :green_heart: | checkstyle | 0m 54s | | the patch passed | | +1 :green_heart: | mvnsite | 1m 14s | | the patch passed | | +1 :green_heart: | javadoc | 0m 47s | | the patch passed with JDK Ubuntu-11.0.11+9-Ubuntu-0ubuntu2.20.04 | | +1 :green_heart: | javadoc | 1m 17s | | the patch passed with JDK Private Build-1.8.0_292-8u292-b10-0ubuntu1~20.04-b10 | | +1 :green_heart: | spotbugs | 3m 19s | | the patch passed | | +1 :green_heart: | shadedclient | 26m 58s | | patch has no errors when building and testing our client artifacts. | _ Other Tests _ | | -1 :x: | unit | 348m 51s | [/patch-unit-hadoop-hdfs-project_hadoop-hdfs.txt](https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-3602/2/artifact/out/patch-unit-hadoop-hdfs-project_hadoop-hdfs.txt) | hadoop-hdfs in the patch passed. | | +1 :green_heart: | asflicense | 0m 39s | | The patch does not generate ASF License warnings. | | | | 457m 26s | | | | Reason | Tests | |---:|:--| | Failed junit tests | hadoop.hdfs.TestHDFSFileSystemContract | | Subsystem | Report/Notes | |--:|:-| | Docker | ClientAPI=1.41 ServerAPI=1.41 base: https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-3602/2/artifact/out/Dockerfile | | GITHUB PR | https://github.com/apache/hadoop/pull/3602 | | Optional Tests | dupname asflicense compile javac javadoc mvninstall mvnsite unit shadedclient spotbugs checkstyle codespell | | uname | Linux 5d5d26c42b34 4.15.0-147-generic #151-Ubuntu SMP Fri Jun 18 19:21:19 UTC 2021 x86_64 x86_64 x86_64 GNU/Linux | | Build tool | maven | | Personality | dev-support/bin/hadoop.sh | | git revision | trunk / 673f55d0883ee7bf09e70202f14d4e334adc3cc5 | | Default Java | Private Build-1.8.0_292-8u292-b10-0ubuntu1~20.04-b10 | | Multi-JDK versions | /usr/lib/jvm/java-11-openjdk-amd64:Ubuntu-11.0.11+9-Ubuntu-0ubuntu2.20.04 /usr/lib/jvm/java-8-openjdk-amd64:Private Build-1.8.0_292-8u292-b10-0ubuntu1~20.04-b10 | | Test Results | https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-3602/2/testReport/ | | Max. process+thread count | 1906 (vs. ulimit of 5500) | | modules | C: hadoop-hdfs-project/hadoop-hdfs U: hadoop-hdfs-project/hadoop-hdfs | | Console output | https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-3602/2/console | | versions | git=2.25.1 maven=3.6.3 spotbugs=4.2.2 | | Powered by | Apache Yetus 0.14.0-SNAPSHOT https://yetus.apache.org | This message was aut
[GitHub] [hadoop] goiri commented on a change in pull request #3553: HDFS-16273. RBF: RouterRpcFairnessPolicyController add availableHandl…
goiri commented on a change in pull request #3553: URL: https://github.com/apache/hadoop/pull/3553#discussion_r742198185 ## File path: hadoop-hdfs-project/hadoop-hdfs-rbf/src/main/java/org/apache/hadoop/hdfs/server/federation/fairness/NoRouterRpcFairnessPolicyController.java ## @@ -46,4 +46,9 @@ public void releasePermit(String nsId) { public void shutdown() { // Nothing for now. } + + @Override + public String getAvailableHandlerOnPerNs(){ +return "N/A"; Review comment: Should we test for this? ## File path: hadoop-hdfs-project/hadoop-hdfs-rbf/src/main/java/org/apache/hadoop/hdfs/server/federation/fairness/AbstractRouterRpcFairnessPolicyController.java ## @@ -75,4 +77,17 @@ protected void insertNameServiceWithPermits(String nsId, int maxPermits) { protected int getAvailablePermits(String nsId) { return this.permits.get(nsId).availablePermits(); } + + @Override + public String getAvailableHandlerOnPerNs() { +JSONObject json = new JSONObject(); +for (Map.Entry entry : permits.entrySet()) { + try { +json.put(entry.getKey(), entry.getValue().availablePermits()); Review comment: Let's extract entry.getKey() and entry.getValue() to have a particular name. -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[GitHub] [hadoop] haiyang1987 commented on a change in pull request #3596: HDFS-16287. Support to make dfs.namenode.avoid.read.slow.datanode reconfigurable
haiyang1987 commented on a change in pull request #3596: URL: https://github.com/apache/hadoop/pull/3596#discussion_r741696525 ## File path: hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/DatanodeManager.java ## @@ -260,17 +257,14 @@ final Timer timer = new Timer(); this.slowPeerTracker = dataNodePeerStatsEnabled ? new SlowPeerTracker(conf, timer) : null; -this.excludeSlowNodesEnabled = conf.getBoolean( -DFS_NAMENODE_BLOCKPLACEMENTPOLICY_EXCLUDE_SLOW_NODES_ENABLED_KEY, -DFS_NAMENODE_BLOCKPLACEMENTPOLICY_EXCLUDE_SLOW_NODES_ENABLED_DEFAULT); this.maxSlowPeerReportNodes = conf.getInt( DFSConfigKeys.DFS_NAMENODE_MAX_SLOWPEER_COLLECT_NODES_KEY, DFSConfigKeys.DFS_NAMENODE_MAX_SLOWPEER_COLLECT_NODES_DEFAULT); this.slowPeerCollectionInterval = conf.getTimeDuration( DFSConfigKeys.DFS_NAMENODE_SLOWPEER_COLLECT_INTERVAL_KEY, DFSConfigKeys.DFS_NAMENODE_SLOWPEER_COLLECT_INTERVAL_DEFAULT, TimeUnit.MILLISECONDS); -if (slowPeerTracker != null && excludeSlowNodesEnabled) { Review comment: @tomscut Thank you for your review. 1.Current parameter 'dataNodePeerStatsEnabled' and 'excludeSlowNodesEnabled' decision SlowPeerCollector thread whether to start ,But it didn't take into account avoid SlowDataNodesForRead logic 2.So think about two phases: a.The first is to start SlowPeerCollector thread b.Second, you can control whether to enable read/write avoid slow datanode according to dynamic parameters ## File path: hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/DatanodeManager.java ## @@ -511,7 +505,16 @@ private boolean isInactive(DatanodeInfo datanode) { private boolean isSlowNode(String dnUuid) { return avoidSlowDataNodesForRead && slowNodesUuidSet.contains(dnUuid); } - + + public void setAvoidSlowDataNodesForReadEnabled(boolean enable) { Review comment: Consider slowNodesUuidSet is generated when the SlowPeerCollector thread is started,therefore it is logical to judge Therefore, it is logical to judge whether the dnUuid exists in the slowNodesUuidSet? ## File path: hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/DatanodeManager.java ## @@ -511,7 +505,16 @@ private boolean isInactive(DatanodeInfo datanode) { private boolean isSlowNode(String dnUuid) { return avoidSlowDataNodesForRead && slowNodesUuidSet.contains(dnUuid); } - + + public void setAvoidSlowDataNodesForReadEnabled(boolean enable) { Review comment: Consider slowNodesUuidSet is generated when the SlowPeerCollector thread is started,therefore it is logical to judge whether the dnUuid exists in the slowNodesUuidSet? -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Work logged] (HADOOP-17124) Support LZO using aircompressor
[ https://issues.apache.org/jira/browse/HADOOP-17124?focusedWorklogId=675999&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-675999 ] ASF GitHub Bot logged work on HADOOP-17124: --- Author: ASF GitHub Bot Created on: 04/Nov/21 01:49 Start Date: 04/Nov/21 01:49 Worklog Time Spent: 10m Work Description: hadoop-yetus commented on pull request #3612: URL: https://github.com/apache/hadoop/pull/3612#issuecomment-958607089 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org Issue Time Tracking --- Worklog Id: (was: 675999) Time Spent: 6h 40m (was: 6.5h) > Support LZO using aircompressor > --- > > Key: HADOOP-17124 > URL: https://issues.apache.org/jira/browse/HADOOP-17124 > Project: Hadoop Common > Issue Type: New Feature > Components: common >Affects Versions: 3.3.0 >Reporter: DB Tsai >Priority: Major > Labels: pull-request-available > Time Spent: 6h 40m > Remaining Estimate: 0h > > LZO codec was removed in HADOOP-4874 because the original LZO binding is GPL > which is problematic. However, many legacy data is still compressed by LZO > codec, and companies often use vendor's GPL LZO codec in the classpath which > might cause GPL contamination. > Presro and ORC-77 use [aircompressor| > [https://github.com/airlift/aircompressor]] (Apache V2 licensed) to compress > and decompress LZO data. Hadoop can add back LZO support using aircompressor > without GPL violation. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[GitHub] [hadoop] cndaimin commented on a change in pull request #3593: HDFS-16286. Add a debug tool to verify the correctness of erasure coding on file
cndaimin commented on a change in pull request #3593: URL: https://github.com/apache/hadoop/pull/3593#discussion_r741582271 ## File path: hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/tools/TestDebugAdmin.java ## @@ -166,8 +179,91 @@ public void testComputeMetaCommand() throws Exception { @Test(timeout = 6) public void testRecoverLeaseforFileNotFound() throws Exception { +cluster = new MiniDFSCluster.Builder(conf).numDataNodes(1).build(); +cluster.waitActive(); assertTrue(runCmd(new String[] { "recoverLease", "-path", "/foo", "-retries", "2" }).contains( "Giving up on recoverLease for /foo after 1 try")); } + + @Test(timeout = 6) + public void testVerifyECCommand() throws Exception { +final ErasureCodingPolicy ecPolicy = SystemErasureCodingPolicies.getByID( +SystemErasureCodingPolicies.RS_3_2_POLICY_ID); +cluster = DFSTestUtil.setupCluster(conf, 6, 5, 0); +cluster.waitActive(); +DistributedFileSystem fs = cluster.getFileSystem(); + +assertEquals("ret: 1, verifyEC -file Verify HDFS erasure coding on " + +"all block groups of the file.", runCmd(new String[]{"verifyEC"})); + +assertEquals("ret: 1, File /bar does not exist.", +runCmd(new String[]{"verifyEC", "-file", "/bar"})); + +fs.create(new Path("/bar")).close(); +assertEquals("ret: 1, File /bar is not erasure coded.", +runCmd(new String[]{"verifyEC", "-file", "/bar"})); + + +final Path ecDir = new Path("/ec"); +fs.mkdir(ecDir, FsPermission.getDirDefault()); +fs.enableErasureCodingPolicy(ecPolicy.getName()); +fs.setErasureCodingPolicy(ecDir, ecPolicy.getName()); + +assertEquals("ret: 1, File /ec is not a regular file.", +runCmd(new String[]{"verifyEC", "-file", "/ec"})); + +fs.create(new Path(ecDir, "foo")); +assertEquals("ret: 1, File /ec/foo is not closed.", +runCmd(new String[]{"verifyEC", "-file", "/ec/foo"})); + +final short repl = 1; +final long k = 1024; +final long m = k * k; +final long seed = 0x1234567L; +DFSTestUtil.createFile(fs, new Path(ecDir, "foo_65535"), 65535, repl, seed); +assertTrue(runCmd(new String[]{"verifyEC", "-file", "/ec/foo_65535"}) +.contains("All EC block group status: OK")); +DFSTestUtil.createFile(fs, new Path(ecDir, "foo_256k"), 256 * k, repl, seed); +assertTrue(runCmd(new String[]{"verifyEC", "-file", "/ec/foo_256k"}) +.contains("All EC block group status: OK")); +DFSTestUtil.createFile(fs, new Path(ecDir, "foo_1m"), m, repl, seed); +assertTrue(runCmd(new String[]{"verifyEC", "-file", "/ec/foo_1m"}) +.contains("All EC block group status: OK")); +DFSTestUtil.createFile(fs, new Path(ecDir, "foo_2m"), 2 * m, repl, seed); +assertTrue(runCmd(new String[]{"verifyEC", "-file", "/ec/foo_2m"}) +.contains("All EC block group status: OK")); +DFSTestUtil.createFile(fs, new Path(ecDir, "foo_3m"), 3 * m, repl, seed); +assertTrue(runCmd(new String[]{"verifyEC", "-file", "/ec/foo_3m"}) +.contains("All EC block group status: OK")); +DFSTestUtil.createFile(fs, new Path(ecDir, "foo_5m"), 5 * m, repl, seed); +assertTrue(runCmd(new String[]{"verifyEC", "-file", "/ec/foo_5m"}) +.contains("All EC block group status: OK")); + Review comment: Thanks, that's a good advice, updated. -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[GitHub] [hadoop] hadoop-yetus commented on pull request #3612: WIP. HADOOP-17124. Support LZO Codec using aircompressor
hadoop-yetus commented on pull request #3612: URL: https://github.com/apache/hadoop/pull/3612#issuecomment-958607089 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Work logged] (HADOOP-17124) Support LZO using aircompressor
[ https://issues.apache.org/jira/browse/HADOOP-17124?focusedWorklogId=675927&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-675927 ] ASF GitHub Bot logged work on HADOOP-17124: --- Author: ASF GitHub Bot Created on: 04/Nov/21 01:41 Start Date: 04/Nov/21 01:41 Worklog Time Spent: 10m Work Description: sunchao commented on pull request #3612: URL: https://github.com/apache/hadoop/pull/3612#issuecomment-959727355 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org Issue Time Tracking --- Worklog Id: (was: 675927) Time Spent: 6.5h (was: 6h 20m) > Support LZO using aircompressor > --- > > Key: HADOOP-17124 > URL: https://issues.apache.org/jira/browse/HADOOP-17124 > Project: Hadoop Common > Issue Type: New Feature > Components: common >Affects Versions: 3.3.0 >Reporter: DB Tsai >Priority: Major > Labels: pull-request-available > Time Spent: 6.5h > Remaining Estimate: 0h > > LZO codec was removed in HADOOP-4874 because the original LZO binding is GPL > which is problematic. However, many legacy data is still compressed by LZO > codec, and companies often use vendor's GPL LZO codec in the classpath which > might cause GPL contamination. > Presro and ORC-77 use [aircompressor| > [https://github.com/airlift/aircompressor]] (Apache V2 licensed) to compress > and decompress LZO data. Hadoop can add back LZO support using aircompressor > without GPL violation. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[GitHub] [hadoop] tasanuma merged pull request #3538: HDFS-16266. Add remote port information to HDFS audit log
tasanuma merged pull request #3538: URL: https://github.com/apache/hadoop/pull/3538 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[GitHub] [hadoop] haiyang1987 closed pull request #3596: HDFS-16287. Support to make dfs.namenode.avoid.read.slow.datanode reconfigurable
haiyang1987 closed pull request #3596: URL: https://github.com/apache/hadoop/pull/3596 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[GitHub] [hadoop] hadoop-yetus commented on pull request #3593: HDFS-16286. Add a debug tool to verify the correctness of erasure coding on file
hadoop-yetus commented on pull request #3593: URL: https://github.com/apache/hadoop/pull/3593#issuecomment-958791127 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Work logged] (HADOOP-17981) Support etag-assisted renames in FileOutputCommitter
[ https://issues.apache.org/jira/browse/HADOOP-17981?focusedWorklogId=675938&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-675938 ] ASF GitHub Bot logged work on HADOOP-17981: --- Author: ASF GitHub Bot Created on: 04/Nov/21 01:42 Start Date: 04/Nov/21 01:42 Worklog Time Spent: 10m Work Description: steveloughran commented on pull request #3597: URL: https://github.com/apache/hadoop/pull/3597#issuecomment-959193924 > EOD - this ends up being a new API (almost on the FileSystem), which is used by the committer first; then someone discovers it and decides to make use of it. yes, and both hive and hbase are known to do that, often ending forcing hdfs only apis to get pulled up without concern for the other stores (see git history of FileSystem there). a builder rename would be a big job with 1. rename/3 public with tests and performant object store impls 2. default builder to delegate to that with same semantics 3. add options for etag etc on a store by store or cross store basis -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org Issue Time Tracking --- Worklog Id: (was: 675938) Time Spent: 7h 50m (was: 7h 40m) > Support etag-assisted renames in FileOutputCommitter > > > Key: HADOOP-17981 > URL: https://issues.apache.org/jira/browse/HADOOP-17981 > Project: Hadoop Common > Issue Type: New Feature > Components: fs, fs/azure >Affects Versions: 3.4.0 >Reporter: Steve Loughran >Assignee: Steve Loughran >Priority: Major > Labels: pull-request-available > Time Spent: 7h 50m > Remaining Estimate: 0h > > To deal with some throttling/retry issues in object stores, > pass the FileStatus entries retrieved during listing > into a private interface ResilientCommitByRename which filesystems > may implement to use extra attributes in the listing (etag, version) > to constrain and validate the operation. > Although targeting azure, GCS and others could use. no point in S3A as they > shouldn't use this committer. > # And we are not going to do any changes to FileSystem as there are explicit > guarantees of public use and stability. > I am not going to make that change as the hive thing that will suddenly start > expecting it to work forever. > # I'm not planning to merge this in, as the manifest committer is going to > include this and more (MAPREDUCE-7341) > However, I do need to get this in on a branch, so am doing this work on trunk > for dev & test and for others to review -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[GitHub] [hadoop] steveloughran commented on pull request #3597: HADOOP-17981 Support etag-assisted renames in FileOutputCommitter
steveloughran commented on pull request #3597: URL: https://github.com/apache/hadoop/pull/3597#issuecomment-959193924 > EOD - this ends up being a new API (almost on the FileSystem), which is used by the committer first; then someone discovers it and decides to make use of it. yes, and both hive and hbase are known to do that, often ending forcing hdfs only apis to get pulled up without concern for the other stores (see git history of FileSystem there). a builder rename would be a big job with 1. rename/3 public with tests and performant object store impls 2. default builder to delegate to that with same semantics 3. add options for etag etc on a store by store or cross store basis -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[GitHub] [hadoop] sunchao commented on pull request #3612: WIP. HADOOP-17124. Support LZO Codec using aircompressor
sunchao commented on pull request #3612: URL: https://github.com/apache/hadoop/pull/3612#issuecomment-959727355 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Work logged] (HADOOP-17124) Support LZO using aircompressor
[ https://issues.apache.org/jira/browse/HADOOP-17124?focusedWorklogId=675901&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-675901 ] ASF GitHub Bot logged work on HADOOP-17124: --- Author: ASF GitHub Bot Created on: 04/Nov/21 01:39 Start Date: 04/Nov/21 01:39 Worklog Time Spent: 10m Work Description: viirya commented on pull request #3612: URL: https://github.com/apache/hadoop/pull/3612#issuecomment-958606536 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org Issue Time Tracking --- Worklog Id: (was: 675901) Time Spent: 6h 20m (was: 6h 10m) > Support LZO using aircompressor > --- > > Key: HADOOP-17124 > URL: https://issues.apache.org/jira/browse/HADOOP-17124 > Project: Hadoop Common > Issue Type: New Feature > Components: common >Affects Versions: 3.3.0 >Reporter: DB Tsai >Priority: Major > Labels: pull-request-available > Time Spent: 6h 20m > Remaining Estimate: 0h > > LZO codec was removed in HADOOP-4874 because the original LZO binding is GPL > which is problematic. However, many legacy data is still compressed by LZO > codec, and companies often use vendor's GPL LZO codec in the classpath which > might cause GPL contamination. > Presro and ORC-77 use [aircompressor| > [https://github.com/airlift/aircompressor]] (Apache V2 licensed) to compress > and decompress LZO data. Hadoop can add back LZO support using aircompressor > without GPL violation. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[GitHub] [hadoop-site] GauthamBanasandra commented on pull request #28: Add Gautham Banasandra to Hadoop committers' list
GauthamBanasandra commented on pull request #28: URL: https://github.com/apache/hadoop-site/pull/28#issuecomment-959811956 @aajisaka could you please review this PR? -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[GitHub] [hadoop] viirya commented on pull request #3612: WIP. HADOOP-17124. Support LZO Codec using aircompressor
viirya commented on pull request #3612: URL: https://github.com/apache/hadoop/pull/3612#issuecomment-958606536 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Work logged] (HADOOP-17124) Support LZO using aircompressor
[ https://issues.apache.org/jira/browse/HADOOP-17124?focusedWorklogId=675889&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-675889 ] ASF GitHub Bot logged work on HADOOP-17124: --- Author: ASF GitHub Bot Created on: 04/Nov/21 01:38 Start Date: 04/Nov/21 01:38 Worklog Time Spent: 10m Work Description: viirya edited a comment on pull request #3612: URL: https://github.com/apache/hadoop/pull/3612#issuecomment-959725363 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org Issue Time Tracking --- Worklog Id: (was: 675889) Time Spent: 6h 10m (was: 6h) > Support LZO using aircompressor > --- > > Key: HADOOP-17124 > URL: https://issues.apache.org/jira/browse/HADOOP-17124 > Project: Hadoop Common > Issue Type: New Feature > Components: common >Affects Versions: 3.3.0 >Reporter: DB Tsai >Priority: Major > Labels: pull-request-available > Time Spent: 6h 10m > Remaining Estimate: 0h > > LZO codec was removed in HADOOP-4874 because the original LZO binding is GPL > which is problematic. However, many legacy data is still compressed by LZO > codec, and companies often use vendor's GPL LZO codec in the classpath which > might cause GPL contamination. > Presro and ORC-77 use [aircompressor| > [https://github.com/airlift/aircompressor]] (Apache V2 licensed) to compress > and decompress LZO data. Hadoop can add back LZO support using aircompressor > without GPL violation. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[GitHub] [hadoop] viirya edited a comment on pull request #3612: WIP. HADOOP-17124. Support LZO Codec using aircompressor
viirya edited a comment on pull request #3612: URL: https://github.com/apache/hadoop/pull/3612#issuecomment-959725363 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[GitHub] [hadoop] hadoop-yetus commented on pull request #3596: HDFS-16287. Support to make dfs.namenode.avoid.read.slow.datanode reconfigurable
hadoop-yetus commented on pull request #3596: URL: https://github.com/apache/hadoop/pull/3596#issuecomment-959651079 :broken_heart: **-1 overall** | Vote | Subsystem | Runtime | Logfile | Comment | |::|--:|:|::|:---:| | +0 :ok: | reexec | 1m 24s | | Docker mode activated. | _ Prechecks _ | | +1 :green_heart: | dupname | 0m 0s | | No case conflicting files found. | | +0 :ok: | codespell | 0m 1s | | codespell was not available. | | +1 :green_heart: | @author | 0m 0s | | The patch does not contain any @author tags. | | +1 :green_heart: | test4tests | 0m 0s | | The patch appears to include 2 new or modified test files. | _ trunk Compile Tests _ | | +1 :green_heart: | mvninstall | 36m 58s | | trunk passed | | +1 :green_heart: | compile | 1m 35s | | trunk passed with JDK Ubuntu-11.0.11+9-Ubuntu-0ubuntu2.20.04 | | +1 :green_heart: | compile | 1m 22s | | trunk passed with JDK Private Build-1.8.0_292-8u292-b10-0ubuntu1~20.04-b10 | | +1 :green_heart: | checkstyle | 1m 4s | | trunk passed | | +1 :green_heart: | mvnsite | 1m 33s | | trunk passed | | +1 :green_heart: | javadoc | 1m 3s | | trunk passed with JDK Ubuntu-11.0.11+9-Ubuntu-0ubuntu2.20.04 | | +1 :green_heart: | javadoc | 1m 28s | | trunk passed with JDK Private Build-1.8.0_292-8u292-b10-0ubuntu1~20.04-b10 | | +1 :green_heart: | spotbugs | 3m 41s | | trunk passed | | +1 :green_heart: | shadedclient | 25m 39s | | branch has no errors when building and testing our client artifacts. | _ Patch Compile Tests _ | | +1 :green_heart: | mvninstall | 1m 27s | | the patch passed | | +1 :green_heart: | compile | 1m 28s | | the patch passed with JDK Ubuntu-11.0.11+9-Ubuntu-0ubuntu2.20.04 | | +1 :green_heart: | javac | 1m 28s | | the patch passed | | +1 :green_heart: | compile | 1m 18s | | the patch passed with JDK Private Build-1.8.0_292-8u292-b10-0ubuntu1~20.04-b10 | | +1 :green_heart: | javac | 1m 18s | | the patch passed | | +1 :green_heart: | blanks | 0m 0s | | The patch has no blanks issues. | | +1 :green_heart: | checkstyle | 0m 57s | | the patch passed | | +1 :green_heart: | mvnsite | 1m 25s | | the patch passed | | +1 :green_heart: | javadoc | 0m 54s | | the patch passed with JDK Ubuntu-11.0.11+9-Ubuntu-0ubuntu2.20.04 | | +1 :green_heart: | javadoc | 1m 24s | | the patch passed with JDK Private Build-1.8.0_292-8u292-b10-0ubuntu1~20.04-b10 | | +1 :green_heart: | spotbugs | 3m 44s | | the patch passed | | +1 :green_heart: | shadedclient | 25m 53s | | patch has no errors when building and testing our client artifacts. | _ Other Tests _ | | -1 :x: | unit | 374m 11s | [/patch-unit-hadoop-hdfs-project_hadoop-hdfs.txt](https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-3596/4/artifact/out/patch-unit-hadoop-hdfs-project_hadoop-hdfs.txt) | hadoop-hdfs in the patch passed. | | +1 :green_heart: | asflicense | 0m 43s | | The patch does not generate ASF License warnings. | | | | 486m 5s | | | | Reason | Tests | |---:|:--| | Failed junit tests | hadoop.hdfs.TestHDFSFileSystemContract | | Subsystem | Report/Notes | |--:|:-| | Docker | ClientAPI=1.41 ServerAPI=1.41 base: https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-3596/4/artifact/out/Dockerfile | | GITHUB PR | https://github.com/apache/hadoop/pull/3596 | | Optional Tests | dupname asflicense compile javac javadoc mvninstall mvnsite unit shadedclient spotbugs checkstyle codespell | | uname | Linux 1fb2ee0e949b 4.15.0-143-generic #147-Ubuntu SMP Wed Apr 14 16:10:11 UTC 2021 x86_64 x86_64 x86_64 GNU/Linux | | Build tool | maven | | Personality | dev-support/bin/hadoop.sh | | git revision | trunk / 2ec82e1c420789afb326f4ebb451522a8a4e2358 | | Default Java | Private Build-1.8.0_292-8u292-b10-0ubuntu1~20.04-b10 | | Multi-JDK versions | /usr/lib/jvm/java-11-openjdk-amd64:Ubuntu-11.0.11+9-Ubuntu-0ubuntu2.20.04 /usr/lib/jvm/java-8-openjdk-amd64:Private Build-1.8.0_292-8u292-b10-0ubuntu1~20.04-b10 | | Test Results | https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-3596/4/testReport/ | | Max. process+thread count | 2022 (vs. ulimit of 5500) | | modules | C: hadoop-hdfs-project/hadoop-hdfs U: hadoop-hdfs-project/hadoop-hdfs | | Console output | https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-3596/4/console | | versions | git=2.25.1 maven=3.6.3 spotbugs=4.2.2 | | Powered by | Apache Yetus 0.14.0-SNAPSHOT https://yetus.apache.org | This message was automatically generated. -- This is an automated message from the Apache Git Service. To respond to the message, please
[GitHub] [hadoop] goiri commented on a change in pull request #3613: HDFS-16296. RouterRpcFairnessPolicyController add rejected permits for each nameservice
goiri commented on a change in pull request #3613: URL: https://github.com/apache/hadoop/pull/3613#discussion_r742196019 ## File path: hadoop-hdfs-project/hadoop-hdfs-rbf/src/test/java/org/apache/hadoop/hdfs/server/federation/fairness/TestRouterHandlersFairness.java ## @@ -208,4 +212,15 @@ private void invokeConcurrent(ClientProtocol routerProto, String clientName) routerProto.renewLease(clientName); } + private int getTotalRejectedPermits(RouterContext routerContext) { +int totalRejectedPermits = 0; +for (String ns : cluster.getNameservices()) { + totalRejectedPermits += routerContext.getRouter().getRpcServer() Review comment: We may want to extract: routerContext.getRouter().getRpcServer().getRPCClient() -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[GitHub] [hadoop] goiri commented on a change in pull request #3595: HDFS-16283: RBF: improve renewLease() to call only a specific NameNod…
goiri commented on a change in pull request #3595: URL: https://github.com/apache/hadoop/pull/3595#discussion_r742199623 ## File path: hadoop-hdfs-project/hadoop-hdfs-client/src/main/java/org/apache/hadoop/hdfs/protocol/ClientProtocol.java ## @@ -765,6 +765,14 @@ BatchedDirectoryListing getBatchedListing( @Idempotent void renewLease(String clientName) throws IOException; + /** + * The functionality is the same as renewLease(clientName). This is to support + * router based FileSystem to newLease against a specific target FileSystem instead + * of all the target FileSystems in each call. + */ + @Idempotent + void renewLease(String clientName, String nsId) throws IOException; Review comment: That's a good point. ClientProtocol shouldn't care about subclusters. The whole abstraction is based on paths and that would make more sense. -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[GitHub] [hadoop-site] aajisaka commented on pull request #28: Add Gautham Banasandra to Hadoop committers' list
aajisaka commented on pull request #28: URL: https://github.com/apache/hadoop-site/pull/28#issuecomment-959830761 Would you run `hugo` to generate the html files and add them to the commit? -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[GitHub] [hadoop] jianghuazhu commented on pull request #3602: HDFS-16291.Make the comment of INode#ReclaimContext more standardized.
jianghuazhu commented on pull request #3602: URL: https://github.com/apache/hadoop/pull/3602#issuecomment-958751029 Thank you very much. @ferhui @virajjasani -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Work logged] (HADOOP-17981) Support etag-assisted renames in FileOutputCommitter
[ https://issues.apache.org/jira/browse/HADOOP-17981?focusedWorklogId=675761&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-675761 ] ASF GitHub Bot logged work on HADOOP-17981: --- Author: ASF GitHub Bot Created on: 04/Nov/21 01:25 Start Date: 04/Nov/21 01:25 Worklog Time Spent: 10m Work Description: hadoop-yetus removed a comment on pull request #3611: URL: https://github.com/apache/hadoop/pull/3611#issuecomment-958583977 :confetti_ball: **+1 overall** | Vote | Subsystem | Runtime | Logfile | Comment | |::|--:|:|::|:---:| | +0 :ok: | reexec | 1m 3s | | Docker mode activated. | _ Prechecks _ | | +1 :green_heart: | dupname | 0m 0s | | No case conflicting files found. | | +0 :ok: | codespell | 0m 0s | | codespell was not available. | | +0 :ok: | markdownlint | 0m 0s | | markdownlint was not available. | | +1 :green_heart: | @author | 0m 0s | | The patch does not contain any @author tags. | | +1 :green_heart: | test4tests | 0m 0s | | The patch appears to include 4 new or modified test files. | _ trunk Compile Tests _ | | +0 :ok: | mvndep | 12m 46s | | Maven dependency ordering for branch | | +1 :green_heart: | mvninstall | 29m 49s | | trunk passed | | +1 :green_heart: | compile | 29m 33s | | trunk passed with JDK Ubuntu-11.0.11+9-Ubuntu-0ubuntu2.20.04 | | +1 :green_heart: | compile | 23m 25s | | trunk passed with JDK Private Build-1.8.0_292-8u292-b10-0ubuntu1~20.04-b10 | | +1 :green_heart: | checkstyle | 4m 23s | | trunk passed | | +1 :green_heart: | mvnsite | 4m 16s | | trunk passed | | +1 :green_heart: | javadoc | 3m 7s | | trunk passed with JDK Ubuntu-11.0.11+9-Ubuntu-0ubuntu2.20.04 | | +1 :green_heart: | javadoc | 3m 22s | | trunk passed with JDK Private Build-1.8.0_292-8u292-b10-0ubuntu1~20.04-b10 | | +1 :green_heart: | spotbugs | 6m 50s | | trunk passed | | +1 :green_heart: | shadedclient | 27m 13s | | branch has no errors when building and testing our client artifacts. | _ Patch Compile Tests _ | | +0 :ok: | mvndep | 0m 26s | | Maven dependency ordering for patch | | +1 :green_heart: | mvninstall | 3m 21s | | the patch passed | | +1 :green_heart: | compile | 29m 29s | | the patch passed with JDK Ubuntu-11.0.11+9-Ubuntu-0ubuntu2.20.04 | | +1 :green_heart: | javac | 29m 29s | | the patch passed | | +1 :green_heart: | compile | 25m 28s | | the patch passed with JDK Private Build-1.8.0_292-8u292-b10-0ubuntu1~20.04-b10 | | +1 :green_heart: | javac | 25m 28s | | the patch passed | | +1 :green_heart: | blanks | 0m 0s | | The patch has no blanks issues. | | -0 :warning: | checkstyle | 4m 20s | [/results-checkstyle-root.txt](https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-3611/1/artifact/out/results-checkstyle-root.txt) | root: The patch generated 14 new + 40 unchanged - 0 fixed = 54 total (was 40) | | +1 :green_heart: | mvnsite | 4m 24s | | the patch passed | | +1 :green_heart: | javadoc | 2m 47s | | the patch passed with JDK Ubuntu-11.0.11+9-Ubuntu-0ubuntu2.20.04 | | +1 :green_heart: | javadoc | 3m 15s | | the patch passed with JDK Private Build-1.8.0_292-8u292-b10-0ubuntu1~20.04-b10 | | +1 :green_heart: | spotbugs | 6m 35s | | the patch passed | | +1 :green_heart: | shadedclient | 23m 34s | | patch has no errors when building and testing our client artifacts. | _ Other Tests _ | | +1 :green_heart: | unit | 17m 8s | | hadoop-common in the patch passed. | | +1 :green_heart: | unit | 5m 58s | | hadoop-mapreduce-client-core in the patch passed. | | +1 :green_heart: | unit | 139m 4s | | hadoop-mapreduce-client-jobclient in the patch passed. | | +1 :green_heart: | unit | 2m 13s | | hadoop-azure in the patch passed. | | +1 :green_heart: | asflicense | 0m 53s | | The patch does not generate ASF License warnings. | | | | 418m 20s | | | | Subsystem | Report/Notes | |--:|:-| | Docker | ClientAPI=1.41 ServerAPI=1.41 base: https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-3611/1/artifact/out/Dockerfile | | GITHUB PR | https://github.com/apache/hadoop/pull/3611 | | Optional Tests | dupname asflicense compile javac javadoc mvninstall mvnsite unit shadedclient spotbugs checkstyle codespell markdownlint | | uname | Linux ec598da7f750 4.15.0-147-generic #151-Ubuntu SMP Fri Jun 18 19:21:19 UTC 2021 x86_64 x86_64 x86_64 GNU/Linux | | Build tool | maven | | Personality | dev-support/bin/hadoop.sh | | git revision | trunk / cf88c3518ce9386aeb89c7fcfe62586669b1e429
[GitHub] [hadoop] hadoop-yetus removed a comment on pull request #3611: HADOOP-17981. resilient commit through etag validation
hadoop-yetus removed a comment on pull request #3611: URL: https://github.com/apache/hadoop/pull/3611#issuecomment-958583977 :confetti_ball: **+1 overall** | Vote | Subsystem | Runtime | Logfile | Comment | |::|--:|:|::|:---:| | +0 :ok: | reexec | 1m 3s | | Docker mode activated. | _ Prechecks _ | | +1 :green_heart: | dupname | 0m 0s | | No case conflicting files found. | | +0 :ok: | codespell | 0m 0s | | codespell was not available. | | +0 :ok: | markdownlint | 0m 0s | | markdownlint was not available. | | +1 :green_heart: | @author | 0m 0s | | The patch does not contain any @author tags. | | +1 :green_heart: | test4tests | 0m 0s | | The patch appears to include 4 new or modified test files. | _ trunk Compile Tests _ | | +0 :ok: | mvndep | 12m 46s | | Maven dependency ordering for branch | | +1 :green_heart: | mvninstall | 29m 49s | | trunk passed | | +1 :green_heart: | compile | 29m 33s | | trunk passed with JDK Ubuntu-11.0.11+9-Ubuntu-0ubuntu2.20.04 | | +1 :green_heart: | compile | 23m 25s | | trunk passed with JDK Private Build-1.8.0_292-8u292-b10-0ubuntu1~20.04-b10 | | +1 :green_heart: | checkstyle | 4m 23s | | trunk passed | | +1 :green_heart: | mvnsite | 4m 16s | | trunk passed | | +1 :green_heart: | javadoc | 3m 7s | | trunk passed with JDK Ubuntu-11.0.11+9-Ubuntu-0ubuntu2.20.04 | | +1 :green_heart: | javadoc | 3m 22s | | trunk passed with JDK Private Build-1.8.0_292-8u292-b10-0ubuntu1~20.04-b10 | | +1 :green_heart: | spotbugs | 6m 50s | | trunk passed | | +1 :green_heart: | shadedclient | 27m 13s | | branch has no errors when building and testing our client artifacts. | _ Patch Compile Tests _ | | +0 :ok: | mvndep | 0m 26s | | Maven dependency ordering for patch | | +1 :green_heart: | mvninstall | 3m 21s | | the patch passed | | +1 :green_heart: | compile | 29m 29s | | the patch passed with JDK Ubuntu-11.0.11+9-Ubuntu-0ubuntu2.20.04 | | +1 :green_heart: | javac | 29m 29s | | the patch passed | | +1 :green_heart: | compile | 25m 28s | | the patch passed with JDK Private Build-1.8.0_292-8u292-b10-0ubuntu1~20.04-b10 | | +1 :green_heart: | javac | 25m 28s | | the patch passed | | +1 :green_heart: | blanks | 0m 0s | | The patch has no blanks issues. | | -0 :warning: | checkstyle | 4m 20s | [/results-checkstyle-root.txt](https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-3611/1/artifact/out/results-checkstyle-root.txt) | root: The patch generated 14 new + 40 unchanged - 0 fixed = 54 total (was 40) | | +1 :green_heart: | mvnsite | 4m 24s | | the patch passed | | +1 :green_heart: | javadoc | 2m 47s | | the patch passed with JDK Ubuntu-11.0.11+9-Ubuntu-0ubuntu2.20.04 | | +1 :green_heart: | javadoc | 3m 15s | | the patch passed with JDK Private Build-1.8.0_292-8u292-b10-0ubuntu1~20.04-b10 | | +1 :green_heart: | spotbugs | 6m 35s | | the patch passed | | +1 :green_heart: | shadedclient | 23m 34s | | patch has no errors when building and testing our client artifacts. | _ Other Tests _ | | +1 :green_heart: | unit | 17m 8s | | hadoop-common in the patch passed. | | +1 :green_heart: | unit | 5m 58s | | hadoop-mapreduce-client-core in the patch passed. | | +1 :green_heart: | unit | 139m 4s | | hadoop-mapreduce-client-jobclient in the patch passed. | | +1 :green_heart: | unit | 2m 13s | | hadoop-azure in the patch passed. | | +1 :green_heart: | asflicense | 0m 53s | | The patch does not generate ASF License warnings. | | | | 418m 20s | | | | Subsystem | Report/Notes | |--:|:-| | Docker | ClientAPI=1.41 ServerAPI=1.41 base: https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-3611/1/artifact/out/Dockerfile | | GITHUB PR | https://github.com/apache/hadoop/pull/3611 | | Optional Tests | dupname asflicense compile javac javadoc mvninstall mvnsite unit shadedclient spotbugs checkstyle codespell markdownlint | | uname | Linux ec598da7f750 4.15.0-147-generic #151-Ubuntu SMP Fri Jun 18 19:21:19 UTC 2021 x86_64 x86_64 x86_64 GNU/Linux | | Build tool | maven | | Personality | dev-support/bin/hadoop.sh | | git revision | trunk / cf88c3518ce9386aeb89c7fcfe62586669b1e429 | | Default Java | Private Build-1.8.0_292-8u292-b10-0ubuntu1~20.04-b10 | | Multi-JDK versions | /usr/lib/jvm/java-11-openjdk-amd64:Ubuntu-11.0.11+9-Ubuntu-0ubuntu2.20.04 /usr/lib/jvm/java-8-openjdk-amd64:Private Build-1.8.0_292-8u292-b10-0ubuntu1~20.04-b10 | | Test Results | https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-3611/1/testReport/ | | Max. process+thread count | 1375 (vs. ulimit of 5500) |
[GitHub] [hadoop] hadoop-yetus commented on pull request #3614: YARN-10999. Make NodeQueueLoadMonitor pluggable in ResourceManager
hadoop-yetus commented on pull request #3614: URL: https://github.com/apache/hadoop/pull/3614#issuecomment-958944662 :broken_heart: **-1 overall** | Vote | Subsystem | Runtime | Logfile | Comment | |::|--:|:|::|:---:| | +0 :ok: | reexec | 0m 56s | | Docker mode activated. | _ Prechecks _ | | +1 :green_heart: | dupname | 0m 1s | | No case conflicting files found. | | +0 :ok: | codespell | 0m 1s | | codespell was not available. | | +1 :green_heart: | @author | 0m 0s | | The patch does not contain any @author tags. | | +1 :green_heart: | test4tests | 0m 0s | | The patch appears to include 2 new or modified test files. | _ trunk Compile Tests _ | | +0 :ok: | mvndep | 12m 51s | | Maven dependency ordering for branch | | +1 :green_heart: | mvninstall | 24m 16s | | trunk passed | | +1 :green_heart: | compile | 10m 26s | | trunk passed with JDK Ubuntu-11.0.11+9-Ubuntu-0ubuntu2.20.04 | | +1 :green_heart: | compile | 8m 35s | | trunk passed with JDK Private Build-1.8.0_292-8u292-b10-0ubuntu1~20.04-b10 | | +1 :green_heart: | checkstyle | 1m 42s | | trunk passed | | +1 :green_heart: | mvnsite | 2m 2s | | trunk passed | | +1 :green_heart: | javadoc | 1m 40s | | trunk passed with JDK Ubuntu-11.0.11+9-Ubuntu-0ubuntu2.20.04 | | +1 :green_heart: | javadoc | 1m 31s | | trunk passed with JDK Private Build-1.8.0_292-8u292-b10-0ubuntu1~20.04-b10 | | +1 :green_heart: | spotbugs | 4m 8s | | trunk passed | | +1 :green_heart: | shadedclient | 24m 12s | | branch has no errors when building and testing our client artifacts. | _ Patch Compile Tests _ | | +0 :ok: | mvndep | 0m 24s | | Maven dependency ordering for patch | | +1 :green_heart: | mvninstall | 1m 30s | | the patch passed | | +1 :green_heart: | compile | 9m 37s | | the patch passed with JDK Ubuntu-11.0.11+9-Ubuntu-0ubuntu2.20.04 | | +1 :green_heart: | javac | 9m 37s | | the patch passed | | +1 :green_heart: | compile | 8m 40s | | the patch passed with JDK Private Build-1.8.0_292-8u292-b10-0ubuntu1~20.04-b10 | | +1 :green_heart: | javac | 8m 40s | | the patch passed | | +1 :green_heart: | blanks | 0m 0s | | The patch has no blanks issues. | | -0 :warning: | checkstyle | 1m 35s | [/results-checkstyle-hadoop-yarn-project_hadoop-yarn.txt](https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-3614/1/artifact/out/results-checkstyle-hadoop-yarn-project_hadoop-yarn.txt) | hadoop-yarn-project/hadoop-yarn: The patch generated 2 new + 179 unchanged - 0 fixed = 181 total (was 179) | | +1 :green_heart: | mvnsite | 1m 54s | | the patch passed | | +1 :green_heart: | javadoc | 1m 31s | | the patch passed with JDK Ubuntu-11.0.11+9-Ubuntu-0ubuntu2.20.04 | | -1 :x: | javadoc | 0m 47s | [/results-javadoc-javadoc-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-server_hadoop-yarn-server-resourcemanager-jdkPrivateBuild-1.8.0_292-8u292-b10-0ubuntu1~20.04-b10.txt](https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-3614/1/artifact/out/results-javadoc-javadoc-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-server_hadoop-yarn-server-resourcemanager-jdkPrivateBuild-1.8.0_292-8u292-b10-0ubuntu1~20.04-b10.txt) | hadoop-yarn-project_hadoop-yarn_hadoop-yarn-server_hadoop-yarn-server-resourcemanager-jdkPrivateBuild-1.8.0_292-8u292-b10-0ubuntu1~20.04-b10 with JDK Private Build-1.8.0_292-8u292-b10-0ubuntu1~20.04-b10 generated 1 new + 343 unchanged - 1 fixed = 344 total (was 344) | | +1 :green_heart: | spotbugs | 4m 21s | | the patch passed | | +1 :green_heart: | shadedclient | 23m 53s | | patch has no errors when building and testing our client artifacts. | _ Other Tests _ | | +1 :green_heart: | unit | 1m 0s | | hadoop-yarn-api in the patch passed. | | +1 :green_heart: | unit | 99m 19s | | hadoop-yarn-server-resourcemanager in the patch passed. | | +1 :green_heart: | asflicense | 0m 45s | | The patch does not generate ASF License warnings. | | | | 248m 47s | | | | Subsystem | Report/Notes | |--:|:-| | Docker | ClientAPI=1.41 ServerAPI=1.41 base: https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-3614/1/artifact/out/Dockerfile | | GITHUB PR | https://github.com/apache/hadoop/pull/3614 | | Optional Tests | dupname asflicense compile javac javadoc mvninstall mvnsite unit shadedclient spotbugs checkstyle codespell | | uname | Linux ac7a28954f73 4.15.0-147-generic #151-Ubuntu SMP Fri Jun 18 19:21:19 UTC 2021 x86_64 x86_64 x86_64 GNU/Linux | | Build tool | maven | | Personality | dev-support/bin/hadoop.sh | | git revision | trunk / 88b64231ca176da9f1902e752ddcde4955b216e6 | | Default Java | Private Build-1
[GitHub] [hadoop] mariosmeim-db commented on a change in pull request #3440: ABFS: Support for Encryption Context
mariosmeim-db commented on a change in pull request #3440: URL: https://github.com/apache/hadoop/pull/3440#discussion_r741897486 ## File path: hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/security/EncryptionAdapter.java ## @@ -0,0 +1,145 @@ +/** + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +package org.apache.hadoop.fs.azurebfs.security; + +import java.io.IOException; +import java.nio.charset.StandardCharsets; +import java.security.MessageDigest; +import java.security.NoSuchAlgorithmException; +import java.util.Arrays; +import java.util.Base64; +import javax.crypto.SecretKey; +import javax.security.auth.DestroyFailedException; +import javax.security.auth.Destroyable; + +import org.apache.hadoop.thirdparty.com.google.common.base.Preconditions; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; + +import org.apache.hadoop.fs.azurebfs.extensions.EncryptionContextProvider; + +public class EncryptionAdapter implements Destroyable { + private final String path; + private SecretKey encryptionContext; + private SecretKey encryptionKey; + private final EncryptionContextProvider provider; + private String encodedKey = null; + private String encodedKeySHA = null; + private static final Logger LOG = + LoggerFactory.getLogger(EncryptionAdapter.class); + + public EncryptionAdapter(EncryptionContextProvider provider, String path, + byte[] encryptionContext) throws IOException { +this(provider, path); +Preconditions.checkNotNull(encryptionContext, +"Encryption context should not be null."); +this.encryptionContext = new ABFSSecretKey(encryptionContext); + } + + public EncryptionAdapter(EncryptionContextProvider provider, String path) + throws IOException { +this.provider = provider; +this.path = path; + } + + public SecretKey getEncryptionKey() throws IOException { +if (encryptionKey != null) { + return encryptionKey; +} +encryptionKey = provider.getEncryptionKey(path, encryptionContext); +return encryptionKey; + } + + public SecretKey fetchEncryptionContextAndComputeKeys() throws IOException { Review comment: Perhaps it should be more explicit in the naming that this is used to create a new encryption context. ## File path: hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/services/AbfsClient.java ## @@ -221,16 +223,50 @@ SharedKeyCredentials getSharedKeyCredentials() { return requestHeaders; } - private void addCustomerProvidedKeyHeaders( - final List requestHeaders) { -if (clientProvidedEncryptionKey != null) { - requestHeaders.add( - new AbfsHttpHeader(X_MS_ENCRYPTION_KEY, clientProvidedEncryptionKey)); - requestHeaders.add(new AbfsHttpHeader(X_MS_ENCRYPTION_KEY_SHA256, - clientProvidedEncryptionKeySHA)); - requestHeaders.add(new AbfsHttpHeader(X_MS_ENCRYPTION_ALGORITHM, - SERVER_SIDE_ENCRYPTION_ALGORITHM)); + private void addEncryptionKeyRequestHeaders(String path, + List requestHeaders, boolean isCreateFileRequest, + EncryptionAdapter encryptionAdapter, TracingContext tracingContext) + throws IOException { +String encodedKey, encodedKeySHA256; +switch (encryptionType) { +case GLOBAL_KEY: + encodedKey = clientProvidedEncryptionKey; + encodedKeySHA256 = clientProvidedEncryptionKeySHA; + break; + +case ENCRYPTION_CONTEXT: + if (isCreateFileRequest) { +// get new context for create file request +SecretKey encryptionContext = +encryptionAdapter.fetchEncryptionContextAndComputeKeys(); +requestHeaders.add(new AbfsHttpHeader(X_MS_ENCRYPTION_CONTEXT, +new String(encryptionContext.getEncoded(), +StandardCharsets.UTF_8))); +try { + encryptionContext.destroy(); +} catch (DestroyFailedException e) { + throw new IOException( + "Could not destroy encryptionContext: " + e.getMessage()); +} + } else if (encryptionAdapter == null) { +// get encryption context from GetPathStatus response header +encryptionAdapter = new EncryptionAdapter(encryption
[jira] [Work logged] (HADOOP-17975) Fallback to simple auth does not work for a secondary DistributedFileSystem instance
[ https://issues.apache.org/jira/browse/HADOOP-17975?focusedWorklogId=675743&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-675743 ] ASF GitHub Bot logged work on HADOOP-17975: --- Author: ASF GitHub Bot Created on: 04/Nov/21 01:24 Start Date: 04/Nov/21 01:24 Worklog Time Spent: 10m Work Description: symious commented on pull request #3579: URL: https://github.com/apache/hadoop/pull/3579#issuecomment-958605853 @fapifta Thanks for the explanation. Is the current situation as follows? Correct me if I'm wrong. Server S has two clients, client A of non-secure settings with "fallbackToSimpleAuth = false", client B of secure settings with "fallbackToSimpleAuth=true", client A and B both connecting to Datanode D, since client A first connect to Datanode D, so the connection is created in _connections_, and the fallbackToSimpleAuth is set to false. Then client B tries to connect Datanode D, since the _ConnectionId_ is the same, it will reuse the connection created by Client A, then in _setupIOStream_, since the socket is created, Client will skip the other process, and the "fallbackToSimpleAuth" won't be changed to true. If the issue is caused by the above situation, I think if we distinguish the connections of Client A and Client B will solve the issue? Currently the ConnectionId of Client A and Client B is same, if we add "fallbackToSimpleAuth" to ConnectionId, since Client A has "false" and Client B has "true", they will generate different ConnectionId, then different connections will be created, so they won't affect each other. -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org Issue Time Tracking --- Worklog Id: (was: 675743) Time Spent: 3h 10m (was: 3h) > Fallback to simple auth does not work for a secondary DistributedFileSystem > instance > > > Key: HADOOP-17975 > URL: https://issues.apache.org/jira/browse/HADOOP-17975 > Project: Hadoop Common > Issue Type: Bug > Components: ipc >Reporter: István Fajth >Assignee: István Fajth >Priority: Major > Labels: pull-request-available > Time Spent: 3h 10m > Remaining Estimate: 0h > > The following code snippet demonstrates what is necessary to cause a failure > in connection to a non secure cluster with fallback to SIMPLE auth allowed > from a secure cluster. > {code:java} > Configuration conf = new Configuration(); > conf.setBoolean("ipc.client.fallback-to-simple-auth-allowed", true); > URI fsUri = new URI("hdfs://"); > conf.setBoolean("fs.hdfs.impl.disable.cache", true); > FileSystem fs = FileSystem.get(fsUri, conf); > FSDataInputStream src = fs.open(new Path("/path/to/a/file")); > FileOutputStream dst = new FileOutputStream(File.createTempFile("foo", > "bar")); > IOUtils.copyBytes(src, dst, 1024); > // The issue happens even if we re-enable cache at this point > //conf.setBoolean("fs.hdfs.impl.disable.cache", false); > // The issue does not happen when we close the first FileSystem object > // before creating the second. > //fs.close(); > FileSystem fs2 = FileSystem.get(fsUri, conf); > FSDataInputStream src2 = fs2.open(new Path("/path/to/a/file")); > FileOutputStream dst2 = new FileOutputStream(File.createTempFile("foo", > "bar")); > IOUtils.copyBytes(src2, dst2, 1024); > {code} > The problem is that when the DfsClient is created it creates an instance of > AtomicBoolean, which is propagated down into the IPC layer, where the > Client.Connection instance in setupIOStreams sets its value. This connection > object is cached and re-used to multiplex requests against the same DataNode. > In case of creating a second DfsClient, the AtomicBoolean reference in the > client is a new AtomicBoolean, but the Client.Connection instance is the > same, and as it has a socket already open to the DataNode, it returns > immediatelly from setupIOStreams, leaving the fallbackToSimpleAuth > AtomicBoolean false as it is created in the DfsClient. > This AtomicBoolean on the other hand controls how the SaslDataTransferClient > handles the connection in the above level, and with this value left on the > default false, the SaslDataTransferClient of the second DfsClient will not > fall back to SIMPLE authentication but will try to send a SASL handshake when > connecting to the DataNode. > > The access to the FileSystem via the
[GitHub] [hadoop] sodonnel commented on pull request #3593: HDFS-16286. Add a debug tool to verify the correctness of erasure coding on file
sodonnel commented on pull request #3593: URL: https://github.com/apache/hadoop/pull/3593#issuecomment-958887599 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[GitHub] [hadoop] symious commented on pull request #3579: HADOOP-17975 Fallback to simple auth does not work for a secondary DistributedFileSystem instance.
symious commented on pull request #3579: URL: https://github.com/apache/hadoop/pull/3579#issuecomment-958605853 @fapifta Thanks for the explanation. Is the current situation as follows? Correct me if I'm wrong. Server S has two clients, client A of non-secure settings with "fallbackToSimpleAuth = false", client B of secure settings with "fallbackToSimpleAuth=true", client A and B both connecting to Datanode D, since client A first connect to Datanode D, so the connection is created in _connections_, and the fallbackToSimpleAuth is set to false. Then client B tries to connect Datanode D, since the _ConnectionId_ is the same, it will reuse the connection created by Client A, then in _setupIOStream_, since the socket is created, Client will skip the other process, and the "fallbackToSimpleAuth" won't be changed to true. If the issue is caused by the above situation, I think if we distinguish the connections of Client A and Client B will solve the issue? Currently the ConnectionId of Client A and Client B is same, if we add "fallbackToSimpleAuth" to ConnectionId, since Client A has "false" and Client B has "true", they will generate different ConnectionId, then different connections will be created, so they won't affect each other. -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[GitHub] [hadoop] hadoop-yetus commented on pull request #3613: HDFS-16296. RouterRpcFairnessPolicyController add rejected permits for each nameservice
hadoop-yetus commented on pull request #3613: URL: https://github.com/apache/hadoop/pull/3613#issuecomment-958690121 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Work logged] (HADOOP-17981) Support etag-assisted renames in FileOutputCommitter
[ https://issues.apache.org/jira/browse/HADOOP-17981?focusedWorklogId=675735&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-675735 ] ASF GitHub Bot logged work on HADOOP-17981: --- Author: ASF GitHub Bot Created on: 04/Nov/21 01:23 Start Date: 04/Nov/21 01:23 Worklog Time Spent: 10m Work Description: hadoop-yetus commented on pull request #3611: URL: https://github.com/apache/hadoop/pull/3611#issuecomment-959854489 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org Issue Time Tracking --- Worklog Id: (was: 675735) Time Spent: 7.5h (was: 7h 20m) > Support etag-assisted renames in FileOutputCommitter > > > Key: HADOOP-17981 > URL: https://issues.apache.org/jira/browse/HADOOP-17981 > Project: Hadoop Common > Issue Type: New Feature > Components: fs, fs/azure >Affects Versions: 3.4.0 >Reporter: Steve Loughran >Assignee: Steve Loughran >Priority: Major > Labels: pull-request-available > Time Spent: 7.5h > Remaining Estimate: 0h > > To deal with some throttling/retry issues in object stores, > pass the FileStatus entries retrieved during listing > into a private interface ResilientCommitByRename which filesystems > may implement to use extra attributes in the listing (etag, version) > to constrain and validate the operation. > Although targeting azure, GCS and others could use. no point in S3A as they > shouldn't use this committer. > # And we are not going to do any changes to FileSystem as there are explicit > guarantees of public use and stability. > I am not going to make that change as the hive thing that will suddenly start > expecting it to work forever. > # I'm not planning to merge this in, as the manifest committer is going to > include this and more (MAPREDUCE-7341) > However, I do need to get this in on a branch, so am doing this work on trunk > for dev & test and for others to review -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[GitHub] [hadoop] hadoop-yetus commented on pull request #3611: HADOOP-17981. resilient commit through etag validation
hadoop-yetus commented on pull request #3611: URL: https://github.com/apache/hadoop/pull/3611#issuecomment-959854489 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[GitHub] [hadoop] tomscut commented on pull request #3538: HDFS-16266. Add remote port information to HDFS audit log
tomscut commented on pull request #3538: URL: https://github.com/apache/hadoop/pull/3538#issuecomment-958648376 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[GitHub] [hadoop] cndaimin commented on pull request #3593: HDFS-16286. Add a debug tool to verify the correctness of erasure coding on file
cndaimin commented on pull request #3593: URL: https://github.com/apache/hadoop/pull/3593#issuecomment-958610440 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[GitHub] [hadoop] hadoop-yetus commented on pull request #3603: [YARN-10998] Add YARN_ROUTER_HEAPSIZE to yarn-env for routers
hadoop-yetus commented on pull request #3603: URL: https://github.com/apache/hadoop/pull/3603#issuecomment-958752318 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Work logged] (HADOOP-17124) Support LZO using aircompressor
[ https://issues.apache.org/jira/browse/HADOOP-17124?focusedWorklogId=675711&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-675711 ] ASF GitHub Bot logged work on HADOOP-17124: --- Author: ASF GitHub Bot Created on: 04/Nov/21 01:20 Start Date: 04/Nov/21 01:20 Worklog Time Spent: 10m Work Description: viirya commented on a change in pull request #3612: URL: https://github.com/apache/hadoop/pull/3612#discussion_r741581302 ## File path: hadoop-common-project/hadoop-common/pom.xml ## @@ -375,6 +375,17 @@ lz4-java provided + + io.airlift + aircompressor + provided + + + com.hadoop.gplcompression + hadoop-lzo + 0.4.21-SNAPSHOT + test + Review comment: Put it here for verifying the new Lzo codec only. We will remove it before merging. ## File path: hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/compress/LzoCodec2.java ## @@ -0,0 +1,186 @@ +/* + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +package org.apache.hadoop.io.compress; + +import java.io.IOException; +import java.io.InputStream; +import java.io.OutputStream; + +import org.apache.hadoop.conf.Configurable; +import org.apache.hadoop.conf.Configuration; +import org.apache.hadoop.fs.CommonConfigurationKeys; +import org.apache.hadoop.io.compress.lzo.LzoCompressor; +import org.apache.hadoop.io.compress.lzo.LzoDecompressor; + +/** + * This class creates lzo compressors/decompressors. + */ +public class LzoCodec2 implements Configurable, CompressionCodec { Review comment: We will rename to `LzoCodec` before merging. ## File path: hadoop-project/pom.xml ## @@ -2457,5 +2463,9 @@ + + twitter + https://maven.twttr.com/ + Review comment: We will remove this before merging. ## File path: hadoop-common-project/hadoop-common/pom.xml ## @@ -375,6 +375,17 @@ lz4-java provided + + io.airlift + aircompressor + provided + + + com.hadoop.gplcompression + hadoop-lzo + 0.4.21-SNAPSHOT + test + Review comment: Put it here for verifying the new Lzo codec only. We will remove it before merging. ## File path: hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/compress/LzoCodec2.java ## @@ -0,0 +1,186 @@ +/* + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +package org.apache.hadoop.io.compress; + +import java.io.IOException; +import java.io.InputStream; +import java.io.OutputStream; + +import org.apache.hadoop.conf.Configurable; +import org.apache.hadoop.conf.Configuration; +import org.apache.hadoop.fs.CommonConfigurationKeys; +import org.apache.hadoop.io.compress.lzo.LzoCompressor; +import org.apache.hadoop.io.compress.lzo.LzoDecompressor; + +/** + * This class creates lzo compressors/decompressors. + */ +public class LzoCodec2 implements Configurable, CompressionCodec { Review comment: We will rename to `LzoCodec` before merging. ## File path: hadoop-project/pom.xml ## @@ -2457,5 +2463,9 @@ + + twitter + https://maven.twttr.com/ + Review comment: We will remove this before merging. -- This is an automated message from the Apache Git Service
[GitHub] [hadoop] viirya commented on a change in pull request #3612: WIP. HADOOP-17124. Support LZO Codec using aircompressor
viirya commented on a change in pull request #3612: URL: https://github.com/apache/hadoop/pull/3612#discussion_r741581302 ## File path: hadoop-common-project/hadoop-common/pom.xml ## @@ -375,6 +375,17 @@ lz4-java provided + + io.airlift + aircompressor + provided + + + com.hadoop.gplcompression + hadoop-lzo + 0.4.21-SNAPSHOT + test + Review comment: Put it here for verifying the new Lzo codec only. We will remove it before merging. ## File path: hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/compress/LzoCodec2.java ## @@ -0,0 +1,186 @@ +/* + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +package org.apache.hadoop.io.compress; + +import java.io.IOException; +import java.io.InputStream; +import java.io.OutputStream; + +import org.apache.hadoop.conf.Configurable; +import org.apache.hadoop.conf.Configuration; +import org.apache.hadoop.fs.CommonConfigurationKeys; +import org.apache.hadoop.io.compress.lzo.LzoCompressor; +import org.apache.hadoop.io.compress.lzo.LzoDecompressor; + +/** + * This class creates lzo compressors/decompressors. + */ +public class LzoCodec2 implements Configurable, CompressionCodec { Review comment: We will rename to `LzoCodec` before merging. ## File path: hadoop-project/pom.xml ## @@ -2457,5 +2463,9 @@ + + twitter + https://maven.twttr.com/ + Review comment: We will remove this before merging. ## File path: hadoop-common-project/hadoop-common/pom.xml ## @@ -375,6 +375,17 @@ lz4-java provided + + io.airlift + aircompressor + provided + + + com.hadoop.gplcompression + hadoop-lzo + 0.4.21-SNAPSHOT + test + Review comment: Put it here for verifying the new Lzo codec only. We will remove it before merging. ## File path: hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/compress/LzoCodec2.java ## @@ -0,0 +1,186 @@ +/* + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +package org.apache.hadoop.io.compress; + +import java.io.IOException; +import java.io.InputStream; +import java.io.OutputStream; + +import org.apache.hadoop.conf.Configurable; +import org.apache.hadoop.conf.Configuration; +import org.apache.hadoop.fs.CommonConfigurationKeys; +import org.apache.hadoop.io.compress.lzo.LzoCompressor; +import org.apache.hadoop.io.compress.lzo.LzoDecompressor; + +/** + * This class creates lzo compressors/decompressors. + */ +public class LzoCodec2 implements Configurable, CompressionCodec { Review comment: We will rename to `LzoCodec` before merging. ## File path: hadoop-project/pom.xml ## @@ -2457,5 +2463,9 @@ + + twitter + https://maven.twttr.com/ + Review comment: We will remove this before merging. -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-
[GitHub] [hadoop] ferhui commented on a change in pull request #3602: HDFS-16291.Make the comment of INode#ReclaimContext more standardized.
ferhui commented on a change in pull request #3602: URL: https://github.com/apache/hadoop/pull/3602#discussion_r741639733 ## File path: hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/INode.java ## @@ -993,15 +993,13 @@ public long getNsDelta() { private final QuotaDelta quotaDelta; /** - * @param bsps - * block storage policy suite to calculate intended storage type Review comment: How about just adding the same blanks on line 997 and below wrong format lines as line 996, and do not change other lines. ## File path: hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/INode.java ## @@ -993,15 +993,13 @@ public long getNsDelta() { private final QuotaDelta quotaDelta; /** - * @param bsps - * block storage policy suite to calculate intended storage type Review comment: How about just adding the same blanks on line 997 and below wrong format lines as line 996, and do not change other lines. -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[GitHub] [hadoop] haiyang1987 removed a comment on pull request #3596: HDFS-16287. Support to make dfs.namenode.avoid.read.slow.datanode reconfigurable
haiyang1987 removed a comment on pull request #3596: URL: https://github.com/apache/hadoop/pull/3596#issuecomment-958731868 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[GitHub] [hadoop] symious commented on pull request #3613: HDFS-16296. RouterRpcFairnessPolicyController add rejected permits for each nameservice
symious commented on pull request #3613: URL: https://github.com/apache/hadoop/pull/3613#issuecomment-958652432 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Work logged] (HADOOP-17124) Support LZO using aircompressor
[ https://issues.apache.org/jira/browse/HADOOP-17124?focusedWorklogId=675678&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-675678 ] ASF GitHub Bot logged work on HADOOP-17124: --- Author: ASF GitHub Bot Created on: 04/Nov/21 01:16 Start Date: 04/Nov/21 01:16 Worklog Time Spent: 10m Work Description: viirya opened a new pull request #3612: URL: https://github.com/apache/hadoop/pull/3612 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org Issue Time Tracking --- Worklog Id: (was: 675678) Time Spent: 5h 50m (was: 5h 40m) > Support LZO using aircompressor > --- > > Key: HADOOP-17124 > URL: https://issues.apache.org/jira/browse/HADOOP-17124 > Project: Hadoop Common > Issue Type: New Feature > Components: common >Affects Versions: 3.3.0 >Reporter: DB Tsai >Priority: Major > Labels: pull-request-available > Time Spent: 5h 50m > Remaining Estimate: 0h > > LZO codec was removed in HADOOP-4874 because the original LZO binding is GPL > which is problematic. However, many legacy data is still compressed by LZO > codec, and companies often use vendor's GPL LZO codec in the classpath which > might cause GPL contamination. > Presro and ORC-77 use [aircompressor| > [https://github.com/airlift/aircompressor]] (Apache V2 licensed) to compress > and decompress LZO data. Hadoop can add back LZO support using aircompressor > without GPL violation. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[GitHub] [hadoop] haiyang1987 commented on pull request #3596: HDFS-16287. Support to make dfs.namenode.avoid.read.slow.datanode reconfigurable
haiyang1987 commented on pull request #3596: URL: https://github.com/apache/hadoop/pull/3596#issuecomment-958731868 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[GitHub] [hadoop] jianghuazhu commented on a change in pull request #3602: HDFS-16291.Make the comment of INode#ReclaimContext more standardized.
jianghuazhu commented on a change in pull request #3602: URL: https://github.com/apache/hadoop/pull/3602#discussion_r741662220 ## File path: hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/INode.java ## @@ -993,15 +993,13 @@ public long getNsDelta() { private final QuotaDelta quotaDelta; /** - * @param bsps - * block storage policy suite to calculate intended storage type Review comment: Thanks @ferhui for the comment and review. I will update it later. The new style will look like this: ![image](https://user-images.githubusercontent.com/6416939/140021341-f0016022-613a-4de2-a658-a63c336b2b14.png) -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Work logged] (HADOOP-17981) Support etag-assisted renames in FileOutputCommitter
[ https://issues.apache.org/jira/browse/HADOOP-17981?focusedWorklogId=675640&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-675640 ] ASF GitHub Bot logged work on HADOOP-17981: --- Author: ASF GitHub Bot Created on: 04/Nov/21 01:11 Start Date: 04/Nov/21 01:11 Worklog Time Spent: 10m Work Description: sidseth commented on pull request #3597: URL: https://github.com/apache/hadoop/pull/3597#issuecomment-958658416 > > This mechanism becomes very FileSystem specific. Implemented by Azure right now. > > I agree, which is why the API is restricted for its uses to mr-client-core only. as abfs is the only one which needs it for correctness under load, And I'm not worried about that specifity. Can I point to how much of the hadoop fs api are hdfs-only -and they are public. > > > Other users of rename will not see the benefits without changing interfaces, which in turn requires shimming etc. > > Please don't try and use this particular interface in Hive. > Was referring to any potential usage - including Hive. > > Would it be better for AzureFileSystem rename itself to add a config parameter which can lookup the src etag (at the cost of a performance hit for consistency), so that downstream components / any users of the rename operation can benefit from this change without having to change interfaces. > > We are going straight from a listing (1 request/500 entries) to that rename. doing a HEAD first cuts the throughtput in half. so no. > In the scenario where this is encountered. Would not be the default behaviour, and limits the change to Abfs. Could also have the less consistent version which is not etag based, and responds only on failures. Again - limited to Abfs. > > Also, if the performance penalty is a big problem - Abfs could create very short-lived caches for FileStatus objects, and handle errors on discrepancies with the cached copy. > > Possible but convoluted. > Agree. Quite convoluted. Tossing in potential options - to avoid a new public API. > > Essentially - don't force usage of the new interface to get the benefits. > > I understand the interests of the hive team, but this fix is not the place to do a better API. > > Briefly cacheing the source FS entries is something to consider though. Not this week. > > What I could do with is some help getting #2735 in, then we can start on a public rename() builder API which will take a file status, as openFile does. > This particular change would be FSImpl agnostic, and potentially remove the need for the new interface here? > > Side note: The fs.getStatus within ResilientCommitByRenameHelper for FileSystems where this new functionality is not supported will lead to a performance penalty for the other FileSystems (performing a getFileStatus on src). > > There is an option to say "i know it is not there"; this skips the check. the committer passes this option down because it issues a delete call first. > EOD - this ends up being a new API (almost on the FileSystem), which is used by the committer first; then someone discovers it and decides to make use of it. > FWIW the manifest committer will make that pre-rename commit optional, saving that IO request. I am curious as to how well that will work I went executed on well formed tables. -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org Issue Time Tracking --- Worklog Id: (was: 675640) Time Spent: 7h 20m (was: 7h 10m) > Support etag-assisted renames in FileOutputCommitter > > > Key: HADOOP-17981 > URL: https://issues.apache.org/jira/browse/HADOOP-17981 > Project: Hadoop Common > Issue Type: New Feature > Components: fs, fs/azure >Affects Versions: 3.4.0 >Reporter: Steve Loughran >Assignee: Steve Loughran >Priority: Major > Labels: pull-request-available > Time Spent: 7h 20m > Remaining Estimate: 0h > > To deal with some throttling/retry issues in object stores, > pass the FileStatus entries retrieved during listing > into a private interface ResilientCommitByRename which filesystems > may implement to use extra attributes in the listing (etag, version) > to constrain and validate the operation. > Although targeting azure, GCS and others could use. no point in S3A as they > shouldn't use this committer. > # And we are not going to do any changes to FileSyste
[GitHub] [hadoop] sidseth commented on pull request #3597: HADOOP-17981 Support etag-assisted renames in FileOutputCommitter
sidseth commented on pull request #3597: URL: https://github.com/apache/hadoop/pull/3597#issuecomment-958658416 > > This mechanism becomes very FileSystem specific. Implemented by Azure right now. > > I agree, which is why the API is restricted for its uses to mr-client-core only. as abfs is the only one which needs it for correctness under load, And I'm not worried about that specifity. Can I point to how much of the hadoop fs api are hdfs-only -and they are public. > > > Other users of rename will not see the benefits without changing interfaces, which in turn requires shimming etc. > > Please don't try and use this particular interface in Hive. > Was referring to any potential usage - including Hive. > > Would it be better for AzureFileSystem rename itself to add a config parameter which can lookup the src etag (at the cost of a performance hit for consistency), so that downstream components / any users of the rename operation can benefit from this change without having to change interfaces. > > We are going straight from a listing (1 request/500 entries) to that rename. doing a HEAD first cuts the throughtput in half. so no. > In the scenario where this is encountered. Would not be the default behaviour, and limits the change to Abfs. Could also have the less consistent version which is not etag based, and responds only on failures. Again - limited to Abfs. > > Also, if the performance penalty is a big problem - Abfs could create very short-lived caches for FileStatus objects, and handle errors on discrepancies with the cached copy. > > Possible but convoluted. > Agree. Quite convoluted. Tossing in potential options - to avoid a new public API. > > Essentially - don't force usage of the new interface to get the benefits. > > I understand the interests of the hive team, but this fix is not the place to do a better API. > > Briefly cacheing the source FS entries is something to consider though. Not this week. > > What I could do with is some help getting #2735 in, then we can start on a public rename() builder API which will take a file status, as openFile does. > This particular change would be FSImpl agnostic, and potentially remove the need for the new interface here? > > Side note: The fs.getStatus within ResilientCommitByRenameHelper for FileSystems where this new functionality is not supported will lead to a performance penalty for the other FileSystems (performing a getFileStatus on src). > > There is an option to say "i know it is not there"; this skips the check. the committer passes this option down because it issues a delete call first. > EOD - this ends up being a new API (almost on the FileSystem), which is used by the committer first; then someone discovers it and decides to make use of it. > FWIW the manifest committer will make that pre-rename commit optional, saving that IO request. I am curious as to how well that will work I went executed on well formed tables. -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[GitHub] [hadoop] haiyang1987 commented on a change in pull request #3596: HDFS-16287. Support to make dfs.namenode.avoid.read.slow.datanode reconfigurable
haiyang1987 commented on a change in pull request #3596: URL: https://github.com/apache/hadoop/pull/3596#discussion_r741696525 ## File path: hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/DatanodeManager.java ## @@ -260,17 +257,14 @@ final Timer timer = new Timer(); this.slowPeerTracker = dataNodePeerStatsEnabled ? new SlowPeerTracker(conf, timer) : null; -this.excludeSlowNodesEnabled = conf.getBoolean( -DFS_NAMENODE_BLOCKPLACEMENTPOLICY_EXCLUDE_SLOW_NODES_ENABLED_KEY, -DFS_NAMENODE_BLOCKPLACEMENTPOLICY_EXCLUDE_SLOW_NODES_ENABLED_DEFAULT); this.maxSlowPeerReportNodes = conf.getInt( DFSConfigKeys.DFS_NAMENODE_MAX_SLOWPEER_COLLECT_NODES_KEY, DFSConfigKeys.DFS_NAMENODE_MAX_SLOWPEER_COLLECT_NODES_DEFAULT); this.slowPeerCollectionInterval = conf.getTimeDuration( DFSConfigKeys.DFS_NAMENODE_SLOWPEER_COLLECT_INTERVAL_KEY, DFSConfigKeys.DFS_NAMENODE_SLOWPEER_COLLECT_INTERVAL_DEFAULT, TimeUnit.MILLISECONDS); -if (slowPeerTracker != null && excludeSlowNodesEnabled) { Review comment: @tomscut Thank you for your review. 1.Current parameter 'dataNodePeerStatsEnabled' and 'excludeSlowNodesEnabled' decision SlowPeerCollector thread whether to start ,But it didn't take into account avoid SlowDataNodesForRead logic 2.So think about two phases: a.The first is to start SlowPeerCollector thread b.Second, you can control whether to enable read/write avoid slow datanode according to dynamic parameters ## File path: hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/DatanodeManager.java ## @@ -511,7 +505,16 @@ private boolean isInactive(DatanodeInfo datanode) { private boolean isSlowNode(String dnUuid) { return avoidSlowDataNodesForRead && slowNodesUuidSet.contains(dnUuid); } - + + public void setAvoidSlowDataNodesForReadEnabled(boolean enable) { Review comment: Consider slowNodesUuidSet is generated when the SlowPeerCollector thread is started,therefore it is logical to judge Therefore, it is logical to judge whether the dnUuid exists in the slowNodesUuidSet? ## File path: hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/DatanodeManager.java ## @@ -511,7 +505,16 @@ private boolean isInactive(DatanodeInfo datanode) { private boolean isSlowNode(String dnUuid) { return avoidSlowDataNodesForRead && slowNodesUuidSet.contains(dnUuid); } - + + public void setAvoidSlowDataNodesForReadEnabled(boolean enable) { Review comment: Consider slowNodesUuidSet is generated when the SlowPeerCollector thread is started,therefore it is logical to judge whether the dnUuid exists in the slowNodesUuidSet? -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Work logged] (HADOOP-17124) Support LZO using aircompressor
[ https://issues.apache.org/jira/browse/HADOOP-17124?focusedWorklogId=675608&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-675608 ] ASF GitHub Bot logged work on HADOOP-17124: --- Author: ASF GitHub Bot Created on: 04/Nov/21 01:07 Start Date: 04/Nov/21 01:07 Worklog Time Spent: 10m Work Description: viirya commented on pull request #3612: URL: https://github.com/apache/hadoop/pull/3612#issuecomment-958606536 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org Issue Time Tracking --- Worklog Id: (was: 675608) Time Spent: 5h 40m (was: 5.5h) > Support LZO using aircompressor > --- > > Key: HADOOP-17124 > URL: https://issues.apache.org/jira/browse/HADOOP-17124 > Project: Hadoop Common > Issue Type: New Feature > Components: common >Affects Versions: 3.3.0 >Reporter: DB Tsai >Priority: Major > Labels: pull-request-available > Time Spent: 5h 40m > Remaining Estimate: 0h > > LZO codec was removed in HADOOP-4874 because the original LZO binding is GPL > which is problematic. However, many legacy data is still compressed by LZO > codec, and companies often use vendor's GPL LZO codec in the classpath which > might cause GPL contamination. > Presro and ORC-77 use [aircompressor| > [https://github.com/airlift/aircompressor]] (Apache V2 licensed) to compress > and decompress LZO data. Hadoop can add back LZO support using aircompressor > without GPL violation. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[GitHub] [hadoop] viirya commented on pull request #3612: WIP. HADOOP-17124. Support LZO Codec using aircompressor
viirya commented on pull request #3612: URL: https://github.com/apache/hadoop/pull/3612#issuecomment-958606536 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[GitHub] [hadoop] cndaimin commented on a change in pull request #3593: HDFS-16286. Add a debug tool to verify the correctness of erasure coding on file
cndaimin commented on a change in pull request #3593: URL: https://github.com/apache/hadoop/pull/3593#discussion_r741582271 ## File path: hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/tools/TestDebugAdmin.java ## @@ -166,8 +179,91 @@ public void testComputeMetaCommand() throws Exception { @Test(timeout = 6) public void testRecoverLeaseforFileNotFound() throws Exception { +cluster = new MiniDFSCluster.Builder(conf).numDataNodes(1).build(); +cluster.waitActive(); assertTrue(runCmd(new String[] { "recoverLease", "-path", "/foo", "-retries", "2" }).contains( "Giving up on recoverLease for /foo after 1 try")); } + + @Test(timeout = 6) + public void testVerifyECCommand() throws Exception { +final ErasureCodingPolicy ecPolicy = SystemErasureCodingPolicies.getByID( +SystemErasureCodingPolicies.RS_3_2_POLICY_ID); +cluster = DFSTestUtil.setupCluster(conf, 6, 5, 0); +cluster.waitActive(); +DistributedFileSystem fs = cluster.getFileSystem(); + +assertEquals("ret: 1, verifyEC -file Verify HDFS erasure coding on " + +"all block groups of the file.", runCmd(new String[]{"verifyEC"})); + +assertEquals("ret: 1, File /bar does not exist.", +runCmd(new String[]{"verifyEC", "-file", "/bar"})); + +fs.create(new Path("/bar")).close(); +assertEquals("ret: 1, File /bar is not erasure coded.", +runCmd(new String[]{"verifyEC", "-file", "/bar"})); + + +final Path ecDir = new Path("/ec"); +fs.mkdir(ecDir, FsPermission.getDirDefault()); +fs.enableErasureCodingPolicy(ecPolicy.getName()); +fs.setErasureCodingPolicy(ecDir, ecPolicy.getName()); + +assertEquals("ret: 1, File /ec is not a regular file.", +runCmd(new String[]{"verifyEC", "-file", "/ec"})); + +fs.create(new Path(ecDir, "foo")); +assertEquals("ret: 1, File /ec/foo is not closed.", +runCmd(new String[]{"verifyEC", "-file", "/ec/foo"})); + +final short repl = 1; +final long k = 1024; +final long m = k * k; +final long seed = 0x1234567L; +DFSTestUtil.createFile(fs, new Path(ecDir, "foo_65535"), 65535, repl, seed); +assertTrue(runCmd(new String[]{"verifyEC", "-file", "/ec/foo_65535"}) +.contains("All EC block group status: OK")); +DFSTestUtil.createFile(fs, new Path(ecDir, "foo_256k"), 256 * k, repl, seed); +assertTrue(runCmd(new String[]{"verifyEC", "-file", "/ec/foo_256k"}) +.contains("All EC block group status: OK")); +DFSTestUtil.createFile(fs, new Path(ecDir, "foo_1m"), m, repl, seed); +assertTrue(runCmd(new String[]{"verifyEC", "-file", "/ec/foo_1m"}) +.contains("All EC block group status: OK")); +DFSTestUtil.createFile(fs, new Path(ecDir, "foo_2m"), 2 * m, repl, seed); +assertTrue(runCmd(new String[]{"verifyEC", "-file", "/ec/foo_2m"}) +.contains("All EC block group status: OK")); +DFSTestUtil.createFile(fs, new Path(ecDir, "foo_3m"), 3 * m, repl, seed); +assertTrue(runCmd(new String[]{"verifyEC", "-file", "/ec/foo_3m"}) +.contains("All EC block group status: OK")); +DFSTestUtil.createFile(fs, new Path(ecDir, "foo_5m"), 5 * m, repl, seed); +assertTrue(runCmd(new String[]{"verifyEC", "-file", "/ec/foo_5m"}) +.contains("All EC block group status: OK")); + Review comment: Thanks, that's a good advice, updated. -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Work logged] (HADOOP-17124) Support LZO using aircompressor
[ https://issues.apache.org/jira/browse/HADOOP-17124?focusedWorklogId=675599&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-675599 ] ASF GitHub Bot logged work on HADOOP-17124: --- Author: ASF GitHub Bot Created on: 04/Nov/21 01:06 Start Date: 04/Nov/21 01:06 Worklog Time Spent: 10m Work Description: hadoop-yetus commented on pull request #3612: URL: https://github.com/apache/hadoop/pull/3612#issuecomment-958607089 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org Issue Time Tracking --- Worklog Id: (was: 675599) Time Spent: 5.5h (was: 5h 20m) > Support LZO using aircompressor > --- > > Key: HADOOP-17124 > URL: https://issues.apache.org/jira/browse/HADOOP-17124 > Project: Hadoop Common > Issue Type: New Feature > Components: common >Affects Versions: 3.3.0 >Reporter: DB Tsai >Priority: Major > Labels: pull-request-available > Time Spent: 5.5h > Remaining Estimate: 0h > > LZO codec was removed in HADOOP-4874 because the original LZO binding is GPL > which is problematic. However, many legacy data is still compressed by LZO > codec, and companies often use vendor's GPL LZO codec in the classpath which > might cause GPL contamination. > Presro and ORC-77 use [aircompressor| > [https://github.com/airlift/aircompressor]] (Apache V2 licensed) to compress > and decompress LZO data. Hadoop can add back LZO support using aircompressor > without GPL violation. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[GitHub] [hadoop] hadoop-yetus commented on pull request #3612: WIP. HADOOP-17124. Support LZO Codec using aircompressor
hadoop-yetus commented on pull request #3612: URL: https://github.com/apache/hadoop/pull/3612#issuecomment-958607089 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[GitHub] [hadoop] hadoop-yetus commented on pull request #3593: HDFS-16286. Add a debug tool to verify the correctness of erasure coding on file
hadoop-yetus commented on pull request #3593: URL: https://github.com/apache/hadoop/pull/3593#issuecomment-958791127 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[GitHub] [hadoop] haiyang1987 closed pull request #3596: HDFS-16287. Support to make dfs.namenode.avoid.read.slow.datanode reconfigurable
haiyang1987 closed pull request #3596: URL: https://github.com/apache/hadoop/pull/3596 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[GitHub] [hadoop] jianghuazhu commented on pull request #3602: HDFS-16291.Make the comment of INode#ReclaimContext more standardized.
jianghuazhu commented on pull request #3602: URL: https://github.com/apache/hadoop/pull/3602#issuecomment-958751029 Thank you very much. @ferhui @virajjasani -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Work logged] (HADOOP-17981) Support etag-assisted renames in FileOutputCommitter
[ https://issues.apache.org/jira/browse/HADOOP-17981?focusedWorklogId=675542&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-675542 ] ASF GitHub Bot logged work on HADOOP-17981: --- Author: ASF GitHub Bot Created on: 04/Nov/21 01:00 Start Date: 04/Nov/21 01:00 Worklog Time Spent: 10m Work Description: hadoop-yetus commented on pull request #3611: URL: https://github.com/apache/hadoop/pull/3611#issuecomment-958583977 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org Issue Time Tracking --- Worklog Id: (was: 675542) Time Spent: 7h 10m (was: 7h) > Support etag-assisted renames in FileOutputCommitter > > > Key: HADOOP-17981 > URL: https://issues.apache.org/jira/browse/HADOOP-17981 > Project: Hadoop Common > Issue Type: New Feature > Components: fs, fs/azure >Affects Versions: 3.4.0 >Reporter: Steve Loughran >Assignee: Steve Loughran >Priority: Major > Labels: pull-request-available > Time Spent: 7h 10m > Remaining Estimate: 0h > > To deal with some throttling/retry issues in object stores, > pass the FileStatus entries retrieved during listing > into a private interface ResilientCommitByRename which filesystems > may implement to use extra attributes in the listing (etag, version) > to constrain and validate the operation. > Although targeting azure, GCS and others could use. no point in S3A as they > shouldn't use this committer. > # And we are not going to do any changes to FileSystem as there are explicit > guarantees of public use and stability. > I am not going to make that change as the hive thing that will suddenly start > expecting it to work forever. > # I'm not planning to merge this in, as the manifest committer is going to > include this and more (MAPREDUCE-7341) > However, I do need to get this in on a branch, so am doing this work on trunk > for dev & test and for others to review -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[GitHub] [hadoop] ferhui commented on pull request #3613: HDFS-16296. RouterRpcFairnessPolicyController add rejected permits for each nameservice
ferhui commented on pull request #3613: URL: https://github.com/apache/hadoop/pull/3613#issuecomment-958680152 @symious Thanks for contribution, it looks good. Let's wait for the CI reports. -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Work logged] (HADOOP-17975) Fallback to simple auth does not work for a secondary DistributedFileSystem instance
[ https://issues.apache.org/jira/browse/HADOOP-17975?focusedWorklogId=675554&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-675554 ] ASF GitHub Bot logged work on HADOOP-17975: --- Author: ASF GitHub Bot Created on: 04/Nov/21 01:01 Start Date: 04/Nov/21 01:01 Worklog Time Spent: 10m Work Description: symious commented on pull request #3579: URL: https://github.com/apache/hadoop/pull/3579#issuecomment-958605853 @fapifta Thanks for the explanation. Is the current situation as follows? Correct me if I'm wrong. Server S has two clients, client A of non-secure settings with "fallbackToSimpleAuth = false", client B of secure settings with "fallbackToSimpleAuth=true", client A and B both connecting to Datanode D, since client A first connect to Datanode D, so the connection is created in _connections_, and the fallbackToSimpleAuth is set to false. Then client B tries to connect Datanode D, since the _ConnectionId_ is the same, it will reuse the connection created by Client A, then in _setupIOStream_, since the socket is created, Client will skip the other process, and the "fallbackToSimpleAuth" won't be changed to true. If the issue is caused by the above situation, I think if we distinguish the connections of Client A and Client B will solve the issue? Currently the ConnectionId of Client A and Client B is same, if we add "fallbackToSimpleAuth" to ConnectionId, since Client A has "false" and Client B has "true", they will generate different ConnectionId, then different connections will be created, so they won't affect each other. -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org Issue Time Tracking --- Worklog Id: (was: 675554) Time Spent: 3h (was: 2h 50m) > Fallback to simple auth does not work for a secondary DistributedFileSystem > instance > > > Key: HADOOP-17975 > URL: https://issues.apache.org/jira/browse/HADOOP-17975 > Project: Hadoop Common > Issue Type: Bug > Components: ipc >Reporter: István Fajth >Assignee: István Fajth >Priority: Major > Labels: pull-request-available > Time Spent: 3h > Remaining Estimate: 0h > > The following code snippet demonstrates what is necessary to cause a failure > in connection to a non secure cluster with fallback to SIMPLE auth allowed > from a secure cluster. > {code:java} > Configuration conf = new Configuration(); > conf.setBoolean("ipc.client.fallback-to-simple-auth-allowed", true); > URI fsUri = new URI("hdfs://"); > conf.setBoolean("fs.hdfs.impl.disable.cache", true); > FileSystem fs = FileSystem.get(fsUri, conf); > FSDataInputStream src = fs.open(new Path("/path/to/a/file")); > FileOutputStream dst = new FileOutputStream(File.createTempFile("foo", > "bar")); > IOUtils.copyBytes(src, dst, 1024); > // The issue happens even if we re-enable cache at this point > //conf.setBoolean("fs.hdfs.impl.disable.cache", false); > // The issue does not happen when we close the first FileSystem object > // before creating the second. > //fs.close(); > FileSystem fs2 = FileSystem.get(fsUri, conf); > FSDataInputStream src2 = fs2.open(new Path("/path/to/a/file")); > FileOutputStream dst2 = new FileOutputStream(File.createTempFile("foo", > "bar")); > IOUtils.copyBytes(src2, dst2, 1024); > {code} > The problem is that when the DfsClient is created it creates an instance of > AtomicBoolean, which is propagated down into the IPC layer, where the > Client.Connection instance in setupIOStreams sets its value. This connection > object is cached and re-used to multiplex requests against the same DataNode. > In case of creating a second DfsClient, the AtomicBoolean reference in the > client is a new AtomicBoolean, but the Client.Connection instance is the > same, and as it has a socket already open to the DataNode, it returns > immediatelly from setupIOStreams, leaving the fallbackToSimpleAuth > AtomicBoolean false as it is created in the DfsClient. > This AtomicBoolean on the other hand controls how the SaslDataTransferClient > handles the connection in the above level, and with this value left on the > default false, the SaslDataTransferClient of the second DfsClient will not > fall back to SIMPLE authentication but will try to send a SASL handshake when > connecting to the DataNode. > > The access to the FileSystem via the sec
[GitHub] [hadoop] symious commented on pull request #3579: HADOOP-17975 Fallback to simple auth does not work for a secondary DistributedFileSystem instance.
symious commented on pull request #3579: URL: https://github.com/apache/hadoop/pull/3579#issuecomment-958605853 @fapifta Thanks for the explanation. Is the current situation as follows? Correct me if I'm wrong. Server S has two clients, client A of non-secure settings with "fallbackToSimpleAuth = false", client B of secure settings with "fallbackToSimpleAuth=true", client A and B both connecting to Datanode D, since client A first connect to Datanode D, so the connection is created in _connections_, and the fallbackToSimpleAuth is set to false. Then client B tries to connect Datanode D, since the _ConnectionId_ is the same, it will reuse the connection created by Client A, then in _setupIOStream_, since the socket is created, Client will skip the other process, and the "fallbackToSimpleAuth" won't be changed to true. If the issue is caused by the above situation, I think if we distinguish the connections of Client A and Client B will solve the issue? Currently the ConnectionId of Client A and Client B is same, if we add "fallbackToSimpleAuth" to ConnectionId, since Client A has "false" and Client B has "true", they will generate different ConnectionId, then different connections will be created, so they won't affect each other. -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[GitHub] [hadoop] hadoop-yetus commented on pull request #3613: HDFS-16296. RouterRpcFairnessPolicyController add rejected permits for each nameservice
hadoop-yetus commented on pull request #3613: URL: https://github.com/apache/hadoop/pull/3613#issuecomment-958690121 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[GitHub] [hadoop] hadoop-yetus commented on pull request #3611: HADOOP-17981. resilient commit through etag validation
hadoop-yetus commented on pull request #3611: URL: https://github.com/apache/hadoop/pull/3611#issuecomment-958583977 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Work logged] (HADOOP-17981) Support etag-assisted renames in FileOutputCommitter
[ https://issues.apache.org/jira/browse/HADOOP-17981?focusedWorklogId=675522&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-675522 ] ASF GitHub Bot logged work on HADOOP-17981: --- Author: ASF GitHub Bot Created on: 04/Nov/21 00:58 Start Date: 04/Nov/21 00:58 Worklog Time Spent: 10m Work Description: steveloughran commented on pull request #3597: URL: https://github.com/apache/hadoop/pull/3597#issuecomment-959193924 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org Issue Time Tracking --- Worklog Id: (was: 675522) Time Spent: 7h (was: 6h 50m) > Support etag-assisted renames in FileOutputCommitter > > > Key: HADOOP-17981 > URL: https://issues.apache.org/jira/browse/HADOOP-17981 > Project: Hadoop Common > Issue Type: New Feature > Components: fs, fs/azure >Affects Versions: 3.4.0 >Reporter: Steve Loughran >Assignee: Steve Loughran >Priority: Major > Labels: pull-request-available > Time Spent: 7h > Remaining Estimate: 0h > > To deal with some throttling/retry issues in object stores, > pass the FileStatus entries retrieved during listing > into a private interface ResilientCommitByRename which filesystems > may implement to use extra attributes in the listing (etag, version) > to constrain and validate the operation. > Although targeting azure, GCS and others could use. no point in S3A as they > shouldn't use this committer. > # And we are not going to do any changes to FileSystem as there are explicit > guarantees of public use and stability. > I am not going to make that change as the hive thing that will suddenly start > expecting it to work forever. > # I'm not planning to merge this in, as the manifest committer is going to > include this and more (MAPREDUCE-7341) > However, I do need to get this in on a branch, so am doing this work on trunk > for dev & test and for others to review -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[GitHub] [hadoop] steveloughran commented on pull request #3597: HADOOP-17981 Support etag-assisted renames in FileOutputCommitter
steveloughran commented on pull request #3597: URL: https://github.com/apache/hadoop/pull/3597#issuecomment-959193924 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Work logged] (HADOOP-17124) Support LZO using aircompressor
[ https://issues.apache.org/jira/browse/HADOOP-17124?focusedWorklogId=675504&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-675504 ] ASF GitHub Bot logged work on HADOOP-17124: --- Author: ASF GitHub Bot Created on: 04/Nov/21 00:56 Start Date: 04/Nov/21 00:56 Worklog Time Spent: 10m Work Description: sunchao commented on pull request #3612: URL: https://github.com/apache/hadoop/pull/3612#issuecomment-959727355 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org Issue Time Tracking --- Worklog Id: (was: 675504) Time Spent: 5h 20m (was: 5h 10m) > Support LZO using aircompressor > --- > > Key: HADOOP-17124 > URL: https://issues.apache.org/jira/browse/HADOOP-17124 > Project: Hadoop Common > Issue Type: New Feature > Components: common >Affects Versions: 3.3.0 >Reporter: DB Tsai >Priority: Major > Labels: pull-request-available > Time Spent: 5h 20m > Remaining Estimate: 0h > > LZO codec was removed in HADOOP-4874 because the original LZO binding is GPL > which is problematic. However, many legacy data is still compressed by LZO > codec, and companies often use vendor's GPL LZO codec in the classpath which > might cause GPL contamination. > Presro and ORC-77 use [aircompressor| > [https://github.com/airlift/aircompressor]] (Apache V2 licensed) to compress > and decompress LZO data. Hadoop can add back LZO support using aircompressor > without GPL violation. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Work logged] (HADOOP-17981) Support etag-assisted renames in FileOutputCommitter
[ https://issues.apache.org/jira/browse/HADOOP-17981?focusedWorklogId=675502&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-675502 ] ASF GitHub Bot logged work on HADOOP-17981: --- Author: ASF GitHub Bot Created on: 04/Nov/21 00:56 Start Date: 04/Nov/21 00:56 Worklog Time Spent: 10m Work Description: hadoop-yetus removed a comment on pull request #3597: URL: https://github.com/apache/hadoop/pull/3597#issuecomment-956975430 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org Issue Time Tracking --- Worklog Id: (was: 675502) Time Spent: 6h 50m (was: 6h 40m) > Support etag-assisted renames in FileOutputCommitter > > > Key: HADOOP-17981 > URL: https://issues.apache.org/jira/browse/HADOOP-17981 > Project: Hadoop Common > Issue Type: New Feature > Components: fs, fs/azure >Affects Versions: 3.4.0 >Reporter: Steve Loughran >Assignee: Steve Loughran >Priority: Major > Labels: pull-request-available > Time Spent: 6h 50m > Remaining Estimate: 0h > > To deal with some throttling/retry issues in object stores, > pass the FileStatus entries retrieved during listing > into a private interface ResilientCommitByRename which filesystems > may implement to use extra attributes in the listing (etag, version) > to constrain and validate the operation. > Although targeting azure, GCS and others could use. no point in S3A as they > shouldn't use this committer. > # And we are not going to do any changes to FileSystem as there are explicit > guarantees of public use and stability. > I am not going to make that change as the hive thing that will suddenly start > expecting it to work forever. > # I'm not planning to merge this in, as the manifest committer is going to > include this and more (MAPREDUCE-7341) > However, I do need to get this in on a branch, so am doing this work on trunk > for dev & test and for others to review -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[GitHub] [hadoop] hadoop-yetus removed a comment on pull request #3597: HADOOP-17981 Support etag-assisted renames in FileOutputCommitter
hadoop-yetus removed a comment on pull request #3597: URL: https://github.com/apache/hadoop/pull/3597#issuecomment-956975430 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[GitHub] [hadoop] sunchao commented on pull request #3612: WIP. HADOOP-17124. Support LZO Codec using aircompressor
sunchao commented on pull request #3612: URL: https://github.com/apache/hadoop/pull/3612#issuecomment-959727355 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[GitHub] [hadoop] sodonnel merged pull request #3593: HDFS-16286. Add a debug tool to verify the correctness of erasure coding on file
sodonnel merged pull request #3593: URL: https://github.com/apache/hadoop/pull/3593 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[GitHub] [hadoop-site] GauthamBanasandra commented on pull request #28: Add Gautham Banasandra to Hadoop committers' list
GauthamBanasandra commented on pull request #28: URL: https://github.com/apache/hadoop-site/pull/28#issuecomment-959811956 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Work logged] (HADOOP-17124) Support LZO using aircompressor
[ https://issues.apache.org/jira/browse/HADOOP-17124?focusedWorklogId=675444&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-675444 ] ASF GitHub Bot logged work on HADOOP-17124: --- Author: ASF GitHub Bot Created on: 04/Nov/21 00:51 Start Date: 04/Nov/21 00:51 Worklog Time Spent: 10m Work Description: viirya edited a comment on pull request #3612: URL: https://github.com/apache/hadoop/pull/3612#issuecomment-959725363 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org Issue Time Tracking --- Worklog Id: (was: 675444) Time Spent: 5h 10m (was: 5h) > Support LZO using aircompressor > --- > > Key: HADOOP-17124 > URL: https://issues.apache.org/jira/browse/HADOOP-17124 > Project: Hadoop Common > Issue Type: New Feature > Components: common >Affects Versions: 3.3.0 >Reporter: DB Tsai >Priority: Major > Labels: pull-request-available > Time Spent: 5h 10m > Remaining Estimate: 0h > > LZO codec was removed in HADOOP-4874 because the original LZO binding is GPL > which is problematic. However, many legacy data is still compressed by LZO > codec, and companies often use vendor's GPL LZO codec in the classpath which > might cause GPL contamination. > Presro and ORC-77 use [aircompressor| > [https://github.com/airlift/aircompressor]] (Apache V2 licensed) to compress > and decompress LZO data. Hadoop can add back LZO support using aircompressor > without GPL violation. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[GitHub] [hadoop] viirya edited a comment on pull request #3612: WIP. HADOOP-17124. Support LZO Codec using aircompressor
viirya edited a comment on pull request #3612: URL: https://github.com/apache/hadoop/pull/3612#issuecomment-959725363 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[GitHub] [hadoop] goiri commented on a change in pull request #3553: HDFS-16273. RBF: RouterRpcFairnessPolicyController add availableHandl…
goiri commented on a change in pull request #3553: URL: https://github.com/apache/hadoop/pull/3553#discussion_r742198185 ## File path: hadoop-hdfs-project/hadoop-hdfs-rbf/src/main/java/org/apache/hadoop/hdfs/server/federation/fairness/NoRouterRpcFairnessPolicyController.java ## @@ -46,4 +46,9 @@ public void releasePermit(String nsId) { public void shutdown() { // Nothing for now. } + + @Override + public String getAvailableHandlerOnPerNs(){ +return "N/A"; Review comment: Should we test for this? ## File path: hadoop-hdfs-project/hadoop-hdfs-rbf/src/main/java/org/apache/hadoop/hdfs/server/federation/fairness/AbstractRouterRpcFairnessPolicyController.java ## @@ -75,4 +77,17 @@ protected void insertNameServiceWithPermits(String nsId, int maxPermits) { protected int getAvailablePermits(String nsId) { return this.permits.get(nsId).availablePermits(); } + + @Override + public String getAvailableHandlerOnPerNs() { +JSONObject json = new JSONObject(); +for (Map.Entry entry : permits.entrySet()) { + try { +json.put(entry.getKey(), entry.getValue().availablePermits()); Review comment: Let's extract entry.getKey() and entry.getValue() to have a particular name. ## File path: hadoop-hdfs-project/hadoop-hdfs-rbf/src/main/java/org/apache/hadoop/hdfs/server/federation/fairness/NoRouterRpcFairnessPolicyController.java ## @@ -46,4 +46,9 @@ public void releasePermit(String nsId) { public void shutdown() { // Nothing for now. } + + @Override + public String getAvailableHandlerOnPerNs(){ +return "N/A"; Review comment: Should we test for this? ## File path: hadoop-hdfs-project/hadoop-hdfs-rbf/src/main/java/org/apache/hadoop/hdfs/server/federation/fairness/AbstractRouterRpcFairnessPolicyController.java ## @@ -75,4 +77,17 @@ protected void insertNameServiceWithPermits(String nsId, int maxPermits) { protected int getAvailablePermits(String nsId) { return this.permits.get(nsId).availablePermits(); } + + @Override + public String getAvailableHandlerOnPerNs() { +JSONObject json = new JSONObject(); +for (Map.Entry entry : permits.entrySet()) { + try { +json.put(entry.getKey(), entry.getValue().availablePermits()); Review comment: Let's extract entry.getKey() and entry.getValue() to have a particular name. -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Work logged] (HADOOP-17981) Support etag-assisted renames in FileOutputCommitter
[ https://issues.apache.org/jira/browse/HADOOP-17981?focusedWorklogId=675410&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-675410 ] ASF GitHub Bot logged work on HADOOP-17981: --- Author: ASF GitHub Bot Created on: 04/Nov/21 00:48 Start Date: 04/Nov/21 00:48 Worklog Time Spent: 10m Work Description: steveloughran closed pull request #3597: URL: https://github.com/apache/hadoop/pull/3597 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org Issue Time Tracking --- Worklog Id: (was: 675410) Time Spent: 6h 40m (was: 6.5h) > Support etag-assisted renames in FileOutputCommitter > > > Key: HADOOP-17981 > URL: https://issues.apache.org/jira/browse/HADOOP-17981 > Project: Hadoop Common > Issue Type: New Feature > Components: fs, fs/azure >Affects Versions: 3.4.0 >Reporter: Steve Loughran >Assignee: Steve Loughran >Priority: Major > Labels: pull-request-available > Time Spent: 6h 40m > Remaining Estimate: 0h > > To deal with some throttling/retry issues in object stores, > pass the FileStatus entries retrieved during listing > into a private interface ResilientCommitByRename which filesystems > may implement to use extra attributes in the listing (etag, version) > to constrain and validate the operation. > Although targeting azure, GCS and others could use. no point in S3A as they > shouldn't use this committer. > # And we are not going to do any changes to FileSystem as there are explicit > guarantees of public use and stability. > I am not going to make that change as the hive thing that will suddenly start > expecting it to work forever. > # I'm not planning to merge this in, as the manifest committer is going to > include this and more (MAPREDUCE-7341) > However, I do need to get this in on a branch, so am doing this work on trunk > for dev & test and for others to review -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[GitHub] [hadoop] steveloughran closed pull request #3597: HADOOP-17981 Support etag-assisted renames in FileOutputCommitter
steveloughran closed pull request #3597: URL: https://github.com/apache/hadoop/pull/3597 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[GitHub] [hadoop] hadoop-yetus commented on pull request #3596: HDFS-16287. Support to make dfs.namenode.avoid.read.slow.datanode reconfigurable
hadoop-yetus commented on pull request #3596: URL: https://github.com/apache/hadoop/pull/3596#issuecomment-959651079 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[GitHub] [hadoop] tomscut commented on a change in pull request #3596: HDFS-16287. Support to make dfs.namenode.avoid.read.slow.datanode reconfigurable
tomscut commented on a change in pull request #3596: URL: https://github.com/apache/hadoop/pull/3596#discussion_r742015162 ## File path: hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/DatanodeManager.java ## @@ -511,7 +505,16 @@ private boolean isInactive(DatanodeInfo datanode) { private boolean isSlowNode(String dnUuid) { return avoidSlowDataNodesForRead && slowNodesUuidSet.contains(dnUuid); } - + + public void setAvoidSlowDataNodesForReadEnabled(boolean enable) { Review comment: Thanks @haiyang1987 for your comment. I think the logic right now is no problem. I mean that when ```excludeSlowNodesEnabled``` is set to true we ```startSlowPeerCollector```, and ```stopSlowPeerCollector``` when ```excludeSlowNodesEnabled``` is set to false. There is no extra overhead. What do you think? ## File path: hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/DatanodeManager.java ## @@ -511,7 +505,16 @@ private boolean isInactive(DatanodeInfo datanode) { private boolean isSlowNode(String dnUuid) { return avoidSlowDataNodesForRead && slowNodesUuidSet.contains(dnUuid); } - + + public void setAvoidSlowDataNodesForReadEnabled(boolean enable) { Review comment: Thanks @haiyang1987 for your comment. I think the logic right now is no problem. I mean that only when ```excludeSlowNodesEnabled``` is set to true we ```startSlowPeerCollector```, and ```stopSlowPeerCollector``` when ```excludeSlowNodesEnabled``` is set to false. There is no extra overhead. What do you think? ## File path: hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/DatanodeManager.java ## @@ -511,7 +505,16 @@ private boolean isInactive(DatanodeInfo datanode) { private boolean isSlowNode(String dnUuid) { return avoidSlowDataNodesForRead && slowNodesUuidSet.contains(dnUuid); } - + + public void setAvoidSlowDataNodesForReadEnabled(boolean enable) { Review comment: Thanks @haiyang1987 for your comment. I think the logic right now is no problem. I mean that only when `excludeSlowNodesEnabled` is set to `true` we `startSlowPeerCollector`, and `stopSlowPeerCollector` when `excludeSlowNodesEnabled` is set to `false`. There is no extra overhead. What do you think? ## File path: hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/DatanodeManager.java ## @@ -511,7 +505,16 @@ private boolean isInactive(DatanodeInfo datanode) { private boolean isSlowNode(String dnUuid) { return avoidSlowDataNodesForRead && slowNodesUuidSet.contains(dnUuid); } - + + public void setAvoidSlowDataNodesForReadEnabled(boolean enable) { Review comment: Thanks @haiyang1987 for your comment. I think the logic right now is no problem. I mean that when ```excludeSlowNodesEnabled``` is set to true we ```startSlowPeerCollector```, and ```stopSlowPeerCollector``` when ```excludeSlowNodesEnabled``` is set to false. There is no extra overhead. What do you think? ## File path: hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/DatanodeManager.java ## @@ -511,7 +505,16 @@ private boolean isInactive(DatanodeInfo datanode) { private boolean isSlowNode(String dnUuid) { return avoidSlowDataNodesForRead && slowNodesUuidSet.contains(dnUuid); } - + + public void setAvoidSlowDataNodesForReadEnabled(boolean enable) { Review comment: Thanks @haiyang1987 for your comment. I think the logic right now is no problem. I mean that only when ```excludeSlowNodesEnabled``` is set to true we ```startSlowPeerCollector```, and ```stopSlowPeerCollector``` when ```excludeSlowNodesEnabled``` is set to false. There is no extra overhead. What do you think? ## File path: hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/DatanodeManager.java ## @@ -511,7 +505,16 @@ private boolean isInactive(DatanodeInfo datanode) { private boolean isSlowNode(String dnUuid) { return avoidSlowDataNodesForRead && slowNodesUuidSet.contains(dnUuid); } - + + public void setAvoidSlowDataNodesForReadEnabled(boolean enable) { Review comment: Thanks @haiyang1987 for your comment. I think the logic right now is no problem. I mean that only when `excludeSlowNodesEnabled` is set to `true` we `startSlowPeerCollector`, and `stopSlowPeerCollector` when `excludeSlowNodesEnabled` is set to `false`. There is no extra overhead. What do you think? -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, pl
[GitHub] [hadoop] goiri commented on a change in pull request #3613: HDFS-16296. RouterRpcFairnessPolicyController add rejected permits for each nameservice
goiri commented on a change in pull request #3613: URL: https://github.com/apache/hadoop/pull/3613#discussion_r742196019 ## File path: hadoop-hdfs-project/hadoop-hdfs-rbf/src/test/java/org/apache/hadoop/hdfs/server/federation/fairness/TestRouterHandlersFairness.java ## @@ -208,4 +212,15 @@ private void invokeConcurrent(ClientProtocol routerProto, String clientName) routerProto.renewLease(clientName); } + private int getTotalRejectedPermits(RouterContext routerContext) { +int totalRejectedPermits = 0; +for (String ns : cluster.getNameservices()) { + totalRejectedPermits += routerContext.getRouter().getRpcServer() Review comment: We may want to extract: routerContext.getRouter().getRpcServer().getRPCClient() ## File path: hadoop-hdfs-project/hadoop-hdfs-rbf/src/test/java/org/apache/hadoop/hdfs/server/federation/fairness/TestRouterHandlersFairness.java ## @@ -208,4 +212,15 @@ private void invokeConcurrent(ClientProtocol routerProto, String clientName) routerProto.renewLease(clientName); } + private int getTotalRejectedPermits(RouterContext routerContext) { +int totalRejectedPermits = 0; +for (String ns : cluster.getNameservices()) { + totalRejectedPermits += routerContext.getRouter().getRpcServer() Review comment: We may want to extract: routerContext.getRouter().getRpcServer().getRPCClient() -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[GitHub] [hadoop] hadoop-yetus commented on pull request #3602: HDFS-16291.Make the comment of INode#ReclaimContext more standardized.
hadoop-yetus commented on pull request #3602: URL: https://github.com/apache/hadoop/pull/3602#issuecomment-959335456 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org