[jira] [Commented] (HDFS-12890) Ozone: XceiverClient should have upper bound on async requests
[ https://issues.apache.org/jira/browse/HDFS-12890?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16283176#comment-16283176 ] Mukul Kumar Singh commented on HDFS-12890: -- Thanks for the updated patch [~shashikant] 1) In case of an exception, the ctx is being closed. I feel that on an exception, all the futures which are enqueued in the Hashmap should be completed exceptionally and the semaphore count should be decremented appropriately. > Ozone: XceiverClient should have upper bound on async requests > -- > > Key: HDFS-12890 > URL: https://issues.apache.org/jira/browse/HDFS-12890 > Project: Hadoop HDFS > Issue Type: Sub-task > Components: HDFS-7240 >Affects Versions: HDFS-7240 >Reporter: Shashikant Banerjee >Assignee: Shashikant Banerjee > Fix For: HDFS-7240 > > Attachments: HDFS-12890-HDFS-7240.001.patch, > HDFS-12890-HDFS-7240.002.patch, HDFS-12890-HDFS-7240.003.patch > > > XceiverClient-ratis maintains upper bound on the no of outstanding async > requests . XceiverClient > should also impose an upper bound on the no of outstanding async requests > received from client > for write. -- This message was sent by Atlassian JIRA (v6.4.14#64029) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-12907) Allow read-only access to reserved raw for non-superusers
[ https://issues.apache.org/jira/browse/HDFS-12907?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16283173#comment-16283173 ] Xiao Chen commented on HDFS-12907: -- Thanks for the explanation Daryn. The rage block is what's painful for us as well, when supporting different downstream components and partners. I hope the kms clients were designed / documented / implemented / reviewed perfectly too. Interestingly if you see the history of KMSCP class you'll see quite a few attempts to make it 'work for the case of xxx'. Maybe what you described can be fixed in a new jira, and consider some of the past behaviors simply wrong so we don't worry about compatibility. Agree letting the DN to pass through raw bytes would be great. I hope this can be rolled into a simple design doc for HDFS-12355 so other people can figure out easily. > Allow read-only access to reserved raw for non-superusers > - > > Key: HDFS-12907 > URL: https://issues.apache.org/jira/browse/HDFS-12907 > Project: Hadoop HDFS > Issue Type: Bug > Components: namenode >Affects Versions: 2.6.0 >Reporter: Daryn Sharp >Assignee: Rushabh S Shah > Attachments: HDFS-12907.001.patch, HDFS-12907.patch > > > HDFS-6509 added a special /.reserved/raw path prefix to access the raw file > contents of EZ files. In the simplest sense it doesn't return the FE info in > the {{LocatedBlocks}} so the dfs client doesn't try to decrypt the data. > This facilitates allowing tools like distcp to copy raw bytes. > Access to the raw hierarchy is restricted to superusers. This seems like an > overly broad restriction designed to prevent non-admins from munging the EZ > related xattrs. I believe we should relax the restriction to allow > non-admins to perform read-only operations. Allowing non-superusers to > easily read the raw bytes will be extremely useful for regular users, esp. > for enabling webhdfs client-side encryption. -- This message was sent by Atlassian JIRA (v6.4.14#64029) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Comment Edited] (HDFS-10285) Storage Policy Satisfier in Namenode
[ https://issues.apache.org/jira/browse/HDFS-10285?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16283102#comment-16283102 ] Uma Maheswara Rao G edited comment on HDFS-10285 at 12/8/17 7:35 AM: - {quote} Here's a rhetorical question: If managing multiple services is hard, why not bundle oozie, spark, storm, sqoop, kafka, ranger, knox, hive server, etc in the same process? Or ZK so HA is easier to deploy/manage? {quote} Few of my thoughts on this question, each of these projects build for their own purpose, with its own spec, not for just for helping HDFS or any other single project. And none of that projects need to access other project internal data structures. Where as SPS is only functions for HDFS and access internal data structures. Even forcibly separated out, we need to expose ‘for SPS only’ RPC APIs. This strikes me to put a question in other way as well, is it make sense to separate ReplicationMonitor as one separate process? is it fine to start EDEK as one separate? is it ok to start other threads (like decommissioning task) as separate processes and co-ordinate via RPC? so that NameSystem class may become very light weight? I think its the value vs cost will decide whether to separate or merge into single. Coming to ZK part, As ZK is not build only for HDFS, I don’t think to have any such thoughts. Its general purpose co-ordination system. Technically we can’t keep monitoring services inside NN, because the worry itself is, NN may die, need failover and so need external process to monitor. Anyway. I think the whole discussion is about services inside a project, but not cross projects itself, IMHO. Here SPS providing only the missing functionality of HSM, that is end-end policy satisfaction. So, IMV, for users it may not worth to manage additional process to achieve that missing functionality for particular feature. {quote} Today, I looked at the code more closely. It can hold the lock (read lock, but still) way too long. Notably, but not limited to, you can’t hold the lock while doing block placement. {quote} Appreciate your review Daryn. I think it should be easy to address. We will make sure to address the comment before merge? is that make sense. {quote} I’m curious why it isn’t just part of the standard replication monitoring. If the DN is told to replicate to itself, it just does the storage movement. {quote} That's a good question. Overall approach is exactly same as RM. RM is has its own q build up for redundancy blocks, and Underreplication scan/check happens at block level, it make sense. Where as in SPS, policy changes for file, so all blocks in that file needs movement and policy check should happen in-co-ordination with replication blocks where they stored currently. So, we track the queues at file level here and scan/check all block for that files together at once. Also , We wanted to provide, on the fly reconfigure feature and we carefully thought that, we don’t want to interfere replication logic should be given more priority than SPS work. While scheduling blocks, we respect xmits counts, they are shared between, RM, SPS for controlling DN load. Assignment priority given to replication/EC blocks, then SPS blocks, when sending tasks to DN. So, as part of impact analysis, we thought of keeping SPS in it's own thread than RM thread would be clean and safer than running in that same loop of RM. was (Author: umamaheswararao): {quote} Here's a rhetorical question: If managing multiple services is hard, why not bundle oozie, spark, storm, sqoop, kafka, ranger, knox, hive server, etc in the same process? Or ZK so HA is easier to deploy/manage? {quote} Few of my thoughts on this question, each of these projects build for their own purpose, with its own spec, not for just for helping HDFS or any other single project. And none of that projects need to access other project internal data structures. Where as SPS is only functions for HDFS and access internal data structures. Even forcibly separated out, we need to expose ‘for SPS only’ RPC APIs. This strikes me to put a question in other way as well, is it make sense to separate ReplicationMonitor as one separate process? is it fine to start EDEK as one separate? is it ok to start other threads (like decommissioning task) as separate processes and co-ordinate via RPC? so that NameSystem class may become very light weight? I think its the value vs cost will decide whether to separate or merge into single. Coming to ZK part, As ZK is not build only for HDFS, I don’t think to have any such thoughts. Its general purpose co-ordination system. Technically we can’t keep monitoring services inside NN, because the worry itself is, NN may die, need failover and so need external process to monitor. Anyway. I think the whole discussion is about services inside a project, but not cross
[jira] [Commented] (HDFS-10285) Storage Policy Satisfier in Namenode
[ https://issues.apache.org/jira/browse/HDFS-10285?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16283102#comment-16283102 ] Uma Maheswara Rao G commented on HDFS-10285: {quote} Here's a rhetorical question: If managing multiple services is hard, why not bundle oozie, spark, storm, sqoop, kafka, ranger, knox, hive server, etc in the same process? Or ZK so HA is easier to deploy/manage? {quote} Few of my thoughts on this question, each of these projects build for their own purpose, with its own spec, not for just for helping HDFS or any other single project. And none of that projects need to access other project internal data structures. Where as SPS is only functions for HDFS and access internal data structures. Even forcibly separated out, we need to expose ‘for SPS only’ RPC APIs. This strikes me to put a question in other way as well, is it make sense to separate ReplicationMonitor as one separate process? is it fine to start EDEK as one separate? is it ok to start other threads (like decommissioning task) as separate processes and co-ordinate via RPC? so that NameSystem class may become very light weight? I think its the value vs cost will decide whether to separate or merge into single. Coming to ZK part, As ZK is not build only for HDFS, I don’t think to have any such thoughts. Its general purpose co-ordination system. Technically we can’t keep monitoring services inside NN, because the worry itself is, NN may die, need failover and so need external process to monitor. Anyway. I think the whole discussion is about services inside a project, but not cross projects itself, IMHO. Here SPS providing only the missing functionality of HSM, that is end-end policy satisfaction. So, IMV, for users it may not worth to manage additional process to achieve that missing functionality for particular feature. {quote} Today, I looked at the code more closely. It can hold the lock (read lock, but still) way too long. Notably, but not limited to, you can’t hold the lock while doing block placement. {quote} Appreciate your review Daryn. I think it should be easy to address. We will make sure to address the comment before merge? is that make sense. {quote} I’m curious why it isn’t just part of the standard replication monitoring. If the DN is told to replicate to itself, it just does the storage movement. {quote} That's a good question. Overall approach is exactly same as RM. As RM is has its own q build up for redundancy blocks. Underreplication happens at block level. Where as in SPS, policy changes for file, so all blocks in that file needs movement. So, we track the queues at file level. We wanted to provide, on the fly reconfigure feature and we carefully don’t want to interfere replication logic as that is more critical that SPS work. So, as part of impact analysis, we thought keeping SPS separate would be safer than running inn that same loop of RM. > Storage Policy Satisfier in Namenode > > > Key: HDFS-10285 > URL: https://issues.apache.org/jira/browse/HDFS-10285 > Project: Hadoop HDFS > Issue Type: New Feature > Components: datanode, namenode >Affects Versions: HDFS-10285 >Reporter: Uma Maheswara Rao G >Assignee: Uma Maheswara Rao G > Attachments: HDFS-10285-consolidated-merge-patch-00.patch, > HDFS-10285-consolidated-merge-patch-01.patch, > HDFS-10285-consolidated-merge-patch-02.patch, > HDFS-10285-consolidated-merge-patch-03.patch, > HDFS-SPS-TestReport-20170708.pdf, > Storage-Policy-Satisfier-in-HDFS-June-20-2017.pdf, > Storage-Policy-Satisfier-in-HDFS-May10.pdf, > Storage-Policy-Satisfier-in-HDFS-Oct-26-2017.pdf > > > Heterogeneous storage in HDFS introduced the concept of storage policy. These > policies can be set on directory/file to specify the user preference, where > to store the physical block. When user set the storage policy before writing > data, then the blocks could take advantage of storage policy preferences and > stores physical block accordingly. > If user set the storage policy after writing and completing the file, then > the blocks would have been written with default storage policy (nothing but > DISK). User has to run the ‘Mover tool’ explicitly by specifying all such > file names as a list. In some distributed system scenarios (ex: HBase) it > would be difficult to collect all the files and run the tool as different > nodes can write files separately and file can have different paths. > Another scenarios is, when user rename the files from one effected storage > policy file (inherited policy from parent directory) to another storage > policy effected directory, it will not copy inherited storage policy from > source. So it will take effect from destination file/dir parent storage > policy. This rename operation is
[jira] [Commented] (HDFS-12886) Ignore minReplication for block recovery
[ https://issues.apache.org/jira/browse/HDFS-12886?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16283101#comment-16283101 ] Daryn Sharp commented on HDFS-12886: Thanks for the ping, will make time to review in the next few days. Just based on cursory read, agree recovery should be able to succeed regardless of min replication. If the number of recovered/available replicas is less than desired then it’s just under replicated. You can’t wait for min replicas when there physically aren’t that many - think rack loss. Waiting only increases odds of data loss. That said, rarely is anything easy to fix as it seems with block management... > Ignore minReplication for block recovery > > > Key: HDFS-12886 > URL: https://issues.apache.org/jira/browse/HDFS-12886 > Project: Hadoop HDFS > Issue Type: Bug > Components: hdfs, namenode >Reporter: Lukas Majercak >Assignee: Lukas Majercak > Attachments: HDFS-12886.001.patch, HDFS-12886.002.patch > > > Ignore minReplication for blocks that went through recovery, and allow NN to > complete them and replicate. -- This message was sent by Atlassian JIRA (v6.4.14#64029) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-12875) RBF: Complete logic for -readonly option of dfsrouteradmin add command
[ https://issues.apache.org/jira/browse/HDFS-12875?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16283086#comment-16283086 ] genericqa commented on HDFS-12875: -- | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 11s{color} | {color:blue} Docker mode activated. {color} | || || || || {color:brown} Prechecks {color} || | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green} The patch does not contain any @author tags. {color} | | {color:green}+1{color} | {color:green} test4tests {color} | {color:green} 0m 0s{color} | {color:green} The patch appears to include 3 new or modified test files. {color} | || || || || {color:brown} trunk Compile Tests {color} || | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 18m 58s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 1m 0s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 43s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 1m 8s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 12m 15s{color} | {color:green} branch has no errors when building and testing our client artifacts. {color} | | {color:red}-1{color} | {color:red} findbugs {color} | {color:red} 2m 16s{color} | {color:red} hadoop-hdfs-project/hadoop-hdfs in trunk has 1 extant Findbugs warnings. {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 52s{color} | {color:green} trunk passed {color} | || || || || {color:brown} Patch Compile Tests {color} || | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 1m 5s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 56s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 0m 56s{color} | {color:green} the patch passed {color} | | {color:orange}-0{color} | {color:orange} checkstyle {color} | {color:orange} 0m 37s{color} | {color:orange} hadoop-hdfs-project/hadoop-hdfs: The patch generated 1 new + 0 unchanged - 0 fixed = 1 total (was 0) {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 1m 0s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 11m 28s{color} | {color:green} patch has no errors when building and testing our client artifacts. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 2m 11s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 53s{color} | {color:green} the patch passed {color} | || || || || {color:brown} Other Tests {color} || | {color:red}-1{color} | {color:red} unit {color} | {color:red} 97m 9s{color} | {color:red} hadoop-hdfs in the patch failed. {color} | | {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 22s{color} | {color:green} The patch does not generate ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black}152m 48s{color} | {color:black} {color} | \\ \\ || Reason || Tests || | Failed junit tests | hadoop.hdfs.qjournal.server.TestJournalNodeSync | | | hadoop.hdfs.TestDFSStripedOutputStreamWithFailure070 | | | hadoop.hdfs.TestDFSStripedOutputStreamWithFailure090 | | | hadoop.hdfs.TestDFSStripedOutputStreamWithFailure080 | | | hadoop.hdfs.TestDFSStripedInputStream | | | hadoop.hdfs.server.datanode.TestDataNodeErasureCodingMetrics | | | hadoop.hdfs.server.federation.router.TestRouterMountTable | | | hadoop.hdfs.TestLeaseRecovery2 | | | hadoop.hdfs.TestUnsetAndChangeDirectoryEcPolicy | \\ \\ || Subsystem || Report/Notes || | Docker | Client=17.05.0-ce Server=17.05.0-ce Image:yetus/hadoop:5b98639 | | JIRA Issue | HDFS-12875 | | JIRA Patch URL | https://issues.apache.org/jira/secure/attachment/12901187/HDFS-12875.005.patch | | Optional Tests | asflicense compile javac javadoc mvninstall mvnsite unit shadedclient findbugs checkstyle | | uname | Linux 5a9155d45c7f 3.13.0-135-generic #184-Ubuntu SMP Wed Oct 18 11:55:51 UTC 2017 x86_64 x86_64 x86_64 GNU/Linux | | Build tool | maven | | Personality | /testptch/patchprocess/precommit/personality/provided.sh | | git revision | trunk / d6c31a3 | | maven | version:
[jira] [Commented] (HDFS-12882) Support full open(PathHandle) contract in HDFS
[ https://issues.apache.org/jira/browse/HDFS-12882?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16283079#comment-16283079 ] genericqa commented on HDFS-12882: -- | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 22s{color} | {color:blue} Docker mode activated. {color} | || || || || {color:brown} Prechecks {color} || | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green} The patch does not contain any @author tags. {color} | | {color:green}+1{color} | {color:green} test4tests {color} | {color:green} 0m 0s{color} | {color:green} The patch appears to include 13 new or modified test files. {color} | || || || || {color:brown} trunk Compile Tests {color} || | {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 1m 30s{color} | {color:blue} Maven dependency ordering for branch {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 15m 14s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 14m 15s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 1m 56s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 2m 38s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 13m 45s{color} | {color:green} branch has no errors when building and testing our client artifacts. {color} | | {color:red}-1{color} | {color:red} findbugs {color} | {color:red} 1m 53s{color} | {color:red} hadoop-hdfs-project/hadoop-hdfs in trunk has 1 extant Findbugs warnings. {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 2m 13s{color} | {color:green} trunk passed {color} | || || || || {color:brown} Patch Compile Tests {color} || | {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 0m 18s{color} | {color:blue} Maven dependency ordering for patch {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 2m 17s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 13m 19s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} cc {color} | {color:green} 13m 19s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 13m 19s{color} | {color:green} the patch passed {color} | | {color:orange}-0{color} | {color:orange} checkstyle {color} | {color:orange} 2m 3s{color} | {color:orange} root: The patch generated 25 new + 715 unchanged - 10 fixed = 740 total (was 725) {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 2m 39s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} xml {color} | {color:green} 0m 2s{color} | {color:green} The patch has no ill-formed XML file. {color} | | {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 8m 53s{color} | {color:green} patch has no errors when building and testing our client artifacts. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 5m 14s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 2m 13s{color} | {color:green} the patch passed {color} | || || || || {color:brown} Other Tests {color} || | {color:green}+1{color} | {color:green} unit {color} | {color:green} 8m 43s{color} | {color:green} hadoop-common in the patch passed. {color} | | {color:green}+1{color} | {color:green} unit {color} | {color:green} 1m 29s{color} | {color:green} hadoop-hdfs-client in the patch passed. {color} | | {color:red}-1{color} | {color:red} unit {color} | {color:red}118m 21s{color} | {color:red} hadoop-hdfs in the patch failed. {color} | | {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 37s{color} | {color:green} The patch does not generate ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black}220m 42s{color} | {color:black} {color} | \\ \\ || Reason || Tests || | Failed junit tests | hadoop.hdfs.server.namenode.ha.TestRetryCacheWithHA | | | hadoop.hdfs.web.TestWebHdfsTimeouts | | | hadoop.hdfs.TestDFSStripedOutputStreamWithFailure | \\ \\ || Subsystem || Report/Notes || | Docker | Client=17.05.0-ce Server=17.05.0-ce Image:yetus/hadoop:5b98639 | | JIRA Issue | HDFS-12882 | | JIRA
[jira] [Commented] (HDFS-12893) [READ] Support replication of Provided blocks with non-default topologies.
[ https://issues.apache.org/jira/browse/HDFS-12893?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16283065#comment-16283065 ] genericqa commented on HDFS-12893: -- | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 9m 55s{color} | {color:blue} Docker mode activated. {color} | || || || || {color:brown} Prechecks {color} || | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green} The patch does not contain any @author tags. {color} | | {color:green}+1{color} | {color:green} test4tests {color} | {color:green} 0m 0s{color} | {color:green} The patch appears to include 1 new or modified test files. {color} | || || || || {color:brown} HDFS-9806 Compile Tests {color} || | {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 1m 50s{color} | {color:blue} Maven dependency ordering for branch {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 18m 44s{color} | {color:green} HDFS-9806 passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 13m 58s{color} | {color:green} HDFS-9806 passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 2m 13s{color} | {color:green} HDFS-9806 passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 1m 45s{color} | {color:green} HDFS-9806 passed {color} | | {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 14m 29s{color} | {color:green} branch has no errors when building and testing our client artifacts. {color} | | {color:red}-1{color} | {color:red} findbugs {color} | {color:red} 0m 33s{color} | {color:red} hadoop-tools/hadoop-fs2img in HDFS-9806 has 1 extant Findbugs warnings. {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 1m 22s{color} | {color:green} HDFS-9806 passed {color} | || || || || {color:brown} Patch Compile Tests {color} || | {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 0m 16s{color} | {color:blue} Maven dependency ordering for patch {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 1m 33s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 15m 12s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 15m 12s{color} | {color:green} the patch passed {color} | | {color:orange}-0{color} | {color:orange} checkstyle {color} | {color:orange} 2m 33s{color} | {color:orange} root: The patch generated 1 new + 140 unchanged - 0 fixed = 141 total (was 140) {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 1m 48s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 10m 49s{color} | {color:green} patch has no errors when building and testing our client artifacts. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 3m 28s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 1m 28s{color} | {color:green} the patch passed {color} | || || || || {color:brown} Other Tests {color} || | {color:red}-1{color} | {color:red} unit {color} | {color:red} 98m 9s{color} | {color:red} hadoop-hdfs in the patch failed. {color} | | {color:green}+1{color} | {color:green} unit {color} | {color:green} 3m 30s{color} | {color:green} hadoop-fs2img in the patch passed. {color} | | {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 37s{color} | {color:green} The patch does not generate ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black}205m 46s{color} | {color:black} {color} | \\ \\ || Reason || Tests || | Failed junit tests | hadoop.hdfs.server.namenode.TestFsck | | | hadoop.fs.TestUnbuffer | | | hadoop.hdfs.server.blockmanagement.TestUnderReplicatedBlocks | | | hadoop.hdfs.TestLeaseRecovery2 | | | hadoop.hdfs.server.namenode.TestDiskspaceQuotaUpdate | | | hadoop.hdfs.server.namenode.TestAclConfigFlag | | | hadoop.hdfs.server.namenode.TestBackupNode | | | hadoop.hdfs.server.namenode.snapshot.TestUpdatePipelineWithSnapshots | | | hadoop.hdfs.server.blockmanagement.TestBlockStatsMXBean | | | hadoop.hdfs.server.namenode.snapshot.TestSnapshotDeletion | | | hadoop.hdfs.server.namenode.TestFSImageWithSnapshot | \\ \\ || Subsystem || Report/Notes || | Docker | Client=17.05.0-ce Server=17.05.0-ce
[jira] [Commented] (HDFS-12905) [READ] Handle decommissioning and under-maintenance Datanodes with Provided storage.
[ https://issues.apache.org/jira/browse/HDFS-12905?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16283045#comment-16283045 ] genericqa commented on HDFS-12905: -- | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 9m 41s{color} | {color:blue} Docker mode activated. {color} | || || || || {color:brown} Prechecks {color} || | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green} The patch does not contain any @author tags. {color} | | {color:green}+1{color} | {color:green} test4tests {color} | {color:green} 0m 0s{color} | {color:green} The patch appears to include 1 new or modified test files. {color} | || || || || {color:brown} HDFS-9806 Compile Tests {color} || | {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 1m 28s{color} | {color:blue} Maven dependency ordering for branch {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 16m 23s{color} | {color:green} HDFS-9806 passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 12m 43s{color} | {color:green} HDFS-9806 passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 2m 2s{color} | {color:green} HDFS-9806 passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 1m 34s{color} | {color:green} HDFS-9806 passed {color} | | {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 14m 5s{color} | {color:green} branch has no errors when building and testing our client artifacts. {color} | | {color:red}-1{color} | {color:red} findbugs {color} | {color:red} 0m 32s{color} | {color:red} hadoop-tools/hadoop-fs2img in HDFS-9806 has 1 extant Findbugs warnings. {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 1m 19s{color} | {color:green} HDFS-9806 passed {color} | || || || || {color:brown} Patch Compile Tests {color} || | {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 0m 15s{color} | {color:blue} Maven dependency ordering for patch {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 1m 17s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 11m 36s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 11m 36s{color} | {color:green} the patch passed {color} | | {color:orange}-0{color} | {color:orange} checkstyle {color} | {color:orange} 2m 5s{color} | {color:orange} root: The patch generated 3 new + 16 unchanged - 0 fixed = 19 total (was 16) {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 1m 32s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 9m 56s{color} | {color:green} patch has no errors when building and testing our client artifacts. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 2m 45s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 1m 19s{color} | {color:green} the patch passed {color} | || || || || {color:brown} Other Tests {color} || | {color:red}-1{color} | {color:red} unit {color} | {color:red} 88m 42s{color} | {color:red} hadoop-hdfs in the patch failed. {color} | | {color:green}+1{color} | {color:green} unit {color} | {color:green} 2m 44s{color} | {color:green} hadoop-fs2img in the patch passed. {color} | | {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 35s{color} | {color:green} The patch does not generate ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black}183m 45s{color} | {color:black} {color} | \\ \\ || Reason || Tests || | Failed junit tests | hadoop.fs.TestUnbuffer | | | hadoop.hdfs.qjournal.server.TestJournalNodeSync | \\ \\ || Subsystem || Report/Notes || | Docker | Client=17.05.0-ce Server=17.05.0-ce Image:yetus/hadoop:5b98639 | | JIRA Issue | HDFS-12905 | | JIRA Patch URL | https://issues.apache.org/jira/secure/attachment/12901165/HDFS-12905-HDFS-9806.002.patch | | Optional Tests | asflicense compile javac javadoc mvninstall mvnsite unit shadedclient findbugs checkstyle | | uname | Linux cdb25ae44bda 3.13.0-135-generic #184-Ubuntu SMP Wed Oct 18 11:55:51 UTC 2017 x86_64 x86_64 x86_64 GNU/Linux | | Build tool | maven | | Personality | /testptch/patchprocess/precommit/personality/provided.sh | | git revision | HDFS-9806 / 2143767
[jira] [Commented] (HDFS-12907) Allow read-only access to reserved raw for non-superusers
[ https://issues.apache.org/jira/browse/HDFS-12907?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16283037#comment-16283037 ] genericqa commented on HDFS-12907: -- | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 10m 20s{color} | {color:blue} Docker mode activated. {color} | || || || || {color:brown} Prechecks {color} || | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green} The patch does not contain any @author tags. {color} | | {color:green}+1{color} | {color:green} test4tests {color} | {color:green} 0m 0s{color} | {color:green} The patch appears to include 2 new or modified test files. {color} | || || || || {color:brown} trunk Compile Tests {color} || | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 19m 5s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 58s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 43s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 1m 14s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 12m 6s{color} | {color:green} branch has no errors when building and testing our client artifacts. {color} | | {color:red}-1{color} | {color:red} findbugs {color} | {color:red} 2m 14s{color} | {color:red} hadoop-hdfs-project/hadoop-hdfs in trunk has 1 extant Findbugs warnings. {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 51s{color} | {color:green} trunk passed {color} | || || || || {color:brown} Patch Compile Tests {color} || | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 0m 59s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 1m 0s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 1m 0s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 37s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 1m 2s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 10m 46s{color} | {color:green} patch has no errors when building and testing our client artifacts. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 2m 18s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 49s{color} | {color:green} the patch passed {color} | || || || || {color:brown} Other Tests {color} || | {color:red}-1{color} | {color:red} unit {color} | {color:red}102m 7s{color} | {color:red} hadoop-hdfs in the patch failed. {color} | | {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 27s{color} | {color:green} The patch does not generate ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black}167m 17s{color} | {color:black} {color} | \\ \\ || Reason || Tests || | Failed junit tests | hadoop.hdfs.TestErasureCodingPolicies | | | hadoop.hdfs.TestFileChecksum | | | hadoop.fs.TestUnbuffer | | | hadoop.hdfs.TestSafeModeWithStripedFileWithRandomECPolicy | | | hadoop.hdfs.server.blockmanagement.TestBlockStatsMXBean | | | hadoop.hdfs.TestDFSStripedOutputStreamWithFailure050 | | | hadoop.hdfs.server.namenode.TestSecurityTokenEditLog | | | hadoop.hdfs.TestDFSStripedOutputStreamWithFailure080 | | | hadoop.hdfs.TestDFSStripedOutputStreamWithFailure | | | hadoop.hdfs.server.datanode.TestDataNodeHotSwapVolumes | | | hadoop.hdfs.TestDFSStripedOutputStreamWithFailure160 | \\ \\ || Subsystem || Report/Notes || | Docker | Client=17.05.0-ce Server=17.05.0-ce Image:yetus/hadoop:5b98639 | | JIRA Issue | HDFS-12907 | | JIRA Patch URL | https://issues.apache.org/jira/secure/attachment/12901178/HDFS-12907.001.patch | | Optional Tests | asflicense compile javac javadoc mvninstall mvnsite unit shadedclient findbugs checkstyle | | uname | Linux 3d02135f23b4 3.13.0-129-generic #178-Ubuntu SMP Fri Aug 11 12:48:20 UTC 2017 x86_64 x86_64 x86_64 GNU/Linux | | Build tool | maven | | Personality | /testptch/patchprocess/precommit/personality/provided.sh | | git revision | trunk / d6c31a3 | | maven |
[jira] [Commented] (HDFS-12875) RBF: Complete logic for -readonly option of dfsrouteradmin add command
[ https://issues.apache.org/jira/browse/HDFS-12875?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16283034#comment-16283034 ] genericqa commented on HDFS-12875: -- | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 22s{color} | {color:blue} Docker mode activated. {color} | || || || || {color:brown} Prechecks {color} || | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green} The patch does not contain any @author tags. {color} | | {color:green}+1{color} | {color:green} test4tests {color} | {color:green} 0m 0s{color} | {color:green} The patch appears to include 3 new or modified test files. {color} | || || || || {color:brown} trunk Compile Tests {color} || | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 29m 20s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 51s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 36s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 56s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 11m 24s{color} | {color:green} branch has no errors when building and testing our client artifacts. {color} | | {color:red}-1{color} | {color:red} findbugs {color} | {color:red} 1m 52s{color} | {color:red} hadoop-hdfs-project/hadoop-hdfs in trunk has 1 extant Findbugs warnings. {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 49s{color} | {color:green} trunk passed {color} | || || || || {color:brown} Patch Compile Tests {color} || | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 0m 55s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 49s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 0m 49s{color} | {color:green} the patch passed {color} | | {color:orange}-0{color} | {color:orange} checkstyle {color} | {color:orange} 0m 33s{color} | {color:orange} hadoop-hdfs-project/hadoop-hdfs: The patch generated 2 new + 0 unchanged - 0 fixed = 2 total (was 0) {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 54s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 10m 40s{color} | {color:green} patch has no errors when building and testing our client artifacts. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 1m 54s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 48s{color} | {color:green} the patch passed {color} | || || || || {color:brown} Other Tests {color} || | {color:red}-1{color} | {color:red} unit {color} | {color:red} 93m 17s{color} | {color:red} hadoop-hdfs in the patch failed. {color} | | {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 23s{color} | {color:green} The patch does not generate ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black}156m 18s{color} | {color:black} {color} | \\ \\ || Reason || Tests || | Failed junit tests | hadoop.hdfs.server.namenode.TestFSEditLogLoader | | | hadoop.hdfs.TestDFSStripedOutputStreamWithFailureWithRandomECPolicy | | | hadoop.fs.TestUnbuffer | | | hadoop.hdfs.server.namenode.TestReencryption | | | hadoop.hdfs.server.namenode.TestListCorruptFileBlocks | | | hadoop.hdfs.TestPersistBlocks | | | hadoop.hdfs.server.balancer.TestBalancerRPCDelay | | | hadoop.hdfs.server.namenode.TestCacheDirectives | | | hadoop.hdfs.server.namenode.TestLargeDirectoryDelete | | | hadoop.hdfs.server.datanode.TestDirectoryScanner | | | hadoop.hdfs.server.federation.router.TestRouterMountTable | \\ \\ || Subsystem || Report/Notes || | Docker | Client=17.05.0-ce Server=17.05.0-ce Image:yetus/hadoop:5b98639 | | JIRA Issue | HDFS-12875 | | JIRA Patch URL | https://issues.apache.org/jira/secure/attachment/12901181/HDFS-12875.004.patch | | Optional Tests | asflicense compile javac javadoc mvninstall mvnsite unit shadedclient findbugs checkstyle | | uname | Linux 895c67f38cdd 3.13.0-135-generic #184-Ubuntu SMP Wed Oct 18 11:55:51 UTC 2017 x86_64 x86_64 x86_64 GNU/Linux | | Build tool | maven | | Personality |
[jira] [Commented] (HDFS-12907) Allow read-only access to reserved raw for non-superusers
[ https://issues.apache.org/jira/browse/HDFS-12907?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16282860#comment-16282860 ] Rushabh S Shah commented on HDFS-12907: --- Following are the tests that failed. {noformat} TestUnbuffer TestJournalNodeSync TestDataNodeVolumeFailureReporting TestNameNodeXAttr TestWebHDFSXAttr TestFileContextXAttr {noformat} Last 3 tests {{TestNameNodeXAttr,TestWebHDFSXAttr,TestFileContextXAttr}} are failed by my patch. Remaining are just flaky tests. {{TestUnbuffer}} -- Tracked by HADOOP-15056. This jira is almost closed to being resolved. Remaining 2 tests passed locally. {noformat} [INFO] Running org.apache.hadoop.hdfs.qjournal.server.TestJournalNodeSync [INFO] Tests run: 10, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 52.376 s - in org.apache.hadoop.hdfs.qjournal.server.TestJournalNodeSync [INFO] Running org.apache.hadoop.hdfs.server.datanode.TestDataNodeVolumeFailureReporting [INFO] Tests run: 7, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 86.746 s - in org.apache.hadoop.hdfs.server.datanode.TestDataNodeVolumeFailureReporting {noformat} Will attach a new patch soon which fixes the 3 test failures. > Allow read-only access to reserved raw for non-superusers > - > > Key: HDFS-12907 > URL: https://issues.apache.org/jira/browse/HDFS-12907 > Project: Hadoop HDFS > Issue Type: Bug > Components: namenode >Affects Versions: 2.6.0 >Reporter: Daryn Sharp >Assignee: Rushabh S Shah > Attachments: HDFS-12907.patch > > > HDFS-6509 added a special /.reserved/raw path prefix to access the raw file > contents of EZ files. In the simplest sense it doesn't return the FE info in > the {{LocatedBlocks}} so the dfs client doesn't try to decrypt the data. > This facilitates allowing tools like distcp to copy raw bytes. > Access to the raw hierarchy is restricted to superusers. This seems like an > overly broad restriction designed to prevent non-admins from munging the EZ > related xattrs. I believe we should relax the restriction to allow > non-admins to perform read-only operations. Allowing non-superusers to > easily read the raw bytes will be extremely useful for regular users, esp. > for enabling webhdfs client-side encryption. -- This message was sent by Atlassian JIRA (v6.4.14#64029) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-12874) [READ] Documentation for provided storage
[ https://issues.apache.org/jira/browse/HDFS-12874?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Chris Douglas updated HDFS-12874: - Hadoop Flags: Reviewed Status: Patch Available (was: Open) > [READ] Documentation for provided storage > - > > Key: HDFS-12874 > URL: https://issues.apache.org/jira/browse/HDFS-12874 > Project: Hadoop HDFS > Issue Type: Sub-task >Reporter: Chris Douglas >Assignee: Virajith Jalaparti > Attachments: HDFS-12874-HDFS-9806.00.patch, > HDFS-12874-HDFS-9806.01.patch > > > The configuration and deployment of provided storage should be documented for > end-users. -- This message was sent by Atlassian JIRA (v6.4.14#64029) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-12907) Allow read-only access to reserved raw for non-superusers
[ https://issues.apache.org/jira/browse/HDFS-12907?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Rushabh S Shah updated HDFS-12907: -- Attachment: HDFS-12907.001.patch [~daryn], [~andrew.wang]: can you please review. > Allow read-only access to reserved raw for non-superusers > - > > Key: HDFS-12907 > URL: https://issues.apache.org/jira/browse/HDFS-12907 > Project: Hadoop HDFS > Issue Type: Bug > Components: namenode >Affects Versions: 2.6.0 >Reporter: Daryn Sharp >Assignee: Rushabh S Shah > Attachments: HDFS-12907.001.patch, HDFS-12907.patch > > > HDFS-6509 added a special /.reserved/raw path prefix to access the raw file > contents of EZ files. In the simplest sense it doesn't return the FE info in > the {{LocatedBlocks}} so the dfs client doesn't try to decrypt the data. > This facilitates allowing tools like distcp to copy raw bytes. > Access to the raw hierarchy is restricted to superusers. This seems like an > overly broad restriction designed to prevent non-admins from munging the EZ > related xattrs. I believe we should relax the restriction to allow > non-admins to perform read-only operations. Allowing non-superusers to > easily read the raw bytes will be extremely useful for regular users, esp. > for enabling webhdfs client-side encryption. -- This message was sent by Atlassian JIRA (v6.4.14#64029) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Comment Edited] (HDFS-12907) Allow read-only access to reserved raw for non-superusers
[ https://issues.apache.org/jira/browse/HDFS-12907?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16282860#comment-16282860 ] Rushabh S Shah edited comment on HDFS-12907 at 12/8/17 1:10 AM: Following are the tests that failed. {noformat} TestUnbuffer TestJournalNodeSync TestDataNodeVolumeFailureReporting TestNameNodeXAttr TestWebHDFSXAttr TestFileContextXAttr {noformat} Last 3 tests {{TestNameNodeXAttr,TestWebHDFSXAttr,TestFileContextXAttr}} are failed by my patch. Remaining are just flaky tests. {{TestUnbuffer}} -- Tracked by HADOOP-15056. This jira is almost closed to being resolved. Remaining 2 tests passed locally. {noformat} [INFO] Running org.apache.hadoop.hdfs.qjournal.server.TestJournalNodeSync [INFO] Tests run: 10, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 52.376 s - in org.apache.hadoop.hdfs.qjournal.server.TestJournalNodeSync [INFO] Running org.apache.hadoop.hdfs.server.datanode.TestDataNodeVolumeFailureReporting [INFO] Tests run: 7, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 86.746 s - in org.apache.hadoop.hdfs.server.datanode.TestDataNodeVolumeFailureReporting {noformat} Will attach a new patch soon which fixes the 3 test failures. Regarding the findbugs warning, it is not related to my patch. was (Author: shahrs87): Following are the tests that failed. {noformat} TestUnbuffer TestJournalNodeSync TestDataNodeVolumeFailureReporting TestNameNodeXAttr TestWebHDFSXAttr TestFileContextXAttr {noformat} Last 3 tests {{TestNameNodeXAttr,TestWebHDFSXAttr,TestFileContextXAttr}} are failed by my patch. Remaining are just flaky tests. {{TestUnbuffer}} -- Tracked by HADOOP-15056. This jira is almost closed to being resolved. Remaining 2 tests passed locally. {noformat} [INFO] Running org.apache.hadoop.hdfs.qjournal.server.TestJournalNodeSync [INFO] Tests run: 10, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 52.376 s - in org.apache.hadoop.hdfs.qjournal.server.TestJournalNodeSync [INFO] Running org.apache.hadoop.hdfs.server.datanode.TestDataNodeVolumeFailureReporting [INFO] Tests run: 7, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 86.746 s - in org.apache.hadoop.hdfs.server.datanode.TestDataNodeVolumeFailureReporting {noformat} Will attach a new patch soon which fixes the 3 test failures. > Allow read-only access to reserved raw for non-superusers > - > > Key: HDFS-12907 > URL: https://issues.apache.org/jira/browse/HDFS-12907 > Project: Hadoop HDFS > Issue Type: Bug > Components: namenode >Affects Versions: 2.6.0 >Reporter: Daryn Sharp >Assignee: Rushabh S Shah > Attachments: HDFS-12907.001.patch, HDFS-12907.patch > > > HDFS-6509 added a special /.reserved/raw path prefix to access the raw file > contents of EZ files. In the simplest sense it doesn't return the FE info in > the {{LocatedBlocks}} so the dfs client doesn't try to decrypt the data. > This facilitates allowing tools like distcp to copy raw bytes. > Access to the raw hierarchy is restricted to superusers. This seems like an > overly broad restriction designed to prevent non-admins from munging the EZ > related xattrs. I believe we should relax the restriction to allow > non-admins to perform read-only operations. Allowing non-superusers to > easily read the raw bytes will be extremely useful for regular users, esp. > for enabling webhdfs client-side encryption. -- This message was sent by Atlassian JIRA (v6.4.14#64029) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-12875) RBF: Complete logic for -readonly option of dfsrouteradmin add command
[ https://issues.apache.org/jira/browse/HDFS-12875?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Íñigo Goiri updated HDFS-12875: --- Attachment: HDFS-12875.005.patch Cleaned up unit tests. > RBF: Complete logic for -readonly option of dfsrouteradmin add command > -- > > Key: HDFS-12875 > URL: https://issues.apache.org/jira/browse/HDFS-12875 > Project: Hadoop HDFS > Issue Type: Sub-task >Affects Versions: 3.0.0-alpha3 >Reporter: Yiqun Lin >Assignee: Íñigo Goiri > Labels: RBF > Attachments: HDFS-12875.001.patch, HDFS-12875.002.patch, > HDFS-12875.003.patch, HDFS-12875.004.patch, HDFS-12875.005.patch > > > The dfsrouteradmin has an option for readonly mount points but this is not > implemented. We should add an special mount point which allows reading but > not writing. -- This message was sent by Atlassian JIRA (v6.4.14#64029) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-12882) Support full open(PathHandle) contract in HDFS
[ https://issues.apache.org/jira/browse/HDFS-12882?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16282831#comment-16282831 ] genericqa commented on HDFS-12882: -- | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 18s{color} | {color:blue} Docker mode activated. {color} | || || || || {color:brown} Prechecks {color} || | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green} The patch does not contain any @author tags. {color} | | {color:green}+1{color} | {color:green} test4tests {color} | {color:green} 0m 0s{color} | {color:green} The patch appears to include 41 new or modified test files. {color} | || || || || {color:brown} trunk Compile Tests {color} || | {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 0m 17s{color} | {color:blue} Maven dependency ordering for branch {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 14m 25s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 13m 40s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 2m 57s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 3m 54s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 17m 47s{color} | {color:green} branch has no errors when building and testing our client artifacts. {color} | | {color:red}-1{color} | {color:red} findbugs {color} | {color:red} 3m 2s{color} | {color:red} hadoop-hdfs-project/hadoop-hdfs in trunk has 1 extant Findbugs warnings. {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 4m 38s{color} | {color:green} trunk passed {color} | || || || || {color:brown} Patch Compile Tests {color} || | {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 0m 31s{color} | {color:blue} Maven dependency ordering for patch {color} | | {color:red}-1{color} | {color:red} mvninstall {color} | {color:red} 1m 12s{color} | {color:red} hadoop-hdfs-client in the patch failed. {color} | | {color:red}-1{color} | {color:red} mvninstall {color} | {color:red} 0m 53s{color} | {color:red} hadoop-hdfs in the patch failed. {color} | | {color:red}-1{color} | {color:red} mvninstall {color} | {color:red} 0m 27s{color} | {color:red} hadoop-hdfs-nfs in the patch failed. {color} | | {color:red}-1{color} | {color:red} compile {color} | {color:red} 4m 31s{color} | {color:red} root in the patch failed. {color} | | {color:red}-1{color} | {color:red} cc {color} | {color:red} 4m 31s{color} | {color:red} root in the patch failed. {color} | | {color:red}-1{color} | {color:red} javac {color} | {color:red} 4m 31s{color} | {color:red} root in the patch failed. {color} | | {color:orange}-0{color} | {color:orange} checkstyle {color} | {color:orange} 4m 0s{color} | {color:orange} root: The patch generated 34 new + 2069 unchanged - 13 fixed = 2103 total (was 2082) {color} | | {color:red}-1{color} | {color:red} mvnsite {color} | {color:red} 0m 21s{color} | {color:red} hadoop-common in the patch failed. {color} | | {color:red}-1{color} | {color:red} mvnsite {color} | {color:red} 0m 19s{color} | {color:red} hadoop-hdfs-client in the patch failed. {color} | | {color:red}-1{color} | {color:red} mvnsite {color} | {color:red} 0m 19s{color} | {color:red} hadoop-hdfs in the patch failed. {color} | | {color:red}-1{color} | {color:red} mvnsite {color} | {color:red} 0m 26s{color} | {color:red} hadoop-hdfs-nfs in the patch failed. {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} xml {color} | {color:green} 0m 2s{color} | {color:green} The patch has no ill-formed XML file. {color} | | {color:red}-1{color} | {color:red} shadedclient {color} | {color:red} 0m 33s{color} | {color:red} patch has errors when building and testing our client artifacts. {color} | | {color:red}-1{color} | {color:red} findbugs {color} | {color:red} 0m 22s{color} | {color:red} hadoop-common in the patch failed. {color} | | {color:red}-1{color} | {color:red} findbugs {color} | {color:red} 0m 18s{color} | {color:red} hadoop-hdfs-client in the patch failed. {color} | | {color:red}-1{color} | {color:red} findbugs {color} | {color:red} 0m 19s{color} | {color:red} hadoop-hdfs in the patch failed. {color} | | {color:red}-1{color} | {color:red} findbugs {color} | {color:red} 0m 17s{color} | {color:red} hadoop-hdfs-nfs in the patch failed. {color} | | {color:red}-1{color} | {color:red} javadoc {color} | {color:red}
[jira] [Commented] (HDFS-12874) [READ] Documentation for provided storage
[ https://issues.apache.org/jira/browse/HDFS-12874?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16282935#comment-16282935 ] Íñigo Goiri commented on HDFS-12874: For reference, the outcome [here|https://github.com/apache/hadoop/blob/HDFS-9806/hadoop-hdfs-project/hadoop-hdfs/src/site/markdown/HdfsProvidedStorage.md]. > [READ] Documentation for provided storage > - > > Key: HDFS-12874 > URL: https://issues.apache.org/jira/browse/HDFS-12874 > Project: Hadoop HDFS > Issue Type: Sub-task >Reporter: Chris Douglas >Assignee: Virajith Jalaparti > Attachments: HDFS-12874-HDFS-9806.00.patch, > HDFS-12874-HDFS-9806.01.patch > > > The configuration and deployment of provided storage should be documented for > end-users. -- This message was sent by Atlassian JIRA (v6.4.14#64029) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-12893) [READ] Support replication of Provided blocks with non-default topologies.
[ https://issues.apache.org/jira/browse/HDFS-12893?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16282924#comment-16282924 ] Íñigo Goiri commented on HDFS-12893: I like this better. For this: {code} if (storage.getStorageType() == StorageType.PROVIDED) { storage = new DatanodeStorageInfo(node, storage.getStorageID(), storage.getStorageType(), storage.getState()); } {code} You are pretty much just copying the storage. Should we have a DatanodeProvidedInfo which just gets: {code} public class DatanodeProvidedInfo extends DatanodeStorageInfo { public DatanodeProvidedInfo (node, storage) { super(node, storage.getStorageID(), StorageType.PROVIDED, storage.getState()); } } {code} This may allow adding a couple more overrides. > [READ] Support replication of Provided blocks with non-default topologies. > -- > > Key: HDFS-12893 > URL: https://issues.apache.org/jira/browse/HDFS-12893 > Project: Hadoop HDFS > Issue Type: Sub-task >Reporter: Virajith Jalaparti >Assignee: Virajith Jalaparti > Attachments: HDFS-12893-HDFS-9806.001.patch, > HDFS-12893-HDFS-9806.002.patch, HDFS-12893-HDFS-9806.003.patch > > > {{chooseSourceDatanodes}} returns the {{ProvidedDatanodeDescriptor}} as the > source of Provided blocks. As this isn't a physical datanode and doesn't > exist the topology, {{ReplicationWork.chooseTargets}} might fail depending on > the chosen {{BlockPlacementPolicy}} implementation. This JIRA aims to fix > this issue. -- This message was sent by Atlassian JIRA (v6.4.14#64029) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-12882) Support full open(PathHandle) contract in HDFS
[ https://issues.apache.org/jira/browse/HDFS-12882?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16282916#comment-16282916 ] Íñigo Goiri commented on HDFS-12882: The approach in [^HDFS-12882.04.patch] is cleaner. Can we document a little more what the token means? One has to go all the way to {{FSNamesystem}} to track it. > Support full open(PathHandle) contract in HDFS > -- > > Key: HDFS-12882 > URL: https://issues.apache.org/jira/browse/HDFS-12882 > Project: Hadoop HDFS > Issue Type: Improvement > Components: hdfs-client >Reporter: Chris Douglas >Assignee: Chris Douglas > Attachments: HDFS-12882.00.patch, HDFS-12882.00.salient.txt, > HDFS-12882.01.patch, HDFS-12882.02.patch, HDFS-12882.03.patch, > HDFS-12882.04.patch > > > HDFS-7878 added support for {{open(PathHandle)}}, but it only partially > implemented the semantics specified in the contract (i.e., open-by-inodeID). > HDFS should implement all permutations of the default options for > {{PathHandle}}. -- This message was sent by Atlassian JIRA (v6.4.14#64029) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-12875) RBF: Complete logic for -readonly option of dfsrouteradmin add command
[ https://issues.apache.org/jira/browse/HDFS-12875?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Íñigo Goiri updated HDFS-12875: --- Attachment: HDFS-12875.004.patch > RBF: Complete logic for -readonly option of dfsrouteradmin add command > -- > > Key: HDFS-12875 > URL: https://issues.apache.org/jira/browse/HDFS-12875 > Project: Hadoop HDFS > Issue Type: Sub-task >Affects Versions: 3.0.0-alpha3 >Reporter: Yiqun Lin >Assignee: Íñigo Goiri > Labels: RBF > Attachments: HDFS-12875.001.patch, HDFS-12875.002.patch, > HDFS-12875.003.patch, HDFS-12875.004.patch > > > The dfsrouteradmin has an option for readonly mount points but this is not > implemented. We should add an special mount point which allows reading but > not writing. -- This message was sent by Atlassian JIRA (v6.4.14#64029) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-12893) [READ] Support replication of Provided blocks with non-default topologies.
[ https://issues.apache.org/jira/browse/HDFS-12893?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Virajith Jalaparti updated HDFS-12893: -- Attachment: HDFS-12893-HDFS-9806.003.patch > [READ] Support replication of Provided blocks with non-default topologies. > -- > > Key: HDFS-12893 > URL: https://issues.apache.org/jira/browse/HDFS-12893 > Project: Hadoop HDFS > Issue Type: Sub-task >Reporter: Virajith Jalaparti >Assignee: Virajith Jalaparti > Attachments: HDFS-12893-HDFS-9806.001.patch, > HDFS-12893-HDFS-9806.002.patch, HDFS-12893-HDFS-9806.003.patch > > > {{chooseSourceDatanodes}} returns the {{ProvidedDatanodeDescriptor}} as the > source of Provided blocks. As this isn't a physical datanode and doesn't > exist the topology, {{ReplicationWork.chooseTargets}} might fail depending on > the chosen {{BlockPlacementPolicy}} implementation. This JIRA aims to fix > this issue. -- This message was sent by Atlassian JIRA (v6.4.14#64029) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-12893) [READ] Support replication of Provided blocks with non-default topologies.
[ https://issues.apache.org/jira/browse/HDFS-12893?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Virajith Jalaparti updated HDFS-12893: -- Status: Open (was: Patch Available) > [READ] Support replication of Provided blocks with non-default topologies. > -- > > Key: HDFS-12893 > URL: https://issues.apache.org/jira/browse/HDFS-12893 > Project: Hadoop HDFS > Issue Type: Sub-task >Reporter: Virajith Jalaparti >Assignee: Virajith Jalaparti > Attachments: HDFS-12893-HDFS-9806.001.patch, > HDFS-12893-HDFS-9806.002.patch, HDFS-12893-HDFS-9806.003.patch > > > {{chooseSourceDatanodes}} returns the {{ProvidedDatanodeDescriptor}} as the > source of Provided blocks. As this isn't a physical datanode and doesn't > exist the topology, {{ReplicationWork.chooseTargets}} might fail depending on > the chosen {{BlockPlacementPolicy}} implementation. This JIRA aims to fix > this issue. -- This message was sent by Atlassian JIRA (v6.4.14#64029) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-12874) [READ] Documentation for provided storage
[ https://issues.apache.org/jira/browse/HDFS-12874?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Chris Douglas updated HDFS-12874: - Resolution: Fixed Status: Resolved (was: Patch Available) I committed this. Thanks Virajith > [READ] Documentation for provided storage > - > > Key: HDFS-12874 > URL: https://issues.apache.org/jira/browse/HDFS-12874 > Project: Hadoop HDFS > Issue Type: Sub-task >Reporter: Chris Douglas >Assignee: Virajith Jalaparti > Attachments: HDFS-12874-HDFS-9806.00.patch, > HDFS-12874-HDFS-9806.01.patch > > > The configuration and deployment of provided storage should be documented for > end-users. -- This message was sent by Atlassian JIRA (v6.4.14#64029) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-12874) [READ] Documentation for provided storage
[ https://issues.apache.org/jira/browse/HDFS-12874?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16282904#comment-16282904 ] genericqa commented on HDFS-12874: -- | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 26s{color} | {color:blue} Docker mode activated. {color} | | {color:red}-1{color} | {color:red} patch {color} | {color:red} 0m 6s{color} | {color:red} HDFS-12874 does not apply to HDFS-9806. Rebase required? Wrong Branch? See https://wiki.apache.org/hadoop/HowToContribute for help. {color} | \\ \\ || Subsystem || Report/Notes || | Docker | Client=17.05.0-ce Server=17.05.0-ce Image:yetus/hadoop:5b98639 | | JIRA Issue | HDFS-12874 | | JIRA Patch URL | https://issues.apache.org/jira/secure/attachment/12901143/HDFS-12874-HDFS-9806.01.patch | | Console output | https://builds.apache.org/job/PreCommit-HDFS-Build/22325/console | | Powered by | Apache Yetus 0.7.0-SNAPSHOT http://yetus.apache.org | This message was automatically generated. > [READ] Documentation for provided storage > - > > Key: HDFS-12874 > URL: https://issues.apache.org/jira/browse/HDFS-12874 > Project: Hadoop HDFS > Issue Type: Sub-task >Reporter: Chris Douglas >Assignee: Virajith Jalaparti > Attachments: HDFS-12874-HDFS-9806.00.patch, > HDFS-12874-HDFS-9806.01.patch > > > The configuration and deployment of provided storage should be documented for > end-users. -- This message was sent by Atlassian JIRA (v6.4.14#64029) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-12905) [READ] Handle decommissioning and under-maintenance Datanodes with Provided storage.
[ https://issues.apache.org/jira/browse/HDFS-12905?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Chris Douglas updated HDFS-12905: - Status: Patch Available (was: Open) > [READ] Handle decommissioning and under-maintenance Datanodes with Provided > storage. > > > Key: HDFS-12905 > URL: https://issues.apache.org/jira/browse/HDFS-12905 > Project: Hadoop HDFS > Issue Type: Sub-task >Reporter: Virajith Jalaparti >Assignee: Virajith Jalaparti > Attachments: HDFS-12905-HDFS-9806.001.patch, > HDFS-12905-HDFS-9806.002.patch > > > {{ProvidedStorageMap}} doesn't keep track of the state of the datanodes with > Provided storage. As a result, it can return nodes that are being > decommissioned or under-maintenance even when live datanodes exist. This JIRA > is to prefer live datanodes to datanodes in other states. -- This message was sent by Atlassian JIRA (v6.4.14#64029) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-11915) Sync rbw dir on the first hsync() to avoid file lost on power failure
[ https://issues.apache.org/jira/browse/HDFS-11915?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16282867#comment-16282867 ] genericqa commented on HDFS-11915: -- | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 24m 32s{color} | {color:blue} Docker mode activated. {color} | || || || || {color:brown} Prechecks {color} || | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green} The patch does not contain any @author tags. {color} | | {color:red}-1{color} | {color:red} test4tests {color} | {color:red} 0m 0s{color} | {color:red} The patch doesn't appear to include any new or modified tests. Please justify why no new tests are needed for this patch. Also please list what manual steps were performed to verify this patch. {color} | || || || || {color:brown} branch-2 Compile Tests {color} || | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 9m 43s{color} | {color:green} branch-2 passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 47s{color} | {color:green} branch-2 passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 29s{color} | {color:green} branch-2 passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 56s{color} | {color:green} branch-2 passed {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 1m 59s{color} | {color:green} branch-2 passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 1m 11s{color} | {color:green} branch-2 passed {color} | || || || || {color:brown} Patch Compile Tests {color} || | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 0m 49s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 45s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 0m 45s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 23s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 53s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 2m 13s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 1m 3s{color} | {color:green} the patch passed {color} | || || || || {color:brown} Other Tests {color} || | {color:red}-1{color} | {color:red} unit {color} | {color:red} 91m 58s{color} | {color:red} hadoop-hdfs in the patch failed. {color} | | {color:red}-1{color} | {color:red} asflicense {color} | {color:red} 1m 6s{color} | {color:red} The patch generated 154 ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black}140m 29s{color} | {color:black} {color} | \\ \\ || Reason || Tests || | Unreaped Processes | hadoop-hdfs:22 | | Failed junit tests | hadoop.hdfs.TestEncryptedTransfer | | | hadoop.hdfs.web.TestWebHdfsTimeouts | | | hadoop.hdfs.TestDFSClientSocketSize | | | hadoop.metrics2.sink.TestRollingFileSystemSinkWithSecureHdfs | | | hadoop.hdfs.TestListFilesInDFS | | | hadoop.hdfs.TestSetrepDecreasing | | | hadoop.hdfs.TestParallelUnixDomainRead | | | hadoop.hdfs.TestHDFSServerPorts | | | hadoop.hdfs.TestMaintenanceState | | | hadoop.hdfs.TestFileConcurrentReader | | Timed out junit tests | org.apache.hadoop.hdfs.TestFileAppend | | | org.apache.hadoop.hdfs.TestSafeMode | | | org.apache.hadoop.hdfs.TestRollingUpgradeDowngrade | | | org.apache.hadoop.hdfs.web.TestWebHdfsWithRestCsrfPreventionFilter | | | org.apache.hadoop.hdfs.TestDFSUpgrade | | | org.apache.hadoop.hdfs.web.TestWebHDFS | | | org.apache.hadoop.hdfs.web.TestWebHDFSXAttr | | | org.apache.hadoop.hdfs.web.TestWebHdfsWithMultipleNameNodes | | | org.apache.hadoop.hdfs.TestRenameWhileOpen | | | org.apache.hadoop.metrics2.sink.TestRollingFileSystemSinkWithHdfs | | | org.apache.hadoop.hdfs.TestFSOutputSummer | | | org.apache.hadoop.hdfs.TestExternalBlockReader | | | org.apache.hadoop.hdfs.TestHFlush | | | org.apache.hadoop.hdfs.TestTrashWithEncryptionZones | | | org.apache.hadoop.hdfs.TestDFSShell | | | org.apache.hadoop.hdfs.TestReplaceDatanodeFailureReplication | | | org.apache.hadoop.hdfs.TestDFSRename | | | org.apache.hadoop.hdfs.web.TestWebHDFSAcl | \\ \\ || Subsystem || Report/Notes || | Docker |
[jira] [Commented] (HDFS-12893) [READ] Support replication of Provided blocks with non-default topologies.
[ https://issues.apache.org/jira/browse/HDFS-12893?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16282889#comment-16282889 ] Virajith Jalaparti commented on HDFS-12893: --- [~elgoiri] - patch v3 wraps the choosing of the DatanodeDescriptor from the storage. > [READ] Support replication of Provided blocks with non-default topologies. > -- > > Key: HDFS-12893 > URL: https://issues.apache.org/jira/browse/HDFS-12893 > Project: Hadoop HDFS > Issue Type: Sub-task >Reporter: Virajith Jalaparti >Assignee: Virajith Jalaparti > Attachments: HDFS-12893-HDFS-9806.001.patch, > HDFS-12893-HDFS-9806.002.patch, HDFS-12893-HDFS-9806.003.patch > > > {{chooseSourceDatanodes}} returns the {{ProvidedDatanodeDescriptor}} as the > source of Provided blocks. As this isn't a physical datanode and doesn't > exist the topology, {{ReplicationWork.chooseTargets}} might fail depending on > the chosen {{BlockPlacementPolicy}} implementation. This JIRA aims to fix > this issue. -- This message was sent by Atlassian JIRA (v6.4.14#64029) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-12893) [READ] Support replication of Provided blocks with non-default topologies.
[ https://issues.apache.org/jira/browse/HDFS-12893?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Virajith Jalaparti updated HDFS-12893: -- Status: Patch Available (was: Open) > [READ] Support replication of Provided blocks with non-default topologies. > -- > > Key: HDFS-12893 > URL: https://issues.apache.org/jira/browse/HDFS-12893 > Project: Hadoop HDFS > Issue Type: Sub-task >Reporter: Virajith Jalaparti >Assignee: Virajith Jalaparti > Attachments: HDFS-12893-HDFS-9806.001.patch, > HDFS-12893-HDFS-9806.002.patch, HDFS-12893-HDFS-9806.003.patch > > > {{chooseSourceDatanodes}} returns the {{ProvidedDatanodeDescriptor}} as the > source of Provided blocks. As this isn't a physical datanode and doesn't > exist the topology, {{ReplicationWork.chooseTargets}} might fail depending on > the chosen {{BlockPlacementPolicy}} implementation. This JIRA aims to fix > this issue. -- This message was sent by Atlassian JIRA (v6.4.14#64029) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-12825) Fsck report shows config key name for min replication issues
[ https://issues.apache.org/jira/browse/HDFS-12825?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Manoj Govindassamy updated HDFS-12825: -- Summary: Fsck report shows config key name for min replication issues (was: After Block Corrupted, FSCK Report printing the Direct configuration. ) > Fsck report shows config key name for min replication issues > > > Key: HDFS-12825 > URL: https://issues.apache.org/jira/browse/HDFS-12825 > Project: Hadoop HDFS > Issue Type: Bug > Components: hdfs >Affects Versions: 3.0.0-alpha1 >Reporter: Harshakiran Reddy >Assignee: Gabor Bota >Priority: Minor > Labels: newbie > Attachments: HDFS-12825.001.patch, error.JPG > > > Scenario: > Corrupt the Block in any datanode > Take the *FSCK *Report for that file. > Actual Output: > == > printing the direct configuration in fsck report > {{dfs.namenode.replication.min}} > Expected Output: > > it should be {{MINIMAL BLOCK REPLICATION}} -- This message was sent by Atlassian JIRA (v6.4.14#64029) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-12907) Allow read-only access to reserved raw for non-superusers
[ https://issues.apache.org/jira/browse/HDFS-12907?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16282868#comment-16282868 ] Daryn Sharp commented on HDFS-12907: bq. My understanding is after HDFS-12355 \[...\] , Is this remotely correct? No, the goal is all encryption/decryption is done on the client. The DN will not be given KMS tokens. Ever. It will not talk to the KMS. Ever. The DN will never encrypt/decrypt. The KMS client completely breaks all the ugi-semantics to enable DNs to do the encrypt/decrypt. How? Why? First off, the KMS client morphs based on the caller's ugi context. Clients are expected to always be who they were when created. Imagine if an IPC client was user1, disconnected, and magically became user2 just because it was in another context. That's the KMS client. It gets worse. If the current ugi is a proxy user, the KMS client will try to authenticate as the real user. That's fine when the ugi real user is the login user of a service (ex. oozie). But if there are _no credentials_, ex. proxy ugi from a token, the client willfully decides to use the login user's credentials! And for good measure, let's proxy as the effective ugi! That's super bizarre. So let's put it together with the DN. I use webhdfs as "daryn (token)", but the DN connects to the KMS as "daryn via dn (kerberos)". Or I submit a job with oozie, so I'm "daryn via oozie (token)" to the DN but it connects to the KMS as "daryn via dn (kerberos)". Wow, that would never work right? It would it you told users to make all their DNs be proxy users on the KMS! And since most people map their DNs to the hdfs superuser, which is a really bad idea, you now have let admins have the ability to decrypt any file. Both Cloudera and Hortonworks actually documented this security insanity. Cloudera's docs appear to be gone now, but used to acknowledge with a yellow box like "this is a bad idea, but if you really want to...". HortonWorks docs still exist with a footnote like "oh yeah, if you are still paying attention after clicking all the ui buttons, all your nodes now have access to all your keys, might want to consider changing your superuser". If you allow every node in your cluster the ability to decrypt everything on your cluster. Why did you even enable security let alone EZ? It's a rotten idea that should have never been implemented or passed a review. It's what happens when a feature is rushed. –– Phew. I value my security and data. I'm sure as hell not making my DNs be proxy users, but we're stuck not breaking all the people that go to sleep with a false sense of cluster security. So in the new design in progress, the DN is used as a dumb passthrough of encrypted bytes. Never encrypts/decrypts or even talks to the KMS. It can be done in a compatible way by the client sending a header to the NN that it knows how to handle EZ. A new NN gives back the feinfo and prefaces the redirect path with /.reserved/raw. That works across both old and new nodes and clusters. Should be beautiful. Stay tuned. > Allow read-only access to reserved raw for non-superusers > - > > Key: HDFS-12907 > URL: https://issues.apache.org/jira/browse/HDFS-12907 > Project: Hadoop HDFS > Issue Type: Bug > Components: namenode >Affects Versions: 2.6.0 >Reporter: Daryn Sharp >Assignee: Rushabh S Shah > Attachments: HDFS-12907.001.patch, HDFS-12907.patch > > > HDFS-6509 added a special /.reserved/raw path prefix to access the raw file > contents of EZ files. In the simplest sense it doesn't return the FE info in > the {{LocatedBlocks}} so the dfs client doesn't try to decrypt the data. > This facilitates allowing tools like distcp to copy raw bytes. > Access to the raw hierarchy is restricted to superusers. This seems like an > overly broad restriction designed to prevent non-admins from munging the EZ > related xattrs. I believe we should relax the restriction to allow > non-admins to perform read-only operations. Allowing non-superusers to > easily read the raw bytes will be extremely useful for regular users, esp. > for enabling webhdfs client-side encryption. -- This message was sent by Atlassian JIRA (v6.4.14#64029) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-12825) Fsck report shows config key name for min replication issues
[ https://issues.apache.org/jira/browse/HDFS-12825?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Manoj Govindassamy updated HDFS-12825: -- Labels: incompatibleChange newbie (was: newbie) > Fsck report shows config key name for min replication issues > > > Key: HDFS-12825 > URL: https://issues.apache.org/jira/browse/HDFS-12825 > Project: Hadoop HDFS > Issue Type: Bug > Components: hdfs >Affects Versions: 3.0.0-alpha1 >Reporter: Harshakiran Reddy >Assignee: Gabor Bota >Priority: Minor > Labels: incompatibleChange, newbie > Attachments: HDFS-12825.001.patch, error.JPG > > > Scenario: > Corrupt the Block in any datanode > Take the *FSCK *Report for that file. > Actual Output: > == > printing the direct configuration in fsck report > {{dfs.namenode.replication.min}} > Expected Output: > > it should be {{MINIMAL BLOCK REPLICATION}} -- This message was sent by Atlassian JIRA (v6.4.14#64029) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-12874) [READ] Documentation for provided storage
[ https://issues.apache.org/jira/browse/HDFS-12874?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16282854#comment-16282854 ] Chris Douglas commented on HDFS-12874: -- +1 I prefer this version. > [READ] Documentation for provided storage > - > > Key: HDFS-12874 > URL: https://issues.apache.org/jira/browse/HDFS-12874 > Project: Hadoop HDFS > Issue Type: Sub-task >Reporter: Chris Douglas > Attachments: HDFS-12874-HDFS-9806.00.patch, > HDFS-12874-HDFS-9806.01.patch > > > The configuration and deployment of provided storage should be documented for > end-users. -- This message was sent by Atlassian JIRA (v6.4.14#64029) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Assigned] (HDFS-12874) [READ] Documentation for provided storage
[ https://issues.apache.org/jira/browse/HDFS-12874?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Chris Douglas reassigned HDFS-12874: Assignee: Virajith Jalaparti > [READ] Documentation for provided storage > - > > Key: HDFS-12874 > URL: https://issues.apache.org/jira/browse/HDFS-12874 > Project: Hadoop HDFS > Issue Type: Sub-task >Reporter: Chris Douglas >Assignee: Virajith Jalaparti > Attachments: HDFS-12874-HDFS-9806.00.patch, > HDFS-12874-HDFS-9806.01.patch > > > The configuration and deployment of provided storage should be documented for > end-users. -- This message was sent by Atlassian JIRA (v6.4.14#64029) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-12882) Support full open(PathHandle) contract in HDFS
[ https://issues.apache.org/jira/browse/HDFS-12882?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Chris Douglas updated HDFS-12882: - Attachment: HDFS-12882.04.patch Alternative approach, adding {{ClientProtocol::getLocatedFileInfo}} > Support full open(PathHandle) contract in HDFS > -- > > Key: HDFS-12882 > URL: https://issues.apache.org/jira/browse/HDFS-12882 > Project: Hadoop HDFS > Issue Type: Improvement > Components: hdfs-client >Reporter: Chris Douglas >Assignee: Chris Douglas > Attachments: HDFS-12882.00.patch, HDFS-12882.00.salient.txt, > HDFS-12882.01.patch, HDFS-12882.02.patch, HDFS-12882.03.patch, > HDFS-12882.04.patch > > > HDFS-7878 added support for {{open(PathHandle)}}, but it only partially > implemented the semantics specified in the contract (i.e., open-by-inodeID). > HDFS should implement all permutations of the default options for > {{PathHandle}}. -- This message was sent by Atlassian JIRA (v6.4.14#64029) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-12893) [READ] Support replication of Provided blocks with non-default topologies.
[ https://issues.apache.org/jira/browse/HDFS-12893?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16282815#comment-16282815 ] Íñigo Goiri commented on HDFS-12893: Can we wrap the functions where we return the regular descriptor or the random one? The creation of the new PROVIDED DN could also be wrapped. > [READ] Support replication of Provided blocks with non-default topologies. > -- > > Key: HDFS-12893 > URL: https://issues.apache.org/jira/browse/HDFS-12893 > Project: Hadoop HDFS > Issue Type: Sub-task >Reporter: Virajith Jalaparti >Assignee: Virajith Jalaparti > Attachments: HDFS-12893-HDFS-9806.001.patch, > HDFS-12893-HDFS-9806.002.patch > > > {{chooseSourceDatanodes}} returns the {{ProvidedDatanodeDescriptor}} as the > source of Provided blocks. As this isn't a physical datanode and doesn't > exist the topology, {{ReplicationWork.chooseTargets}} might fail depending on > the chosen {{BlockPlacementPolicy}} implementation. This JIRA aims to fix > this issue. -- This message was sent by Atlassian JIRA (v6.4.14#64029) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-12907) Allow read-only access to reserved raw for non-superusers
[ https://issues.apache.org/jira/browse/HDFS-12907?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16282559#comment-16282559 ] Rushabh S Shah commented on HDFS-12907: --- bq. On the topic of webhdfs client-side encryption, could you talk a little more about your usecase? The use case is read/write to encrypted directory via WebhdfsfileSystem. If we follow the current community implementation, we have to create a new user to run datanode as (lets say {{dn}} separate from the user with which the namenode runs: {{hdfs}}). Also we need to whitelist {{dn}} user to be able to decrypt {{edek}}. If someone gets access to {{dn}} user, they can easily decrypt all client's edek. > Allow read-only access to reserved raw for non-superusers > - > > Key: HDFS-12907 > URL: https://issues.apache.org/jira/browse/HDFS-12907 > Project: Hadoop HDFS > Issue Type: Bug > Components: namenode >Affects Versions: 2.6.0 >Reporter: Daryn Sharp >Assignee: Rushabh S Shah > Attachments: HDFS-12907.patch > > > HDFS-6509 added a special /.reserved/raw path prefix to access the raw file > contents of EZ files. In the simplest sense it doesn't return the FE info in > the {{LocatedBlocks}} so the dfs client doesn't try to decrypt the data. > This facilitates allowing tools like distcp to copy raw bytes. > Access to the raw hierarchy is restricted to superusers. This seems like an > overly broad restriction designed to prevent non-admins from munging the EZ > related xattrs. I believe we should relax the restriction to allow > non-admins to perform read-only operations. Allowing non-superusers to > easily read the raw bytes will be extremely useful for regular users, esp. > for enabling webhdfs client-side encryption. -- This message was sent by Atlassian JIRA (v6.4.14#64029) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-12905) [READ] Handle decommissioning and under-maintenance Datanodes with Provided storage.
[ https://issues.apache.org/jira/browse/HDFS-12905?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16282801#comment-16282801 ] Íñigo Goiri commented on HDFS-12905: Thanks [~virajith], I personally prefer the approach in [^HDFS-12905-HDFS-9806.002.patch]. +1 on it pending Jenkins. > [READ] Handle decommissioning and under-maintenance Datanodes with Provided > storage. > > > Key: HDFS-12905 > URL: https://issues.apache.org/jira/browse/HDFS-12905 > Project: Hadoop HDFS > Issue Type: Sub-task >Reporter: Virajith Jalaparti >Assignee: Virajith Jalaparti > Attachments: HDFS-12905-HDFS-9806.001.patch, > HDFS-12905-HDFS-9806.002.patch > > > {{ProvidedStorageMap}} doesn't keep track of the state of the datanodes with > Provided storage. As a result, it can return nodes that are being > decommissioned or under-maintenance even when live datanodes exist. This JIRA > is to prefer live datanodes to datanodes in other states. -- This message was sent by Atlassian JIRA (v6.4.14#64029) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Comment Edited] (HDFS-12905) [READ] Handle decommissioning and under-maintenance Datanodes with Provided storage.
[ https://issues.apache.org/jira/browse/HDFS-12905?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16282794#comment-16282794 ] Virajith Jalaparti edited comment on HDFS-12905 at 12/8/17 12:23 AM: - A 2nd attempt at this following offline discussions with [~elgoiri] - the concern with patch v1 was the changes involved in the {{HeartBeatManager}}. This creates an implicit dependency between the {{ProvidedStorageMap}} and the {{HeartBeatManager}}, and any future changes to the {{HeartBeatManager}} has to reason about how it affects the {{ProvidedStorageMap}}. Patch v2 limits the changes to the {{ProvidedStorageMap}}, which is now fully responsible to reason about the state of the Datanodes. This increases the complexity in the implementation of the {{ProvidedStorageMap}}. In particular, patch v2 has a potentially higher compute complexity -- to choose a random node, it first looks for nodes that are in {{AdminStates.NORMAL}} state. If none are found, it then tries to find nodes whose state is not {{AdminStates.NORMAL}}. An alternate approach could be to maintain 2 lists, one for nodes whose state is {{AdminStates.NORMAL}} and one for nodes whose state is not {{AdminStates.NORMAL}}. While this would be more efficient at choosing nodes (avoid looking at a node twice), comes with higher bookkeeping costs (e.g., use a background thread to move Datanodes between these lists). For now, I think this approach seems reasonable assuming most Datanodes are in the state {{AdminStates.NORMAL}} and few are not. Once HDFS-12848 is completed, this could be made pluggable and other efficient implementations are possible. [~elgoiri], [~chris.douglas] - can you take a look? was (Author: virajith): A 2nd attempt at this following offline discussions with [~elgoiri] - the concern with patch v1 was the changes involved in the {{HeartBeatManager}}. This creates an implicit dependency between the {{ProvidedStorageMap}} and the {{HeartBeatManager}}, and any future changes to the {{HeartBeatManager}} has to reason about how it affects the {{ProvidedStorageMap}}. Patch v2 limits the changes to the {{ProvidedStorageMap}}, which is now responsible to reason about Datanodes whose state is not {{AdminStates.NORMAL}}. This increases the complexity in the implementation of the {{ProvidedStorageMap}}. In particular, patch v2 has a potentially higher compute complexity -- to choose a random node, it first looks for nodes that are in {{AdminStates.NORMAL}} state. If none are found, it then tries to find nodes whose state is not {{AdminStates.NORMAL}}. An alternate approach could be to maintain 2 lists, one for nodes whose state is {{AdminStates.NORMAL}} and one for nodes whose state is not {{AdminStates.NORMAL}}. While this would be more efficient at choosing nodes (avoid looking at a node twice), leads to more complex bookkeeping (e.g., use a background thread to move Datanodes between these lists). For now, I think this approach seems reasonable assuming most Datanodes are in the state {{AdminStates.NORMAL}} and few are not. Once HDFS-12848 is completed, this could be made pluggable and other efficient implementations are possible. [~elgoiri], [~chris.douglas] - can you take a look? > [READ] Handle decommissioning and under-maintenance Datanodes with Provided > storage. > > > Key: HDFS-12905 > URL: https://issues.apache.org/jira/browse/HDFS-12905 > Project: Hadoop HDFS > Issue Type: Sub-task >Reporter: Virajith Jalaparti >Assignee: Virajith Jalaparti > Attachments: HDFS-12905-HDFS-9806.001.patch, > HDFS-12905-HDFS-9806.002.patch > > > {{ProvidedStorageMap}} doesn't keep track of the state of the datanodes with > Provided storage. As a result, it can return nodes that are being > decommissioned or under-maintenance even when live datanodes exist. This JIRA > is to prefer live datanodes to datanodes in other states. -- This message was sent by Atlassian JIRA (v6.4.14#64029) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-12905) [READ] Handle decommissioning and under-maintenance Datanodes with Provided storage.
[ https://issues.apache.org/jira/browse/HDFS-12905?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16282794#comment-16282794 ] Virajith Jalaparti commented on HDFS-12905: --- A 2nd attempt at this following offline discussions with [~elgoiri] - the concern with patch v1 was the changes involved in the {{HeartBeatManager}}. This creates an implicit dependency between the {{ProvidedStorageMap}} and the {{HeartBeatManager}}, and any future changes to the {{HeartBeatManager}} has to reason about how it affects the {{ProvidedStorageMap}}. Patch v2 limits the changes to the {{ProvidedStorageMap}}, which is now responsible to reason about Datanodes whose state is not {{AdminStates.NORMAL}}. This increases the complexity in the implementation of the {{ProvidedStorageMap}}. In particular, patch v2 has a potentially higher compute complexity -- to choose a random node, it first looks for nodes that are in {{AdminStates.NORMAL}} state. If none are found, it then tries to find nodes whose state is not {{AdminStates.NORMAL}}. An alternate approach could be to maintain 2 lists, one for nodes whose state is {{AdminStates.NORMAL}} and one for nodes whose state is not {{AdminStates.NORMAL}}. While this would be more efficient at choosing nodes (avoid looking at a node twice), leads to more complex bookkeeping (e.g., use a background thread to move Datanodes between these lists). For now, I think this approach seems reasonable assuming most Datanodes are in the state {{AdminStates.NORMAL}} and few are not. Once HDFS-12848 is completed, this could be made pluggable and other efficient implementations are possible. [~elgoiri], [~chris.douglas] - can you take a look? > [READ] Handle decommissioning and under-maintenance Datanodes with Provided > storage. > > > Key: HDFS-12905 > URL: https://issues.apache.org/jira/browse/HDFS-12905 > Project: Hadoop HDFS > Issue Type: Sub-task >Reporter: Virajith Jalaparti >Assignee: Virajith Jalaparti > Attachments: HDFS-12905-HDFS-9806.001.patch, > HDFS-12905-HDFS-9806.002.patch > > > {{ProvidedStorageMap}} doesn't keep track of the state of the datanodes with > Provided storage. As a result, it can return nodes that are being > decommissioned or under-maintenance even when live datanodes exist. This JIRA > is to prefer live datanodes to datanodes in other states. -- This message was sent by Atlassian JIRA (v6.4.14#64029) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-12905) [READ] Handle decommissioning and under-maintenance Datanodes with Provided storage.
[ https://issues.apache.org/jira/browse/HDFS-12905?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Virajith Jalaparti updated HDFS-12905: -- Attachment: HDFS-12905-HDFS-9806.002.patch > [READ] Handle decommissioning and under-maintenance Datanodes with Provided > storage. > > > Key: HDFS-12905 > URL: https://issues.apache.org/jira/browse/HDFS-12905 > Project: Hadoop HDFS > Issue Type: Sub-task >Reporter: Virajith Jalaparti >Assignee: Virajith Jalaparti > Attachments: HDFS-12905-HDFS-9806.001.patch, > HDFS-12905-HDFS-9806.002.patch > > > {{ProvidedStorageMap}} doesn't keep track of the state of the datanodes with > Provided storage. As a result, it can return nodes that are being > decommissioned or under-maintenance even when live datanodes exist. This JIRA > is to prefer live datanodes to datanodes in other states. -- This message was sent by Atlassian JIRA (v6.4.14#64029) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-12905) [READ] Handle decommissioning and under-maintenance Datanodes with Provided storage.
[ https://issues.apache.org/jira/browse/HDFS-12905?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Virajith Jalaparti updated HDFS-12905: -- Status: Open (was: Patch Available) > [READ] Handle decommissioning and under-maintenance Datanodes with Provided > storage. > > > Key: HDFS-12905 > URL: https://issues.apache.org/jira/browse/HDFS-12905 > Project: Hadoop HDFS > Issue Type: Sub-task >Reporter: Virajith Jalaparti >Assignee: Virajith Jalaparti > Attachments: HDFS-12905-HDFS-9806.001.patch > > > {{ProvidedStorageMap}} doesn't keep track of the state of the datanodes with > Provided storage. As a result, it can return nodes that are being > decommissioned or under-maintenance even when live datanodes exist. This JIRA > is to prefer live datanodes to datanodes in other states. -- This message was sent by Atlassian JIRA (v6.4.14#64029) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-11733) TestGetBlocks.getBlocksWithException() ignores datanode and size parameters.
[ https://issues.apache.org/jira/browse/HDFS-11733?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16282763#comment-16282763 ] Konstantin Shvachko commented on HDFS-11733: Hi [~yuanbo]. Could you please check if the test failures are related to the changes here? > TestGetBlocks.getBlocksWithException() ignores datanode and size parameters. > > > Key: HDFS-11733 > URL: https://issues.apache.org/jira/browse/HDFS-11733 > Project: Hadoop HDFS > Issue Type: Bug > Components: balancer & mover, test >Affects Versions: 2.6.1 >Reporter: Konstantin Shvachko >Assignee: Yuanbo Liu > Labels: newbie++ > Attachments: HDFS-11733.001.patch > > > {{TestGetBlocks.getBlocksWithException()}} has 3 parameters, but uses only > one. So whatever callers think they pass in, it is ignored. > Looks like we should change it to use the parameters, but I am not sure how > this will affect the test. -- This message was sent by Atlassian JIRA (v6.4.14#64029) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-11915) Sync rbw dir on the first hsync() to avoid file lost on power failure
[ https://issues.apache.org/jira/browse/HDFS-11915?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16282669#comment-16282669 ] Wei-Chiu Chuang commented on HDFS-11915: Pushed the patch to trunk. Thanks [~vinayrpet] > Sync rbw dir on the first hsync() to avoid file lost on power failure > - > > Key: HDFS-11915 > URL: https://issues.apache.org/jira/browse/HDFS-11915 > Project: Hadoop HDFS > Issue Type: Bug >Reporter: Kanaka Kumar Avvaru >Assignee: Vinayakumar B >Priority: Critical > Fix For: 3.1.0 > > Attachments: HDFS-11915-01.patch, HDFS-11915-branch-2-01.patch > > > As discussed in HDFS-5042, there is a chance to lose blocks on power failure > if rbw file creation entry is not yet sync to device. Then the block created > is nowhere exists on disk. Neither in rbw nor in finalized. > As suggested by [~kihwal], will discuss and track it in this JIRA. > As suggested by [~vinayrpet], May be first hsync() request on block file can > call fsync on its parent directory (rbw) directory. -- This message was sent by Atlassian JIRA (v6.4.14#64029) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-11915) Sync rbw dir on the first hsync() to avoid file lost on power failure
[ https://issues.apache.org/jira/browse/HDFS-11915?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Wei-Chiu Chuang updated HDFS-11915: --- Fix Version/s: 3.1.0 > Sync rbw dir on the first hsync() to avoid file lost on power failure > - > > Key: HDFS-11915 > URL: https://issues.apache.org/jira/browse/HDFS-11915 > Project: Hadoop HDFS > Issue Type: Bug >Reporter: Kanaka Kumar Avvaru >Assignee: Vinayakumar B >Priority: Critical > Fix For: 3.1.0 > > Attachments: HDFS-11915-01.patch, HDFS-11915-branch-2-01.patch > > > As discussed in HDFS-5042, there is a chance to lose blocks on power failure > if rbw file creation entry is not yet sync to device. Then the block created > is nowhere exists on disk. Neither in rbw nor in finalized. > As suggested by [~kihwal], will discuss and track it in this JIRA. > As suggested by [~vinayrpet], May be first hsync() request on block file can > call fsync on its parent directory (rbw) directory. -- This message was sent by Atlassian JIRA (v6.4.14#64029) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-12848) [READ] Add a pluggable policy for selecting locations for Provided files.
[ https://issues.apache.org/jira/browse/HDFS-12848?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Íñigo Goiri updated HDFS-12848: --- Description: Add a pluggable policy for selecting locations for Provided files. > [READ] Add a pluggable policy for selecting locations for Provided files. > - > > Key: HDFS-12848 > URL: https://issues.apache.org/jira/browse/HDFS-12848 > Project: Hadoop HDFS > Issue Type: Sub-task >Reporter: Virajith Jalaparti > > Add a pluggable policy for selecting locations for Provided files. -- This message was sent by Atlassian JIRA (v6.4.14#64029) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-12907) Allow read-only access to reserved raw for non-superusers
[ https://issues.apache.org/jira/browse/HDFS-12907?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16282731#comment-16282731 ] Xiao Chen commented on HDFS-12907: -- Thanks Andrew for the ping and Rushabh / Daryn for discussions. Sorry I did not fully understand the intent here, and probably misunderstood some part of HDFS-12355. Could you help elaborating? My understanding is after HDFS-12355, webhdfs eventually works with encryption by: - user gets the DT from hdfs and kms. - user read / write a file, auths with HDFS using DT, get file status, then gets redirected to a DN - user passes the DTs along to the DN, where read/write a file with the crypto streams happens. - CryptoStreams auths with KMS using kms DT. The data is then read, decrypted and returned. - user cancels the DT. Is this remotely correct? Why do we need to run datanode as a separate user? (I think I understood Daryn's comment, and agree it would be another jira. Not 100% sure I see the relation here, are we trying to write raw bytes to the DN and decrypt at the client-side instead of on the DN on HDFS-12355?) > Allow read-only access to reserved raw for non-superusers > - > > Key: HDFS-12907 > URL: https://issues.apache.org/jira/browse/HDFS-12907 > Project: Hadoop HDFS > Issue Type: Bug > Components: namenode >Affects Versions: 2.6.0 >Reporter: Daryn Sharp >Assignee: Rushabh S Shah > Attachments: HDFS-12907.patch > > > HDFS-6509 added a special /.reserved/raw path prefix to access the raw file > contents of EZ files. In the simplest sense it doesn't return the FE info in > the {{LocatedBlocks}} so the dfs client doesn't try to decrypt the data. > This facilitates allowing tools like distcp to copy raw bytes. > Access to the raw hierarchy is restricted to superusers. This seems like an > overly broad restriction designed to prevent non-admins from munging the EZ > related xattrs. I believe we should relax the restriction to allow > non-admins to perform read-only operations. Allowing non-superusers to > easily read the raw bytes will be extremely useful for regular users, esp. > for enabling webhdfs client-side encryption. -- This message was sent by Atlassian JIRA (v6.4.14#64029) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-12875) RBF: Complete logic for -readonly option of dfsrouteradmin add command
[ https://issues.apache.org/jira/browse/HDFS-12875?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16282730#comment-16282730 ] Íñigo Goiri commented on HDFS-12875: I need to fix: * TestMountTableResolver * TestRouterMountTable The FindBug is unrelated (already in trunk). > RBF: Complete logic for -readonly option of dfsrouteradmin add command > -- > > Key: HDFS-12875 > URL: https://issues.apache.org/jira/browse/HDFS-12875 > Project: Hadoop HDFS > Issue Type: Sub-task >Affects Versions: 3.0.0-alpha3 >Reporter: Yiqun Lin >Assignee: Íñigo Goiri > Labels: RBF > Attachments: HDFS-12875.001.patch, HDFS-12875.002.patch, > HDFS-12875.003.patch > > > The dfsrouteradmin has an option for readonly mount points but this is not > implemented. We should add an special mount point which allows reading but > not writing. -- This message was sent by Atlassian JIRA (v6.4.14#64029) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-11915) Sync rbw dir on the first hsync() to avoid file lost on power failure
[ https://issues.apache.org/jira/browse/HDFS-11915?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16282695#comment-16282695 ] Hudson commented on HDFS-11915: --- SUCCESS: Integrated in Jenkins build Hadoop-trunk-Commit #13343 (See [https://builds.apache.org/job/Hadoop-trunk-Commit/13343/]) HDFS-11915. Sync rbw dir on the first hsync() to avoid file lost on (weichiu: rev d6c31a3e6b60c4b8af9ae4661f16614805654e59) * (edit) hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/BlockReceiver.java > Sync rbw dir on the first hsync() to avoid file lost on power failure > - > > Key: HDFS-11915 > URL: https://issues.apache.org/jira/browse/HDFS-11915 > Project: Hadoop HDFS > Issue Type: Bug >Reporter: Kanaka Kumar Avvaru >Assignee: Vinayakumar B >Priority: Critical > Fix For: 3.1.0 > > Attachments: HDFS-11915-01.patch, HDFS-11915-branch-2-01.patch > > > As discussed in HDFS-5042, there is a chance to lose blocks on power failure > if rbw file creation entry is not yet sync to device. Then the block created > is nowhere exists on disk. Neither in rbw nor in finalized. > As suggested by [~kihwal], will discuss and track it in this JIRA. > As suggested by [~vinayrpet], May be first hsync() request on block file can > call fsync on its parent directory (rbw) directory. -- This message was sent by Atlassian JIRA (v6.4.14#64029) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-12882) Support full open(PathHandle) contract in HDFS
[ https://issues.apache.org/jira/browse/HDFS-12882?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Chris Douglas updated HDFS-12882: - Attachment: HDFS-12882.03.patch This adds a separate flag to {{getFileInfo}} that (eventually) controls whether {{FSDirStatAndListingOp::createFileStatus}} also generates block tokens for its {{LocatedBlocks}}. While a single flag would be sufficient for this issue, if one implemented {{getLocatedFileStatus}} then we'd need a similar flag. This also allows for correct metrics/audit, since this is an {{open}} call only if the client requested block tokens. If one does not request locations, then {{needBlockToken}} has no effect. Alternatively, we could follow the pattern for {{getFileLinkInfo}}, and add {{ClientProtocol::getLocatedFileInfo}} that requests a {{HdfsLocatedFileStatus}} and only includes the block token flag. Internally (as with {{getFileLinkInfo}}) the changes are basically the same. Fewer, existing calls to {{getFileInfo}} (particularly in tests) would require updates. v03 also fixes related unit test failures, checkstyle, and findbugs serialization warnings. > Support full open(PathHandle) contract in HDFS > -- > > Key: HDFS-12882 > URL: https://issues.apache.org/jira/browse/HDFS-12882 > Project: Hadoop HDFS > Issue Type: Improvement > Components: hdfs-client >Reporter: Chris Douglas >Assignee: Chris Douglas > Attachments: HDFS-12882.00.patch, HDFS-12882.00.salient.txt, > HDFS-12882.01.patch, HDFS-12882.02.patch, HDFS-12882.03.patch > > > HDFS-7878 added support for {{open(PathHandle)}}, but it only partially > implemented the semantics specified in the contract (i.e., open-by-inodeID). > HDFS should implement all permutations of the default options for > {{PathHandle}}. -- This message was sent by Atlassian JIRA (v6.4.14#64029) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-12907) Allow read-only access to reserved raw for non-superusers
[ https://issues.apache.org/jira/browse/HDFS-12907?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16282676#comment-16282676 ] genericqa commented on HDFS-12907: -- | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 11m 56s{color} | {color:blue} Docker mode activated. {color} | || || || || {color:brown} Prechecks {color} || | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green} The patch does not contain any @author tags. {color} | | {color:green}+1{color} | {color:green} test4tests {color} | {color:green} 0m 0s{color} | {color:green} The patch appears to include 1 new or modified test files. {color} | || || || || {color:brown} trunk Compile Tests {color} || | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 16m 35s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 52s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 39s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 57s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 11m 20s{color} | {color:green} branch has no errors when building and testing our client artifacts. {color} | | {color:red}-1{color} | {color:red} findbugs {color} | {color:red} 1m 50s{color} | {color:red} hadoop-hdfs-project/hadoop-hdfs in trunk has 1 extant Findbugs warnings. {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 48s{color} | {color:green} trunk passed {color} | || || || || {color:brown} Patch Compile Tests {color} || | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 0m 55s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 50s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 0m 50s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 35s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 54s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 10m 51s{color} | {color:green} patch has no errors when building and testing our client artifacts. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 1m 56s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 47s{color} | {color:green} the patch passed {color} | || || || || {color:brown} Other Tests {color} || | {color:red}-1{color} | {color:red} unit {color} | {color:red} 92m 42s{color} | {color:red} hadoop-hdfs in the patch failed. {color} | | {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 23s{color} | {color:green} The patch does not generate ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black}154m 38s{color} | {color:black} {color} | \\ \\ || Reason || Tests || | Failed junit tests | hadoop.fs.TestUnbuffer | | | hadoop.hdfs.qjournal.server.TestJournalNodeSync | | | hadoop.hdfs.web.TestWebHDFSXAttr | | | hadoop.hdfs.server.namenode.TestFileContextXAttr | | | hadoop.hdfs.server.datanode.TestDataNodeVolumeFailureReporting | | | hadoop.hdfs.server.namenode.TestNameNodeXAttr | \\ \\ || Subsystem || Report/Notes || | Docker | Client=17.05.0-ce Server=17.05.0-ce Image:yetus/hadoop:5b98639 | | JIRA Issue | HDFS-12907 | | JIRA Patch URL | https://issues.apache.org/jira/secure/attachment/12901126/HDFS-12907.patch | | Optional Tests | asflicense compile javac javadoc mvninstall mvnsite unit shadedclient findbugs checkstyle | | uname | Linux 67f1b5e640eb 3.13.0-135-generic #184-Ubuntu SMP Wed Oct 18 11:55:51 UTC 2017 x86_64 x86_64 x86_64 GNU/Linux | | Build tool | maven | | Personality | /testptch/patchprocess/precommit/personality/provided.sh | | git revision | trunk / acb9290 | | maven | version: Apache Maven 3.3.9 | | Default Java | 1.8.0_151 | | findbugs | v3.1.0-RC1 | | findbugs | https://builds.apache.org/job/PreCommit-HDFS-Build/22316/artifact/out/branch-findbugs-hadoop-hdfs-project_hadoop-hdfs-warnings.html | | unit |
[jira] [Updated] (HDFS-12907) Allow read-only access to reserved raw for non-superusers
[ https://issues.apache.org/jira/browse/HDFS-12907?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Rushabh S Shah updated HDFS-12907: -- Status: Patch Available (was: Open) > Allow read-only access to reserved raw for non-superusers > - > > Key: HDFS-12907 > URL: https://issues.apache.org/jira/browse/HDFS-12907 > Project: Hadoop HDFS > Issue Type: Bug > Components: namenode >Affects Versions: 2.6.0 >Reporter: Daryn Sharp >Assignee: Rushabh S Shah > Attachments: HDFS-12907.patch > > > HDFS-6509 added a special /.reserved/raw path prefix to access the raw file > contents of EZ files. In the simplest sense it doesn't return the FE info in > the {{LocatedBlocks}} so the dfs client doesn't try to decrypt the data. > This facilitates allowing tools like distcp to copy raw bytes. > Access to the raw hierarchy is restricted to superusers. This seems like an > overly broad restriction designed to prevent non-admins from munging the EZ > related xattrs. I believe we should relax the restriction to allow > non-admins to perform read-only operations. Allowing non-superusers to > easily read the raw bytes will be extremely useful for regular users, esp. > for enabling webhdfs client-side encryption. -- This message was sent by Atlassian JIRA (v6.4.14#64029) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-11915) Sync rbw dir on the first hsync() to avoid file lost on power failure
[ https://issues.apache.org/jira/browse/HDFS-11915?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16282648#comment-16282648 ] Wei-Chiu Chuang commented on HDFS-11915: The patch for trunk is good +1. Will commit the patch later. The branch-2 patch needs some work. It duplicates the fsyncDirectory() method added in HDFS-5042. > Sync rbw dir on the first hsync() to avoid file lost on power failure > - > > Key: HDFS-11915 > URL: https://issues.apache.org/jira/browse/HDFS-11915 > Project: Hadoop HDFS > Issue Type: Bug >Reporter: Kanaka Kumar Avvaru >Assignee: Vinayakumar B >Priority: Critical > Attachments: HDFS-11915-01.patch, HDFS-11915-branch-2-01.patch > > > As discussed in HDFS-5042, there is a chance to lose blocks on power failure > if rbw file creation entry is not yet sync to device. Then the block created > is nowhere exists on disk. Neither in rbw nor in finalized. > As suggested by [~kihwal], will discuss and track it in this JIRA. > As suggested by [~vinayrpet], May be first hsync() request on block file can > call fsync on its parent directory (rbw) directory. -- This message was sent by Atlassian JIRA (v6.4.14#64029) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-12875) RBF: Complete logic for -readonly option of dfsrouteradmin add command
[ https://issues.apache.org/jira/browse/HDFS-12875?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16282641#comment-16282641 ] genericqa commented on HDFS-12875: -- | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 44s{color} | {color:blue} Docker mode activated. {color} | || || || || {color:brown} Prechecks {color} || | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green} The patch does not contain any @author tags. {color} | | {color:green}+1{color} | {color:green} test4tests {color} | {color:green} 0m 0s{color} | {color:green} The patch appears to include 3 new or modified test files. {color} | || || || || {color:brown} trunk Compile Tests {color} || | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 14m 47s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 50s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 33s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 53s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 9m 51s{color} | {color:green} branch has no errors when building and testing our client artifacts. {color} | | {color:red}-1{color} | {color:red} findbugs {color} | {color:red} 1m 43s{color} | {color:red} hadoop-hdfs-project/hadoop-hdfs in trunk has 1 extant Findbugs warnings. {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 48s{color} | {color:green} trunk passed {color} | || || || || {color:brown} Patch Compile Tests {color} || | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 0m 51s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 46s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 0m 46s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 30s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 50s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 9m 35s{color} | {color:green} patch has no errors when building and testing our client artifacts. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 1m 49s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 46s{color} | {color:green} the patch passed {color} | || || || || {color:brown} Other Tests {color} || | {color:red}-1{color} | {color:red} unit {color} | {color:red}127m 0s{color} | {color:red} hadoop-hdfs in the patch failed. {color} | | {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 23s{color} | {color:green} The patch does not generate ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black}172m 28s{color} | {color:black} {color} | \\ \\ || Reason || Tests || | Failed junit tests | hadoop.hdfs.TestDFSStripedOutputStreamWithFailure200 | | | hadoop.hdfs.TestDFSStripedOutputStreamWithFailure000 | | | hadoop.hdfs.TestDFSPermission | | | hadoop.hdfs.server.namenode.ha.TestRetryCacheWithHA | | | hadoop.hdfs.TestDFSUpgradeFromImage | | | hadoop.hdfs.web.TestWebHdfsTimeouts | | | hadoop.hdfs.TestErasureCodingPolicies | | | hadoop.hdfs.client.impl.TestBlockReaderLocal | | | hadoop.hdfs.TestSafeModeWithStripedFile | | | hadoop.hdfs.TestDFSStripedOutputStreamWithFailure | | | hadoop.hdfs.TestReadStripedFileWithMissingBlocks | | | hadoop.hdfs.TestErasureCodingMultipleRacks | | | hadoop.hdfs.server.datanode.TestDataNodeVolumeFailureReporting | | | hadoop.hdfs.TestDFSStorageStateRecovery | | | hadoop.hdfs.TestDFSStripedOutputStreamWithFailure030 | | | hadoop.hdfs.TestDFSStripedOutputStreamWithFailure140 | | | hadoop.hdfs.TestDFSStripedOutputStreamWithFailure100 | | | hadoop.hdfs.server.federation.resolver.TestMountTableResolver | | | hadoop.hdfs.server.federation.router.TestRouterMountTable | | | hadoop.hdfs.TestErasureCodingPoliciesWithRandomECPolicy | | | hadoop.hdfs.TestDFSStripedOutputStreamWithFailure050 | | | hadoop.hdfs.TestFileCreationDelete | | |
[jira] [Commented] (HDFS-12818) Support multiple storages in DataNodeCluster / SimulatedFSDataset
[ https://issues.apache.org/jira/browse/HDFS-12818?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16282607#comment-16282607 ] genericqa commented on HDFS-12818: -- | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 19s{color} | {color:blue} Docker mode activated. {color} | || || || || {color:brown} Prechecks {color} || | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green} The patch does not contain any @author tags. {color} | | {color:green}+1{color} | {color:green} test4tests {color} | {color:green} 0m 0s{color} | {color:green} The patch appears to include 3 new or modified test files. {color} | || || || || {color:brown} trunk Compile Tests {color} || | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 15m 8s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 50s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 39s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 54s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 10m 53s{color} | {color:green} branch has no errors when building and testing our client artifacts. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 1m 40s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 50s{color} | {color:green} trunk passed {color} | || || || || {color:brown} Patch Compile Tests {color} || | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 0m 51s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 46s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 0m 46s{color} | {color:green} the patch passed {color} | | {color:orange}-0{color} | {color:orange} checkstyle {color} | {color:orange} 0m 35s{color} | {color:orange} hadoop-hdfs-project/hadoop-hdfs: The patch generated 1 new + 206 unchanged - 8 fixed = 207 total (was 214) {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 52s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 10m 12s{color} | {color:green} patch has no errors when building and testing our client artifacts. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 1m 48s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 48s{color} | {color:green} the patch passed {color} | || || || || {color:brown} Other Tests {color} || | {color:red}-1{color} | {color:red} unit {color} | {color:red}142m 13s{color} | {color:red} hadoop-hdfs in the patch failed. {color} | | {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 20s{color} | {color:green} The patch does not generate ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black}189m 29s{color} | {color:black} {color} | \\ \\ || Reason || Tests || | Failed junit tests | hadoop.hdfs.server.balancer.TestBalancerWithSaslDataTransfer | | | hadoop.hdfs.server.blockmanagement.TestBlockTokenWithDFSStriped | | | hadoop.hdfs.server.datanode.TestDataNodeVolumeFailure | | | hadoop.hdfs.TestFileCreation | | | hadoop.hdfs.server.namenode.ha.TestDNFencing | | | hadoop.hdfs.server.balancer.TestBalancerWithHANameNodes | | | hadoop.hdfs.web.TestWebHdfsTimeouts | | | hadoop.hdfs.server.namenode.ha.TestStandbyCheckpoints | | | hadoop.hdfs.TestDFSStripedOutputStreamWithFailure | | | hadoop.hdfs.tools.offlineImageViewer.TestOfflineImageViewerWithStripedBlocks | | | hadoop.hdfs.server.namenode.ha.TestBootstrapStandbyWithQJM | | | hadoop.hdfs.server.federation.router.TestRouterRpcMultiDestination | | | hadoop.hdfs.server.namenode.TestNameNodeMXBean | | | hadoop.hdfs.server.blockmanagement.TestBlockTokenWithDFS | | | hadoop.hdfs.tools.TestDFSAdminWithHA | | | hadoop.hdfs.server.balancer.TestBalancerWithEncryptedTransfer | | | hadoop.hdfs.server.namenode.ha.TestPipelinesFailover | | | hadoop.hdfs.TestDFSStripedOutputStreamWithFailure070 | | | hadoop.hdfs.server.datanode.TestDataNodeInitStorage | | |
[jira] [Updated] (HDFS-12848) [READ] Add a pluggable policy for selecting locations for Provided files.
[ https://issues.apache.org/jira/browse/HDFS-12848?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Virajith Jalaparti updated HDFS-12848: -- Parent Issue: HDFS-12090 (was: HDFS-9806) > [READ] Add a pluggable policy for selecting locations for Provided files. > - > > Key: HDFS-12848 > URL: https://issues.apache.org/jira/browse/HDFS-12848 > Project: Hadoop HDFS > Issue Type: Sub-task >Reporter: Virajith Jalaparti > -- This message was sent by Atlassian JIRA (v6.4.14#64029) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-12903) [READ] Fix closing streams in ImageWriter
[ https://issues.apache.org/jira/browse/HDFS-12903?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Virajith Jalaparti updated HDFS-12903: -- Resolution: Fixed Status: Resolved (was: Patch Available) > [READ] Fix closing streams in ImageWriter > - > > Key: HDFS-12903 > URL: https://issues.apache.org/jira/browse/HDFS-12903 > Project: Hadoop HDFS > Issue Type: Sub-task >Reporter: Íñigo Goiri > Attachments: HDFS-12903-HDFS-9806.001.patch > > > HDFS-12894 showed a FindBug in HDFS-9806. This seems related to HDFS-12881 > when using {{IOUtils.cleanupWithLogger()}}. -- This message was sent by Atlassian JIRA (v6.4.14#64029) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Assigned] (HDFS-12903) [READ] Fix closing streams in ImageWriter
[ https://issues.apache.org/jira/browse/HDFS-12903?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Virajith Jalaparti reassigned HDFS-12903: - Assignee: Virajith Jalaparti > [READ] Fix closing streams in ImageWriter > - > > Key: HDFS-12903 > URL: https://issues.apache.org/jira/browse/HDFS-12903 > Project: Hadoop HDFS > Issue Type: Sub-task >Reporter: Íñigo Goiri >Assignee: Virajith Jalaparti > Attachments: HDFS-12903-HDFS-9806.001.patch > > > HDFS-12894 showed a FindBug in HDFS-9806. This seems related to HDFS-12881 > when using {{IOUtils.cleanupWithLogger()}}. -- This message was sent by Atlassian JIRA (v6.4.14#64029) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-12874) [READ] Documentation for provided storage
[ https://issues.apache.org/jira/browse/HDFS-12874?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Virajith Jalaparti updated HDFS-12874: -- Attachment: HDFS-12874-HDFS-9806.01.patch Modified documentation attached (configuration options specified more in xml). > [READ] Documentation for provided storage > - > > Key: HDFS-12874 > URL: https://issues.apache.org/jira/browse/HDFS-12874 > Project: Hadoop HDFS > Issue Type: Sub-task >Reporter: Chris Douglas > Attachments: HDFS-12874-HDFS-9806.00.patch, > HDFS-12874-HDFS-9806.01.patch > > > The configuration and deployment of provided storage should be documented for > end-users. -- This message was sent by Atlassian JIRA (v6.4.14#64029) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-12903) [READ] Fix closing streams in ImageWriter
[ https://issues.apache.org/jira/browse/HDFS-12903?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16282620#comment-16282620 ] Virajith Jalaparti commented on HDFS-12903: --- Thanks [~elgoiri] and [~chris.douglas]. Committed this to feature branch. > [READ] Fix closing streams in ImageWriter > - > > Key: HDFS-12903 > URL: https://issues.apache.org/jira/browse/HDFS-12903 > Project: Hadoop HDFS > Issue Type: Sub-task >Reporter: Íñigo Goiri > Attachments: HDFS-12903-HDFS-9806.001.patch > > > HDFS-12894 showed a FindBug in HDFS-9806. This seems related to HDFS-12881 > when using {{IOUtils.cleanupWithLogger()}}. -- This message was sent by Atlassian JIRA (v6.4.14#64029) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-12890) Ozone: XceiverClient should have upper bound on async requests
[ https://issues.apache.org/jira/browse/HDFS-12890?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16282586#comment-16282586 ] genericqa commented on HDFS-12890: -- | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 14m 24s{color} | {color:blue} Docker mode activated. {color} | || || || || {color:brown} Prechecks {color} || | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green} The patch does not contain any @author tags. {color} | | {color:red}-1{color} | {color:red} test4tests {color} | {color:red} 0m 0s{color} | {color:red} The patch doesn't appear to include any new or modified tests. Please justify why no new tests are needed for this patch. Also please list what manual steps were performed to verify this patch. {color} | || || || || {color:brown} HDFS-7240 Compile Tests {color} || | {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 0m 37s{color} | {color:blue} Maven dependency ordering for branch {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 19m 19s{color} | {color:green} HDFS-7240 passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 1m 53s{color} | {color:green} HDFS-7240 passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 47s{color} | {color:green} HDFS-7240 passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 2m 3s{color} | {color:green} HDFS-7240 passed {color} | | {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 14m 15s{color} | {color:green} branch has no errors when building and testing our client artifacts. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 4m 11s{color} | {color:green} HDFS-7240 passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 2m 3s{color} | {color:green} HDFS-7240 passed {color} | || || || || {color:brown} Patch Compile Tests {color} || | {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 0m 9s{color} | {color:blue} Maven dependency ordering for patch {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 1m 59s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 1m 56s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 1m 56s{color} | {color:green} the patch passed {color} | | {color:orange}-0{color} | {color:orange} checkstyle {color} | {color:orange} 0m 45s{color} | {color:orange} hadoop-hdfs-project: The patch generated 1 new + 0 unchanged - 0 fixed = 1 total (was 0) {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 1m 55s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} xml {color} | {color:green} 0m 2s{color} | {color:green} The patch has no ill-formed XML file. {color} | | {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 12m 44s{color} | {color:green} patch has no errors when building and testing our client artifacts. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 4m 22s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 2m 3s{color} | {color:green} the patch passed {color} | || || || || {color:brown} Other Tests {color} || | {color:green}+1{color} | {color:green} unit {color} | {color:green} 1m 49s{color} | {color:green} hadoop-hdfs-client in the patch passed. {color} | | {color:red}-1{color} | {color:red} unit {color} | {color:red}155m 56s{color} | {color:red} hadoop-hdfs in the patch failed. {color} | | {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 25s{color} | {color:green} The patch does not generate ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black}242m 25s{color} | {color:black} {color} | \\ \\ || Reason || Tests || | Failed junit tests | hadoop.ozone.web.client.TestKeysRatis | | | hadoop.hdfs.server.namenode.TestFSEditLogLoader | | | hadoop.hdfs.server.namenode.TestQuotaByStorageType | | | hadoop.hdfs.server.namenode.TestFsck | | | hadoop.ozone.client.rpc.TestOzoneRpcClient | | | hadoop.hdfs.server.datanode.TestDataNodeVolumeFailure | | | hadoop.hdfs.server.namenode.TestLeaseManager | | | hadoop.hdfs.server.namenode.TestReencryptionWithKMS | | | hadoop.hdfs.server.namenode.TestEditLog |
[jira] [Commented] (HDFS-12907) Allow read-only access to reserved raw for non-superusers
[ https://issues.apache.org/jira/browse/HDFS-12907?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16282583#comment-16282583 ] Rushabh S Shah commented on HDFS-12907: --- HDFS-12355 is tracking all the efforts for adding EZ support to WebhdfsFileSystem. > Allow read-only access to reserved raw for non-superusers > - > > Key: HDFS-12907 > URL: https://issues.apache.org/jira/browse/HDFS-12907 > Project: Hadoop HDFS > Issue Type: Bug > Components: namenode >Affects Versions: 2.6.0 >Reporter: Daryn Sharp >Assignee: Rushabh S Shah > Attachments: HDFS-12907.patch > > > HDFS-6509 added a special /.reserved/raw path prefix to access the raw file > contents of EZ files. In the simplest sense it doesn't return the FE info in > the {{LocatedBlocks}} so the dfs client doesn't try to decrypt the data. > This facilitates allowing tools like distcp to copy raw bytes. > Access to the raw hierarchy is restricted to superusers. This seems like an > overly broad restriction designed to prevent non-admins from munging the EZ > related xattrs. I believe we should relax the restriction to allow > non-admins to perform read-only operations. Allowing non-superusers to > easily read the raw bytes will be extremely useful for regular users, esp. > for enabling webhdfs client-side encryption. -- This message was sent by Atlassian JIRA (v6.4.14#64029) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-12907) Allow read-only access to reserved raw for non-superusers
[ https://issues.apache.org/jira/browse/HDFS-12907?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16282579#comment-16282579 ] Andrew Wang commented on HDFS-12907: Solving this for WebHdfsFileSystem is a lot more tractable than Hue, so this makes sense to me. FYI [~xiaochen] also, since he's a KMS expert and was also involved in the internal discussions. > Allow read-only access to reserved raw for non-superusers > - > > Key: HDFS-12907 > URL: https://issues.apache.org/jira/browse/HDFS-12907 > Project: Hadoop HDFS > Issue Type: Bug > Components: namenode >Affects Versions: 2.6.0 >Reporter: Daryn Sharp >Assignee: Rushabh S Shah > Attachments: HDFS-12907.patch > > > HDFS-6509 added a special /.reserved/raw path prefix to access the raw file > contents of EZ files. In the simplest sense it doesn't return the FE info in > the {{LocatedBlocks}} so the dfs client doesn't try to decrypt the data. > This facilitates allowing tools like distcp to copy raw bytes. > Access to the raw hierarchy is restricted to superusers. This seems like an > overly broad restriction designed to prevent non-admins from munging the EZ > related xattrs. I believe we should relax the restriction to allow > non-admins to perform read-only operations. Allowing non-superusers to > easily read the raw bytes will be extremely useful for regular users, esp. > for enabling webhdfs client-side encryption. -- This message was sent by Atlassian JIRA (v6.4.14#64029) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-12903) [READ] Fix closing streams in ImageWriter
[ https://issues.apache.org/jira/browse/HDFS-12903?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16282501#comment-16282501 ] Íñigo Goiri commented on HDFS-12903: Yetus doesn't show the FindBug even in HDFS-9806. I think this fix should be it. +1 > [READ] Fix closing streams in ImageWriter > - > > Key: HDFS-12903 > URL: https://issues.apache.org/jira/browse/HDFS-12903 > Project: Hadoop HDFS > Issue Type: Sub-task >Reporter: Íñigo Goiri > Attachments: HDFS-12903-HDFS-9806.001.patch > > > HDFS-12894 showed a FindBug in HDFS-9806. This seems related to HDFS-12881 > when using {{IOUtils.cleanupWithLogger()}}. -- This message was sent by Atlassian JIRA (v6.4.14#64029) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-12907) Allow read-only access to reserved raw for non-superusers
[ https://issues.apache.org/jira/browse/HDFS-12907?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16282487#comment-16282487 ] Andrew Wang commented on HDFS-12907: I think that would work, though of course I'd prefer not to open up internal state representation if it can be avoided. On the topic of webhdfs client-side encryption, could you talk a little more about your usecase? We discussed this internally before in the context of Hue, and there didn't seem to be a great solution. They have a very simple Python WebHDFS client built around effectively curl, and they'd need to add their own KMS client and encryption routines. Really though, we'd want to move this all the way to the browser, and write the KMS client and encryption routines in Javascript. Ouch. A way of scoping the KMS delegation token to limit what keys could be accessed would also be an improvement, e.g. a "key token" similar to the HDFS block token. It addresses some of the issues with webhdfs and encryption. > Allow read-only access to reserved raw for non-superusers > - > > Key: HDFS-12907 > URL: https://issues.apache.org/jira/browse/HDFS-12907 > Project: Hadoop HDFS > Issue Type: Bug > Components: namenode >Affects Versions: 2.6.0 >Reporter: Daryn Sharp >Assignee: Rushabh S Shah > Attachments: HDFS-12907.patch > > > HDFS-6509 added a special /.reserved/raw path prefix to access the raw file > contents of EZ files. In the simplest sense it doesn't return the FE info in > the {{LocatedBlocks}} so the dfs client doesn't try to decrypt the data. > This facilitates allowing tools like distcp to copy raw bytes. > Access to the raw hierarchy is restricted to superusers. This seems like an > overly broad restriction designed to prevent non-admins from munging the EZ > related xattrs. I believe we should relax the restriction to allow > non-admins to perform read-only operations. Allowing non-superusers to > easily read the raw bytes will be extremely useful for regular users, esp. > for enabling webhdfs client-side encryption. -- This message was sent by Atlassian JIRA (v6.4.14#64029) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-7240) Object store in HDFS
[ https://issues.apache.org/jira/browse/HDFS-7240?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16282441#comment-16282441 ] Andrew Wang commented on HDFS-7240: --- Hi Sanjay, Thanks for writing up that summary. It's clear there's still disagreement on the merge. How should we proceed on reaching consensus? On the last call you suggested making a document, or we could do another call. > Object store in HDFS > > > Key: HDFS-7240 > URL: https://issues.apache.org/jira/browse/HDFS-7240 > Project: Hadoop HDFS > Issue Type: New Feature >Reporter: Jitendra Nath Pandey >Assignee: Jitendra Nath Pandey > Attachments: HDFS Scalability and Ozone.pdf, HDFS-7240.001.patch, > HDFS-7240.002.patch, HDFS-7240.003.patch, HDFS-7240.003.patch, > HDFS-7240.004.patch, HDFS-7240.005.patch, HDFS-7240.006.patch, > MeetingMinutes.pdf, Ozone-architecture-v1.pdf, Ozonedesignupdate.pdf, > ozone_user_v0.pdf > > > This jira proposes to add object store capabilities into HDFS. > As part of the federation work (HDFS-1052) we separated block storage as a > generic storage layer. Using the Block Pool abstraction, new kinds of > namespaces can be built on top of the storage layer i.e. datanodes. > In this jira I will explore building an object store using the datanode > storage, but independent of namespace metadata. > I will soon update with a detailed design document. -- This message was sent by Atlassian JIRA (v6.4.14#64029) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-12907) Allow read-only access to reserved raw for non-superusers
[ https://issues.apache.org/jira/browse/HDFS-12907?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16282451#comment-16282451 ] Daryn Sharp commented on HDFS-12907: bq. It's so they don't accidentally write data without xattrs or with the wrong xattrs, which would be essentially corrupt. This morning, Rushabh and I discussed/debated non-superuser write access, and my position too is non-superusers should not have access to raw attrs like feinfo. I see no use case except allowing users to (un)intentionally lose access to their data. Except... I am inclined to believe that create should have been extended to allow an optional feinfo. NN verifies the key name is correct, output stream doesn't encrypt. Or creating a raw path requires a mandatory feinfo. Lots of messy backwards compat issues to consider which is why it's a topic for another jira. > Allow read-only access to reserved raw for non-superusers > - > > Key: HDFS-12907 > URL: https://issues.apache.org/jira/browse/HDFS-12907 > Project: Hadoop HDFS > Issue Type: Bug > Components: namenode >Affects Versions: 2.6.0 >Reporter: Daryn Sharp >Assignee: Rushabh S Shah > Attachments: HDFS-12907.patch > > > HDFS-6509 added a special /.reserved/raw path prefix to access the raw file > contents of EZ files. In the simplest sense it doesn't return the FE info in > the {{LocatedBlocks}} so the dfs client doesn't try to decrypt the data. > This facilitates allowing tools like distcp to copy raw bytes. > Access to the raw hierarchy is restricted to superusers. This seems like an > overly broad restriction designed to prevent non-admins from munging the EZ > related xattrs. I believe we should relax the restriction to allow > non-admins to perform read-only operations. Allowing non-superusers to > easily read the raw bytes will be extremely useful for regular users, esp. > for enabling webhdfs client-side encryption. -- This message was sent by Atlassian JIRA (v6.4.14#64029) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-10285) Storage Policy Satisfier in Namenode
[ https://issues.apache.org/jira/browse/HDFS-10285?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16282419#comment-16282419 ] Daryn Sharp commented on HDFS-10285: bq. Im coming at this from the standpoint of supporting Cloudera's Hadoop customers. \[…\] the average Hadoop admin who wants this to be turnkey If a cluster is a magical turnkey, it’s just an implementation detail whether it’s a monolithic service or collection of managed services. I understand the support burden of telling customers “run this, run that”, but isn’t that a deficiency of Ambari, Cloudera Manager, etc? Here's a rhetorical question: If managing multiple services is hard, why not bundle oozie, spark, storm, sqoop, kafka, ranger, knox, hive server, etc in the same process? Or ZK so HA is easier to deploy/manage? bq. For a large, sophisticated Hadoop user like Yahoo, it may not be a big cost to deploy a new service, but in relative terms a much bigger cost for a small user. Cluster upgrades are formal. It can take weeks or months to reach critical production clusters. If a scale level bug is found near the end of the runway, it’s hard to short-circuit the restarting and rescheduling the entire runway. On the other hand, an adjunct service like the kms has an extremely short runway. The balancer, being generally non-critical, has a lot of leeway and when necessary can be tinkered on w/o a deployment. Tech support asking a user to start a process costs less? Anyway, fast review. *Locking* Yesterday, I was going to say I'm not overly worried with locking other than correctness and doesn't impact whether it should be in or out of the NN. Today, I looked at the code more closely. It can hold the lock (read lock, but still) way too long. Notably, but not limited to, _you can’t hold the lock while doing block placement_. Being in the NN makes it too easy to abuse the lock in subtle ways. *Memory* Bounded queues are not a panacea for memory concerns. I’m more concerned with GC issues. Throttling via queues is going to result in promotion to oldgen where collection is much more expensive. The memory estimate is narrowly focused and assumes a 32-bit jvm. It omits the all ancillary heavyweight data structures, futures, etc. *CPU* Yesterday, not too worried based on misconception that very little locking is occurring. Today, I see there’s an incredible amount of computation occurring which often appears to be within the fsn lock. There’s a lot of garbage generation which invisibly saps cpu too. *Other* bq. This feature is switched OFF by default and no impact to HDFS. I should start sending bills to everyone who makes this fraudulent claim. :). {{FSDirectory#addToInodeMap}} imposes a nontrivial performance penalty even when SPS is not enabled. We had to hack out the similar EZ check because it had a noticeable performance impact esp. on startup. However now that we support EZ, I need to revisit optimizing it. There’s likely more performance hits if I looked harder at where it’s spliced in. bq. NN has existing feature EDEK which also does scanning and we reuses the same code in SPS. Yes, and I’m not very happy about that feature’s implementation but it was jammed in. –– I’m torn on this issue. I think the HSM experience is lackluster and needs to be improved. I haven’t looked at the Mover so no idea how well it works or doesn’t work. If it works ok, then perhaps it should have an rpc service to poke something at the front of the queue for those that don’t want to wait like hbase. If it’s an internal service, I’d rather it work in a dumbed-down background fashion. Otherwise it’s going to be a real problem as it becomes too smart and bloated. I’m curious why it isn’t just part of the standard replication monitoring. If the DN is told to replicate to itself, it just does the storage movement. > Storage Policy Satisfier in Namenode > > > Key: HDFS-10285 > URL: https://issues.apache.org/jira/browse/HDFS-10285 > Project: Hadoop HDFS > Issue Type: New Feature > Components: datanode, namenode >Affects Versions: HDFS-10285 >Reporter: Uma Maheswara Rao G >Assignee: Uma Maheswara Rao G > Attachments: HDFS-10285-consolidated-merge-patch-00.patch, > HDFS-10285-consolidated-merge-patch-01.patch, > HDFS-10285-consolidated-merge-patch-02.patch, > HDFS-10285-consolidated-merge-patch-03.patch, > HDFS-SPS-TestReport-20170708.pdf, > Storage-Policy-Satisfier-in-HDFS-June-20-2017.pdf, > Storage-Policy-Satisfier-in-HDFS-May10.pdf, > Storage-Policy-Satisfier-in-HDFS-Oct-26-2017.pdf > > > Heterogeneous storage in HDFS introduced the concept of storage policy. These > policies can be set on directory/file to specify the user preference, where > to store the physical block. When
[jira] [Updated] (HDFS-12875) RBF: Complete logic for -readonly option of dfsrouteradmin add command
[ https://issues.apache.org/jira/browse/HDFS-12875?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Íñigo Goiri updated HDFS-12875: --- Description: The dfsrouteradmin has an option for readonly mount points but this is not implemented. We should add an special mount point which allows reading but not writing. (was: Currently the option -readonly of command {{dfsrouteradmin -add}} doesn't make any sense.The desired behavior is that read-only mount table that be set in add command cannot be removed.) > RBF: Complete logic for -readonly option of dfsrouteradmin add command > -- > > Key: HDFS-12875 > URL: https://issues.apache.org/jira/browse/HDFS-12875 > Project: Hadoop HDFS > Issue Type: Sub-task >Affects Versions: 3.0.0-alpha3 >Reporter: Yiqun Lin >Assignee: Íñigo Goiri > Labels: RBF > Attachments: HDFS-12875.001.patch, HDFS-12875.002.patch, > HDFS-12875.003.patch > > > The dfsrouteradmin has an option for readonly mount points but this is not > implemented. We should add an special mount point which allows reading but > not writing. -- This message was sent by Atlassian JIRA (v6.4.14#64029) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-12840) Creating a file with non-default EC policy in a EC zone is not correctly serialized in the editlog
[ https://issues.apache.org/jira/browse/HDFS-12840?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16282421#comment-16282421 ] Hudson commented on HDFS-12840: --- SUCCESS: Integrated in Jenkins build Hadoop-trunk-Commit #13341 (See [https://builds.apache.org/job/Hadoop-trunk-Commit/13341/]) HDFS-12840. Creating a file with non-default EC policy in a EC zone is (lei: rev 67662e2ac9e68f32b725c8118cf2be79a662fca5) * (edit) hadoop-hdfs-project/hadoop-hdfs/src/test/resources/editsStored.xml * (edit) hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/DFSTestUtil.java * (edit) hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSEditLog.java * (edit) hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/INodeFile.java * (edit) hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSEditLogOp.java * (edit) hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestDistributedFileSystemWithECFile.java * (edit) hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestNamenodeRetryCache.java * (edit) hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/ha/TestRetryCacheWithHA.java * (edit) hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/OfflineEditsViewerHelper.java * (edit) hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/erasurecode/ErasureCodeConstants.java * (edit) hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSEditLogLoader.java * (edit) hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSDirWriteFileOp.java * (edit) hadoop-hdfs-project/hadoop-hdfs/src/test/resources/editsStored > Creating a file with non-default EC policy in a EC zone is not correctly > serialized in the editlog > -- > > Key: HDFS-12840 > URL: https://issues.apache.org/jira/browse/HDFS-12840 > Project: Hadoop HDFS > Issue Type: Bug > Components: erasure-coding >Affects Versions: 3.0.0-beta1 >Reporter: Lei (Eddy) Xu >Assignee: Lei (Eddy) Xu >Priority: Blocker > Labels: hdfs-ec-3.0-must-do > Fix For: 3.0.0 > > Attachments: HDFS-12840.00.patch, HDFS-12840.01.patch, > HDFS-12840.02.patch, HDFS-12840.03.patch, HDFS-12840.04.patch, > HDFS-12840.05.patch, HDFS-12840.reprod.patch, editsStored, editsStored, > editsStored.03, editsStored.05 > > > When create a replicated file in an existing EC zone, the edit logs does not > differentiate it from an EC file. When {{FSEditLogLoader}} to replay edits, > this file is treated as EC file, as a results, it crashes the NN because the > blocks of this file are replicated, which does not match with {{INode}}. > {noformat} > ERROR org.apache.hadoop.hdfs.server.namenode.FSEditLogLoader: Encountered > exception on operation AddBlockOp [path=/system/balancer.id, > penultimateBlock=NULL, lastBlock=blk_1073743259_2455, RpcClientId=, > RpcCallId=-2] > java.lang.IllegalArgumentException: reportedBlock is not striped > at > com.google.common.base.Preconditions.checkArgument(Preconditions.java:88) > at > org.apache.hadoop.hdfs.server.blockmanagement.BlockInfoStriped.addStorage(BlockInfoStriped.java:118) > at > org.apache.hadoop.hdfs.server.blockmanagement.DatanodeStorageInfo.addBlock(DatanodeStorageInfo.java:256) > at > org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.addStoredBlock(BlockManager.java:3141) > at > org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.addStoredBlockUnderConstruction(BlockManager.java:3068) > at > org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.processAndHandleReportedBlock(BlockManager.java:3864) > at > org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.processQueuedMessages(BlockManager.java:2916) > at > org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.processQueuedMessagesForBlock(BlockManager.java:2903) > at > org.apache.hadoop.hdfs.server.namenode.FSEditLogLoader.addNewBlock(FSEditLogLoader.java:1069) > at > org.apache.hadoop.hdfs.server.namenode.FSEditLogLoader.applyEditLogOp(FSEditLogLoader.java:532) > at > org.apache.hadoop.hdfs.server.namenode.FSEditLogLoader.loadEditRecords(FSEditLogLoader.java:249) > at > org.apache.hadoop.hdfs.server.namenode.FSEditLogLoader.loadFSEdits(FSEditLogLoader.java:158) > at > org.apache.hadoop.hdfs.server.namenode.FSImage.loadEdits(FSImage.java:882) > at > org.apache.hadoop.hdfs.server.namenode.FSImage.loadEdits(FSImage.java:863) > at >
[jira] [Updated] (HDFS-12875) RBF: Complete logic for -readonly option of dfsrouteradmin add command
[ https://issues.apache.org/jira/browse/HDFS-12875?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Íñigo Goiri updated HDFS-12875: --- Attachment: HDFS-12875.003.patch > RBF: Complete logic for -readonly option of dfsrouteradmin add command > -- > > Key: HDFS-12875 > URL: https://issues.apache.org/jira/browse/HDFS-12875 > Project: Hadoop HDFS > Issue Type: Sub-task >Affects Versions: 3.0.0-alpha3 >Reporter: Yiqun Lin >Assignee: Íñigo Goiri > Labels: RBF > Attachments: HDFS-12875.001.patch, HDFS-12875.002.patch, > HDFS-12875.003.patch > > > Currently the option -readonly of command {{dfsrouteradmin -add}} doesn't > make any sense.The desired behavior is that read-only mount table that be set > in add command cannot be removed. -- This message was sent by Atlassian JIRA (v6.4.14#64029) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-12840) Creating a file with non-default EC policy in a EC zone is not correctly serialized in the editlog
[ https://issues.apache.org/jira/browse/HDFS-12840?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Lei (Eddy) Xu updated HDFS-12840: - Resolution: Fixed Hadoop Flags: Reviewed Fix Version/s: 3.0.0 Release Note: Add ErasureCodingPolicyId to each OP_ADD edit log op. Status: Resolved (was: Patch Available) Thanks [~Sammi] and [~rakesh_r] for the reviews! Part of the checkstyles and findbugs warnigns are false positive. Fixed the rest warnings in {{DFSTestUtil.java}}. The test failures listed above will pass after applying the new {{editStored}}. Committed to trunk and branch-3.0 / 3.0.0. > Creating a file with non-default EC policy in a EC zone is not correctly > serialized in the editlog > -- > > Key: HDFS-12840 > URL: https://issues.apache.org/jira/browse/HDFS-12840 > Project: Hadoop HDFS > Issue Type: Bug > Components: erasure-coding >Affects Versions: 3.0.0-beta1 >Reporter: Lei (Eddy) Xu >Assignee: Lei (Eddy) Xu >Priority: Blocker > Labels: hdfs-ec-3.0-must-do > Fix For: 3.0.0 > > Attachments: HDFS-12840.00.patch, HDFS-12840.01.patch, > HDFS-12840.02.patch, HDFS-12840.03.patch, HDFS-12840.04.patch, > HDFS-12840.05.patch, HDFS-12840.reprod.patch, editsStored, editsStored, > editsStored.03, editsStored.05 > > > When create a replicated file in an existing EC zone, the edit logs does not > differentiate it from an EC file. When {{FSEditLogLoader}} to replay edits, > this file is treated as EC file, as a results, it crashes the NN because the > blocks of this file are replicated, which does not match with {{INode}}. > {noformat} > ERROR org.apache.hadoop.hdfs.server.namenode.FSEditLogLoader: Encountered > exception on operation AddBlockOp [path=/system/balancer.id, > penultimateBlock=NULL, lastBlock=blk_1073743259_2455, RpcClientId=, > RpcCallId=-2] > java.lang.IllegalArgumentException: reportedBlock is not striped > at > com.google.common.base.Preconditions.checkArgument(Preconditions.java:88) > at > org.apache.hadoop.hdfs.server.blockmanagement.BlockInfoStriped.addStorage(BlockInfoStriped.java:118) > at > org.apache.hadoop.hdfs.server.blockmanagement.DatanodeStorageInfo.addBlock(DatanodeStorageInfo.java:256) > at > org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.addStoredBlock(BlockManager.java:3141) > at > org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.addStoredBlockUnderConstruction(BlockManager.java:3068) > at > org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.processAndHandleReportedBlock(BlockManager.java:3864) > at > org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.processQueuedMessages(BlockManager.java:2916) > at > org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.processQueuedMessagesForBlock(BlockManager.java:2903) > at > org.apache.hadoop.hdfs.server.namenode.FSEditLogLoader.addNewBlock(FSEditLogLoader.java:1069) > at > org.apache.hadoop.hdfs.server.namenode.FSEditLogLoader.applyEditLogOp(FSEditLogLoader.java:532) > at > org.apache.hadoop.hdfs.server.namenode.FSEditLogLoader.loadEditRecords(FSEditLogLoader.java:249) > at > org.apache.hadoop.hdfs.server.namenode.FSEditLogLoader.loadFSEdits(FSEditLogLoader.java:158) > at > org.apache.hadoop.hdfs.server.namenode.FSImage.loadEdits(FSImage.java:882) > at > org.apache.hadoop.hdfs.server.namenode.FSImage.loadEdits(FSImage.java:863) > at > org.apache.hadoop.hdfs.server.namenode.ha.EditLogTailer.doTailEdits(EditLogTailer.java:293) > at > org.apache.hadoop.hdfs.server.namenode.ha.EditLogTailer$EditLogTailerThread.doWork(EditLogTailer.java:427) > at > org.apache.hadoop.hdfs.server.namenode.ha.EditLogTailer$EditLogTailerThread.access$400(EditLogTailer.java:380) > at > org.apache.hadoop.hdfs.server.namenode.ha.EditLogTailer$EditLogTailerThread$1.run(EditLogTailer.java:397) > {noformat} -- This message was sent by Atlassian JIRA (v6.4.14#64029) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-12907) Allow read-only access to reserved raw for non-superusers
[ https://issues.apache.org/jira/browse/HDFS-12907?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Rushabh S Shah updated HDFS-12907: -- Attachment: HDFS-12907.patch Attaching a simple patch. > Allow read-only access to reserved raw for non-superusers > - > > Key: HDFS-12907 > URL: https://issues.apache.org/jira/browse/HDFS-12907 > Project: Hadoop HDFS > Issue Type: Bug > Components: namenode >Affects Versions: 2.6.0 >Reporter: Daryn Sharp >Assignee: Rushabh S Shah > Attachments: HDFS-12907.patch > > > HDFS-6509 added a special /.reserved/raw path prefix to access the raw file > contents of EZ files. In the simplest sense it doesn't return the FE info in > the {{LocatedBlocks}} so the dfs client doesn't try to decrypt the data. > This facilitates allowing tools like distcp to copy raw bytes. > Access to the raw hierarchy is restricted to superusers. This seems like an > overly broad restriction designed to prevent non-admins from munging the EZ > related xattrs. I believe we should relax the restriction to allow > non-admins to perform read-only operations. Allowing non-superusers to > easily read the raw bytes will be extremely useful for regular users, esp. > for enabling webhdfs client-side encryption. -- This message was sent by Atlassian JIRA (v6.4.14#64029) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-12903) [READ] Fix closing streams in ImageWriter
[ https://issues.apache.org/jira/browse/HDFS-12903?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16282355#comment-16282355 ] Chris Douglas commented on HDFS-12903: -- Duh, misread the Yetus output. This did fix the warning. +1 > [READ] Fix closing streams in ImageWriter > - > > Key: HDFS-12903 > URL: https://issues.apache.org/jira/browse/HDFS-12903 > Project: Hadoop HDFS > Issue Type: Sub-task >Reporter: Íñigo Goiri > Attachments: HDFS-12903-HDFS-9806.001.patch > > > HDFS-12894 showed a FindBug in HDFS-9806. This seems related to HDFS-12881 > when using {{IOUtils.cleanupWithLogger()}}. -- This message was sent by Atlassian JIRA (v6.4.14#64029) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-12751) Ozone: SCM: update container allocated size to container db for all the open containers in ContainerStateManager#close
[ https://issues.apache.org/jira/browse/HDFS-12751?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16282358#comment-16282358 ] Chen Liang commented on HDFS-12751: --- Thanks [~nandakumar131] for the followup! bq. We do not update allocatedBytes here But looks to me it does update allocated bytes, as in the follow code from {{ContainerStateManager#updateContainerState}}. The {{info}} variable is the new container info passed in from {{updateContainerState}}. {code} ContainerInfo containerInfo = new ContainerInfo.Builder() .setContainerName(info.getContainerName()) .setState(newState) .setPipeline(info.getPipeline()) .setAllocatedBytes(info.getAllocatedBytes()) .setUsedBytes(info.getUsedBytes()) .setNumberOfKeys(info.getNumberOfKeys()) .setStateEnterTime(Time.monotonicNow()) .setOwner(info.getOwner()) .build(); {code} bq. Not exactly. Whenever we allocate block... I did miss this part earlier, thanks for point out! But appears to me that this is what BlockManager perceives as the bytes that might get allocated, not necessarily the actual allocated bytes on the container. e.g. client may then terminate before talking to container etc. While the allocatedBytes we got from block report seems to be the more accurate number, precisely the number of bytes the container sees when sending the report. And block report is the trigger of the {{updateContainerState}} code path. Namely, I feel that persisting the number we got from updateContainerState is already the better thing. Additionally, I'm under the impression that {{ContainerMapping}} is the class that interacts the container store, while {{ContainerStateManager}} is purely an in-memory state representation that (currently) does not read/write to container meta store at all. Seems to me that we should keep this abstraction, leave {{ContainerStateManager}} away from container.db and only let {{ContainerMapping}} do the container metadata management. So although we are indeed missing the {{allocatedSize}} from allocateBlock code path, I would prefer to leave this part as-is. What do you think? Nonetheless, containerStateManager.close() not being does look like a bug, will upload a patch later. > Ozone: SCM: update container allocated size to container db for all the open > containers in ContainerStateManager#close > -- > > Key: HDFS-12751 > URL: https://issues.apache.org/jira/browse/HDFS-12751 > Project: Hadoop HDFS > Issue Type: Sub-task > Components: ozone >Reporter: Nanda kumar >Assignee: Chen Liang > > Container allocated size is maintained in memory by > {{ContainerStateManager}}, this has to be updated in container db when we > shutdown SCM. {{ContainerStateManager#close}} will be called during SCM > shutdown, so updating allocated size for all the open containers should be > done here. -- This message was sent by Atlassian JIRA (v6.4.14#64029) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-12818) Support multiple storages in DataNodeCluster / SimulatedFSDataset
[ https://issues.apache.org/jira/browse/HDFS-12818?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Erik Krogen updated HDFS-12818: --- Attachment: HDFS-12818.002.patch > Support multiple storages in DataNodeCluster / SimulatedFSDataset > - > > Key: HDFS-12818 > URL: https://issues.apache.org/jira/browse/HDFS-12818 > Project: Hadoop HDFS > Issue Type: Improvement > Components: datanode, test >Reporter: Erik Krogen >Assignee: Erik Krogen >Priority: Minor > Attachments: HDFS-12818.000.patch, HDFS-12818.001.patch, > HDFS-12818.002.patch > > > Currently {{SimulatedFSDataset}} (and thus, {{DataNodeCluster}} with > {{-simulated}}) only supports a single storage per {{DataNode}}. Given that > the number of storages can have important implications on the performance of > block report processing, it would be useful for these classes to support a > multiple storage configuration. -- This message was sent by Atlassian JIRA (v6.4.14#64029) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-12818) Support multiple storages in DataNodeCluster / SimulatedFSDataset
[ https://issues.apache.org/jira/browse/HDFS-12818?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16282322#comment-16282322 ] Erik Krogen commented on HDFS-12818: Thanks for looking [~shv]! {{getStorage()}} can never return null: {code} private SimulatedStorage getStorage(Block b) { return storages.get(LongMath.mod(b.getBlockId(), storages.size())); } {code} This will always return one of the values contained within {{storages}} which does not contain any null values. {{getBlockMap(b, bpid)}} also cannot return null; it simply passes through to {{SimulatedStorage#getBlockMap(bpid)}}: {code} MapgetBlockMap(String bpid) throws IOException { SimulatedBPStorage bpStorage = map.get(bpid); if (bpStorage == null) { throw new IOException("Nonexistent block pool: " + bpid); } return bpStorage.getBlockMap(); } {code} {{SimulatedBPStorage#getBlockMap()}} returns a {{final}} non-null field so we are good to go throughout. I updated the Javadoc comments to make this more clear; attaching v002 patch. I believe the 1 new checkstyle issue is a long import from a static import in the test; nothing I can do hence why I did not fix it between v000 and v001. > Support multiple storages in DataNodeCluster / SimulatedFSDataset > - > > Key: HDFS-12818 > URL: https://issues.apache.org/jira/browse/HDFS-12818 > Project: Hadoop HDFS > Issue Type: Improvement > Components: datanode, test >Reporter: Erik Krogen >Assignee: Erik Krogen >Priority: Minor > Attachments: HDFS-12818.000.patch, HDFS-12818.001.patch, > HDFS-12818.002.patch > > > Currently {{SimulatedFSDataset}} (and thus, {{DataNodeCluster}} with > {{-simulated}}) only supports a single storage per {{DataNode}}. Given that > the number of storages can have important implications on the performance of > block report processing, it would be useful for these classes to support a > multiple storage configuration. -- This message was sent by Atlassian JIRA (v6.4.14#64029) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-12510) RBF: Add security to UI
[ https://issues.apache.org/jira/browse/HDFS-12510?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16282220#comment-16282220 ] Ajay Kumar commented on HDFS-12510: --- [~elgoiri],[~ywskycn] Thanks for the information. > RBF: Add security to UI > --- > > Key: HDFS-12510 > URL: https://issues.apache.org/jira/browse/HDFS-12510 > Project: Hadoop HDFS > Issue Type: Sub-task >Reporter: Íñigo Goiri > Labels: RBF > > HDFS-12273 implemented the UI for Router Based Federation without security. -- This message was sent by Atlassian JIRA (v6.4.14#64029) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-12907) Allow read-only access to reserved raw for non-superusers
[ https://issues.apache.org/jira/browse/HDFS-12907?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16282250#comment-16282250 ] Andrew Wang commented on HDFS-12907: It's so they don't accidentally write data without xattrs or with the wrong xattrs, which would be essentially corrupt. We also don't want plaintext getting written accidentally. The NameNode does a fair amount of work at create time to provision an EDEK for the file. This logic is beyond the ability of most clients. > Allow read-only access to reserved raw for non-superusers > - > > Key: HDFS-12907 > URL: https://issues.apache.org/jira/browse/HDFS-12907 > Project: Hadoop HDFS > Issue Type: Bug > Components: namenode >Affects Versions: 2.6.0 >Reporter: Daryn Sharp >Assignee: Rushabh S Shah > > HDFS-6509 added a special /.reserved/raw path prefix to access the raw file > contents of EZ files. In the simplest sense it doesn't return the FE info in > the {{LocatedBlocks}} so the dfs client doesn't try to decrypt the data. > This facilitates allowing tools like distcp to copy raw bytes. > Access to the raw hierarchy is restricted to superusers. This seems like an > overly broad restriction designed to prevent non-admins from munging the EZ > related xattrs. I believe we should relax the restriction to allow > non-admins to perform read-only operations. Allowing non-superusers to > easily read the raw bytes will be extremely useful for regular users, esp. > for enabling webhdfs client-side encryption. -- This message was sent by Atlassian JIRA (v6.4.14#64029) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-12825) Fsck report shows config key name for min replication issues
[ https://issues.apache.org/jira/browse/HDFS-12825?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Manoj Govindassamy updated HDFS-12825: -- Labels: newbie (was: incompatibleChange newbie) Hadoop Flags: Incompatible change > Fsck report shows config key name for min replication issues > > > Key: HDFS-12825 > URL: https://issues.apache.org/jira/browse/HDFS-12825 > Project: Hadoop HDFS > Issue Type: Bug > Components: hdfs >Affects Versions: 3.0.0-alpha1 >Reporter: Harshakiran Reddy >Assignee: Gabor Bota >Priority: Minor > Labels: newbie > Attachments: HDFS-12825.001.patch, error.JPG > > > Scenario: > Corrupt the Block in any datanode > Take the *FSCK *Report for that file. > Actual Output: > == > printing the direct configuration in fsck report > {{dfs.namenode.replication.min}} > Expected Output: > > it should be {{MINIMAL BLOCK REPLICATION}} -- This message was sent by Atlassian JIRA (v6.4.14#64029) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-12825) After Block Corrupted, FSCK Report printing the Direct configuration.
[ https://issues.apache.org/jira/browse/HDFS-12825?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16282239#comment-16282239 ] Manoj Govindassamy commented on HDFS-12825: --- Patch looks good to me. +1. Thanks for working on this [~gabor.bota] and thanks for reporting [~Harsha1206], [~usharani]. [~gabor.bota], I would prefer labelling this jira as Incompatible change since it changes the fsck output format. > After Block Corrupted, FSCK Report printing the Direct configuration. > --- > > Key: HDFS-12825 > URL: https://issues.apache.org/jira/browse/HDFS-12825 > Project: Hadoop HDFS > Issue Type: Bug > Components: hdfs >Affects Versions: 3.0.0-alpha1 >Reporter: Harshakiran Reddy >Assignee: Gabor Bota >Priority: Minor > Labels: newbie > Attachments: HDFS-12825.001.patch, error.JPG > > > Scenario: > Corrupt the Block in any datanode > Take the *FSCK *Report for that file. > Actual Output: > == > printing the direct configuration in fsck report > {{dfs.namenode.replication.min}} > Expected Output: > > it should be {{MINIMAL BLOCK REPLICATION}} -- This message was sent by Atlassian JIRA (v6.4.14#64029) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-12890) Ozone: XceiverClient should have upper bound on async requests
[ https://issues.apache.org/jira/browse/HDFS-12890?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Shashikant Banerjee updated HDFS-12890: --- Attachment: HDFS-12890-HDFS-7240.003.patch Thanks [~msingh], [~anu] for the review comments . [~anu], thanks for pointing out the deadlock scenario. As per our discussion, in patch v3, the semaphore ref count is dropped in the error path as well. Raft client has an upper bound of 100 as default value on the no of async requests. Hence, I am keeping it 100 for Standalone as well. Please have a look. > Ozone: XceiverClient should have upper bound on async requests > -- > > Key: HDFS-12890 > URL: https://issues.apache.org/jira/browse/HDFS-12890 > Project: Hadoop HDFS > Issue Type: Sub-task > Components: HDFS-7240 >Affects Versions: HDFS-7240 >Reporter: Shashikant Banerjee >Assignee: Shashikant Banerjee > Fix For: HDFS-7240 > > Attachments: HDFS-12890-HDFS-7240.001.patch, > HDFS-12890-HDFS-7240.002.patch, HDFS-12890-HDFS-7240.003.patch > > > XceiverClient-ratis maintains upper bound on the no of outstanding async > requests . XceiverClient > should also impose an upper bound on the no of outstanding async requests > received from client > for write. -- This message was sent by Atlassian JIRA (v6.4.14#64029) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-12510) RBF: Add security to UI
[ https://issues.apache.org/jira/browse/HDFS-12510?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16282162#comment-16282162 ] Wei Yan commented on HDFS-12510: [~ajayydv] I'll post a patch for HDFS-12512 soon. We'll have more details about the security after that. > RBF: Add security to UI > --- > > Key: HDFS-12510 > URL: https://issues.apache.org/jira/browse/HDFS-12510 > Project: Hadoop HDFS > Issue Type: Sub-task >Reporter: Íñigo Goiri > Labels: RBF > > HDFS-12273 implemented the UI for Router Based Federation without security. -- This message was sent by Atlassian JIRA (v6.4.14#64029) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-9668) Optimize the locking in FsDatasetImpl
[ https://issues.apache.org/jira/browse/HDFS-9668?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16282112#comment-16282112 ] genericqa commented on HDFS-9668: - | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 0s{color} | {color:blue} Docker mode activated. {color} | | {color:red}-1{color} | {color:red} patch {color} | {color:red} 0m 6s{color} | {color:red} HDFS-9668 does not apply to trunk. Rebase required? Wrong Branch? See https://wiki.apache.org/hadoop/HowToContribute for help. {color} | \\ \\ || Subsystem || Report/Notes || | JIRA Issue | HDFS-9668 | | JIRA Patch URL | https://issues.apache.org/jira/secure/attachment/12841233/HDFS-9668-26.patch | | Console output | https://builds.apache.org/job/PreCommit-HDFS-Build/22312/console | | Powered by | Apache Yetus 0.7.0-SNAPSHOT http://yetus.apache.org | This message was automatically generated. > Optimize the locking in FsDatasetImpl > - > > Key: HDFS-9668 > URL: https://issues.apache.org/jira/browse/HDFS-9668 > Project: Hadoop HDFS > Issue Type: Improvement > Components: datanode >Reporter: Jingcheng Du >Assignee: Jingcheng Du > Attachments: HDFS-9668-1.patch, HDFS-9668-10.patch, > HDFS-9668-11.patch, HDFS-9668-12.patch, HDFS-9668-13.patch, > HDFS-9668-14.patch, HDFS-9668-14.patch, HDFS-9668-15.patch, > HDFS-9668-16.patch, HDFS-9668-17.patch, HDFS-9668-18.patch, > HDFS-9668-19.patch, HDFS-9668-19.patch, HDFS-9668-2.patch, > HDFS-9668-20.patch, HDFS-9668-21.patch, HDFS-9668-22.patch, > HDFS-9668-23.patch, HDFS-9668-23.patch, HDFS-9668-24.patch, > HDFS-9668-25.patch, HDFS-9668-26.patch, HDFS-9668-3.patch, HDFS-9668-4.patch, > HDFS-9668-5.patch, HDFS-9668-6.patch, HDFS-9668-7.patch, HDFS-9668-8.patch, > HDFS-9668-9.patch, execution_time.png > > > During the HBase test on a tiered storage of HDFS (WAL is stored in > SSD/RAMDISK, and all other files are stored in HDD), we observe many > long-time BLOCKED threads on FsDatasetImpl in DataNode. The following is part > of the jstack result: > {noformat} > "DataXceiver for client DFSClient_NONMAPREDUCE_-1626037897_1 at > /192.168.50.16:48521 [Receiving block > BP-1042877462-192.168.50.13-1446173170517:blk_1073779272_40852]" - Thread > t@93336 >java.lang.Thread.State: BLOCKED > at > org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.createRbw(FsDatasetImpl.java:) > - waiting to lock <18324c9> (a > org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl) owned by > "DataXceiver for client DFSClient_NONMAPREDUCE_-1626037897_1 at > /192.168.50.16:48520 [Receiving block > BP-1042877462-192.168.50.13-1446173170517:blk_1073779271_40851]" t@93335 > at > org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.createRbw(FsDatasetImpl.java:113) > at > org.apache.hadoop.hdfs.server.datanode.BlockReceiver.(BlockReceiver.java:183) > at > org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:615) > at > org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:137) > at > org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:74) > at > org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:235) > at java.lang.Thread.run(Thread.java:745) >Locked ownable synchronizers: > - None > > "DataXceiver for client DFSClient_NONMAPREDUCE_-1626037897_1 at > /192.168.50.16:48520 [Receiving block > BP-1042877462-192.168.50.13-1446173170517:blk_1073779271_40851]" - Thread > t@93335 >java.lang.Thread.State: RUNNABLE > at java.io.UnixFileSystem.createFileExclusively(Native Method) > at java.io.File.createNewFile(File.java:1012) > at > org.apache.hadoop.hdfs.server.datanode.DatanodeUtil.createTmpFile(DatanodeUtil.java:66) > at > org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.BlockPoolSlice.createRbwFile(BlockPoolSlice.java:271) > at > org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsVolumeImpl.createRbwFile(FsVolumeImpl.java:286) > at > org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.createRbw(FsDatasetImpl.java:1140) > - locked <18324c9> (a > org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl) > at > org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.createRbw(FsDatasetImpl.java:113) > at > org.apache.hadoop.hdfs.server.datanode.BlockReceiver.(BlockReceiver.java:183) > at > org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:615) > at > org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:137) >
[jira] [Commented] (HDFS-6804) race condition between transferring block and appending block causes "Unexpected checksum mismatch exception"
[ https://issues.apache.org/jira/browse/HDFS-6804?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16282096#comment-16282096 ] Wei-Chiu Chuang commented on HDFS-6804: --- Sorry for coming to it late. I wasn't aware of the last comment. There's a minor conflict in the patch. You could also use try () {} syntax to create FSDataOutputStream and that takes care of closing the output stream. But that's more of a personal taste. Other than that the last patch looks good to me. > race condition between transferring block and appending block causes > "Unexpected checksum mismatch exception" > -- > > Key: HDFS-6804 > URL: https://issues.apache.org/jira/browse/HDFS-6804 > Project: Hadoop HDFS > Issue Type: Bug > Components: datanode >Affects Versions: 2.2.0 >Reporter: Gordon Wang >Assignee: Brahma Reddy Battula > Attachments: HDFS-6804-branch-2.8.patch, > Testcase_append_transfer_block.patch > > > We found some error log in the datanode. like this > {noformat} > 2014-07-22 01:49:51,338 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: > Ex > ception for BP-2072804351-192.168.2.104-1406008383435:blk_1073741997_9248 > java.io.IOException: Terminating due to a checksum error.java.io.IOException: > Unexpected checksum mismatch while writing > BP-2072804351-192.168.2.104-1406008383435:blk_1073741997_9248 from > /192.168.2.101:39495 > at > org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:536) > at > org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:703) > at > org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:575) > at > org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:115) > at > org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:68) > at > org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:221) > at java.lang.Thread.run(Thread.java:744) > {noformat} > While on the source datanode, the log says the block is transmitted. > {noformat} > 2014-07-22 01:49:50,805 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: > Da > taTransfer: Transmitted > BP-2072804351-192.168.2.104-1406008383435:blk_1073741997 > _9248 (numBytes=16188152) to /192.168.2.103:50010 > {noformat} > When the destination datanode gets the checksum mismatch, it reports bad > block to NameNode and NameNode marks the replica on the source datanode as > corrupt. But actually, the replica on the source datanode is valid. Because > the replica can pass the checksum verification. > In all, the replica on the source data is wrongly marked as corrupted. -- This message was sent by Atlassian JIRA (v6.4.14#64029) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-9668) Optimize the locking in FsDatasetImpl
[ https://issues.apache.org/jira/browse/HDFS-9668?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16282090#comment-16282090 ] Wei-Chiu Chuang commented on HDFS-9668: --- I am reviewing the patch. It looks like almost ready. Would you like to rebase the latest patch? There are a number of conflicts against the trunk. > Optimize the locking in FsDatasetImpl > - > > Key: HDFS-9668 > URL: https://issues.apache.org/jira/browse/HDFS-9668 > Project: Hadoop HDFS > Issue Type: Improvement > Components: datanode >Reporter: Jingcheng Du >Assignee: Jingcheng Du > Attachments: HDFS-9668-1.patch, HDFS-9668-10.patch, > HDFS-9668-11.patch, HDFS-9668-12.patch, HDFS-9668-13.patch, > HDFS-9668-14.patch, HDFS-9668-14.patch, HDFS-9668-15.patch, > HDFS-9668-16.patch, HDFS-9668-17.patch, HDFS-9668-18.patch, > HDFS-9668-19.patch, HDFS-9668-19.patch, HDFS-9668-2.patch, > HDFS-9668-20.patch, HDFS-9668-21.patch, HDFS-9668-22.patch, > HDFS-9668-23.patch, HDFS-9668-23.patch, HDFS-9668-24.patch, > HDFS-9668-25.patch, HDFS-9668-26.patch, HDFS-9668-3.patch, HDFS-9668-4.patch, > HDFS-9668-5.patch, HDFS-9668-6.patch, HDFS-9668-7.patch, HDFS-9668-8.patch, > HDFS-9668-9.patch, execution_time.png > > > During the HBase test on a tiered storage of HDFS (WAL is stored in > SSD/RAMDISK, and all other files are stored in HDD), we observe many > long-time BLOCKED threads on FsDatasetImpl in DataNode. The following is part > of the jstack result: > {noformat} > "DataXceiver for client DFSClient_NONMAPREDUCE_-1626037897_1 at > /192.168.50.16:48521 [Receiving block > BP-1042877462-192.168.50.13-1446173170517:blk_1073779272_40852]" - Thread > t@93336 >java.lang.Thread.State: BLOCKED > at > org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.createRbw(FsDatasetImpl.java:) > - waiting to lock <18324c9> (a > org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl) owned by > "DataXceiver for client DFSClient_NONMAPREDUCE_-1626037897_1 at > /192.168.50.16:48520 [Receiving block > BP-1042877462-192.168.50.13-1446173170517:blk_1073779271_40851]" t@93335 > at > org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.createRbw(FsDatasetImpl.java:113) > at > org.apache.hadoop.hdfs.server.datanode.BlockReceiver.(BlockReceiver.java:183) > at > org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:615) > at > org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:137) > at > org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:74) > at > org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:235) > at java.lang.Thread.run(Thread.java:745) >Locked ownable synchronizers: > - None > > "DataXceiver for client DFSClient_NONMAPREDUCE_-1626037897_1 at > /192.168.50.16:48520 [Receiving block > BP-1042877462-192.168.50.13-1446173170517:blk_1073779271_40851]" - Thread > t@93335 >java.lang.Thread.State: RUNNABLE > at java.io.UnixFileSystem.createFileExclusively(Native Method) > at java.io.File.createNewFile(File.java:1012) > at > org.apache.hadoop.hdfs.server.datanode.DatanodeUtil.createTmpFile(DatanodeUtil.java:66) > at > org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.BlockPoolSlice.createRbwFile(BlockPoolSlice.java:271) > at > org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsVolumeImpl.createRbwFile(FsVolumeImpl.java:286) > at > org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.createRbw(FsDatasetImpl.java:1140) > - locked <18324c9> (a > org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl) > at > org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.createRbw(FsDatasetImpl.java:113) > at > org.apache.hadoop.hdfs.server.datanode.BlockReceiver.(BlockReceiver.java:183) > at > org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:615) > at > org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:137) > at > org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:74) > at > org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:235) > at java.lang.Thread.run(Thread.java:745) >Locked ownable synchronizers: > - None > {noformat} > We measured the execution of some operations in FsDatasetImpl during the > test. Here following is the result. > !execution_time.png! > The operations of finalizeBlock, addBlock and createRbw on HDD in a heavy > load take a really long time. > It means one slow operation of finalizeBlock, addBlock and createRbw in a > slow storage can block all the other same
[jira] [Commented] (HDFS-12907) Allow read-only access to reserved raw for non-superusers
[ https://issues.apache.org/jira/browse/HDFS-12907?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16281948#comment-16281948 ] Rushabh S Shah commented on HDFS-12907: --- bq. Allowing non-superusers to easily read the raw bytes will be extremely useful for regular users, esp. for enabling webhdfs client-side encryption. I am wondering why only read access ? If the user has access to write in the encrypted directory, we should not block them to access /.reserved/raw/ directory structure. Any thoughts ? > Allow read-only access to reserved raw for non-superusers > - > > Key: HDFS-12907 > URL: https://issues.apache.org/jira/browse/HDFS-12907 > Project: Hadoop HDFS > Issue Type: Bug > Components: namenode >Affects Versions: 2.6.0 >Reporter: Daryn Sharp >Assignee: Rushabh S Shah > > HDFS-6509 added a special /.reserved/raw path prefix to access the raw file > contents of EZ files. In the simplest sense it doesn't return the FE info in > the {{LocatedBlocks}} so the dfs client doesn't try to decrypt the data. > This facilitates allowing tools like distcp to copy raw bytes. > Access to the raw hierarchy is restricted to superusers. This seems like an > overly broad restriction designed to prevent non-admins from munging the EZ > related xattrs. I believe we should relax the restriction to allow > non-admins to perform read-only operations. Allowing non-superusers to > easily read the raw bytes will be extremely useful for regular users, esp. > for enabling webhdfs client-side encryption. -- This message was sent by Atlassian JIRA (v6.4.14#64029) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Comment Edited] (HDFS-10285) Storage Policy Satisfier in Namenode
[ https://issues.apache.org/jira/browse/HDFS-10285?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16281503#comment-16281503 ] Rakesh R edited comment on HDFS-10285 at 12/7/17 1:17 PM: -- Thanks a lot [~anu] for your time and comments. bq. This is the most critical concern that I have. In one of the discussions with SPS developers, they pointed out to me that they want to make sure an SPS move happens within a reasonable time. Apparently, I was told that this is a requirement from HBase. If you have such a need, then the first thing an admin will do is to increase this queue size. Slowly, but steadily SPS will eat into more and more memory of Namenode Increasing Namenode Q will not help to speedup the block movements. It is the Datanode who does actual block movements and need to tune Datanode bandwidth to speedup the block movements. Hence there is no sense in increasing Namenode Q. Infact, that will simply add up the pending tasks at the Namenode side. Let me try putting the memory usage of Namenode Q: Assume there are 1 million directories and users invoked {{dfs#satisfyStoragePolicy(path)}} API on these directories, which is a huge data movement and it may not be a regular case. Again, assume without knowing the advantage of increasing the Q size if some unpleasant user set the Q size to a higher value 1,000,000. Each API call, will add an {{Xattr}} to represent the pending movement and NN maintains list of pending dir InodeId to satisfy the policy, which is {{Long}} value. Each Xattr takes 15 chars {{"system.hdfs.sps"}} for the marking(Note: in the branch code it uses {{system.hdfs.satisfy.storage.policy}}, we will shorten the no. of chars to {{system.hdfs.sps}}). With that, the total space occupy is (xattr + inodeId) size. *(1) Xattr entry* Xattr: 12bytes(Object overhead) + 4bytes(String reference) + 4bytes(byte array) = aligned 24bytes. String "system.hdfs.sps": 40bytes(String Object) + 15bytes(chars) = 56bytes. Its not creating new String("system.hdfs.sps") object every time, so ideally 56bytes count need not be counted every time. Still, I'm considering this. byte[]: 4bytes 84 bytes = (aligned 88bytes * 1,000,000) = 83.923MB If we keep SPS outside or inside Namenode, this much memory space will be occupied as xattribute is used to mark the pending item. *(2) Namenode Q* LinkedList entry = 24bytes Long object = 12bytes(Object overhead) + 8bytes = aligned 24bytes -- 48bytes * 1000,000 = 45.78MB -- 46MB approax, which I feel is a smaller percentage and this may occur in the misconfgured scenario where many {{InodeIds}} queued up. Default Q size value will be recommended as 1000 or even 10,000 = 48bytes * 10,000 = 468.75KB. = 469KB to keep the memory usage very less. Again open to change default value (increase/decrease) based on the feedback. Please feel free to correct me if I missed anything. Thanks! bq. We have an existing pattern Balancer, Mover, DiskBalancer where we have the "scan and move tools" as an external feature to namenode. I am not able to see any convincing reason for breaking this pattern. - {{Scanning}} - For scanning, CPU is the most consumed resource. IIUC, from your previous comments, I'm glad that you agreed that CPU is not an issue. Hence scanning is not a concern. If we run SPS outside, it has to put additional RPC calls for the SPS work and again switching of SPS-ha service has to blindly scan the entire namespace to figure out the xattrs. Now, for handling the switching scenarios, we have to come up with some kind of unfair tweaking logic like, write xattr somewhere in the file and new active SPS service should read it from there and continue. With this, I feel to keep the scanning logic at NN. FYI, NN has existing feature EDEK which also does scanning and we reuses the same code in SPS. Also, I'm re-iterating the point that, SPS does not scan the files its own, user has to call API to satisfy a particular file. - {{Moving blocks}} - It is something assigning the responsibility to Datanode. Presently, Namenode has several logic which does block movement - ReplicationMonitor, EC-Reconstruction, Decommissioning etc. We have added throttling mechanism for the sps block movements also, not to affect the existing data movements. - AFAIK, DiskBalancer is completely run at the Datanode and it looks like Datanode utility. I don't think to compare it with SPS. Coming to the Balancer, which doesn't need any input file paths and it does balancing HDFS cluster based on the utilization. Balancer can run independently as it doesn't
[jira] [Comment Edited] (HDFS-12618) fsck -includeSnapshots reports wrong amount of total blocks
[ https://issues.apache.org/jira/browse/HDFS-12618?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16281811#comment-16281811 ] Wellington Chevreuil edited comment on HDFS-12618 at 12/7/17 1:01 PM: -- bq. How is FSDirectory.getINodesInPath(filePath, FSDirectory.DirOp.READ) broken? I wouldn't say it's broken, as the "filePath" being passed to it in this case does not actually exist. When dealing with appended or truncated files, the original file path may still exist (if the file has never been deleted), but the file version on the snapshot folders may have different blocks. That adds the need to check if the original file path still exists out of snapshots. Thats the reason behind the *dir.getINodesInPath(inodeFile.getName(),FSDirectory.DirOp.READ).validate()*, as *inodeFile.getName()* returns the original file path. So if the file has been deleted, *inodeFile.getName()* actually refers to an invalid path, that's why *dir.getINodesInPath(inodeFile.getName(),FSDirectory.DirOp.READ).validate()* throws the assertion error. An alternative that I thought then was to go through the INodes array from this IIP, comparing it with the *inodeFile.getName()*, since usage of *validate()* is discouraged. was (Author: wchevreuil): bq. How is FSDirectory.getINodesInPath(filePath, FSDirectory.DirOp.READ) broken? I wouldn't say it's broken, as the "filePath" being passed to it in this case does not actually exist. When dealing with appended or truncated files, the original file path may still exist (if the file has never been deleted), but the file version on the snapshot folders may have different blocks. That adds the need to check if the original file path still exists out of snapshots. Thats the reason behind the *dir.getINodesInPath(inodeFile.getName(),FSDirectory.DirOp.READ).validate();*, as *inodeFile.getName()* returns the original file path. So if the file has been deleted, *inodeFile.getName()* actually refers to an invalid path, that's why *dir.getINodesInPath(inodeFile.getName(),FSDirectory.DirOp.READ).validate();* throws the assertion error. An alternative that I thought then was to go through the INodes array from this IIP, comparing it with the *inodeFile.getName()*, since usage of *validate()* is discouraged. > fsck -includeSnapshots reports wrong amount of total blocks > --- > > Key: HDFS-12618 > URL: https://issues.apache.org/jira/browse/HDFS-12618 > Project: Hadoop HDFS > Issue Type: Bug > Components: tools >Affects Versions: 3.0.0-alpha3 >Reporter: Wellington Chevreuil >Assignee: Wellington Chevreuil >Priority: Minor > Attachments: HDFS-121618.initial, HDFS-12618.001.patch, > HDFS-12618.002.patch, HDFS-12618.003.patch, HDFS-12618.004.patch > > > When snapshot is enabled, if a file is deleted but is contained by a > snapshot, *fsck* will not reported blocks for such file, showing different > number of *total blocks* than what is exposed in the Web UI. > This should be fine, as *fsck* provides *-includeSnapshots* option. The > problem is that *-includeSnapshots* option causes *fsck* to count blocks for > every occurrence of a file on snapshots, which is wrong because these blocks > should be counted only once (for instance, if a 100MB file is present on 3 > snapshots, it would still map to one block only in hdfs). This causes fsck to > report much more blocks than what actually exist in hdfs and is reported in > the Web UI. > Here's an example: > 1) HDFS has two files of 2 blocks each: > {noformat} > $ hdfs dfs -ls -R / > drwxr-xr-x - root supergroup 0 2017-10-07 21:21 /snap-test > -rw-r--r-- 1 root supergroup 209715200 2017-10-07 20:16 /snap-test/file1 > -rw-r--r-- 1 root supergroup 209715200 2017-10-07 20:17 /snap-test/file2 > drwxr-xr-x - root supergroup 0 2017-05-13 13:03 /test > {noformat} > 2) There are two snapshots, with the two files present on each of the > snapshots: > {noformat} > $ hdfs dfs -ls -R /snap-test/.snapshot > drwxr-xr-x - root supergroup 0 2017-10-07 21:21 > /snap-test/.snapshot/snap1 > -rw-r--r-- 1 root supergroup 209715200 2017-10-07 20:16 > /snap-test/.snapshot/snap1/file1 > -rw-r--r-- 1 root supergroup 209715200 2017-10-07 20:17 > /snap-test/.snapshot/snap1/file2 > drwxr-xr-x - root supergroup 0 2017-10-07 21:21 > /snap-test/.snapshot/snap2 > -rw-r--r-- 1 root supergroup 209715200 2017-10-07 20:16 > /snap-test/.snapshot/snap2/file1 > -rw-r--r-- 1 root supergroup 209715200 2017-10-07 20:17 > /snap-test/.snapshot/snap2/file2 > {noformat} > 3) *fsck -includeSnapshots* reports 12 blocks in total (4 blocks for the > normal file path, plus 4 blocks for each snapshot path): > {noformat} > $ hdfs fsck /
[jira] [Comment Edited] (HDFS-12618) fsck -includeSnapshots reports wrong amount of total blocks
[ https://issues.apache.org/jira/browse/HDFS-12618?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16281811#comment-16281811 ] Wellington Chevreuil edited comment on HDFS-12618 at 12/7/17 1:01 PM: -- bq. How is FSDirectory.getINodesInPath(filePath, FSDirectory.DirOp.READ) broken? I wouldn't say it's broken, as the "filePath" being passed to it in this case does not actually exist. When dealing with appended or truncated files, the original file path may still exist (if the file has never been deleted), but the file version on the snapshot folders may have different blocks. That adds the need to check if the original file path still exists out of snapshots. Thats the reason behind the *dir.getINodesInPath(inodeFile.getName(),FSDirectory.DirOp.READ).validate();*, as *inodeFile.getName()* returns the original file path. So if the file has been deleted, *inodeFile.getName()* actually refers to an invalid path, that's why *dir.getINodesInPath(inodeFile.getName(),FSDirectory.DirOp.READ).validate();* throws the assertion error. An alternative that I thought then was to go through the INodes array from this IIP, comparing it with the *inodeFile.getName()*, since usage of *validate()* is discouraged. was (Author: wchevreuil): bq. How is FSDirectory.getINodesInPath(filePath, FSDirectory.DirOp.READ) broken? I wouldn't say it's broken, as the "filePath" being passed to it in this case does not actually exist. When dealing with appended or truncated files, the original file path may still exist (if the file has never been deleted), but the file version on the snapshot folders may have different blocks. That adds the need to check if the original file path still exists out of snapshots. Thats the reason behind the *dir.getINodesInPath(inodeFile.getName(),FSDirectory.DirOp.READ).validate();*, as *inodeFile.getName()* returns the original file path. So if the file has been deleted, *inodeFile.getName()* actually refers to an invalid path, that's why *dir.getINodesInPath(inodeFile.getName(),FSDirectory.DirOp.READ).validate();*. An alternative that I thought then was to go through the INodes array from this IIP, comparing it with the *inodeFile.getName()*, since usage of *validate()* is discouraged. > fsck -includeSnapshots reports wrong amount of total blocks > --- > > Key: HDFS-12618 > URL: https://issues.apache.org/jira/browse/HDFS-12618 > Project: Hadoop HDFS > Issue Type: Bug > Components: tools >Affects Versions: 3.0.0-alpha3 >Reporter: Wellington Chevreuil >Assignee: Wellington Chevreuil >Priority: Minor > Attachments: HDFS-121618.initial, HDFS-12618.001.patch, > HDFS-12618.002.patch, HDFS-12618.003.patch, HDFS-12618.004.patch > > > When snapshot is enabled, if a file is deleted but is contained by a > snapshot, *fsck* will not reported blocks for such file, showing different > number of *total blocks* than what is exposed in the Web UI. > This should be fine, as *fsck* provides *-includeSnapshots* option. The > problem is that *-includeSnapshots* option causes *fsck* to count blocks for > every occurrence of a file on snapshots, which is wrong because these blocks > should be counted only once (for instance, if a 100MB file is present on 3 > snapshots, it would still map to one block only in hdfs). This causes fsck to > report much more blocks than what actually exist in hdfs and is reported in > the Web UI. > Here's an example: > 1) HDFS has two files of 2 blocks each: > {noformat} > $ hdfs dfs -ls -R / > drwxr-xr-x - root supergroup 0 2017-10-07 21:21 /snap-test > -rw-r--r-- 1 root supergroup 209715200 2017-10-07 20:16 /snap-test/file1 > -rw-r--r-- 1 root supergroup 209715200 2017-10-07 20:17 /snap-test/file2 > drwxr-xr-x - root supergroup 0 2017-05-13 13:03 /test > {noformat} > 2) There are two snapshots, with the two files present on each of the > snapshots: > {noformat} > $ hdfs dfs -ls -R /snap-test/.snapshot > drwxr-xr-x - root supergroup 0 2017-10-07 21:21 > /snap-test/.snapshot/snap1 > -rw-r--r-- 1 root supergroup 209715200 2017-10-07 20:16 > /snap-test/.snapshot/snap1/file1 > -rw-r--r-- 1 root supergroup 209715200 2017-10-07 20:17 > /snap-test/.snapshot/snap1/file2 > drwxr-xr-x - root supergroup 0 2017-10-07 21:21 > /snap-test/.snapshot/snap2 > -rw-r--r-- 1 root supergroup 209715200 2017-10-07 20:16 > /snap-test/.snapshot/snap2/file1 > -rw-r--r-- 1 root supergroup 209715200 2017-10-07 20:17 > /snap-test/.snapshot/snap2/file2 > {noformat} > 3) *fsck -includeSnapshots* reports 12 blocks in total (4 blocks for the > normal file path, plus 4 blocks for each snapshot path): > {noformat} > $ hdfs fsck / -includeSnapshots > FSCK started by
[jira] [Commented] (HDFS-12618) fsck -includeSnapshots reports wrong amount of total blocks
[ https://issues.apache.org/jira/browse/HDFS-12618?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16281811#comment-16281811 ] Wellington Chevreuil commented on HDFS-12618: - bq. How is FSDirectory.getINodesInPath(filePath, FSDirectory.DirOp.READ) broken? I wouldn't say it's broken, as the "filePath" being passed to it in this case does not actually exist. When dealing with appended or truncated files, the original file path may still exist (if the file has never been deleted), but the file version on the snapshot folders may have different blocks. That adds the need to check if the original file path still exists out of snapshots. Thats the reason behind the *dir.getINodesInPath(inodeFile.getName(),FSDirectory.DirOp.READ).validate();*, as *inodeFile.getName()* returns the original file path. So if the file has been deleted, *inodeFile.getName()* actually refers to an invalid path, that's why *dir.getINodesInPath(inodeFile.getName(),FSDirectory.DirOp.READ).validate();*. An alternative that I thought then was to go through the INodes array from this IIP, comparing it with the *inodeFile.getName()*, since usage of *validate()* is discouraged. > fsck -includeSnapshots reports wrong amount of total blocks > --- > > Key: HDFS-12618 > URL: https://issues.apache.org/jira/browse/HDFS-12618 > Project: Hadoop HDFS > Issue Type: Bug > Components: tools >Affects Versions: 3.0.0-alpha3 >Reporter: Wellington Chevreuil >Assignee: Wellington Chevreuil >Priority: Minor > Attachments: HDFS-121618.initial, HDFS-12618.001.patch, > HDFS-12618.002.patch, HDFS-12618.003.patch, HDFS-12618.004.patch > > > When snapshot is enabled, if a file is deleted but is contained by a > snapshot, *fsck* will not reported blocks for such file, showing different > number of *total blocks* than what is exposed in the Web UI. > This should be fine, as *fsck* provides *-includeSnapshots* option. The > problem is that *-includeSnapshots* option causes *fsck* to count blocks for > every occurrence of a file on snapshots, which is wrong because these blocks > should be counted only once (for instance, if a 100MB file is present on 3 > snapshots, it would still map to one block only in hdfs). This causes fsck to > report much more blocks than what actually exist in hdfs and is reported in > the Web UI. > Here's an example: > 1) HDFS has two files of 2 blocks each: > {noformat} > $ hdfs dfs -ls -R / > drwxr-xr-x - root supergroup 0 2017-10-07 21:21 /snap-test > -rw-r--r-- 1 root supergroup 209715200 2017-10-07 20:16 /snap-test/file1 > -rw-r--r-- 1 root supergroup 209715200 2017-10-07 20:17 /snap-test/file2 > drwxr-xr-x - root supergroup 0 2017-05-13 13:03 /test > {noformat} > 2) There are two snapshots, with the two files present on each of the > snapshots: > {noformat} > $ hdfs dfs -ls -R /snap-test/.snapshot > drwxr-xr-x - root supergroup 0 2017-10-07 21:21 > /snap-test/.snapshot/snap1 > -rw-r--r-- 1 root supergroup 209715200 2017-10-07 20:16 > /snap-test/.snapshot/snap1/file1 > -rw-r--r-- 1 root supergroup 209715200 2017-10-07 20:17 > /snap-test/.snapshot/snap1/file2 > drwxr-xr-x - root supergroup 0 2017-10-07 21:21 > /snap-test/.snapshot/snap2 > -rw-r--r-- 1 root supergroup 209715200 2017-10-07 20:16 > /snap-test/.snapshot/snap2/file1 > -rw-r--r-- 1 root supergroup 209715200 2017-10-07 20:17 > /snap-test/.snapshot/snap2/file2 > {noformat} > 3) *fsck -includeSnapshots* reports 12 blocks in total (4 blocks for the > normal file path, plus 4 blocks for each snapshot path): > {noformat} > $ hdfs fsck / -includeSnapshots > FSCK started by root (auth:SIMPLE) from /127.0.0.1 for path / at Mon Oct 09 > 15:15:36 BST 2017 > Status: HEALTHY > Number of data-nodes:1 > Number of racks: 1 > Total dirs: 6 > Total symlinks: 0 > Replicated Blocks: > Total size: 1258291200 B > Total files: 6 > Total blocks (validated):12 (avg. block size 104857600 B) > Minimally replicated blocks: 12 (100.0 %) > Over-replicated blocks: 0 (0.0 %) > Under-replicated blocks: 0 (0.0 %) > Mis-replicated blocks: 0 (0.0 %) > Default replication factor: 1 > Average block replication: 1.0 > Missing blocks: 0 > Corrupt blocks: 0 > Missing replicas:0 (0.0 %) > {noformat} > 4) Web UI shows the correct number (4 blocks only): > {noformat} > Security is off. > Safemode is off. > 5 files and directories, 4 blocks = 9 total filesystem object(s). > {noformat} > I would like to work on this solution, will propose an initial solution > shortly. -- This message was sent by Atlassian JIRA (v6.4.14#64029)
[jira] [Comment Edited] (HDFS-10285) Storage Policy Satisfier in Namenode
[ https://issues.apache.org/jira/browse/HDFS-10285?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16281503#comment-16281503 ] Rakesh R edited comment on HDFS-10285 at 12/7/17 11:26 AM: --- Thanks a lot [~anu] for your time and comments. bq. This is the most critical concern that I have. In one of the discussions with SPS developers, they pointed out to me that they want to make sure an SPS move happens within a reasonable time. Apparently, I was told that this is a requirement from HBase. If you have such a need, then the first thing an admin will do is to increase this queue size. Slowly, but steadily SPS will eat into more and more memory of Namenode Increasing Namenode Q will not help to speedup the block movements. It is the Datanode who does actual block movements and need to tune Datanode bandwidth to speedup the block movements. Hence there is no sense in increasing Namenode Q. Infact, that will simply add up the pending tasks at the Namenode side. Let me try putting the memory usage of Namenode Q: Assume there are 1 million directories and users invoked {{dfs#satisfyStoragePolicy(path)}} API on these directories, which is a huge data movement and it may not be a regular case. Again, assume without knowing the advantage of increasing the Q size if some unpleasant user set the Q size to a higher value 1,000,000. Each API call, will add an {{Xattr}} to represent the pending movement and NN maintains list of pending dir InodeId to satisfy the policy, which is {{Long}} value. Each Xattr takes 15 chars {{"system.hdfs.sps"}} for the marking(Note: in the branch code it uses {{system.hdfs.satisfy.storage.policy}}, we will shorten the no. of chars to {{system.hdfs.sps}}). With that, the total space occupy is (xattr + inodeId) size. *(1) Xattr entry* Xattr: 12bytes(Object overhead) + 4bytes(String reference) + 4bytes(byte array) = 24 String "system.hdfs.sps": 40bytes(String Object) + 15bytes(chars) = 56bytes. Its creating new String objects every time ideally 56bytes count need not be counted every time. Still, I'm considering this. byte[]: 4bytes 84 bytes = (aligned 88bytes * 1,000,000) = 83.923MB If we keep SPS outside or inside Namenode, this much memory space will be occupied as xattribute is used to mark the pending item. *(2) Namenode Q* LinkedList entry = 24bytes Long object = 12bytes(Object overhead) + 8bytes = aligned 24bytes -- 48bytes * 1000,000 = 45.78MB -- 45MB approax, which I feel is a smaller percentage and this may occur in the misconfgured scenario where many {{InodeIds}} queued up. Default Q size value will be recommended as 10,000 = 48bytes * 10,000 = 468.75KB. = 469KB. Please feel free to correct me if I missed anything. Thanks! bq. We have an existing pattern Balancer, Mover, DiskBalancer where we have the "scan and move tools" as an external feature to namenode. I am not able to see any convincing reason for breaking this pattern. - {{Scanning}} - For scanning, CPU is the most consumed resource. IIUC, from your previous comments, I'm glad that you agreed that CPU is not an issue. Hence scanning is not a concern. If we run SPS outside, it has to put additional RPC calls for the SPS work and again switching of SPS-ha service has to blindly scan the entire namespace to figure out the xattrs. Now, for handling the switching scenarios, we have to come up with some kind of unfair tweaking logic like, write xattr somewhere in the file and new active SPS service should read it from there and continue. With this, I feel to keep the scanning logic at NN. FYI, NN has existing feature EDEK which also does scanning and we reuses the same code in SPS. Also, I'm re-iterating the point that, SPS does not scan the files its own, user has to call API to satisfy a particular file. - {{Moving blocks}} - It is something assigning the responsibility to Datanode. Presently, Namenode has several logic which does block movement - ReplicationMonitor, EC-Reconstruction, Decommissioning etc. We have added throttling mechanism for the sps block movements also, not to affect the existing data movements. - AFAIK, DiskBalancer is completely run at the Datanode and it looks like Datanode utility. I don't think to compare it with SPS. Coming to the Balancer, which doesn't need any input file paths and it does balancing HDFS cluster based on the utilization. Balancer can run independently as it doesn't take any input file path argument and user may not be waiting to finish the balancing work, whereas SPS is exposed to the user via HSM feature. HSM is completely
[jira] [Commented] (HDFS-10453) ReplicationMonitor thread could stuck for long time due to the race between replication and delete of same file in a large cluster.
[ https://issues.apache.org/jira/browse/HDFS-10453?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16281696#comment-16281696 ] He Xiaoqiao commented on HDFS-10453: [~Octivian] The new patch is ready and update based on you mentioned above, FYI. > ReplicationMonitor thread could stuck for long time due to the race between > replication and delete of same file in a large cluster. > --- > > Key: HDFS-10453 > URL: https://issues.apache.org/jira/browse/HDFS-10453 > Project: Hadoop HDFS > Issue Type: Bug > Components: namenode >Affects Versions: 2.4.1, 2.5.2, 2.7.1, 2.6.4 >Reporter: He Xiaoqiao > Attachments: HDFS-10453-branch-2.001.patch, > HDFS-10453-branch-2.003.patch, HDFS-10453-branch-2.7.004.patch, > HDFS-10453.001.patch > > > ReplicationMonitor thread could stuck for long time and loss data with little > probability. Consider the typical scenario: > (1) create and close a file with the default replicas(3); > (2) increase replication (to 10) of the file. > (3) delete the file while ReplicationMonitor is scheduling blocks belong to > that file for replications. > if ReplicationMonitor stuck reappeared, NameNode will print log as: > {code:xml} > 2016-04-19 10:20:48,083 WARN > org.apache.hadoop.hdfs.server.blockmanagement.BlockPlacementPolicy: Failed to > place enough replicas, still in need of 7 to reach 10 > (unavailableStorages=[], storagePolicy=BlockStoragePolicy{HOT:7, > storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, > newBlock=false) For more information, please enable DEBUG log level on > org.apache.hadoop.hdfs.server.blockmanagement.BlockPlacementPolicy > .. > 2016-04-19 10:21:17,184 WARN > org.apache.hadoop.hdfs.server.blockmanagement.BlockPlacementPolicy: Failed to > place enough replicas, still in need of 7 to reach 10 > (unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, > storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, > newBlock=false) For more information, please enable DEBUG log level on > org.apache.hadoop.hdfs.server.blockmanagement.BlockPlacementPolicy > 2016-04-19 10:21:17,184 WARN > org.apache.hadoop.hdfs.protocol.BlockStoragePolicy: Failed to place enough > replicas: expected size is 7 but only 0 storage types can be selected > (replication=10, selected=[], unavailable=[DISK, ARCHIVE], removed=[DISK, > DISK, DISK, DISK, DISK, DISK, DISK], policy=BlockStoragePolicy{HOT:7, > storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}) > 2016-04-19 10:21:17,184 WARN > org.apache.hadoop.hdfs.server.blockmanagement.BlockPlacementPolicy: Failed to > place enough replicas, still in need of 7 to reach 10 > (unavailableStorages=[DISK, ARCHIVE], storagePolicy=BlockStoragePolicy{HOT:7, > storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, > newBlock=false) All required storage types are unavailable: > unavailableStorages=[DISK, ARCHIVE], storagePolicy=BlockStoragePolicy{HOT:7, > storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]} > {code} > This is because 2 threads (#NameNodeRpcServer and #ReplicationMonitor) > process same block at the same moment. > (1) ReplicationMonitor#computeReplicationWorkForBlocks get blocks to > replicate and leave the global lock. > (2) FSNamesystem#delete invoked to delete blocks then clear the reference in > blocksmap, needReplications, etc. the block's NumBytes will set > NO_ACK(Long.MAX_VALUE) which is used to indicate that the block deletion does > not need explicit ACK from the node. > (3) ReplicationMonitor#computeReplicationWorkForBlocks continue to > chooseTargets for the same blocks and no node will be selected after traverse > whole cluster because no node choice satisfy the goodness criteria > (remaining spaces achieve required size Long.MAX_VALUE). > During of stage#3 ReplicationMonitor stuck for long time, especial in a large > cluster. invalidateBlocks & neededReplications continues to grow and no > consumes. it will loss data at the worst. > This can mostly be avoided by skip chooseTarget for BlockCommand.NO_ACK block > and remove it from neededReplications. -- This message was sent by Atlassian JIRA (v6.4.14#64029) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-10453) ReplicationMonitor thread could stuck for long time due to the race between replication and delete of same file in a large cluster.
[ https://issues.apache.org/jira/browse/HDFS-10453?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] He Xiaoqiao updated HDFS-10453: --- Attachment: HDFS-10453-branch-2.7.004.patch attach new patch for branch-2.7 > ReplicationMonitor thread could stuck for long time due to the race between > replication and delete of same file in a large cluster. > --- > > Key: HDFS-10453 > URL: https://issues.apache.org/jira/browse/HDFS-10453 > Project: Hadoop HDFS > Issue Type: Bug > Components: namenode >Affects Versions: 2.4.1, 2.5.2, 2.7.1, 2.6.4 >Reporter: He Xiaoqiao > Attachments: HDFS-10453-branch-2.001.patch, > HDFS-10453-branch-2.003.patch, HDFS-10453-branch-2.7.004.patch, > HDFS-10453.001.patch > > > ReplicationMonitor thread could stuck for long time and loss data with little > probability. Consider the typical scenario: > (1) create and close a file with the default replicas(3); > (2) increase replication (to 10) of the file. > (3) delete the file while ReplicationMonitor is scheduling blocks belong to > that file for replications. > if ReplicationMonitor stuck reappeared, NameNode will print log as: > {code:xml} > 2016-04-19 10:20:48,083 WARN > org.apache.hadoop.hdfs.server.blockmanagement.BlockPlacementPolicy: Failed to > place enough replicas, still in need of 7 to reach 10 > (unavailableStorages=[], storagePolicy=BlockStoragePolicy{HOT:7, > storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, > newBlock=false) For more information, please enable DEBUG log level on > org.apache.hadoop.hdfs.server.blockmanagement.BlockPlacementPolicy > .. > 2016-04-19 10:21:17,184 WARN > org.apache.hadoop.hdfs.server.blockmanagement.BlockPlacementPolicy: Failed to > place enough replicas, still in need of 7 to reach 10 > (unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, > storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, > newBlock=false) For more information, please enable DEBUG log level on > org.apache.hadoop.hdfs.server.blockmanagement.BlockPlacementPolicy > 2016-04-19 10:21:17,184 WARN > org.apache.hadoop.hdfs.protocol.BlockStoragePolicy: Failed to place enough > replicas: expected size is 7 but only 0 storage types can be selected > (replication=10, selected=[], unavailable=[DISK, ARCHIVE], removed=[DISK, > DISK, DISK, DISK, DISK, DISK, DISK], policy=BlockStoragePolicy{HOT:7, > storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}) > 2016-04-19 10:21:17,184 WARN > org.apache.hadoop.hdfs.server.blockmanagement.BlockPlacementPolicy: Failed to > place enough replicas, still in need of 7 to reach 10 > (unavailableStorages=[DISK, ARCHIVE], storagePolicy=BlockStoragePolicy{HOT:7, > storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, > newBlock=false) All required storage types are unavailable: > unavailableStorages=[DISK, ARCHIVE], storagePolicy=BlockStoragePolicy{HOT:7, > storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]} > {code} > This is because 2 threads (#NameNodeRpcServer and #ReplicationMonitor) > process same block at the same moment. > (1) ReplicationMonitor#computeReplicationWorkForBlocks get blocks to > replicate and leave the global lock. > (2) FSNamesystem#delete invoked to delete blocks then clear the reference in > blocksmap, needReplications, etc. the block's NumBytes will set > NO_ACK(Long.MAX_VALUE) which is used to indicate that the block deletion does > not need explicit ACK from the node. > (3) ReplicationMonitor#computeReplicationWorkForBlocks continue to > chooseTargets for the same blocks and no node will be selected after traverse > whole cluster because no node choice satisfy the goodness criteria > (remaining spaces achieve required size Long.MAX_VALUE). > During of stage#3 ReplicationMonitor stuck for long time, especial in a large > cluster. invalidateBlocks & neededReplications continues to grow and no > consumes. it will loss data at the worst. > This can mostly be avoided by skip chooseTarget for BlockCommand.NO_ACK block > and remove it from neededReplications. -- This message was sent by Atlassian JIRA (v6.4.14#64029) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Work started] (HDFS-12892) TestClusterTopology#testChooseRandom fails intermittently
[ https://issues.apache.org/jira/browse/HDFS-12892?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Work on HDFS-12892 started by Zsolt Venczel. > TestClusterTopology#testChooseRandom fails intermittently > - > > Key: HDFS-12892 > URL: https://issues.apache.org/jira/browse/HDFS-12892 > Project: Hadoop HDFS > Issue Type: Bug > Components: test >Affects Versions: 3.0.0 >Reporter: Zsolt Venczel >Assignee: Zsolt Venczel > Labels: flaky-test > > Flaky test failure: > {code:java} > java.lang.AssertionError > Error > Not choosing nodes randomly > Stack Trace > java.lang.AssertionError: Not choosing nodes randomly > at > org.apache.hadoop.net.TestClusterTopology.testChooseRandom(TestClusterTopology.java:170) > {code} -- This message was sent by Atlassian JIRA (v6.4.14#64029) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-12891) TestDNFencingWithReplication.testFencingStress: java.lang.AssertionError
[ https://issues.apache.org/jira/browse/HDFS-12891?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16281575#comment-16281575 ] Zsolt Venczel commented on HDFS-12891: -- * I have double checked the failing unit tests and they seem to be unrelated (are failing with or without my commit). * No test modification was needed as the issue was identified by a correctly executing test case. > TestDNFencingWithReplication.testFencingStress: java.lang.AssertionError > > > Key: HDFS-12891 > URL: https://issues.apache.org/jira/browse/HDFS-12891 > Project: Hadoop HDFS > Issue Type: Bug >Affects Versions: 3.0.0 >Reporter: Zsolt Venczel >Assignee: Zsolt Venczel > Labels: flaky-test > Attachments: HDFS-12891.01.patch > > > {code:java} > java.lang.AssertionError: Test resulted in an unexpected exit > at > org.apache.hadoop.hdfs.server.namenode.ha.TestDNFencingWithReplication.testFencingStress(TestDNFencingWithReplication.java:147) > : > : > 2017-10-19 21:39:40,068 [main] INFO hdfs.MiniDFSCluster > (MiniDFSCluster.java:shutdown(1965)) - Shutting down the Mini HDFS Cluster > 2017-10-19 21:39:40,068 [main] FATAL hdfs.MiniDFSCluster > (MiniDFSCluster.java:shutdown(1968)) - Test resulted in an unexpected exit > 1: java.lang.AssertionError > at org.apache.hadoop.util.ExitUtil.terminate(ExitUtil.java:265) > at > org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$RedundancyMonitor.run(BlockManager.java:4437) > at java.lang.Thread.run(Thread.java:748) > Caused by: java.lang.AssertionError > at > org.apache.hadoop.hdfs.server.blockmanagement.DatanodeDescriptor.addBlocksToBeInvalidated(DatanodeDescriptor.java:641) > at > org.apache.hadoop.hdfs.server.blockmanagement.InvalidateBlocks.invalidateWork(InvalidateBlocks.java:299) > at > org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.invalidateWorkForOneNode(BlockManager.java:4246) > at > org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.computeInvalidateWork(BlockManager.java:1736) > at > org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.computeDatanodeWork(BlockManager.java:4561) > at > org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$RedundancyMonitor.run(BlockManager.java:4418) > ... 1 more > {code} -- This message was sent by Atlassian JIRA (v6.4.14#64029) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-10453) ReplicationMonitor thread could stuck for long time due to the race between replication and delete of same file in a large cluster.
[ https://issues.apache.org/jira/browse/HDFS-10453?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16281571#comment-16281571 ] He Xiaoqiao commented on HDFS-10453: [~Octivian] [~genericqa] Thanks for your comments and tests. Actually you are right, it needs add lock to {{neededReplications}} exactly. In our production env, this patch has update with synchronized of {{neededReplications}}. I will update this patch for a moment. Thanks again. > ReplicationMonitor thread could stuck for long time due to the race between > replication and delete of same file in a large cluster. > --- > > Key: HDFS-10453 > URL: https://issues.apache.org/jira/browse/HDFS-10453 > Project: Hadoop HDFS > Issue Type: Bug > Components: namenode >Affects Versions: 2.4.1, 2.5.2, 2.7.1, 2.6.4 >Reporter: He Xiaoqiao > Attachments: HDFS-10453-branch-2.001.patch, > HDFS-10453-branch-2.003.patch, HDFS-10453.001.patch > > > ReplicationMonitor thread could stuck for long time and loss data with little > probability. Consider the typical scenario: > (1) create and close a file with the default replicas(3); > (2) increase replication (to 10) of the file. > (3) delete the file while ReplicationMonitor is scheduling blocks belong to > that file for replications. > if ReplicationMonitor stuck reappeared, NameNode will print log as: > {code:xml} > 2016-04-19 10:20:48,083 WARN > org.apache.hadoop.hdfs.server.blockmanagement.BlockPlacementPolicy: Failed to > place enough replicas, still in need of 7 to reach 10 > (unavailableStorages=[], storagePolicy=BlockStoragePolicy{HOT:7, > storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, > newBlock=false) For more information, please enable DEBUG log level on > org.apache.hadoop.hdfs.server.blockmanagement.BlockPlacementPolicy > .. > 2016-04-19 10:21:17,184 WARN > org.apache.hadoop.hdfs.server.blockmanagement.BlockPlacementPolicy: Failed to > place enough replicas, still in need of 7 to reach 10 > (unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, > storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, > newBlock=false) For more information, please enable DEBUG log level on > org.apache.hadoop.hdfs.server.blockmanagement.BlockPlacementPolicy > 2016-04-19 10:21:17,184 WARN > org.apache.hadoop.hdfs.protocol.BlockStoragePolicy: Failed to place enough > replicas: expected size is 7 but only 0 storage types can be selected > (replication=10, selected=[], unavailable=[DISK, ARCHIVE], removed=[DISK, > DISK, DISK, DISK, DISK, DISK, DISK], policy=BlockStoragePolicy{HOT:7, > storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}) > 2016-04-19 10:21:17,184 WARN > org.apache.hadoop.hdfs.server.blockmanagement.BlockPlacementPolicy: Failed to > place enough replicas, still in need of 7 to reach 10 > (unavailableStorages=[DISK, ARCHIVE], storagePolicy=BlockStoragePolicy{HOT:7, > storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, > newBlock=false) All required storage types are unavailable: > unavailableStorages=[DISK, ARCHIVE], storagePolicy=BlockStoragePolicy{HOT:7, > storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]} > {code} > This is because 2 threads (#NameNodeRpcServer and #ReplicationMonitor) > process same block at the same moment. > (1) ReplicationMonitor#computeReplicationWorkForBlocks get blocks to > replicate and leave the global lock. > (2) FSNamesystem#delete invoked to delete blocks then clear the reference in > blocksmap, needReplications, etc. the block's NumBytes will set > NO_ACK(Long.MAX_VALUE) which is used to indicate that the block deletion does > not need explicit ACK from the node. > (3) ReplicationMonitor#computeReplicationWorkForBlocks continue to > chooseTargets for the same blocks and no node will be selected after traverse > whole cluster because no node choice satisfy the goodness criteria > (remaining spaces achieve required size Long.MAX_VALUE). > During of stage#3 ReplicationMonitor stuck for long time, especial in a large > cluster. invalidateBlocks & neededReplications continues to grow and no > consumes. it will loss data at the worst. > This can mostly be avoided by skip chooseTarget for BlockCommand.NO_ACK block > and remove it from neededReplications. -- This message was sent by Atlassian JIRA (v6.4.14#64029) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-10453) ReplicationMonitor thread could stuck for long time due to the race between replication and delete of same file in a large cluster.
[ https://issues.apache.org/jira/browse/HDFS-10453?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16281548#comment-16281548 ] genericqa commented on HDFS-10453: -- | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 19m 0s{color} | {color:blue} Docker mode activated. {color} | || || || || {color:brown} Prechecks {color} || | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green} The patch does not contain any @author tags. {color} | | {color:green}+1{color} | {color:green} test4tests {color} | {color:green} 0m 0s{color} | {color:green} The patch appears to include 1 new or modified test files. {color} | || || || || {color:brown} branch-2 Compile Tests {color} || | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 9m 51s{color} | {color:green} branch-2 passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 49s{color} | {color:green} branch-2 passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 31s{color} | {color:green} branch-2 passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 59s{color} | {color:green} branch-2 passed {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 2m 9s{color} | {color:green} branch-2 passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 1m 10s{color} | {color:green} branch-2 passed {color} | || || || || {color:brown} Patch Compile Tests {color} || | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 0m 55s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 45s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 0m 45s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 28s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 55s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 2m 23s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 1m 7s{color} | {color:green} the patch passed {color} | || || || || {color:brown} Other Tests {color} || | {color:red}-1{color} | {color:red} unit {color} | {color:red}107m 16s{color} | {color:red} hadoop-hdfs in the patch failed. {color} | | {color:red}-1{color} | {color:red} asflicense {color} | {color:red} 1m 16s{color} | {color:red} The patch generated 207 ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black}151m 56s{color} | {color:black} {color} | \\ \\ || Reason || Tests || | Unreaped Processes | hadoop-hdfs:24 | | Failed junit tests | hadoop.hdfs.server.datanode.TestDataNodeUUID | | | hadoop.hdfs.server.datanode.checker.TestThrottledAsyncChecker | | | hadoop.hdfs.TestEncryptedTransfer | | | hadoop.hdfs.server.namenode.snapshot.TestSnapshotMetrics | | | hadoop.hdfs.server.namenode.snapshot.TestSnapshottableDirListing | | | hadoop.hdfs.TestDFSPermission | | | hadoop.hdfs.server.datanode.TestBatchIbr | | | hadoop.hdfs.server.namenode.TestNestedEncryptionZones | | | hadoop.hdfs.server.datanode.fsdataset.impl.TestLazyPersistPolicy | | | hadoop.hdfs.server.datanode.TestBpServiceActorScheduler | | | hadoop.hdfs.server.federation.router.TestRouter | | | hadoop.hdfs.server.datanode.metrics.TestDataNodeOutlierDetectionViaMetrics | | | hadoop.hdfs.server.federation.metrics.TestFederationMetrics | | | hadoop.hdfs.server.datanode.fsdataset.impl.TestInterDatanodeProtocol | | | hadoop.hdfs.server.namenode.snapshot.TestNestedSnapshots | | | hadoop.hdfs.server.blockmanagement.TestNameNodePrunesMissingStorages | | | hadoop.hdfs.TestSetTimes | | | hadoop.hdfs.server.blockmanagement.TestHeartbeatHandling | | | hadoop.hdfs.server.blockmanagement.TestPendingInvalidateBlock | | | hadoop.hdfs.server.blockmanagement.TestBlockTokenWithDFS | | | hadoop.hdfs.server.datanode.fsdataset.impl.TestLazyPersistLockedMemory | | | hadoop.hdfs.TestDatanodeDeath | | | hadoop.hdfs.server.blockmanagement.TestReplicationPolicyConsiderLoad | | | hadoop.hdfs.server.datanode.TestDataNodeTransferSocketSize | | | hadoop.hdfs.server.datanode.TestBlockCountersInPendingIBR | | | hadoop.hdfs.server.datanode.TestDataNodeVolumeMetrics
[jira] [Assigned] (HDFS-12308) Erasure Coding: Provide DistributedFileSystem & DFSClient API to return the effective EC policy on a directory or file, including the replication policy
[ https://issues.apache.org/jira/browse/HDFS-12308?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] chencan reassigned HDFS-12308: -- Assignee: chencan > Erasure Coding: Provide DistributedFileSystem & DFSClient API to return the > effective EC policy on a directory or file, including the replication policy > - > > Key: HDFS-12308 > URL: https://issues.apache.org/jira/browse/HDFS-12308 > Project: Hadoop HDFS > Issue Type: Improvement > Components: erasure-coding > Environment: Provide DistributedFileSystem & DFSClient API to return > the effective EC policy on a directory or file, including the replication > policy. Policy name will like {{getNominalErasureCodingPolicy(PATH)}} and > {{getAllNominalErasureCodingPolicies}}. >Reporter: SammiChen >Assignee: chencan > Labels: hdfs-ec-3.0-nice-to-have > -- This message was sent by Atlassian JIRA (v6.4.14#64029) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-11576) Block recovery will fail indefinitely if recovery time > heartbeat interval
[ https://issues.apache.org/jira/browse/HDFS-11576?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Andrew Wang updated HDFS-11576: --- Fix Version/s: 3.0.0 > Block recovery will fail indefinitely if recovery time > heartbeat interval > --- > > Key: HDFS-11576 > URL: https://issues.apache.org/jira/browse/HDFS-11576 > Project: Hadoop HDFS > Issue Type: Bug > Components: datanode, hdfs, namenode >Affects Versions: 2.7.1, 2.7.2, 2.7.3, 3.0.0-alpha1, 3.0.0-alpha2 >Reporter: Lukas Majercak >Assignee: Lukas Majercak >Priority: Critical > Fix For: 3.0.0, 2.9.1 > > Attachments: HDFS-11576-branch-2.00.patch, > HDFS-11576-branch-2.01.patch, HDFS-11576.001.patch, HDFS-11576.002.patch, > HDFS-11576.003.patch, HDFS-11576.004.patch, HDFS-11576.005.patch, > HDFS-11576.006.patch, HDFS-11576.007.patch, HDFS-11576.008.patch, > HDFS-11576.009.patch, HDFS-11576.010.patch, HDFS-11576.011.patch, > HDFS-11576.012.patch, HDFS-11576.013.patch, HDFS-11576.014.patch, > HDFS-11576.015.patch, HDFS-11576.repro.patch > > > Block recovery will fail indefinitely if the time to recover a block is > always longer than the heartbeat interval. Scenario: > 1. DN sends heartbeat > 2. NN sends a recovery command to DN, recoveryID=X > 3. DN starts recovery > 4. DN sends another heartbeat > 5. NN sends a recovery command to DN, recoveryID=X+1 > 6. DN calls commitBlockSyncronization after succeeding with first recovery to > NN, which fails because X < X+1 > ... -- This message was sent by Atlassian JIRA (v6.4.14#64029) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org