[jira] [Commented] (HDFS-5832) Deadlock found in NN between SafeMode#canLeave and DatanodeManager#handleHeartbeat
[ https://issues.apache.org/jira/browse/HDFS-5832?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13882189#comment-13882189 ] Vinay commented on HDFS-5832: - As mentioned in HDFS-5132, Moving SafemodeMonitor#run() checks under fsn write lock, will solve the issue. 1. handleHeartbeat() is always done under fsn readlock 2. incrementSafeBlockCount() and getNumLivedatanodes() will always will be called under writeLock(). By directly seeing the synchronization order it appears to be deadlock. But its avoided by the fsn lock. I think jcarder will not identify the read-write lock mechanism. For this reason only I have made HDFS-5368 duplicate of HDFS-5132 > Deadlock found in NN between SafeMode#canLeave and > DatanodeManager#handleHeartbeat > -- > > Key: HDFS-5832 > URL: https://issues.apache.org/jira/browse/HDFS-5832 > Project: Hadoop HDFS > Issue Type: Bug > Components: namenode >Affects Versions: 3.0.0 >Reporter: Rakesh R >Assignee: Rakesh R >Priority: Blocker > Attachments: HDFS-5832.patch, jcarder_nn_deadlock.gif > > > Found the deadlock during the Namenode startup. Attached jcarder report which > shows the cycles about the deadlock situation. -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Updated] (HDFS-5832) Deadlock found in NN between SafeMode#canLeave and DatanodeManager#handleHeartbeat
[ https://issues.apache.org/jira/browse/HDFS-5832?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vinay updated HDFS-5832: Summary: Deadlock found in NN between SafeMode#canLeave and DatanodeManager#handleHeartbeat (was: Deadeadlock found in NN between SafeMode#canLeave and DatanodeManager#handleHeartbeat) > Deadlock found in NN between SafeMode#canLeave and > DatanodeManager#handleHeartbeat > -- > > Key: HDFS-5832 > URL: https://issues.apache.org/jira/browse/HDFS-5832 > Project: Hadoop HDFS > Issue Type: Bug > Components: namenode >Affects Versions: 3.0.0 >Reporter: Rakesh R >Assignee: Rakesh R >Priority: Blocker > Attachments: HDFS-5832.patch, jcarder_nn_deadlock.gif > > > Found the deadlock during the Namenode startup. Attached jcarder report which > shows the cycles about the deadlock situation. -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Resolved] (HDFS-5827) Children are not inheriting parent's default ACLs
[ https://issues.apache.org/jira/browse/HDFS-5827?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vinay resolved HDFS-5827. - Resolution: Duplicate This will be implemented in HDFS-5616 And tests are already added as part of HDFS-5702. So closing this as duplicate. Thanks Chris for the update > Children are not inheriting parent's default ACLs > - > > Key: HDFS-5827 > URL: https://issues.apache.org/jira/browse/HDFS-5827 > Project: Hadoop HDFS > Issue Type: Bug >Reporter: Vinay >Assignee: Chris Nauroth > > Children are not inheriting the parent's default ACLs on creation. -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Commented] (HDFS-5709) Improve upgrade with existing files and directories named ".snapshot"
[ https://issues.apache.org/jira/browse/HDFS-5709?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13882103#comment-13882103 ] Suresh Srinivas commented on HDFS-5709: --- bq. This feels procedurally pretty similar to the alternative of a special startup option, and it's only required to be set if we hit a .snapshot path. Most users will never need to worry about this config. The problem is, now we have a configuration that needs to be added to NameNode for one time upgrade. The configuration lingers on for ever, unless and informed user can get rid of it. That is the reason why configuration is perhaps not the best way to do this. The mechanism I and [~jingzhao] talked about addresses .reserved as well. Do you plan to address that in this jira? bq. ...that implies scanning the fsimage and logs initially to enumerate all the .snapshot paths, have the user configure their rename rules... Or, just rename it based on convention as I had proposed. User can either glean this from the log (preferred) or from searching fsimage the renamed paths and rename them as they wish. Given that the likelihood of users running into this conflict in the first place is low, this should be acceptable. > Improve upgrade with existing files and directories named ".snapshot" > - > > Key: HDFS-5709 > URL: https://issues.apache.org/jira/browse/HDFS-5709 > Project: Hadoop HDFS > Issue Type: Improvement > Components: namenode >Affects Versions: 3.0.0, 2.2.0 >Reporter: Andrew Wang >Assignee: Andrew Wang > Labels: snapshots, upgrade > Attachments: hdfs-5709-1.patch, hdfs-5709-2.patch, hdfs-5709-3.patch, > hdfs-5709-4.patch, hdfs-5709-5.patch > > > Right now in trunk, upgrade fails messily if the old fsimage or edits refer > to a directory named ".snapshot". We should at least print a better error > message (which I believe was the original intention in HDFS-4666), and [~atm] > proposed automatically renaming these files and directories. -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Commented] (HDFS-5138) Support HDFS upgrade in HA
[ https://issues.apache.org/jira/browse/HDFS-5138?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13882096#comment-13882096 ] Suresh Srinivas commented on HDFS-5138: --- @Todd, I have had some conversation about this [~atm] related to this jira. I had brought up one issue about potentially losing editlogs on JournalNode. I thought that would be addressed before this jira can be committed. I have been very busy and have not been able to provide all my comments. Reviewing this patch has been quite tricky. Here are my almost complete review comments. While some of the issues are minor nits, I do not think this patch and the documentation is ready. I am adding information about the design, the way I understand it. Let me know if I got it wrong. *Upgrade preparation:* # New bits are installed on the cluster nodes. # The cluster is brought down. *Upgrade:* For HA setup, choose one of the namenodes to initiate upgrade on and start it with -upgrade flag. # NN performs preupgrade for all non shared storage directories by moving current to previous.tmp and creating new current. #* Failure here is fine. NN start up fails. Next attempt at upgrade the storage directories are recovered. # NN performs preupgrade of shared edits (NFS/JournalNodes) over RPC. JournalNodes current moved to previous.tmp and new current is created. #* If one of the JN preupgrade fails and upgrade is reattempted, editlog directory could be lost on the JN. Restarting the JN does not fix the issue. # NN performs upgrade of non shared edits by writing new CTIME to current and moving previous.tmp to previous. #* If one of the JN preupgrade fails and upgrade is reattempted, editlog directory could be lost on the JN. Restarting the JN does not fix the issue. # NN performs upgrade of shared edits (NFS/JournalNodes) over RPC. JournalNodes current has new CTIM and previous.tmp is moved to previous. # We need to document that all the JournalNodes must be up. If a JN is irrecoverably lost, configuration must be changed to exclude the JN. *Rollback:* NN is started with rollback flag # For all the non shared directories, the NN checks for canRollBack, essentially ensures that previous directory with the right layout version exists. # For all the shared directories, the NN checks for canRollBack, essentially ensures that previous directory with the right layout version exists. # NN performs rollback for shared directories (moving previous to current) #* If rollback of one of the JN fails, then directories are in inconsistent state. I think any attempt at retrying rollback will fail and will require manually moving files around. I do not think restarting JN fixes this. # We need to document that all the JournalNodes must be up. If a JN is irrecoverably lost, configuration must be changed to exclude the JN. *Finalize:* DFSAdmin command is run to finalize the upgrade. # Active NN performs finalizing of editlog. If JN's fail to finalize, active NN fails to finalize. However it is possible that standby finalizes, leaving the cluster in an inconsistent state. # We need to document that all the JournalNodes must be up. If a JN is irrecoverably lost, configuration must be changed to exclude the JN. Comments on the code in the patch (this is almost complete): Comments: # Minor nit: there are some white space changes # assertAllResultsEqual - for loop can just start with i = 1? Also if the collection objects is of size zero or one, the method can return early. Is there a need to do object.toArray() for these early checks? With that, perhaps the findbugs exclude may not be necessary. # Unit test can be added for methods isAtLeastOneActive, getRpcAddressesForNameserviceId and getProxiesForAllNameNodesInNameservice (I am okay if this is done in a separate jira) # Finalizing upgrade is quite tricky. Consider the following scenarios: #* One NN is active and the other is standby - works fine #* One NN is active and the other is down or all NNs - finalize command throws exception and the user will not know if it has succeeded or failed and what to do next #* No active NN - throws an exception cannot finalize with no active #* BlockPoolSliceStorage.java change seems unnecessary # Why is {{throw new AssertionError("Unreachable code.");}} in QuorumJournalManager.java methods? # FSImage#doRollBack() - when canRollBack is false after checking if non-share directories can rollback, an exception must be immediately thrown, instead of checking shared editlog. Also printing Log.info when storages can be rolled back will help in debugging. # FSEditlog#canRollBackSharedLog should accept StorageInfo instead of Storage # QuorumJournalManager#canRollBack and getJournalCTime can throw AssertionError (from DFSUtil.assertAllResultsEqual()). Is that the right exception to expose or IOException? # Namenode startup throws AssertionError with -rollback option. I think w
[jira] [Commented] (HDFS-5776) Support 'hedged' reads in DFSClient
[ https://issues.apache.org/jira/browse/HDFS-5776?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13882037#comment-13882037 ] stack commented on HDFS-5776: - bq. I might be misunderstanding, but it seems like this should be a client setting, not a datanode setting. Right? [~cmccabe] You are correct. I had it wrong. s/restart DN/restart client/regionserver/ in the above. Thanks C. > Support 'hedged' reads in DFSClient > --- > > Key: HDFS-5776 > URL: https://issues.apache.org/jira/browse/HDFS-5776 > Project: Hadoop HDFS > Issue Type: Improvement > Components: hdfs-client >Affects Versions: 3.0.0 >Reporter: Liang Xie >Assignee: Liang Xie > Attachments: HDFS-5776-v2.txt, HDFS-5776-v3.txt, HDFS-5776-v4.txt, > HDFS-5776-v5.txt, HDFS-5776-v6.txt, HDFS-5776-v7.txt, HDFS-5776-v8.txt, > HDFS-5776.txt > > > This is a placeholder of hdfs related stuff backport from > https://issues.apache.org/jira/browse/HBASE-7509 > The quorum read ability should be helpful especially to optimize read outliers > we can utilize "dfs.dfsclient.quorum.read.threshold.millis" & > "dfs.dfsclient.quorum.read.threadpool.size" to enable/disable the hedged read > ability from client side(e.g. HBase), and by using DFSQuorumReadMetrics, we > could export the interested metric valus into client system(e.g. HBase's > regionserver metric). > The core logic is in pread code path, we decide to goto the original > fetchBlockByteRange or the new introduced fetchBlockByteRangeSpeculative per > the above config items. -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Commented] (HDFS-5138) Support HDFS upgrade in HA
[ https://issues.apache.org/jira/browse/HDFS-5138?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13882035#comment-13882035 ] Hadoop QA commented on HDFS-5138: - {color:red}-1 overall{color}. Here are the results of testing the latest attachment http://issues.apache.org/jira/secure/attachment/12625212/hdfs-5138-branch-2.txt against trunk revision . {color:red}-1 patch{color}. The patch command could not apply the patch. Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/5944//console This message is automatically generated. > Support HDFS upgrade in HA > -- > > Key: HDFS-5138 > URL: https://issues.apache.org/jira/browse/HDFS-5138 > Project: Hadoop HDFS > Issue Type: Bug >Affects Versions: 2.1.1-beta >Reporter: Kihwal Lee >Assignee: Aaron T. Myers >Priority: Blocker > Attachments: HDFS-5138.patch, HDFS-5138.patch, HDFS-5138.patch, > HDFS-5138.patch, HDFS-5138.patch, HDFS-5138.patch, HDFS-5138.patch, > HDFS-5138.patch, HDFS-5138.patch, HDFS-5138.patch, HDFS-5138.patch, > hdfs-5138-branch-2.txt > > > With HA enabled, NN wo't start with "-upgrade". Since there has been a layout > version change between 2.0.x and 2.1.x, starting NN in upgrade mode was > necessary when deploying 2.1.x to an existing 2.0.x cluster. But the only way > to get around this was to disable HA and upgrade. > The NN and the cluster cannot be flipped back to HA until the upgrade is > finalized. If HA is disabled only on NN for layout upgrade and HA is turned > back on without involving DNs, things will work, but finaliizeUpgrade won't > work (the NN is in HA and it cannot be in upgrade mode) and DN's upgrade > snapshots won't get removed. > We will need a different ways of doing layout upgrade and upgrade snapshot. > I am marking this as a 2.1.1-beta blocker based on feedback from others. If > there is a reasonable workaround that does not increase maintenance window > greatly, we can lower its priority from blocker to critical. -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Updated] (HDFS-5138) Support HDFS upgrade in HA
[ https://issues.apache.org/jira/browse/HDFS-5138?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Todd Lipcon updated HDFS-5138: -- Attachment: hdfs-5138-branch-2.txt I cherry-picked HDFS-5721 and HDFS-5719 to branch-2, since they had some trivial changes which changed a lot of indentation (thus creating conflicts). After doing so, the backport of this JIRA was pretty clean (just imports and some changed context in DFSUtil). I ran all the modified test suites on branch-2 with this backport patch and they passed. ATM, can you take a quick look at this? > Support HDFS upgrade in HA > -- > > Key: HDFS-5138 > URL: https://issues.apache.org/jira/browse/HDFS-5138 > Project: Hadoop HDFS > Issue Type: Bug >Affects Versions: 2.1.1-beta >Reporter: Kihwal Lee >Assignee: Aaron T. Myers >Priority: Blocker > Attachments: HDFS-5138.patch, HDFS-5138.patch, HDFS-5138.patch, > HDFS-5138.patch, HDFS-5138.patch, HDFS-5138.patch, HDFS-5138.patch, > HDFS-5138.patch, HDFS-5138.patch, HDFS-5138.patch, HDFS-5138.patch, > hdfs-5138-branch-2.txt > > > With HA enabled, NN wo't start with "-upgrade". Since there has been a layout > version change between 2.0.x and 2.1.x, starting NN in upgrade mode was > necessary when deploying 2.1.x to an existing 2.0.x cluster. But the only way > to get around this was to disable HA and upgrade. > The NN and the cluster cannot be flipped back to HA until the upgrade is > finalized. If HA is disabled only on NN for layout upgrade and HA is turned > back on without involving DNs, things will work, but finaliizeUpgrade won't > work (the NN is in HA and it cannot be in upgrade mode) and DN's upgrade > snapshots won't get removed. > We will need a different ways of doing layout upgrade and upgrade snapshot. > I am marking this as a 2.1.1-beta blocker based on feedback from others. If > there is a reasonable workaround that does not increase maintenance window > greatly, we can lower its priority from blocker to critical. -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Updated] (HDFS-5721) sharedEditsImage in Namenode#initializeSharedEdits() should be closed before method returns
[ https://issues.apache.org/jira/browse/HDFS-5721?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Todd Lipcon updated HDFS-5721: -- Fix Version/s: 2.4.0 Cherry-picked to branch-2 to make HDFS-5138 apply more cleanly. > sharedEditsImage in Namenode#initializeSharedEdits() should be closed before > method returns > --- > > Key: HDFS-5721 > URL: https://issues.apache.org/jira/browse/HDFS-5721 > Project: Hadoop HDFS > Issue Type: Improvement >Reporter: Ted Yu >Assignee: Ted Yu >Priority: Minor > Fix For: 3.0.0, 2.4.0 > > Attachments: hdfs-5721-v1.txt, hdfs-5721-v2.txt, hdfs-5721-v3.txt > > > At line 901: > {code} > FSImage sharedEditsImage = new FSImage(conf, > Lists.newArrayList(), > sharedEditsDirs); > {code} > sharedEditsImage is not closed before the method returns. -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Updated] (HDFS-5719) FSImage#doRollback() should close prevState before return
[ https://issues.apache.org/jira/browse/HDFS-5719?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Todd Lipcon updated HDFS-5719: -- Fix Version/s: 2.4.0 Cherry-picked this to branch-2 to make HDFS-5138 apply cleaner. > FSImage#doRollback() should close prevState before return > - > > Key: HDFS-5719 > URL: https://issues.apache.org/jira/browse/HDFS-5719 > Project: Hadoop HDFS > Issue Type: Bug > Components: namenode >Reporter: Ted Yu >Assignee: Ted Yu >Priority: Minor > Fix For: 3.0.0, 2.4.0 > > Attachments: hdfs-5719.txt > > > {code} > FSImage prevState = new FSImage(conf); > {code} > prevState should be closed before return from doRollback() -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Commented] (HDFS-5138) Support HDFS upgrade in HA
[ https://issues.apache.org/jira/browse/HDFS-5138?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13882026#comment-13882026 ] Hudson commented on HDFS-5138: -- SUCCESS: Integrated in Hadoop-trunk-Commit #5038 (See [https://builds.apache.org/job/Hadoop-trunk-Commit/5038/]) HDFS-5138. Support HDFS upgrade in HA. Contributed by Aaron T. Myers. (todd: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1561381) * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/dev-support/findbugsExcludeFile.xml * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/contrib/bkjournal/src/main/java/org/apache/hadoop/contrib/bkjournal/BookKeeperJournalManager.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/contrib/bkjournal/src/test/java/org/apache/hadoop/contrib/bkjournal/TestBookKeeperAsHASharedDir.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/DFSUtil.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/HAUtil.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/qjournal/client/AsyncLogger.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/qjournal/client/AsyncLoggerSet.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/qjournal/client/IPCLoggerChannel.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/qjournal/client/QuorumJournalManager.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/qjournal/protocol/QJournalProtocol.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/qjournal/protocolPB/QJournalProtocolServerSideTranslatorPB.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/qjournal/protocolPB/QJournalProtocolTranslatorPB.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/qjournal/server/GetJournalEditServlet.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/qjournal/server/JNStorage.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/qjournal/server/Journal.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/qjournal/server/JournalNode.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/qjournal/server/JournalNodeRpcServer.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/common/Storage.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/common/StorageInfo.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/BlockPoolSliceStorage.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/BackupJournalManager.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/BackupNode.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSEditLog.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSImage.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSNamesystem.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FileJournalManager.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/JournalManager.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/JournalSet.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/NNStorage.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/NNUpgradeUtil.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/NameNode.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/ha/BootstrapStandby.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/tools/DFSAdmin.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/proto/QJournalProtocol.proto * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/site/apt/HDFSHighAvailabilityWi
[jira] [Commented] (HDFS-5138) Support HDFS upgrade in HA
[ https://issues.apache.org/jira/browse/HDFS-5138?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13882019#comment-13882019 ] Todd Lipcon commented on HDFS-5138: --- I committed this to trunk. Looks like the patch has a few conflicts against branch-2, so I didn't commit there yet. Leaving open for branch-2 commit. > Support HDFS upgrade in HA > -- > > Key: HDFS-5138 > URL: https://issues.apache.org/jira/browse/HDFS-5138 > Project: Hadoop HDFS > Issue Type: Bug >Affects Versions: 2.1.1-beta >Reporter: Kihwal Lee >Assignee: Aaron T. Myers >Priority: Blocker > Attachments: HDFS-5138.patch, HDFS-5138.patch, HDFS-5138.patch, > HDFS-5138.patch, HDFS-5138.patch, HDFS-5138.patch, HDFS-5138.patch, > HDFS-5138.patch, HDFS-5138.patch, HDFS-5138.patch, HDFS-5138.patch > > > With HA enabled, NN wo't start with "-upgrade". Since there has been a layout > version change between 2.0.x and 2.1.x, starting NN in upgrade mode was > necessary when deploying 2.1.x to an existing 2.0.x cluster. But the only way > to get around this was to disable HA and upgrade. > The NN and the cluster cannot be flipped back to HA until the upgrade is > finalized. If HA is disabled only on NN for layout upgrade and HA is turned > back on without involving DNs, things will work, but finaliizeUpgrade won't > work (the NN is in HA and it cannot be in upgrade mode) and DN's upgrade > snapshots won't get removed. > We will need a different ways of doing layout upgrade and upgrade snapshot. > I am marking this as a 2.1.1-beta blocker based on feedback from others. If > there is a reasonable workaround that does not increase maintenance window > greatly, we can lower its priority from blocker to critical. -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Commented] (HDFS-5832) Deadeadlock found in NN between SafeMode#canLeave and DatanodeManager#handleHeartbeat
[ https://issues.apache.org/jira/browse/HDFS-5832?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13881997#comment-13881997 ] Uma Maheswara Rao G commented on HDFS-5832: --- I have read the comments from HDFS-5368 and HDFS-5132. It was closed because, safemode checks done under fsn lock and hb also under fsn lock. So, they were protected with fsn lock. I think because of that fix, earlier issue closed. Do you see any cases, where they are running outside of fsn lock? > Deadeadlock found in NN between SafeMode#canLeave and > DatanodeManager#handleHeartbeat > - > > Key: HDFS-5832 > URL: https://issues.apache.org/jira/browse/HDFS-5832 > Project: Hadoop HDFS > Issue Type: Bug > Components: namenode >Affects Versions: 3.0.0 >Reporter: Rakesh R >Assignee: Rakesh R >Priority: Blocker > Attachments: HDFS-5832.patch, jcarder_nn_deadlock.gif > > > Found the deadlock during the Namenode startup. Attached jcarder report which > shows the cycles about the deadlock situation. -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Commented] (HDFS-5832) Deadeadlock found in NN between SafeMode#canLeave and DatanodeManager#handleHeartbeat
[ https://issues.apache.org/jira/browse/HDFS-5832?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13881945#comment-13881945 ] Hadoop QA commented on HDFS-5832: - {color:red}-1 overall{color}. Here are the results of testing the latest attachment http://issues.apache.org/jira/secure/attachment/12625184/HDFS-5832.patch against trunk revision . {color:green}+1 @author{color}. The patch does not contain any @author tags. {color:red}-1 tests included{color}. The patch doesn't appear to include any new or modified tests. Please justify why no new tests are needed for this patch. Also please list what manual steps were performed to verify this patch. {color:green}+1 javac{color}. The applied patch does not increase the total number of javac compiler warnings. {color:green}+1 javadoc{color}. The javadoc tool did not generate any warning messages. {color:green}+1 eclipse:eclipse{color}. The patch built with eclipse:eclipse. {color:green}+1 findbugs{color}. The patch does not introduce any new Findbugs (version 1.3.9) warnings. {color:green}+1 release audit{color}. The applied patch does not increase the total number of release audit warnings. {color:green}+1 core tests{color}. The patch passed unit tests in hadoop-hdfs-project/hadoop-hdfs. {color:green}+1 contrib tests{color}. The patch passed contrib unit tests. Test results: https://builds.apache.org/job/PreCommit-HDFS-Build/5943//testReport/ Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/5943//console This message is automatically generated. > Deadeadlock found in NN between SafeMode#canLeave and > DatanodeManager#handleHeartbeat > - > > Key: HDFS-5832 > URL: https://issues.apache.org/jira/browse/HDFS-5832 > Project: Hadoop HDFS > Issue Type: Bug > Components: namenode >Affects Versions: 3.0.0 >Reporter: Rakesh R >Assignee: Rakesh R >Priority: Blocker > Attachments: HDFS-5832.patch, jcarder_nn_deadlock.gif > > > Found the deadlock during the Namenode startup. Attached jcarder report which > shows the cycles about the deadlock situation. -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Commented] (HDFS-5343) When cat command is issued on snapshot files getting unexpected result
[ https://issues.apache.org/jira/browse/HDFS-5343?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13881932#comment-13881932 ] Hudson commented on HDFS-5343: -- SUCCESS: Integrated in Hadoop-trunk-Commit #5037 (See [https://builds.apache.org/job/Hadoop-trunk-Commit/5037/]) HDFS-5343. When cat command is issued on snapshot files, getting unexpected result. Contributed by Sathish. (umamahesh: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1561325) * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/DFSInputStream.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/snapshot/TestSnapshotFileLength.java > When cat command is issued on snapshot files getting unexpected result > -- > > Key: HDFS-5343 > URL: https://issues.apache.org/jira/browse/HDFS-5343 > Project: Hadoop HDFS > Issue Type: Bug > Components: hdfs-client >Reporter: sathish >Assignee: sathish > Fix For: 3.0.0, 2.3.0 > > Attachments: HDFS-5343-0003.patch, HDFS-5343-0004.patch, > HDFS-5343-0005.patch, HDFS-5343-0006.patch, HDFS-5343-002.patch > > > first if we create one file with some file length and take the snapshot of > that file,and again append some data through append method to that file,then > if we do cat command operation on snapshot of that file,in general it should > dispaly the data what we added with create operation,but it is displaying the > total data i.e. create +_ appended data. > but if we do the same operation and if we read the contents of snapshot file > through input stream it is just displaying the data created in snapshoted > files. > in this the behaviour of cat command and reading through inputstream is > getting different -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Commented] (HDFS-5343) When cat command is issued on snapshot files getting unexpected result
[ https://issues.apache.org/jira/browse/HDFS-5343?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13881931#comment-13881931 ] Uma Maheswara Rao G commented on HDFS-5343: --- Thanks a lot, Sathish for the patch. I have just committed this to trunk and branch-2 > When cat command is issued on snapshot files getting unexpected result > -- > > Key: HDFS-5343 > URL: https://issues.apache.org/jira/browse/HDFS-5343 > Project: Hadoop HDFS > Issue Type: Bug > Components: hdfs-client >Reporter: sathish >Assignee: sathish > Fix For: 3.0.0, 2.3.0 > > Attachments: HDFS-5343-0003.patch, HDFS-5343-0004.patch, > HDFS-5343-0005.patch, HDFS-5343-0006.patch, HDFS-5343-002.patch > > > first if we create one file with some file length and take the snapshot of > that file,and again append some data through append method to that file,then > if we do cat command operation on snapshot of that file,in general it should > dispaly the data what we added with create operation,but it is displaying the > total data i.e. create +_ appended data. > but if we do the same operation and if we read the contents of snapshot file > through input stream it is just displaying the data created in snapshoted > files. > in this the behaviour of cat command and reading through inputstream is > getting different -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Updated] (HDFS-5343) When cat command is issued on snapshot files getting unexpected result
[ https://issues.apache.org/jira/browse/HDFS-5343?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Uma Maheswara Rao G updated HDFS-5343: -- Resolution: Fixed Fix Version/s: 2.3.0 3.0.0 Hadoop Flags: Reviewed Status: Resolved (was: Patch Available) > When cat command is issued on snapshot files getting unexpected result > -- > > Key: HDFS-5343 > URL: https://issues.apache.org/jira/browse/HDFS-5343 > Project: Hadoop HDFS > Issue Type: Bug > Components: hdfs-client >Reporter: sathish >Assignee: sathish > Fix For: 3.0.0, 2.3.0 > > Attachments: HDFS-5343-0003.patch, HDFS-5343-0004.patch, > HDFS-5343-0005.patch, HDFS-5343-0006.patch, HDFS-5343-002.patch > > > first if we create one file with some file length and take the snapshot of > that file,and again append some data through append method to that file,then > if we do cat command operation on snapshot of that file,in general it should > dispaly the data what we added with create operation,but it is displaying the > total data i.e. create +_ appended data. > but if we do the same operation and if we read the contents of snapshot file > through input stream it is just displaying the data created in snapshoted > files. > in this the behaviour of cat command and reading through inputstream is > getting different -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Commented] (HDFS-5832) Deadeadlock found in NN between SafeMode#canLeave and DatanodeManager#handleHeartbeat
[ https://issues.apache.org/jira/browse/HDFS-5832?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13881927#comment-13881927 ] Rakesh R commented on HDFS-5832: Thanks Uma for the interest. Looks like similar, but I could see still there is a chance of reverse locking in the code. I'm seeing the following code flow like, - In one side, when running SafeModeMonitor thread, it is invoking safeMode#canLeave() by acquiring 'safeMode.this' lock and here it is asking for the DataNodeManger#getNumLiveDataNodes by trying to acquire 'datanodeMap' lock. - On the other side, DN heartbeat comes and DatanodeManager#handleHeartbeat() has acquired 'datanodeMap' lock and calling namesystem#isInSafeMode(), which inturn trying to acquire 'safeModeInfo.this' lock > Deadeadlock found in NN between SafeMode#canLeave and > DatanodeManager#handleHeartbeat > - > > Key: HDFS-5832 > URL: https://issues.apache.org/jira/browse/HDFS-5832 > Project: Hadoop HDFS > Issue Type: Bug > Components: namenode >Affects Versions: 3.0.0 >Reporter: Rakesh R >Assignee: Rakesh R >Priority: Blocker > Attachments: HDFS-5832.patch, jcarder_nn_deadlock.gif > > > Found the deadlock during the Namenode startup. Attached jcarder report which > shows the cycles about the deadlock situation. -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Commented] (HDFS-5832) Deadeadlock found in NN between SafeMode#canLeave and DatanodeManager#handleHeartbeat
[ https://issues.apache.org/jira/browse/HDFS-5832?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13881912#comment-13881912 ] Uma Maheswara Rao G commented on HDFS-5832: --- Is this same as HDFS-5368? That was closed as dupe of HDFS-5132 > Deadeadlock found in NN between SafeMode#canLeave and > DatanodeManager#handleHeartbeat > - > > Key: HDFS-5832 > URL: https://issues.apache.org/jira/browse/HDFS-5832 > Project: Hadoop HDFS > Issue Type: Bug > Components: namenode >Affects Versions: 3.0.0 >Reporter: Rakesh R >Assignee: Rakesh R >Priority: Blocker > Attachments: HDFS-5832.patch, jcarder_nn_deadlock.gif > > > Found the deadlock during the Namenode startup. Attached jcarder report which > shows the cycles about the deadlock situation. -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Updated] (HDFS-5832) Deadeadlock found in NN between SafeMode#canLeave and DatanodeManager#handleHeartbeat
[ https://issues.apache.org/jira/browse/HDFS-5832?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Rakesh R updated HDFS-5832: --- Status: Patch Available (was: Open) > Deadeadlock found in NN between SafeMode#canLeave and > DatanodeManager#handleHeartbeat > - > > Key: HDFS-5832 > URL: https://issues.apache.org/jira/browse/HDFS-5832 > Project: Hadoop HDFS > Issue Type: Bug > Components: namenode >Affects Versions: 3.0.0 >Reporter: Rakesh R >Assignee: Rakesh R >Priority: Blocker > Attachments: HDFS-5832.patch, jcarder_nn_deadlock.gif > > > Found the deadlock during the Namenode startup. Attached jcarder report which > shows the cycles about the deadlock situation. -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Updated] (HDFS-5832) Deadeadlock found in NN between SafeMode#canLeave and DatanodeManager#handleHeartbeat
[ https://issues.apache.org/jira/browse/HDFS-5832?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Rakesh R updated HDFS-5832: --- Attachment: HDFS-5832.patch Attaching a proposal where I have moved the namesystem.isInSafeMode(); out of datanodeMap lock. Could someone help me to validate this case. Thanks > Deadeadlock found in NN between SafeMode#canLeave and > DatanodeManager#handleHeartbeat > - > > Key: HDFS-5832 > URL: https://issues.apache.org/jira/browse/HDFS-5832 > Project: Hadoop HDFS > Issue Type: Bug > Components: namenode >Affects Versions: 3.0.0 >Reporter: Rakesh R >Assignee: Rakesh R >Priority: Blocker > Attachments: HDFS-5832.patch, jcarder_nn_deadlock.gif > > > Found the deadlock during the Namenode startup. Attached jcarder report which > shows the cycles about the deadlock situation. -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Updated] (HDFS-5832) Deadeadlock found in NN between SafeMode#canLeave and DatanodeManager#handleHeartbeat
[ https://issues.apache.org/jira/browse/HDFS-5832?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Rakesh R updated HDFS-5832: --- Attachment: jcarder_nn_deadlock.gif > Deadeadlock found in NN between SafeMode#canLeave and > DatanodeManager#handleHeartbeat > - > > Key: HDFS-5832 > URL: https://issues.apache.org/jira/browse/HDFS-5832 > Project: Hadoop HDFS > Issue Type: Bug > Components: namenode >Affects Versions: 3.0.0 >Reporter: Rakesh R >Assignee: Rakesh R >Priority: Blocker > Attachments: jcarder_nn_deadlock.gif > > > Found the deadlock during the Namenode startup. Attached jcarder report which > shows the cycles about the deadlock situation. -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Commented] (HDFS-5343) When cat command is issued on snapshot files getting unexpected result
[ https://issues.apache.org/jira/browse/HDFS-5343?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13881900#comment-13881900 ] Hadoop QA commented on HDFS-5343: - {color:green}+1 overall{color}. Here are the results of testing the latest attachment http://issues.apache.org/jira/secure/attachment/12625029/HDFS-5343-0006.patch against trunk revision . {color:green}+1 @author{color}. The patch does not contain any @author tags. {color:green}+1 tests included{color}. The patch appears to include 1 new or modified test files. {color:green}+1 javac{color}. The applied patch does not increase the total number of javac compiler warnings. {color:green}+1 javadoc{color}. The javadoc tool did not generate any warning messages. {color:green}+1 eclipse:eclipse{color}. The patch built with eclipse:eclipse. {color:green}+1 findbugs{color}. The patch does not introduce any new Findbugs (version 1.3.9) warnings. {color:green}+1 release audit{color}. The applied patch does not increase the total number of release audit warnings. {color:green}+1 core tests{color}. The patch passed unit tests in hadoop-hdfs-project/hadoop-hdfs. {color:green}+1 contrib tests{color}. The patch passed contrib unit tests. Test results: https://builds.apache.org/job/PreCommit-HDFS-Build/5942//testReport/ Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/5942//console This message is automatically generated. > When cat command is issued on snapshot files getting unexpected result > -- > > Key: HDFS-5343 > URL: https://issues.apache.org/jira/browse/HDFS-5343 > Project: Hadoop HDFS > Issue Type: Bug > Components: hdfs-client >Reporter: sathish >Assignee: sathish > Attachments: HDFS-5343-0003.patch, HDFS-5343-0004.patch, > HDFS-5343-0005.patch, HDFS-5343-0006.patch, HDFS-5343-002.patch > > > first if we create one file with some file length and take the snapshot of > that file,and again append some data through append method to that file,then > if we do cat command operation on snapshot of that file,in general it should > dispaly the data what we added with create operation,but it is displaying the > total data i.e. create +_ appended data. > but if we do the same operation and if we read the contents of snapshot file > through input stream it is just displaying the data created in snapshoted > files. > in this the behaviour of cat command and reading through inputstream is > getting different -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Created] (HDFS-5832) Deadeadlock found in NN between SafeMode#canLeave and DatanodeManager#handleHeartbeat
Rakesh R created HDFS-5832: -- Summary: Deadeadlock found in NN between SafeMode#canLeave and DatanodeManager#handleHeartbeat Key: HDFS-5832 URL: https://issues.apache.org/jira/browse/HDFS-5832 Project: Hadoop HDFS Issue Type: Bug Components: namenode Affects Versions: 3.0.0 Reporter: Rakesh R Assignee: Rakesh R Priority: Blocker Found the deadlock during the Namenode startup. Attached jcarder report which shows the cycles about the deadlock situation. -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Commented] (HDFS-5789) Some of snapshot APIs missing checkOperation double check in fsn
[ https://issues.apache.org/jira/browse/HDFS-5789?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13881884#comment-13881884 ] Hudson commented on HDFS-5789: -- SUCCESS: Integrated in Hadoop-Hdfs-trunk #1653 (See [https://builds.apache.org/job/Hadoop-Hdfs-trunk/1653/]) HDFS-5789. Some of snapshot APIs missing checkOperation double check in fsn. Contributed by Uma Maheswara Rao G. (umamahesh: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1560731) * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSNamesystem.java > Some of snapshot APIs missing checkOperation double check in fsn > > > Key: HDFS-5789 > URL: https://issues.apache.org/jira/browse/HDFS-5789 > Project: Hadoop HDFS > Issue Type: Bug > Components: namenode >Affects Versions: 3.0.0 >Reporter: Uma Maheswara Rao G >Assignee: Uma Maheswara Rao G > Fix For: 3.0.0, 2.3.0 > > Attachments: HDFS-5789.patch > > > HDFS-4591 introduced double checked for HA state while taking fsn lock. > checkoperation made before actually taking lock and after the lock again. > This pattern missed in some of the snapshot APIs and cache management related > APIs. -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Commented] (HDFS-5241) Provide alternate queuing audit logger to reduce logging contention
[ https://issues.apache.org/jira/browse/HDFS-5241?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13881878#comment-13881878 ] Hudson commented on HDFS-5241: -- SUCCESS: Integrated in Hadoop-Hdfs-trunk #1653 (See [https://builds.apache.org/job/Hadoop-Hdfs-trunk/1653/]) HDFS-5241. Provide alternate queuing audit logger to reduce logging contention (daryn) (daryn: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1560761) * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/DFSConfigKeys.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSNamesystem.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestAuditLogs.java > Provide alternate queuing audit logger to reduce logging contention > --- > > Key: HDFS-5241 > URL: https://issues.apache.org/jira/browse/HDFS-5241 > Project: Hadoop HDFS > Issue Type: Sub-task > Components: namenode >Affects Versions: 2.0.0-alpha, 3.0.0 >Reporter: Daryn Sharp >Assignee: Daryn Sharp > Attachments: HDFS-5241.patch, HDFS-5241.patch > > > The default audit logger has extremely poor performance. The internal > synchronization of log4j causes massive contention between the call handlers > (100 by default) which drastically limits the throughput of the NN. -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Commented] (HDFS-5728) [Diskfull] Block recovery will fail if the metafile does not have crc for all chunks of the block
[ https://issues.apache.org/jira/browse/HDFS-5728?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13881880#comment-13881880 ] Hudson commented on HDFS-5728: -- SUCCESS: Integrated in Hadoop-Hdfs-trunk #1653 (See [https://builds.apache.org/job/Hadoop-Hdfs-trunk/1653/]) HDFS-5728. Block recovery will fail if the metafile does not have crc for all chunks of the block. Contributed by Vinay. (kihwal: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1561223) * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/BlockPoolSlice.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestLeaseRecovery.java > [Diskfull] Block recovery will fail if the metafile does not have crc for all > chunks of the block > - > > Key: HDFS-5728 > URL: https://issues.apache.org/jira/browse/HDFS-5728 > Project: Hadoop HDFS > Issue Type: Bug > Components: datanode >Affects Versions: 0.23.10, 2.2.0 >Reporter: Vinay >Assignee: Vinay >Priority: Critical > Fix For: 3.0.0, 2.4.0 > > Attachments: HDFS-5728.patch, HDFS-5728.patch, HDFS-5728.patch > > > 1. Client (regionsever) has opened stream to write its WAL to HDFS. This is > not one time upload, data will be written slowly. > 2. One of the DataNode got diskfull ( due to some other data filled up disks) > 3. Unfortunately block was being written to only this datanode in cluster, so > client write has also failed. > 4. After some time disk is made free and all processes are restarted. > 5. Now HMaster try to recover the file by calling recoverLease. > At this time recovery was failing saying file length mismatch. > When checked, > actual block file length: 62484480 > Calculated block length: 62455808 > This was because, metafile was having crc for only 62455808 bytes, and it > considered 62455808 as the block size. > No matter how many times, recovery was continously failing. -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Commented] (HDFS-5703) Add support for HTTPS and swebhdfs to HttpFS
[ https://issues.apache.org/jira/browse/HDFS-5703?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13881879#comment-13881879 ] Hudson commented on HDFS-5703: -- SUCCESS: Integrated in Hadoop-Hdfs-trunk #1653 (See [https://builds.apache.org/job/Hadoop-Hdfs-trunk/1653/]) HDFS-5703. Add support for HTTPS and swebhdfs to HttpFS. (tucu) (tucu: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1560504) * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/pom.xml * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/conf/httpfs-env.sh * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/client/HttpFSFileSystem.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/client/HttpFSKerberosAuthenticator.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/client/HttpFSUtils.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/client/HttpsFSFileSystem.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServerWebApp.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/service/DelegationTokenIdentifier.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/service/security/DelegationTokenManagerService.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/servlet/ServerWebApp.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/libexec/httpfs-config.sh * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/sbin/httpfs.sh * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/tomcat/ssl-server.xml * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/site/apt/ServerSetup.apt.vm * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/test/java/org/apache/hadoop/fs/http/client/BaseTestHttpFSWith.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/test/java/org/apache/hadoop/fs/http/client/TestHttpFSFWithSWebhdfsFileSystem.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/test/java/org/apache/hadoop/fs/http/server/TestHttpFSKerberosAuthenticationHandler.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/test/java/org/apache/hadoop/lib/service/security/TestDelegationTokenManagerService.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/test/java/org/apache/hadoop/test/TestJettyHelper.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt > Add support for HTTPS and swebhdfs to HttpFS > > > Key: HDFS-5703 > URL: https://issues.apache.org/jira/browse/HDFS-5703 > Project: Hadoop HDFS > Issue Type: New Feature > Components: webhdfs >Affects Versions: 2.4.0 >Reporter: Alejandro Abdelnur >Assignee: Alejandro Abdelnur > Fix For: 2.4.0 > > Attachments: HDFS-5703.patch, HDFS-5703.patch > > > HDFS-3987 added HTTPS support to webhdfs, using the new scheme swebhdfs://. > This JIRA is to add HTTPS support to HttpFS as well as supporting the > DelegationTokens required by swebhdfs:// -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Commented] (HDFS-5806) balancer should set SoTimeout to avoid indefinite hangs
[ https://issues.apache.org/jira/browse/HDFS-5806?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13881881#comment-13881881 ] Hudson commented on HDFS-5806: -- SUCCESS: Integrated in Hadoop-Hdfs-trunk #1653 (See [https://builds.apache.org/job/Hadoop-Hdfs-trunk/1653/]) HDFS-5806. Balancer should set soTimeout to avoid indefinite hangs. Contributed by Nathan Roberts. (wang: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1560548) * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/balancer/Balancer.java > balancer should set SoTimeout to avoid indefinite hangs > --- > > Key: HDFS-5806 > URL: https://issues.apache.org/jira/browse/HDFS-5806 > Project: Hadoop HDFS > Issue Type: Bug > Components: balancer >Affects Versions: 3.0.0, 2.2.0 >Reporter: Nathan Roberts >Assignee: Nathan Roberts > Fix For: 2.4.0 > > Attachments: HDFS-5806.patch > > > Simple patch to avoid the balancer hanging when datanode stops responding to > requests. -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Commented] (HDFS-5788) listLocatedStatus response can be very large
[ https://issues.apache.org/jira/browse/HDFS-5788?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13881885#comment-13881885 ] Hudson commented on HDFS-5788: -- SUCCESS: Integrated in Hadoop-Hdfs-trunk #1653 (See [https://builds.apache.org/job/Hadoop-Hdfs-trunk/1653/]) HDFS-5788. listLocatedStatus response can be very large. Contributed by Nathan Roberts. (kihwal: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1560750) * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSDirectory.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestINodeFile.java > listLocatedStatus response can be very large > > > Key: HDFS-5788 > URL: https://issues.apache.org/jira/browse/HDFS-5788 > Project: Hadoop HDFS > Issue Type: Improvement > Components: namenode >Affects Versions: 3.0.0, 0.23.10, 2.2.0 >Reporter: Nathan Roberts >Assignee: Nathan Roberts > Fix For: 3.0.0, 2.4.0 > > Attachments: HDFS-5788.patch > > > Currently we limit the size of listStatus requests to a default of 1000 > entries. This works fine except in the case of listLocatedStatus where the > location information can be quite large. As an example, a directory with 7000 > entries, 4 blocks each, 3 way replication - a listLocatedStatus response is > over 1MB. This can chew up very large amounts of memory in the NN if lots of > clients try to do this simultaneously. > Seems like it would be better if we also considered the amount of location > information being returned when deciding how many files to return. > Patch will follow shortly. -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Commented] (HDFS-4949) Centralized cache management in HDFS
[ https://issues.apache.org/jira/browse/HDFS-4949?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13881875#comment-13881875 ] Hudson commented on HDFS-4949: -- SUCCESS: Integrated in Hadoop-Hdfs-trunk #1653 (See [https://builds.apache.org/job/Hadoop-Hdfs-trunk/1653/]) Move HDFS-4949 subtasks in CHANGES.txt to a new section under 2.4.0 release. (wang: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1560528) * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt > Centralized cache management in HDFS > > > Key: HDFS-4949 > URL: https://issues.apache.org/jira/browse/HDFS-4949 > Project: Hadoop HDFS > Issue Type: New Feature > Components: datanode, namenode >Affects Versions: 3.0.0, 2.4.0 >Reporter: Andrew Wang >Assignee: Andrew Wang > Fix For: 2.4.0 > > Attachments: HDFS-4949-consolidated.patch, > caching-design-doc-2013-07-02.pdf, caching-design-doc-2013-08-09.pdf, > caching-design-doc-2013-10-24.pdf, caching-testplan.pdf, > hdfs-4949-branch-2.patch > > > HDFS currently has no support for managing or exposing in-memory caches at > datanodes. This makes it harder for higher level application frameworks like > Hive, Pig, and Impala to effectively use cluster memory, because they cannot > explicitly cache important datasets or place their tasks for memory locality. -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Commented] (HDFS-5703) Add support for HTTPS and swebhdfs to HttpFS
[ https://issues.apache.org/jira/browse/HDFS-5703?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13881861#comment-13881861 ] Hudson commented on HDFS-5703: -- FAILURE: Integrated in Hadoop-Mapreduce-trunk #1678 (See [https://builds.apache.org/job/Hadoop-Mapreduce-trunk/1678/]) HDFS-5703. Add support for HTTPS and swebhdfs to HttpFS. (tucu) (tucu: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1560504) * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/pom.xml * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/conf/httpfs-env.sh * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/client/HttpFSFileSystem.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/client/HttpFSKerberosAuthenticator.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/client/HttpFSUtils.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/client/HttpsFSFileSystem.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServerWebApp.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/service/DelegationTokenIdentifier.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/service/security/DelegationTokenManagerService.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/servlet/ServerWebApp.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/libexec/httpfs-config.sh * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/sbin/httpfs.sh * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/tomcat/ssl-server.xml * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/site/apt/ServerSetup.apt.vm * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/test/java/org/apache/hadoop/fs/http/client/BaseTestHttpFSWith.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/test/java/org/apache/hadoop/fs/http/client/TestHttpFSFWithSWebhdfsFileSystem.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/test/java/org/apache/hadoop/fs/http/server/TestHttpFSKerberosAuthenticationHandler.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/test/java/org/apache/hadoop/lib/service/security/TestDelegationTokenManagerService.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/test/java/org/apache/hadoop/test/TestJettyHelper.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt > Add support for HTTPS and swebhdfs to HttpFS > > > Key: HDFS-5703 > URL: https://issues.apache.org/jira/browse/HDFS-5703 > Project: Hadoop HDFS > Issue Type: New Feature > Components: webhdfs >Affects Versions: 2.4.0 >Reporter: Alejandro Abdelnur >Assignee: Alejandro Abdelnur > Fix For: 2.4.0 > > Attachments: HDFS-5703.patch, HDFS-5703.patch > > > HDFS-3987 added HTTPS support to webhdfs, using the new scheme swebhdfs://. > This JIRA is to add HTTPS support to HttpFS as well as supporting the > DelegationTokens required by swebhdfs:// -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Commented] (HDFS-5806) balancer should set SoTimeout to avoid indefinite hangs
[ https://issues.apache.org/jira/browse/HDFS-5806?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13881863#comment-13881863 ] Hudson commented on HDFS-5806: -- FAILURE: Integrated in Hadoop-Mapreduce-trunk #1678 (See [https://builds.apache.org/job/Hadoop-Mapreduce-trunk/1678/]) HDFS-5806. Balancer should set soTimeout to avoid indefinite hangs. Contributed by Nathan Roberts. (wang: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1560548) * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/balancer/Balancer.java > balancer should set SoTimeout to avoid indefinite hangs > --- > > Key: HDFS-5806 > URL: https://issues.apache.org/jira/browse/HDFS-5806 > Project: Hadoop HDFS > Issue Type: Bug > Components: balancer >Affects Versions: 3.0.0, 2.2.0 >Reporter: Nathan Roberts >Assignee: Nathan Roberts > Fix For: 2.4.0 > > Attachments: HDFS-5806.patch > > > Simple patch to avoid the balancer hanging when datanode stops responding to > requests. -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Commented] (HDFS-5788) listLocatedStatus response can be very large
[ https://issues.apache.org/jira/browse/HDFS-5788?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13881867#comment-13881867 ] Hudson commented on HDFS-5788: -- FAILURE: Integrated in Hadoop-Mapreduce-trunk #1678 (See [https://builds.apache.org/job/Hadoop-Mapreduce-trunk/1678/]) HDFS-5788. listLocatedStatus response can be very large. Contributed by Nathan Roberts. (kihwal: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1560750) * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSDirectory.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestINodeFile.java > listLocatedStatus response can be very large > > > Key: HDFS-5788 > URL: https://issues.apache.org/jira/browse/HDFS-5788 > Project: Hadoop HDFS > Issue Type: Improvement > Components: namenode >Affects Versions: 3.0.0, 0.23.10, 2.2.0 >Reporter: Nathan Roberts >Assignee: Nathan Roberts > Fix For: 3.0.0, 2.4.0 > > Attachments: HDFS-5788.patch > > > Currently we limit the size of listStatus requests to a default of 1000 > entries. This works fine except in the case of listLocatedStatus where the > location information can be quite large. As an example, a directory with 7000 > entries, 4 blocks each, 3 way replication - a listLocatedStatus response is > over 1MB. This can chew up very large amounts of memory in the NN if lots of > clients try to do this simultaneously. > Seems like it would be better if we also considered the amount of location > information being returned when deciding how many files to return. > Patch will follow shortly. -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Commented] (HDFS-5789) Some of snapshot APIs missing checkOperation double check in fsn
[ https://issues.apache.org/jira/browse/HDFS-5789?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13881866#comment-13881866 ] Hudson commented on HDFS-5789: -- FAILURE: Integrated in Hadoop-Mapreduce-trunk #1678 (See [https://builds.apache.org/job/Hadoop-Mapreduce-trunk/1678/]) HDFS-5789. Some of snapshot APIs missing checkOperation double check in fsn. Contributed by Uma Maheswara Rao G. (umamahesh: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1560731) * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSNamesystem.java > Some of snapshot APIs missing checkOperation double check in fsn > > > Key: HDFS-5789 > URL: https://issues.apache.org/jira/browse/HDFS-5789 > Project: Hadoop HDFS > Issue Type: Bug > Components: namenode >Affects Versions: 3.0.0 >Reporter: Uma Maheswara Rao G >Assignee: Uma Maheswara Rao G > Fix For: 3.0.0, 2.3.0 > > Attachments: HDFS-5789.patch > > > HDFS-4591 introduced double checked for HA state while taking fsn lock. > checkoperation made before actually taking lock and after the lock again. > This pattern missed in some of the snapshot APIs and cache management related > APIs. -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Commented] (HDFS-5241) Provide alternate queuing audit logger to reduce logging contention
[ https://issues.apache.org/jira/browse/HDFS-5241?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13881860#comment-13881860 ] Hudson commented on HDFS-5241: -- FAILURE: Integrated in Hadoop-Mapreduce-trunk #1678 (See [https://builds.apache.org/job/Hadoop-Mapreduce-trunk/1678/]) HDFS-5241. Provide alternate queuing audit logger to reduce logging contention (daryn) (daryn: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1560761) * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/DFSConfigKeys.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSNamesystem.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestAuditLogs.java > Provide alternate queuing audit logger to reduce logging contention > --- > > Key: HDFS-5241 > URL: https://issues.apache.org/jira/browse/HDFS-5241 > Project: Hadoop HDFS > Issue Type: Sub-task > Components: namenode >Affects Versions: 2.0.0-alpha, 3.0.0 >Reporter: Daryn Sharp >Assignee: Daryn Sharp > Attachments: HDFS-5241.patch, HDFS-5241.patch > > > The default audit logger has extremely poor performance. The internal > synchronization of log4j causes massive contention between the call handlers > (100 by default) which drastically limits the throughput of the NN. -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Commented] (HDFS-5728) [Diskfull] Block recovery will fail if the metafile does not have crc for all chunks of the block
[ https://issues.apache.org/jira/browse/HDFS-5728?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13881862#comment-13881862 ] Hudson commented on HDFS-5728: -- FAILURE: Integrated in Hadoop-Mapreduce-trunk #1678 (See [https://builds.apache.org/job/Hadoop-Mapreduce-trunk/1678/]) HDFS-5728. Block recovery will fail if the metafile does not have crc for all chunks of the block. Contributed by Vinay. (kihwal: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1561223) * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/BlockPoolSlice.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestLeaseRecovery.java > [Diskfull] Block recovery will fail if the metafile does not have crc for all > chunks of the block > - > > Key: HDFS-5728 > URL: https://issues.apache.org/jira/browse/HDFS-5728 > Project: Hadoop HDFS > Issue Type: Bug > Components: datanode >Affects Versions: 0.23.10, 2.2.0 >Reporter: Vinay >Assignee: Vinay >Priority: Critical > Fix For: 3.0.0, 2.4.0 > > Attachments: HDFS-5728.patch, HDFS-5728.patch, HDFS-5728.patch > > > 1. Client (regionsever) has opened stream to write its WAL to HDFS. This is > not one time upload, data will be written slowly. > 2. One of the DataNode got diskfull ( due to some other data filled up disks) > 3. Unfortunately block was being written to only this datanode in cluster, so > client write has also failed. > 4. After some time disk is made free and all processes are restarted. > 5. Now HMaster try to recover the file by calling recoverLease. > At this time recovery was failing saying file length mismatch. > When checked, > actual block file length: 62484480 > Calculated block length: 62455808 > This was because, metafile was having crc for only 62455808 bytes, and it > considered 62455808 as the block size. > No matter how many times, recovery was continously failing. -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Commented] (HDFS-4949) Centralized cache management in HDFS
[ https://issues.apache.org/jira/browse/HDFS-4949?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13881857#comment-13881857 ] Hudson commented on HDFS-4949: -- FAILURE: Integrated in Hadoop-Mapreduce-trunk #1678 (See [https://builds.apache.org/job/Hadoop-Mapreduce-trunk/1678/]) Move HDFS-4949 subtasks in CHANGES.txt to a new section under 2.4.0 release. (wang: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1560528) * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt > Centralized cache management in HDFS > > > Key: HDFS-4949 > URL: https://issues.apache.org/jira/browse/HDFS-4949 > Project: Hadoop HDFS > Issue Type: New Feature > Components: datanode, namenode >Affects Versions: 3.0.0, 2.4.0 >Reporter: Andrew Wang >Assignee: Andrew Wang > Fix For: 2.4.0 > > Attachments: HDFS-4949-consolidated.patch, > caching-design-doc-2013-07-02.pdf, caching-design-doc-2013-08-09.pdf, > caching-design-doc-2013-10-24.pdf, caching-testplan.pdf, > hdfs-4949-branch-2.patch > > > HDFS currently has no support for managing or exposing in-memory caches at > datanodes. This makes it harder for higher level application frameworks like > Hive, Pig, and Impala to effectively use cluster memory, because they cannot > explicitly cache important datasets or place their tasks for memory locality. -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Commented] (HDFS-5703) Add support for HTTPS and swebhdfs to HttpFS
[ https://issues.apache.org/jira/browse/HDFS-5703?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13881834#comment-13881834 ] Hudson commented on HDFS-5703: -- FAILURE: Integrated in Hadoop-Yarn-trunk #461 (See [https://builds.apache.org/job/Hadoop-Yarn-trunk/461/]) HDFS-5703. Add support for HTTPS and swebhdfs to HttpFS. (tucu) (tucu: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1560504) * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/pom.xml * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/conf/httpfs-env.sh * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/client/HttpFSFileSystem.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/client/HttpFSKerberosAuthenticator.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/client/HttpFSUtils.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/client/HttpsFSFileSystem.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServerWebApp.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/service/DelegationTokenIdentifier.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/service/security/DelegationTokenManagerService.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/servlet/ServerWebApp.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/libexec/httpfs-config.sh * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/sbin/httpfs.sh * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/tomcat/ssl-server.xml * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/site/apt/ServerSetup.apt.vm * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/test/java/org/apache/hadoop/fs/http/client/BaseTestHttpFSWith.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/test/java/org/apache/hadoop/fs/http/client/TestHttpFSFWithSWebhdfsFileSystem.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/test/java/org/apache/hadoop/fs/http/server/TestHttpFSKerberosAuthenticationHandler.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/test/java/org/apache/hadoop/lib/service/security/TestDelegationTokenManagerService.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/test/java/org/apache/hadoop/test/TestJettyHelper.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt > Add support for HTTPS and swebhdfs to HttpFS > > > Key: HDFS-5703 > URL: https://issues.apache.org/jira/browse/HDFS-5703 > Project: Hadoop HDFS > Issue Type: New Feature > Components: webhdfs >Affects Versions: 2.4.0 >Reporter: Alejandro Abdelnur >Assignee: Alejandro Abdelnur > Fix For: 2.4.0 > > Attachments: HDFS-5703.patch, HDFS-5703.patch > > > HDFS-3987 added HTTPS support to webhdfs, using the new scheme swebhdfs://. > This JIRA is to add HTTPS support to HttpFS as well as supporting the > DelegationTokens required by swebhdfs:// -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Commented] (HDFS-5806) balancer should set SoTimeout to avoid indefinite hangs
[ https://issues.apache.org/jira/browse/HDFS-5806?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13881836#comment-13881836 ] Hudson commented on HDFS-5806: -- FAILURE: Integrated in Hadoop-Yarn-trunk #461 (See [https://builds.apache.org/job/Hadoop-Yarn-trunk/461/]) HDFS-5806. Balancer should set soTimeout to avoid indefinite hangs. Contributed by Nathan Roberts. (wang: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1560548) * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/balancer/Balancer.java > balancer should set SoTimeout to avoid indefinite hangs > --- > > Key: HDFS-5806 > URL: https://issues.apache.org/jira/browse/HDFS-5806 > Project: Hadoop HDFS > Issue Type: Bug > Components: balancer >Affects Versions: 3.0.0, 2.2.0 >Reporter: Nathan Roberts >Assignee: Nathan Roberts > Fix For: 2.4.0 > > Attachments: HDFS-5806.patch > > > Simple patch to avoid the balancer hanging when datanode stops responding to > requests. -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Commented] (HDFS-5728) [Diskfull] Block recovery will fail if the metafile does not have crc for all chunks of the block
[ https://issues.apache.org/jira/browse/HDFS-5728?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13881835#comment-13881835 ] Hudson commented on HDFS-5728: -- FAILURE: Integrated in Hadoop-Yarn-trunk #461 (See [https://builds.apache.org/job/Hadoop-Yarn-trunk/461/]) HDFS-5728. Block recovery will fail if the metafile does not have crc for all chunks of the block. Contributed by Vinay. (kihwal: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1561223) * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/BlockPoolSlice.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestLeaseRecovery.java > [Diskfull] Block recovery will fail if the metafile does not have crc for all > chunks of the block > - > > Key: HDFS-5728 > URL: https://issues.apache.org/jira/browse/HDFS-5728 > Project: Hadoop HDFS > Issue Type: Bug > Components: datanode >Affects Versions: 0.23.10, 2.2.0 >Reporter: Vinay >Assignee: Vinay >Priority: Critical > Fix For: 3.0.0, 2.4.0 > > Attachments: HDFS-5728.patch, HDFS-5728.patch, HDFS-5728.patch > > > 1. Client (regionsever) has opened stream to write its WAL to HDFS. This is > not one time upload, data will be written slowly. > 2. One of the DataNode got diskfull ( due to some other data filled up disks) > 3. Unfortunately block was being written to only this datanode in cluster, so > client write has also failed. > 4. After some time disk is made free and all processes are restarted. > 5. Now HMaster try to recover the file by calling recoverLease. > At this time recovery was failing saying file length mismatch. > When checked, > actual block file length: 62484480 > Calculated block length: 62455808 > This was because, metafile was having crc for only 62455808 bytes, and it > considered 62455808 as the block size. > No matter how many times, recovery was continously failing. -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Commented] (HDFS-5241) Provide alternate queuing audit logger to reduce logging contention
[ https://issues.apache.org/jira/browse/HDFS-5241?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13881833#comment-13881833 ] Hudson commented on HDFS-5241: -- FAILURE: Integrated in Hadoop-Yarn-trunk #461 (See [https://builds.apache.org/job/Hadoop-Yarn-trunk/461/]) HDFS-5241. Provide alternate queuing audit logger to reduce logging contention (daryn) (daryn: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1560761) * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/DFSConfigKeys.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSNamesystem.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestAuditLogs.java > Provide alternate queuing audit logger to reduce logging contention > --- > > Key: HDFS-5241 > URL: https://issues.apache.org/jira/browse/HDFS-5241 > Project: Hadoop HDFS > Issue Type: Sub-task > Components: namenode >Affects Versions: 2.0.0-alpha, 3.0.0 >Reporter: Daryn Sharp >Assignee: Daryn Sharp > Attachments: HDFS-5241.patch, HDFS-5241.patch > > > The default audit logger has extremely poor performance. The internal > synchronization of log4j causes massive contention between the call handlers > (100 by default) which drastically limits the throughput of the NN. -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Commented] (HDFS-5788) listLocatedStatus response can be very large
[ https://issues.apache.org/jira/browse/HDFS-5788?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13881840#comment-13881840 ] Hudson commented on HDFS-5788: -- FAILURE: Integrated in Hadoop-Yarn-trunk #461 (See [https://builds.apache.org/job/Hadoop-Yarn-trunk/461/]) HDFS-5788. listLocatedStatus response can be very large. Contributed by Nathan Roberts. (kihwal: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1560750) * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSDirectory.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestINodeFile.java > listLocatedStatus response can be very large > > > Key: HDFS-5788 > URL: https://issues.apache.org/jira/browse/HDFS-5788 > Project: Hadoop HDFS > Issue Type: Improvement > Components: namenode >Affects Versions: 3.0.0, 0.23.10, 2.2.0 >Reporter: Nathan Roberts >Assignee: Nathan Roberts > Fix For: 3.0.0, 2.4.0 > > Attachments: HDFS-5788.patch > > > Currently we limit the size of listStatus requests to a default of 1000 > entries. This works fine except in the case of listLocatedStatus where the > location information can be quite large. As an example, a directory with 7000 > entries, 4 blocks each, 3 way replication - a listLocatedStatus response is > over 1MB. This can chew up very large amounts of memory in the NN if lots of > clients try to do this simultaneously. > Seems like it would be better if we also considered the amount of location > information being returned when deciding how many files to return. > Patch will follow shortly. -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Commented] (HDFS-5789) Some of snapshot APIs missing checkOperation double check in fsn
[ https://issues.apache.org/jira/browse/HDFS-5789?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13881839#comment-13881839 ] Hudson commented on HDFS-5789: -- FAILURE: Integrated in Hadoop-Yarn-trunk #461 (See [https://builds.apache.org/job/Hadoop-Yarn-trunk/461/]) HDFS-5789. Some of snapshot APIs missing checkOperation double check in fsn. Contributed by Uma Maheswara Rao G. (umamahesh: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1560731) * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSNamesystem.java > Some of snapshot APIs missing checkOperation double check in fsn > > > Key: HDFS-5789 > URL: https://issues.apache.org/jira/browse/HDFS-5789 > Project: Hadoop HDFS > Issue Type: Bug > Components: namenode >Affects Versions: 3.0.0 >Reporter: Uma Maheswara Rao G >Assignee: Uma Maheswara Rao G > Fix For: 3.0.0, 2.3.0 > > Attachments: HDFS-5789.patch > > > HDFS-4591 introduced double checked for HA state while taking fsn lock. > checkoperation made before actually taking lock and after the lock again. > This pattern missed in some of the snapshot APIs and cache management related > APIs. -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Commented] (HDFS-4949) Centralized cache management in HDFS
[ https://issues.apache.org/jira/browse/HDFS-4949?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13881830#comment-13881830 ] Hudson commented on HDFS-4949: -- FAILURE: Integrated in Hadoop-Yarn-trunk #461 (See [https://builds.apache.org/job/Hadoop-Yarn-trunk/461/]) Move HDFS-4949 subtasks in CHANGES.txt to a new section under 2.4.0 release. (wang: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1560528) * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt > Centralized cache management in HDFS > > > Key: HDFS-4949 > URL: https://issues.apache.org/jira/browse/HDFS-4949 > Project: Hadoop HDFS > Issue Type: New Feature > Components: datanode, namenode >Affects Versions: 3.0.0, 2.4.0 >Reporter: Andrew Wang >Assignee: Andrew Wang > Fix For: 2.4.0 > > Attachments: HDFS-4949-consolidated.patch, > caching-design-doc-2013-07-02.pdf, caching-design-doc-2013-08-09.pdf, > caching-design-doc-2013-10-24.pdf, caching-testplan.pdf, > hdfs-4949-branch-2.patch > > > HDFS currently has no support for managing or exposing in-memory caches at > datanodes. This makes it harder for higher level application frameworks like > Hive, Pig, and Impala to effectively use cluster memory, because they cannot > explicitly cache important datasets or place their tasks for memory locality. -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Commented] (HDFS-5776) Support 'hedged' reads in DFSClient
[ https://issues.apache.org/jira/browse/HDFS-5776?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13881827#comment-13881827 ] Hadoop QA commented on HDFS-5776: - {color:red}-1 overall{color}. Here are the results of testing the latest attachment http://issues.apache.org/jira/secure/attachment/12625003/HDFS-5776-v8.txt against trunk revision . {color:green}+1 @author{color}. The patch does not contain any @author tags. {color:green}+1 tests included{color}. The patch appears to include 1 new or modified test files. {color:green}+1 javac{color}. The applied patch does not increase the total number of javac compiler warnings. {color:green}+1 javadoc{color}. The javadoc tool did not generate any warning messages. {color:red}-1 eclipse:eclipse{color}. The patch failed to build with eclipse:eclipse. {color:red}-1 findbugs{color}. The patch appears to introduce 1 new Findbugs (version 1.3.9) warnings. {color:red}-1 release audit{color}. The applied patch generated 1 release audit warnings. {color:green}+1 core tests{color}. The patch passed unit tests in hadoop-hdfs-project/hadoop-hdfs. {color:green}+1 contrib tests{color}. The patch passed contrib unit tests. Test results: https://builds.apache.org/job/PreCommit-HDFS-Build/5941//testReport/ Release audit warnings: https://builds.apache.org/job/PreCommit-HDFS-Build/5941//artifact/trunk/patchprocess/patchReleaseAuditProblems.txt Findbugs warnings: https://builds.apache.org/job/PreCommit-HDFS-Build/5941//artifact/trunk/patchprocess/newPatchFindbugsWarningshadoop-hdfs.html Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/5941//console This message is automatically generated. > Support 'hedged' reads in DFSClient > --- > > Key: HDFS-5776 > URL: https://issues.apache.org/jira/browse/HDFS-5776 > Project: Hadoop HDFS > Issue Type: Improvement > Components: hdfs-client >Affects Versions: 3.0.0 >Reporter: Liang Xie >Assignee: Liang Xie > Attachments: HDFS-5776-v2.txt, HDFS-5776-v3.txt, HDFS-5776-v4.txt, > HDFS-5776-v5.txt, HDFS-5776-v6.txt, HDFS-5776-v7.txt, HDFS-5776-v8.txt, > HDFS-5776.txt > > > This is a placeholder of hdfs related stuff backport from > https://issues.apache.org/jira/browse/HBASE-7509 > The quorum read ability should be helpful especially to optimize read outliers > we can utilize "dfs.dfsclient.quorum.read.threshold.millis" & > "dfs.dfsclient.quorum.read.threadpool.size" to enable/disable the hedged read > ability from client side(e.g. HBase), and by using DFSQuorumReadMetrics, we > could export the interested metric valus into client system(e.g. HBase's > regionserver metric). > The core logic is in pread code path, we decide to goto the original > fetchBlockByteRange or the new introduced fetchBlockByteRangeSpeculative per > the above config items. -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Commented] (HDFS-5343) When cat command is issued on snapshot files getting unexpected result
[ https://issues.apache.org/jira/browse/HDFS-5343?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13881748#comment-13881748 ] Hadoop QA commented on HDFS-5343: - {color:red}-1 overall{color}. Here are the results of testing the latest attachment http://issues.apache.org/jira/secure/attachment/12625029/HDFS-5343-0006.patch against trunk revision . {color:green}+1 @author{color}. The patch does not contain any @author tags. {color:green}+1 tests included{color}. The patch appears to include 1 new or modified test files. {color:green}+1 javac{color}. The applied patch does not increase the total number of javac compiler warnings. {color:red}-1 javadoc{color}. The javadoc tool appears to have generated -14 warning messages. {color:green}+1 eclipse:eclipse{color}. The patch built with eclipse:eclipse. {color:green}+1 findbugs{color}. The patch does not introduce any new Findbugs (version 1.3.9) warnings. {color:red}-1 release audit{color}. The applied patch generated 1 release audit warnings. {color:red}-1 core tests{color}. The patch failed these unit tests in hadoop-hdfs-project/hadoop-hdfs: org.apache.hadoop.hdfs.server.namenode.ha.TestHASafeMode The following test timeouts occurred in hadoop-hdfs-project/hadoop-hdfs: org.apache.hadoop.hdfs.server.namenode.ha.TestDNFencingWithReplication {color:green}+1 contrib tests{color}. The patch passed contrib unit tests. Test results: https://builds.apache.org/job/PreCommit-HDFS-Build/5939//testReport/ Release audit warnings: https://builds.apache.org/job/PreCommit-HDFS-Build/5939//artifact/trunk/patchprocess/patchReleaseAuditProblems.txt Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/5939//console This message is automatically generated. > When cat command is issued on snapshot files getting unexpected result > -- > > Key: HDFS-5343 > URL: https://issues.apache.org/jira/browse/HDFS-5343 > Project: Hadoop HDFS > Issue Type: Bug > Components: hdfs-client >Reporter: sathish >Assignee: sathish > Attachments: HDFS-5343-0003.patch, HDFS-5343-0004.patch, > HDFS-5343-0005.patch, HDFS-5343-0006.patch, HDFS-5343-002.patch > > > first if we create one file with some file length and take the snapshot of > that file,and again append some data through append method to that file,then > if we do cat command operation on snapshot of that file,in general it should > dispaly the data what we added with create operation,but it is displaying the > total data i.e. create +_ appended data. > but if we do the same operation and if we read the contents of snapshot file > through input stream it is just displaying the data created in snapshoted > files. > in this the behaviour of cat command and reading through inputstream is > getting different -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Commented] (HDFS-5631) Expose interfaces required by FsDatasetSpi implementations
[ https://issues.apache.org/jira/browse/HDFS-5631?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13881745#comment-13881745 ] Hadoop QA commented on HDFS-5631: - {color:green}+1 overall{color}. Here are the results of testing the latest attachment http://issues.apache.org/jira/secure/attachment/12625158/HDFS-5631.patch against trunk revision . {color:green}+1 @author{color}. The patch does not contain any @author tags. {color:green}+1 tests included{color}. The patch appears to include 6 new or modified test files. {color:green}+1 javac{color}. The applied patch does not increase the total number of javac compiler warnings. {color:green}+1 javadoc{color}. The javadoc tool did not generate any warning messages. {color:green}+1 eclipse:eclipse{color}. The patch built with eclipse:eclipse. {color:green}+1 findbugs{color}. The patch does not introduce any new Findbugs (version 1.3.9) warnings. {color:green}+1 release audit{color}. The applied patch does not increase the total number of release audit warnings. {color:green}+1 core tests{color}. The patch passed unit tests in hadoop-hdfs-project/hadoop-hdfs. {color:green}+1 contrib tests{color}. The patch passed contrib unit tests. Test results: https://builds.apache.org/job/PreCommit-HDFS-Build/5938//testReport/ Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/5938//console This message is automatically generated. > Expose interfaces required by FsDatasetSpi implementations > -- > > Key: HDFS-5631 > URL: https://issues.apache.org/jira/browse/HDFS-5631 > Project: Hadoop HDFS > Issue Type: Sub-task > Components: datanode >Affects Versions: 3.0.0 >Reporter: David Powell >Priority: Minor > Attachments: HDFS-5631.patch, HDFS-5631.patch > > > This sub-task addresses section 4.1 of the document attached to HDFS-5194, > the exposure of interfaces needed by a FsDatasetSpi implementation. > Specifically it makes ChunkChecksum public and BlockMetadataHeader's > readHeader() and writeHeader() methods public. > The changes to BlockReaderUtil (and related classes) discussed by section > 4.1 are only needed if supporting short-circuit, and should be addressed > as part of an effort to provide such support rather than this JIRA. > To help ensure these changes are complete and are not regressed in the > future, tests that gauge the accessibility (though *not* behavior) > of interfaces needed by a FsDatasetSpi subclass are also included. > These take the form of a dummy FsDatasetSpi subclass -- a successful > compilation is effectively a pass. Trivial unit tests are included so > that there is something tangible to track. -- This message was sent by Atlassian JIRA (v6.1.5#6160)