[jira] [Updated] (HDFS-4114) Remove the CheckpointNode
[ https://issues.apache.org/jira/browse/HDFS-4114?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Eli Collins updated HDFS-4114: -- Summary: Remove the CheckpointNode (was: Remove the BackupNode and CheckpointNode) > Remove the CheckpointNode > - > > Key: HDFS-4114 > URL: https://issues.apache.org/jira/browse/HDFS-4114 > Project: Hadoop HDFS > Issue Type: Bug >Reporter: Eli Collins >Assignee: Eli Collins > > Per the thread on hdfs-dev@ (http://s.apache.org/tMT) let's remove the > BackupNode and CheckpointNode. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HDFS-4114) Remove the BackupNode and CheckpointNode
[ https://issues.apache.org/jira/browse/HDFS-4114?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13488517#comment-13488517 ] Eli Collins commented on HDFS-4114: --- Hey Konst, I'll re-purpose this jira to just remove the CheckpointNode. > Remove the BackupNode and CheckpointNode > > > Key: HDFS-4114 > URL: https://issues.apache.org/jira/browse/HDFS-4114 > Project: Hadoop HDFS > Issue Type: Bug >Reporter: Eli Collins >Assignee: Eli Collins > > Per the thread on hdfs-dev@ (http://s.apache.org/tMT) let's remove the > BackupNode and CheckpointNode. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HDFS-4133) Add testcases for testing basic snapshot functionalities
[ https://issues.apache.org/jira/browse/HDFS-4133?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13488506#comment-13488506 ] Suresh Srinivas commented on HDFS-4133: --- Excellent tests. All the other changes look good. The following comments are for TestSnapshot.java: # File is missing the Apache License at the beginning. # Add javadoc to TestSnapshot class. Please mention that this test tests creation 1 or multiple snapshots. # #composeSnapshotRootPath could be named #getSnapshotRoot # #composeSnapshotFilePath could be named #getSnapshotRath. This method also should #composeSnapshotRootPath. Indentation in this method is messed up. # Please add javadoc to #prepareModifications method, expecially the parameter number # Instead of using names file1, file2, and file3, names such as fileToBeDeleted, fileToBeModified, fileToBeCreated better? # Make the following methods static and move it to a helper class say SnapshotTestHelper: #* createSnapshot, checkSnapshotCreation # Move some of the common utility methods in this class to a new class SnapshotTestHelper. The same utility methods could also be used by TestSnapshotPathInodes. Some methods that should move are: composeSnapshot* methods, createSnapshot, # A lot of member variables of Modification and its subclasses can be final. # Please add very brief javadoc to Modification class methods > Add testcases for testing basic snapshot functionalities > > > Key: HDFS-4133 > URL: https://issues.apache.org/jira/browse/HDFS-4133 > Project: Hadoop HDFS > Issue Type: Sub-task > Components: data-node, name-node >Reporter: Jing Zhao >Assignee: Jing Zhao > Attachments: HDFS-4133.001.patch > > > Add testcase for basic snapshot functionalities. In the test we keep creating > snapshots, modifying original files, and check previous snapshots. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HDFS-2802) Support for RW/RO snapshots in HDFS
[ https://issues.apache.org/jira/browse/HDFS-2802?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13488497#comment-13488497 ] Aaron T. Myers commented on HDFS-2802: -- Hi Suresh, I've now had some more time to examine the document, and I have a few questions/comments: # Regarding the ownership requirement for users to create snapshot - this answers my question above regarding what would happen when a user tries to create a snapshot of a subtree where the user doesn't have read access to all the files. But, why not decouple it completely from Unix permissions? They really have nothing to do with each other now. Perhaps we should just make a snapshottable directory have a list of users that are allowed to create snapshots of it? # Related to the above, the document shows that deleting a snapshot consists of just using the FSShell to delete it. The document also says that even though a non super-user may create a snapshot for a given snapshottable directory, files under that directory that the user cannot read will continue to not be readable in the snapshot. Doesn't this imply that the user may not be able to delete a snapshot the user has just created? If so, I find the asymmetry between being able to create a snapshot, but not necessarily delete that snapshot, a little odd. I think it'd be better to have a deleteSnapshot administrative command to correspond to the createSnapshot command. # Rename only within snapshottable directory - I don't see why this limitation is necessary or desirable. I see that the doc says "This can be revisited" but I'd really like to know why this restriction is proposed at all. # Any other use for the linked list (INodeFileWithLink) besides determining the max replicaiton factor? The document says that this is one of the uses, but I don't see any other uses mentioned. # You mentioned to Konstantin that the design supports nested snapshots, but I don't see any discussion of it in the document. My apologies if I missed this. # Regarding open question #2 - I think that the values of the permission, atime, mtime, etc. attributes should definitely _not_ change when the attributes of the original file changes. That should be part of the guarantee of a snapshot. In fact, I believe this open question is already answered by the #2 and #3 "Operations Supported on Snapshots". I do think it may make sense to treat the replication factor separately from this, however. Is this what you were referring to when you said "...if we decide snapshot of file does not preserve the replication factor at the time of snapshot"? > Support for RW/RO snapshots in HDFS > --- > > Key: HDFS-2802 > URL: https://issues.apache.org/jira/browse/HDFS-2802 > Project: Hadoop HDFS > Issue Type: New Feature > Components: data-node, name-node >Reporter: Hari Mankude >Assignee: Hari Mankude > Attachments: HDFSSnapshotsDesign.pdf, snap.patch, > snapshot-design.pdf, snapshot-design.tex, snapshot-one-pager.pdf, > Snapshots20121018.pdf, Snapshots20121030.pdf > > > Snapshots are point in time images of parts of the filesystem or the entire > filesystem. Snapshots can be a read-only or a read-write point in time copy > of the filesystem. There are several use cases for snapshots in HDFS. I will > post a detailed write-up soon with with more information. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HDFS-2802) Support for RW/RO snapshots in HDFS
[ https://issues.apache.org/jira/browse/HDFS-2802?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13488489#comment-13488489 ] Aaron T. Myers commented on HDFS-2802: -- Hi Konstantin, bq. I like the idea of generating globally unique version ids, and assigning them to snapshots internally rather than letting people invent their own. One can always list available versions and read the desired one. So the -createSnapshot command does not need to pass , but will instead get it in return. Please take a look at the updated design document I just uploaded. It goes into a lot more detail about how snapshots would be named and accessed by users. From my understanding of you're suggestion, I think it describes pretty much exactly what you propose with regard to this. > Support for RW/RO snapshots in HDFS > --- > > Key: HDFS-2802 > URL: https://issues.apache.org/jira/browse/HDFS-2802 > Project: Hadoop HDFS > Issue Type: New Feature > Components: data-node, name-node >Reporter: Hari Mankude >Assignee: Hari Mankude > Attachments: HDFSSnapshotsDesign.pdf, snap.patch, > snapshot-design.pdf, snapshot-design.tex, snapshot-one-pager.pdf, > Snapshots20121018.pdf, Snapshots20121030.pdf > > > Snapshots are point in time images of parts of the filesystem or the entire > filesystem. Snapshots can be a read-only or a read-write point in time copy > of the filesystem. There are several use cases for snapshots in HDFS. I will > post a detailed write-up soon with with more information. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HDFS-2802) Support for RW/RO snapshots in HDFS
[ https://issues.apache.org/jira/browse/HDFS-2802?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Aaron T. Myers updated HDFS-2802: - Attachment: snapshot-design.tex snapshot-design.pdf Hello all, attached please find an updated design proposal which attempts to merge the two designs as I previously described in [this comment|https://issues.apache.org/jira/browse/HDFS-2802?focusedCommentId=13485761&page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-13485761]. This hybrid design merges the concepts of a snapshottable directory and user-initiated snapshots from the designs posted by Nicholas and Suresh, with the NN snapshot representation and client materialization of snapshots from the earlier design I posted. This document also expands upon the CLI commands/API that would be supported, as well as provide more details on how users would be able to access the snapshots available to them. Please have a look and review. I'd appreciate any feedback you have. I'm also uploading the .tex file used to produce this document and have removed the author listing. I hope this facilitates collaborating on a single design that we can eventually all agree to. > Support for RW/RO snapshots in HDFS > --- > > Key: HDFS-2802 > URL: https://issues.apache.org/jira/browse/HDFS-2802 > Project: Hadoop HDFS > Issue Type: New Feature > Components: data-node, name-node >Reporter: Hari Mankude >Assignee: Hari Mankude > Attachments: HDFSSnapshotsDesign.pdf, snap.patch, > snapshot-design.pdf, snapshot-design.tex, snapshot-one-pager.pdf, > Snapshots20121018.pdf, Snapshots20121030.pdf > > > Snapshots are point in time images of parts of the filesystem or the entire > filesystem. Snapshots can be a read-only or a read-write point in time copy > of the filesystem. There are several use cases for snapshots in HDFS. I will > post a detailed write-up soon with with more information. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HDFS-4056) Always start the NN's SecretManager
[ https://issues.apache.org/jira/browse/HDFS-4056?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13488438#comment-13488438 ] Kan Zhang commented on HDFS-4056: - bq. Yes, it should work if you fetch the token yourself. As far as I know, that is not by design. bq. What would be the confusion? Suppose some clients are configured to use tokens, some don't. How do you make sure they did what they are supposed to do? Or you don't care? If a client happens to use a Hadoop conf that says "use tokens", it will fetch and use them; otherwise, it won't. Either way it works. It's hard to reason about cluster behavior, that is what I meant by confusing. bq. I don't recall a problem ever being root-caused to a stale token file. May I quote "Past performance is no guarantee of future results"? :) > Always start the NN's SecretManager > --- > > Key: HDFS-4056 > URL: https://issues.apache.org/jira/browse/HDFS-4056 > Project: Hadoop HDFS > Issue Type: Improvement > Components: name-node >Affects Versions: 0.23.0, 2.0.0-alpha, 3.0.0 >Reporter: Daryn Sharp >Assignee: Daryn Sharp > Attachments: HDFS-4056.patch > > > To support the ability to use tokens regardless of whether kerberos is > enabled, the NN's secret manager should always be started. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HDFS-4133) Add testcases for testing basic snapshot functionalities
[ https://issues.apache.org/jira/browse/HDFS-4133?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jing Zhao updated HDFS-4133: Attachment: HDFS-4133.001.patch Initial patch. Also rename original TestSnapshot.java to TestSnapshotPathINodes.java > Add testcases for testing basic snapshot functionalities > > > Key: HDFS-4133 > URL: https://issues.apache.org/jira/browse/HDFS-4133 > Project: Hadoop HDFS > Issue Type: Sub-task > Components: data-node, name-node >Reporter: Jing Zhao >Assignee: Jing Zhao > Attachments: HDFS-4133.001.patch > > > Add testcase for basic snapshot functionalities. In the test we keep creating > snapshots, modifying original files, and check previous snapshots. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Created] (HDFS-4133) Add testcases for testing basic snapshot functionalities
Jing Zhao created HDFS-4133: --- Summary: Add testcases for testing basic snapshot functionalities Key: HDFS-4133 URL: https://issues.apache.org/jira/browse/HDFS-4133 Project: Hadoop HDFS Issue Type: Sub-task Reporter: Jing Zhao Assignee: Jing Zhao Add testcase for basic snapshot functionalities. In the test we keep creating snapshots, modifying original files, and check previous snapshots. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HDFS-4132) when libwebhdfs is not enabled, nativeMiniDfsClient frees uninitialized memory
[ https://issues.apache.org/jira/browse/HDFS-4132?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13488387#comment-13488387 ] Jing Zhao commented on HDFS-4132: - That's bug brought by HDFS-3923. Thanks for the fix Colin! > when libwebhdfs is not enabled, nativeMiniDfsClient frees uninitialized > memory > --- > > Key: HDFS-4132 > URL: https://issues.apache.org/jira/browse/HDFS-4132 > Project: Hadoop HDFS > Issue Type: Bug > Components: libhdfs >Affects Versions: 2.0.3-alpha >Reporter: Colin Patrick McCabe >Assignee: Colin Patrick McCabe > Attachments: HDFS-4132.001.patch > > > When libwebhdfs is not enabled, nativeMiniDfsClient frees uninitialized > memory. > Details: jconfStr is declared uninitialized... > {code} > struct NativeMiniDfsCluster* nmdCreate(struct NativeMiniDfsConf *conf) > { > struct NativeMiniDfsCluster* cl = NULL; > jobject bld = NULL, bld2 = NULL, cobj = NULL; > jvalue val; > JNIEnv *env = getJNIEnv(); > jthrowable jthr; > jstring jconfStr; > {code} > and only initialized later if conf->webhdfsEnabled: > {code} > ... > if (conf->webhdfsEnabled) { > jthr = newJavaStr(env, DFS_WEBHDFS_ENABLED_KEY, &jconfStr); > if (jthr) { > printExceptionAndFree(env, jthr, PRINT_EXC_ALL, > ... > {code} > Then we try to free this uninitialized memory at the end, usually resulting > in a crash. > {code} > (*env)->DeleteLocalRef(env, jconfStr); > return cl; > {code} -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HDFS-4132) when libwebhdfs is not enabled, nativeMiniDfsClient frees uninitialized memory
[ https://issues.apache.org/jira/browse/HDFS-4132?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13488379#comment-13488379 ] Hadoop QA commented on HDFS-4132: - {color:red}-1 overall{color}. Here are the results of testing the latest attachment http://issues.apache.org/jira/secure/attachment/12551642/HDFS-4132.001.patch against trunk revision . {color:green}+1 @author{color}. The patch does not contain any @author tags. {color:red}-1 tests included{color}. The patch doesn't appear to include any new or modified tests. Please justify why no new tests are needed for this patch. Also please list what manual steps were performed to verify this patch. {color:green}+1 javac{color}. The applied patch does not increase the total number of javac compiler warnings. {color:green}+1 javadoc{color}. The javadoc tool did not generate any warning messages. {color:green}+1 eclipse:eclipse{color}. The patch built with eclipse:eclipse. {color:green}+1 findbugs{color}. The patch does not introduce any new Findbugs (version 1.3.9) warnings. {color:green}+1 release audit{color}. The applied patch does not increase the total number of release audit warnings. {color:green}+1 core tests{color}. The patch passed unit tests in hadoop-hdfs-project/hadoop-hdfs. {color:green}+1 contrib tests{color}. The patch passed contrib unit tests. Test results: https://builds.apache.org/job/PreCommit-HDFS-Build/3436//testReport/ Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/3436//console This message is automatically generated. > when libwebhdfs is not enabled, nativeMiniDfsClient frees uninitialized > memory > --- > > Key: HDFS-4132 > URL: https://issues.apache.org/jira/browse/HDFS-4132 > Project: Hadoop HDFS > Issue Type: Bug > Components: libhdfs >Affects Versions: 2.0.3-alpha >Reporter: Colin Patrick McCabe >Assignee: Colin Patrick McCabe > Attachments: HDFS-4132.001.patch > > > When libwebhdfs is not enabled, nativeMiniDfsClient frees uninitialized > memory. > Details: jconfStr is declared uninitialized... > {code} > struct NativeMiniDfsCluster* nmdCreate(struct NativeMiniDfsConf *conf) > { > struct NativeMiniDfsCluster* cl = NULL; > jobject bld = NULL, bld2 = NULL, cobj = NULL; > jvalue val; > JNIEnv *env = getJNIEnv(); > jthrowable jthr; > jstring jconfStr; > {code} > and only initialized later if conf->webhdfsEnabled: > {code} > ... > if (conf->webhdfsEnabled) { > jthr = newJavaStr(env, DFS_WEBHDFS_ENABLED_KEY, &jconfStr); > if (jthr) { > printExceptionAndFree(env, jthr, PRINT_EXC_ALL, > ... > {code} > Then we try to free this uninitialized memory at the end, usually resulting > in a crash. > {code} > (*env)->DeleteLocalRef(env, jconfStr); > return cl; > {code} -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HDFS-4132) when libwebhdfs is not enabled, nativeMiniDfsClient frees uninitialized memory
[ https://issues.apache.org/jira/browse/HDFS-4132?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Colin Patrick McCabe updated HDFS-4132: --- Attachment: HDFS-4132.001.patch > when libwebhdfs is not enabled, nativeMiniDfsClient frees uninitialized > memory > --- > > Key: HDFS-4132 > URL: https://issues.apache.org/jira/browse/HDFS-4132 > Project: Hadoop HDFS > Issue Type: Bug > Components: libhdfs >Affects Versions: 2.0.3-alpha >Reporter: Colin Patrick McCabe >Assignee: Colin Patrick McCabe > Attachments: HDFS-4132.001.patch > > > When libwebhdfs is not enabled, nativeMiniDfsClient frees uninitialized > memory. > Details: jconfStr is declared uninitialized... > {code} > struct NativeMiniDfsCluster* nmdCreate(struct NativeMiniDfsConf *conf) > { > struct NativeMiniDfsCluster* cl = NULL; > jobject bld = NULL, bld2 = NULL, cobj = NULL; > jvalue val; > JNIEnv *env = getJNIEnv(); > jthrowable jthr; > jstring jconfStr; > {code} > and only initialized later if conf->webhdfsEnabled: > {code} > ... > if (conf->webhdfsEnabled) { > jthr = newJavaStr(env, DFS_WEBHDFS_ENABLED_KEY, &jconfStr); > if (jthr) { > printExceptionAndFree(env, jthr, PRINT_EXC_ALL, > ... > {code} > Then we try to free this uninitialized memory at the end, usually resulting > in a crash. > {code} > (*env)->DeleteLocalRef(env, jconfStr); > return cl; > {code} -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HDFS-4132) when libwebhdfs is not enabled, nativeMiniDfsClient frees uninitialized memory
[ https://issues.apache.org/jira/browse/HDFS-4132?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Colin Patrick McCabe updated HDFS-4132: --- Status: Patch Available (was: Open) > when libwebhdfs is not enabled, nativeMiniDfsClient frees uninitialized > memory > --- > > Key: HDFS-4132 > URL: https://issues.apache.org/jira/browse/HDFS-4132 > Project: Hadoop HDFS > Issue Type: Bug > Components: libhdfs >Affects Versions: 2.0.3-alpha >Reporter: Colin Patrick McCabe >Assignee: Colin Patrick McCabe > Attachments: HDFS-4132.001.patch > > > When libwebhdfs is not enabled, nativeMiniDfsClient frees uninitialized > memory. > Details: jconfStr is declared uninitialized... > {code} > struct NativeMiniDfsCluster* nmdCreate(struct NativeMiniDfsConf *conf) > { > struct NativeMiniDfsCluster* cl = NULL; > jobject bld = NULL, bld2 = NULL, cobj = NULL; > jvalue val; > JNIEnv *env = getJNIEnv(); > jthrowable jthr; > jstring jconfStr; > {code} > and only initialized later if conf->webhdfsEnabled: > {code} > ... > if (conf->webhdfsEnabled) { > jthr = newJavaStr(env, DFS_WEBHDFS_ENABLED_KEY, &jconfStr); > if (jthr) { > printExceptionAndFree(env, jthr, PRINT_EXC_ALL, > ... > {code} > Then we try to free this uninitialized memory at the end, usually resulting > in a crash. > {code} > (*env)->DeleteLocalRef(env, jconfStr); > return cl; > {code} -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Created] (HDFS-4132) when libwebhdfs is not enabled, nativeMiniDfsClient frees uninitialized memory
Colin Patrick McCabe created HDFS-4132: -- Summary: when libwebhdfs is not enabled, nativeMiniDfsClient frees uninitialized memory Key: HDFS-4132 URL: https://issues.apache.org/jira/browse/HDFS-4132 Project: Hadoop HDFS Issue Type: Bug Components: libhdfs Affects Versions: 2.0.3-alpha Reporter: Colin Patrick McCabe Assignee: Colin Patrick McCabe When libwebhdfs is not enabled, nativeMiniDfsClient frees uninitialized memory. Details: jconfStr is declared uninitialized... {code} struct NativeMiniDfsCluster* nmdCreate(struct NativeMiniDfsConf *conf) { struct NativeMiniDfsCluster* cl = NULL; jobject bld = NULL, bld2 = NULL, cobj = NULL; jvalue val; JNIEnv *env = getJNIEnv(); jthrowable jthr; jstring jconfStr; {code} and only initialized later if conf->webhdfsEnabled: {code} ... if (conf->webhdfsEnabled) { jthr = newJavaStr(env, DFS_WEBHDFS_ENABLED_KEY, &jconfStr); if (jthr) { printExceptionAndFree(env, jthr, PRINT_EXC_ALL, ... {code} Then we try to free this uninitialized memory at the end, usually resulting in a crash. {code} (*env)->DeleteLocalRef(env, jconfStr); return cl; {code} -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HDFS-4108) In a secure cluster, in the HDFS WEBUI , clicking on a datanode in the node list , gives an error
[ https://issues.apache.org/jira/browse/HDFS-4108?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13488331#comment-13488331 ] Benoy Antony commented on HDFS-4108: When NN is in safe-mode, the links will not work as no delegation token can be generated. It throws an exception indicating that NN is in safemode. > In a secure cluster, in the HDFS WEBUI , clicking on a datanode in the node > list , gives an error > - > > Key: HDFS-4108 > URL: https://issues.apache.org/jira/browse/HDFS-4108 > Project: Hadoop HDFS > Issue Type: Bug > Components: security, webhdfs >Affects Versions: 1.1.0 >Reporter: Benoy Antony >Assignee: Benoy Antony >Priority: Minor > Attachments: HDFS-4108-1-1.patch, HDFS-4108-1-1.patch > > > This issue happens in secure cluster. > To reproduce : > Go to the NameNode WEB UI. (dfshealth.jsp) > Click to bring up the list of LiveNodes (dfsnodelist.jsp) > Click on a datanode to bring up the filesystem web page ( > browsedirectory.jsp) > The page containing the directory listing does not come up. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HDFS-4108) In a secure cluster, in the HDFS WEBUI , clicking on a datanode in the node list , gives an error
[ https://issues.apache.org/jira/browse/HDFS-4108?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Benoy Antony updated HDFS-4108: --- Attachment: HDFS-4108-1-1.patch Attaching the patch based on the MAPREDUCE-4661 codebase. It also keeps the old behavior with insecure cluster. > In a secure cluster, in the HDFS WEBUI , clicking on a datanode in the node > list , gives an error > - > > Key: HDFS-4108 > URL: https://issues.apache.org/jira/browse/HDFS-4108 > Project: Hadoop HDFS > Issue Type: Bug > Components: security, webhdfs >Affects Versions: 1.1.0 >Reporter: Benoy Antony >Assignee: Benoy Antony >Priority: Minor > Attachments: HDFS-4108-1-1.patch, HDFS-4108-1-1.patch > > > This issue happens in secure cluster. > To reproduce : > Go to the NameNode WEB UI. (dfshealth.jsp) > Click to bring up the list of LiveNodes (dfsnodelist.jsp) > Click on a datanode to bring up the filesystem web page ( > browsedirectory.jsp) > The page containing the directory listing does not come up. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HDFS-4056) Always start the NN's SecretManager
[ https://issues.apache.org/jira/browse/HDFS-4056?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13488252#comment-13488252 ] Daryn Sharp commented on HDFS-4056: --- bq. I'm not sure we support an insecure cluster to talk to a secure one right now. Yes, it should work if you fetch the token yourself. For clarification, the way it currently works: * Insecure cluster: ** NN always returns no token (null) if client asks for one ** Client correctly handles null response as security disabled ** Client passes token, if present, in subsequent connections - tokens currently will never be present, unless fetchdt is used to obtain one from a secure cluster ** If a client attempts to use a token, the NN tells it to revert to SIMPLE * Secure cluster: ** NN returns token if: **# client is kerberos, else throws exception for other auths (like token) **# if the secret manager is running, else returns null like an insecure cluster ** Client passes token, if present, in subsequent connections This allows secure and insecure clusters to interoperate under the right conditions. The main difference is insecure clients only fetch tokens if explicitly requested, whereas secure clients automatically attempt to fetch tokens. Both always send tokens if available, and both interpret no token from an NN to mean security is disabled. bq. That [the secret manager running?] would be very confusing and I certainly don't want to see it happen in my production cluster. I may configure a cluster to use either SIMPLE + SIMPLE or SIMPLE + TOKEN for different purposes, but I don't see why I want to mix them up, especially at production (there is no security benefit, only overhead). What would be the confusion? With this patch alone, there is no change and zero overhead for clients of an insecure cluster. Clients from a secure cluster will however be able to request and use tokens on the insecure cluster. bq. If the conf says SIMPLE + SIMPLE, then even if a token is present, it shouldn't be used. That token may be from a stale token file. RPC Client shouldn't even waste its time looking for tokens. Trust me, I've dealt with so many token issues that my poor wife knows what they are. :) I don't recall a problem ever being root-caused to a stale token file. The overhead of the RPC client looking for a token in an empty collection in the UGI is negligible/moot. Although there is no security benefit for an insecure cluster, the huge benefit is to the hadoop codebase by moving towards the elimination of multiple code paths related to security. We can decrease the complexity and increase the code coverage. Most importantly, we can remove the fragility caused by people accidentally breaking tokens because they have no access to a secure cluster. > Always start the NN's SecretManager > --- > > Key: HDFS-4056 > URL: https://issues.apache.org/jira/browse/HDFS-4056 > Project: Hadoop HDFS > Issue Type: Improvement > Components: name-node >Affects Versions: 0.23.0, 2.0.0-alpha, 3.0.0 >Reporter: Daryn Sharp >Assignee: Daryn Sharp > Attachments: HDFS-4056.patch > > > To support the ability to use tokens regardless of whether kerberos is > enabled, the NN's secret manager should always be started. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HDFS-4045) SecondaryNameNode cannot read from QuorumJournal URI
[ https://issues.apache.org/jira/browse/HDFS-4045?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13488246#comment-13488246 ] Colin Patrick McCabe commented on HDFS-4045: I guess we have to think about whether the overhead of avoiding an extra "deserialize + serialize" pair on the JN is worth violating the EditLogInputStream abstraction. Unfortunately, I can't think of a good way to get the benefits of {{RedundantEditLogInputStream}} without using its intended interface-- {{readOp}}. If you do choose to violate the abstraction, it would be best to do so by creating FileInputStream objects for the underlying edit log files, when appropriate. Then just read directly from the files. You'll still have a hard time dealing with IOExceptions in the middle of reading the file. It wouldn't be *too* hard to do that failover yourself, but you run into sticky situations. What if you get an IOExecption, fail over to the next file, and then discover that it's actually fairly different from the first one? You might have sent bytes over the wire that completely fail the edit log operation checksums. I think it's best to punt on the issue of zero-copy {{serveEdits}} for now. Maybe open up a follow-up JIRA for that. Just do the simple thing of streaming it for now. I also wonder whether serveEdits should serve a range of edits, rather than discrete edit log segments? The API feels very segment-oriented, and I thought we were trying to get away from that when possible. > SecondaryNameNode cannot read from QuorumJournal URI > > > Key: HDFS-4045 > URL: https://issues.apache.org/jira/browse/HDFS-4045 > Project: Hadoop HDFS > Issue Type: Bug > Components: name-node >Affects Versions: 3.0.0 >Reporter: Vinithra Varadharajan >Assignee: Andy Isaacson > Attachments: hdfs-4045-2.txt, hdfs4045-3.txt, hdfs-4045.txt > > > If HDFS is set up in basic mode (non-HA) with QuorumJournal, and the > dfs.namenode.edits.dir is set to only the QuorumJournal URI and no local dir, > the SecondaryNameNode is unable to do a checkpoint. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HDFS-2802) Support for RW/RO snapshots in HDFS
[ https://issues.apache.org/jira/browse/HDFS-2802?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13488234#comment-13488234 ] Suresh Srinivas commented on HDFS-2802: --- bq. Can you send a link? NetApp uses this - search on google, you will find many other file systems such as HP X9000 File System etc. bq. My concern is that you have a directory dr with two subdirectories sd2 and sd3. You create a snapshot of sd2 and sd3 under the same name. Then create a snapshot of dr with the same name. What happens? The design should work fine. Every snapshot has the following: - A directory where the snapshot is created - A name that is unique for a given directory where snapshot is taken - Internally snapshot dir + name is mapped to a unique snapshot ID and the diffs are maintained against that snapshot ID. When snapshot is accessed, the path used provides snapshotted directory and snapshot name. It gets mapped to a unique snapshot ID. For this we can get to snapshotted version. > Support for RW/RO snapshots in HDFS > --- > > Key: HDFS-2802 > URL: https://issues.apache.org/jira/browse/HDFS-2802 > Project: Hadoop HDFS > Issue Type: New Feature > Components: data-node, name-node >Reporter: Hari Mankude >Assignee: Hari Mankude > Attachments: HDFSSnapshotsDesign.pdf, snap.patch, > snapshot-one-pager.pdf, Snapshots20121018.pdf, Snapshots20121030.pdf > > > Snapshots are point in time images of parts of the filesystem or the entire > filesystem. Snapshots can be a read-only or a read-write point in time copy > of the filesystem. There are several use cases for snapshots in HDFS. I will > post a detailed write-up soon with with more information. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HDFS-4056) Always start the NN's SecretManager
[ https://issues.apache.org/jira/browse/HDFS-4056?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13488141#comment-13488141 ] Kan Zhang commented on HDFS-4056: - bq. ... there is no harm in having the secret manager running if a client using SIMPLE auth choses to use tokens. That would be very confusing and I certainly don't want to see it happen in my production cluster. I may configure a cluster to use either SIMPLE + SIMPLE or SIMPLE + TOKEN for different purposes, but I don't see why I want to mix them up, especially at production (there is no security benefit, only overhead). bq. I don't understand this objection. If a token is available, why not use it? If the conf says SIMPLE + SIMPLE, then even if a token is present, it shouldn't be used. That token may be from a stale token file. RPC Client shouldn't even waste its time looking for tokens. bq. If a cluster not using tokens wants to talk to a cluster requiring tokens, then doesn't it have to send the token regardless of the local config? That's a different issue. I'm not sure we support an insecure cluster to talk to a secure one right now. Is that your goal? > Always start the NN's SecretManager > --- > > Key: HDFS-4056 > URL: https://issues.apache.org/jira/browse/HDFS-4056 > Project: Hadoop HDFS > Issue Type: Improvement > Components: name-node >Affects Versions: 0.23.0, 2.0.0-alpha, 3.0.0 >Reporter: Daryn Sharp >Assignee: Daryn Sharp > Attachments: HDFS-4056.patch > > > To support the ability to use tokens regardless of whether kerberos is > enabled, the NN's secret manager should always be started. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HDFS-2802) Support for RW/RO snapshots in HDFS
[ https://issues.apache.org/jira/browse/HDFS-2802?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13488112#comment-13488112 ] Konstantin Shvachko commented on HDFS-2802: --- > After looking at many other file systems, where .snapshot convention Can you send a link? My concern is that you have a directory dr with two subdirectories sd2 and sd3. You create a snapshot of sd2 and sd3 under the same name. Then create a snapshot of dr with the same name. What happens? If the system controls ids this never happens because they are unique. > Support for RW/RO snapshots in HDFS > --- > > Key: HDFS-2802 > URL: https://issues.apache.org/jira/browse/HDFS-2802 > Project: Hadoop HDFS > Issue Type: New Feature > Components: data-node, name-node >Reporter: Hari Mankude >Assignee: Hari Mankude > Attachments: HDFSSnapshotsDesign.pdf, snap.patch, > snapshot-one-pager.pdf, Snapshots20121018.pdf, Snapshots20121030.pdf > > > Snapshots are point in time images of parts of the filesystem or the entire > filesystem. Snapshots can be a read-only or a read-write point in time copy > of the filesystem. There are several use cases for snapshots in HDFS. I will > post a detailed write-up soon with with more information. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HDFS-1331) dfs -test should work like /bin/test
[ https://issues.apache.org/jira/browse/HDFS-1331?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Andy Isaacson updated HDFS-1331: Status: Open (was: Patch Available) re-submitting patch to try to kick hadoopqa into action > dfs -test should work like /bin/test > > > Key: HDFS-1331 > URL: https://issues.apache.org/jira/browse/HDFS-1331 > Project: Hadoop HDFS > Issue Type: Bug > Components: tools >Affects Versions: 2.0.2-alpha, 0.20.2, 3.0.0 >Reporter: Allen Wittenauer >Assignee: Andy Isaacson >Priority: Minor > Attachments: hdfs1331-2.txt, hdfs1331.txt, > hdfs1331-with-hadoop8994.txt > > > hadoop dfs -test doesn't act like its shell equivalent, making it difficult > to actually use if you are used to the real test command: > hadoop: > $hadoop dfs -test -d /nonexist; echo $? > test: File does not exist: /nonexist > 255 > shell: > $ test -d /nonexist; echo $? > 1 > a) Why is it spitting out a message? Even so, why is it saying file instead > of directory when I used -d? > b) Why is the return code 255? I realize this is documented as '0' if true. > But docs basically say the value is undefined if it isn't. > c) where is -f? > d) Why is empty -z instead of -s ? Was it a misunderstanding of the man page? -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HDFS-1331) dfs -test should work like /bin/test
[ https://issues.apache.org/jira/browse/HDFS-1331?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Andy Isaacson updated HDFS-1331: Status: Patch Available (was: Open) > dfs -test should work like /bin/test > > > Key: HDFS-1331 > URL: https://issues.apache.org/jira/browse/HDFS-1331 > Project: Hadoop HDFS > Issue Type: Bug > Components: tools >Affects Versions: 2.0.2-alpha, 0.20.2, 3.0.0 >Reporter: Allen Wittenauer >Assignee: Andy Isaacson >Priority: Minor > Attachments: hdfs1331-2.txt, hdfs1331.txt, > hdfs1331-with-hadoop8994.txt > > > hadoop dfs -test doesn't act like its shell equivalent, making it difficult > to actually use if you are used to the real test command: > hadoop: > $hadoop dfs -test -d /nonexist; echo $? > test: File does not exist: /nonexist > 255 > shell: > $ test -d /nonexist; echo $? > 1 > a) Why is it spitting out a message? Even so, why is it saying file instead > of directory when I used -d? > b) Why is the return code 255? I realize this is documented as '0' if true. > But docs basically say the value is undefined if it isn't. > c) where is -f? > d) Why is empty -z instead of -s ? Was it a misunderstanding of the man page? -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HDFS-2802) Support for RW/RO snapshots in HDFS
[ https://issues.apache.org/jira/browse/HDFS-2802?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13488074#comment-13488074 ] Eli Collins commented on HDFS-2802: --- Hey Suresh, This isn't the one and only discussion on snapshots, a lot of us plan to attend the following one as well. It's OK to have multiple discussions. A bunch of people are going to be around and discussing things on Thursday anyway, so I thought it better to open up the discussion to more people in the community. I'm sorry that you'll be out of town, I've setup a dial in and would love it if you and others that are out could attend. I'll post minutes here so that people who can not attend can see and participate. > Support for RW/RO snapshots in HDFS > --- > > Key: HDFS-2802 > URL: https://issues.apache.org/jira/browse/HDFS-2802 > Project: Hadoop HDFS > Issue Type: New Feature > Components: data-node, name-node >Reporter: Hari Mankude >Assignee: Hari Mankude > Attachments: HDFSSnapshotsDesign.pdf, snap.patch, > snapshot-one-pager.pdf, Snapshots20121018.pdf, Snapshots20121030.pdf > > > Snapshots are point in time images of parts of the filesystem or the entire > filesystem. Snapshots can be a read-only or a read-write point in time copy > of the filesystem. There are several use cases for snapshots in HDFS. I will > post a detailed write-up soon with with more information. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HDFS-3804) TestHftpFileSystem fails intermittently with JDK7
[ https://issues.apache.org/jira/browse/HDFS-3804?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13488075#comment-13488075 ] Hadoop QA commented on HDFS-3804: - {color:green}+1 overall{color}. Here are the results of testing the latest attachment http://issues.apache.org/jira/secure/attachment/12551578/HDFS-3804-3.patch against trunk revision . {color:green}+1 @author{color}. The patch does not contain any @author tags. {color:green}+1 tests included{color}. The patch appears to include 1 new or modified test files. {color:green}+1 javac{color}. The applied patch does not increase the total number of javac compiler warnings. {color:green}+1 javadoc{color}. The javadoc tool did not generate any warning messages. {color:green}+1 eclipse:eclipse{color}. The patch built with eclipse:eclipse. {color:green}+1 findbugs{color}. The patch does not introduce any new Findbugs (version 1.3.9) warnings. {color:green}+1 release audit{color}. The applied patch does not increase the total number of release audit warnings. {color:green}+1 core tests{color}. The patch passed unit tests in hadoop-hdfs-project/hadoop-hdfs. {color:green}+1 contrib tests{color}. The patch passed contrib unit tests. Test results: https://builds.apache.org/job/PreCommit-HDFS-Build/3435//testReport/ Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/3435//console This message is automatically generated. > TestHftpFileSystem fails intermittently with JDK7 > - > > Key: HDFS-3804 > URL: https://issues.apache.org/jira/browse/HDFS-3804 > Project: Hadoop HDFS > Issue Type: Bug > Components: test > Environment: Apache Maven 3.0.4 > Maven home: /usr/share/maven > Java version: 1.7.0_04, vendor: Oracle Corporation > Java home: /usr/lib/jvm/jdk1.7.0_04/jre > Default locale: en_US, platform encoding: ISO-8859-1 > OS name: "linux", version: "3.2.0-25-generic", arch: "amd64", family: "unix" >Reporter: Trevor Robinson >Assignee: Trevor Robinson > Labels: java7 > Attachments: HDFS-3804-2.patch, HDFS-3804-3.patch, HDFS-3804-3.patch, > HDFS-3804.patch, HDFS-3804.patch > > > For example: > testFileNameEncoding(org.apache.hadoop.hdfs.TestHftpFileSystem): Filesystem > closed > testDataNodeRedirect(org.apache.hadoop.hdfs.TestHftpFileSystem): Filesystem > closed > This test case sets up a filesystem that is used by the first half of the > test methods (in declaration order), but the second half of the tests start > by calling {{FileSystem.closeAll}}. With JDK7, test methods are run in an > arbitrary order, so if any first half methods run after any second half > methods, they fail. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HDFS-2802) Support for RW/RO snapshots in HDFS
[ https://issues.apache.org/jira/browse/HDFS-2802?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13487983#comment-13487983 ] Suresh Srinivas commented on HDFS-2802: --- bq. I went ahead and scheduled this... [~eli]I did not schedule it the week of Hadoop World even though it was convenient for me and Nicholas. I have shown sensitivity towards your team's travels. Unfortunately I and Nicholas are scheduled for a travel this week. Even though I had indicated that I will schedule a meeting, instead of giving me your feedback on time, you have gone ahead and scheduled this meeting without consulting the availability of the main authors of this jira. I am not sure what you are trying to achieve here. My suggestion is to cancel the first meeting and stick with the second meeting. If the time for the second meeting does not work, Sanjay could reschedule it. > Support for RW/RO snapshots in HDFS > --- > > Key: HDFS-2802 > URL: https://issues.apache.org/jira/browse/HDFS-2802 > Project: Hadoop HDFS > Issue Type: New Feature > Components: data-node, name-node >Reporter: Hari Mankude >Assignee: Hari Mankude > Attachments: HDFSSnapshotsDesign.pdf, snap.patch, > snapshot-one-pager.pdf, Snapshots20121018.pdf, Snapshots20121030.pdf > > > Snapshots are point in time images of parts of the filesystem or the entire > filesystem. Snapshots can be a read-only or a read-write point in time copy > of the filesystem. There are several use cases for snapshots in HDFS. I will > post a detailed write-up soon with with more information. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HDFS-4056) Always start the NN's SecretManager
[ https://issues.apache.org/jira/browse/HDFS-4056?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13487977#comment-13487977 ] Daryn Sharp commented on HDFS-4056: --- bq. To me, a cluster is configured to run in either token testing mode or production mode. The original goal was to have only one code path so tokens are always used. Ie. there is no testing mode. I've implemented PLAIN as a compromise but there is no harm in having the secret manager running if a client using SIMPLE auth choses to use tokens. bq. IMO, they make the Client and Server less intelligent in the sense that they don't recognize situations they used to recognize. I'm not sure their new behavior is desirable. For example, Client will always look for token and try to use it if found, even if configuration says otherwise. I don't understand this objection. If a token is available, why not use it? Under what scenario do you envision a client, for any external auth, requesting a token and then not wanting to use it? If a cluster not using tokens wants to talk to a cluster requiring tokens, then doesn't it have to send the token regardless of the local config? > Always start the NN's SecretManager > --- > > Key: HDFS-4056 > URL: https://issues.apache.org/jira/browse/HDFS-4056 > Project: Hadoop HDFS > Issue Type: Improvement > Components: name-node >Affects Versions: 0.23.0, 2.0.0-alpha, 3.0.0 >Reporter: Daryn Sharp >Assignee: Daryn Sharp > Attachments: HDFS-4056.patch > > > To support the ability to use tokens regardless of whether kerberos is > enabled, the NN's secret manager should always be started. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HDFS-4075) Reduce recommissioning overhead
[ https://issues.apache.org/jira/browse/HDFS-4075?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13487975#comment-13487975 ] Hadoop QA commented on HDFS-4075: - {color:red}-1 overall{color}. Here are the results of testing the latest attachment http://issues.apache.org/jira/secure/attachment/12551560/hdfs-4075.patch against trunk revision . {color:green}+1 @author{color}. The patch does not contain any @author tags. {color:red}-1 tests included{color}. The patch doesn't appear to include any new or modified tests. Please justify why no new tests are needed for this patch. Also please list what manual steps were performed to verify this patch. {color:green}+1 javac{color}. The applied patch does not increase the total number of javac compiler warnings. {color:green}+1 javadoc{color}. The javadoc tool did not generate any warning messages. {color:green}+1 eclipse:eclipse{color}. The patch built with eclipse:eclipse. {color:green}+1 findbugs{color}. The patch does not introduce any new Findbugs (version 1.3.9) warnings. {color:green}+1 release audit{color}. The applied patch does not increase the total number of release audit warnings. {color:green}+1 core tests{color}. The patch passed unit tests in hadoop-hdfs-project/hadoop-hdfs. {color:green}+1 contrib tests{color}. The patch passed contrib unit tests. Test results: https://builds.apache.org/job/PreCommit-HDFS-Build/3434//testReport/ Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/3434//console This message is automatically generated. > Reduce recommissioning overhead > --- > > Key: HDFS-4075 > URL: https://issues.apache.org/jira/browse/HDFS-4075 > Project: Hadoop HDFS > Issue Type: Bug > Components: name-node >Affects Versions: 0.23.4, 2.0.2-alpha >Reporter: Kihwal Lee >Assignee: Kihwal Lee >Priority: Critical > Attachments: hdfs-4075.patch, hdfs-4075.patch > > > When datanodes are recommissioned, > {BlockManager#processOverReplicatedBlocksOnReCommission()} is called for each > rejoined node and excess blocks are added to the invalidate list. The problem > is this is done while the namesystem write lock is held. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HDFS-2802) Support for RW/RO snapshots in HDFS
[ https://issues.apache.org/jira/browse/HDFS-2802?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13487972#comment-13487972 ] Suresh Srinivas commented on HDFS-2802: --- [~shv]Thanks for your comments. Some answers here: versions seem like an interesting capability to support. I want to think about it more. However snapshots have their own use cases as well. Ability to snapshot and attach a name to it comes handy for writing applications using this functionality. .snapshot is a convention that is used widely and requires no changes to many of the APIs such as rm, ls etc. In fact we had initially proposed using .snap_. After looking at many other file systems, where .snapshot convention is used to identify the snapshot, we decided to go with it. bq. Creating duplicate INodes with a diff, this is sort of COW technique, right? Sounds hard. Hopefully this should not add too much code complexity. But this is being done to avoid taking too much memory for snapshots on the namenode. bq. My dumb question: can I create a snapshot of a subdirectory that is a part of a snapshot above it? This is not a dumb question :-) The design already supports nested snapshots. I will add an update to the document describing this. > Support for RW/RO snapshots in HDFS > --- > > Key: HDFS-2802 > URL: https://issues.apache.org/jira/browse/HDFS-2802 > Project: Hadoop HDFS > Issue Type: New Feature > Components: data-node, name-node >Reporter: Hari Mankude >Assignee: Hari Mankude > Attachments: HDFSSnapshotsDesign.pdf, snap.patch, > snapshot-one-pager.pdf, Snapshots20121018.pdf, Snapshots20121030.pdf > > > Snapshots are point in time images of parts of the filesystem or the entire > filesystem. Snapshots can be a read-only or a read-write point in time copy > of the filesystem. There are several use cases for snapshots in HDFS. I will > post a detailed write-up soon with with more information. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HDFS-4075) Reduce recommissioning overhead
[ https://issues.apache.org/jira/browse/HDFS-4075?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13487969#comment-13487969 ] Daryn Sharp commented on HDFS-4075: --- +1 I think it looks ok. Eli, can you confirm? > Reduce recommissioning overhead > --- > > Key: HDFS-4075 > URL: https://issues.apache.org/jira/browse/HDFS-4075 > Project: Hadoop HDFS > Issue Type: Bug > Components: name-node >Affects Versions: 0.23.4, 2.0.2-alpha >Reporter: Kihwal Lee >Assignee: Kihwal Lee >Priority: Critical > Attachments: hdfs-4075.patch, hdfs-4075.patch > > > When datanodes are recommissioned, > {BlockManager#processOverReplicatedBlocksOnReCommission()} is called for each > rejoined node and excess blocks are added to the invalidate list. The problem > is this is done while the namesystem write lock is held. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HDFS-3804) TestHftpFileSystem fails intermittently with JDK7
[ https://issues.apache.org/jira/browse/HDFS-3804?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Trevor Robinson updated HDFS-3804: -- Attachment: HDFS-3804-3.patch Sure, done. > TestHftpFileSystem fails intermittently with JDK7 > - > > Key: HDFS-3804 > URL: https://issues.apache.org/jira/browse/HDFS-3804 > Project: Hadoop HDFS > Issue Type: Bug > Components: test > Environment: Apache Maven 3.0.4 > Maven home: /usr/share/maven > Java version: 1.7.0_04, vendor: Oracle Corporation > Java home: /usr/lib/jvm/jdk1.7.0_04/jre > Default locale: en_US, platform encoding: ISO-8859-1 > OS name: "linux", version: "3.2.0-25-generic", arch: "amd64", family: "unix" >Reporter: Trevor Robinson >Assignee: Trevor Robinson > Labels: java7 > Attachments: HDFS-3804-2.patch, HDFS-3804-3.patch, HDFS-3804-3.patch, > HDFS-3804.patch, HDFS-3804.patch > > > For example: > testFileNameEncoding(org.apache.hadoop.hdfs.TestHftpFileSystem): Filesystem > closed > testDataNodeRedirect(org.apache.hadoop.hdfs.TestHftpFileSystem): Filesystem > closed > This test case sets up a filesystem that is used by the first half of the > test methods (in declaration order), but the second half of the tests start > by calling {{FileSystem.closeAll}}. With JDK7, test methods are run in an > arbitrary order, so if any first half methods run after any second half > methods, they fail. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HDFS-4080) Add an option to disable block-level state change logging
[ https://issues.apache.org/jira/browse/HDFS-4080?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13487962#comment-13487962 ] Hadoop QA commented on HDFS-4080: - {color:green}+1 overall{color}. Here are the results of testing the latest attachment http://issues.apache.org/jira/secure/attachment/12551558/hdfs-4080.patch against trunk revision . {color:green}+1 @author{color}. The patch does not contain any @author tags. {color:green}+1 tests included{color}. The patch appears to include 2 new or modified test files. {color:green}+1 javac{color}. The applied patch does not increase the total number of javac compiler warnings. {color:green}+1 javadoc{color}. The javadoc tool did not generate any warning messages. {color:green}+1 eclipse:eclipse{color}. The patch built with eclipse:eclipse. {color:green}+1 findbugs{color}. The patch does not introduce any new Findbugs (version 1.3.9) warnings. {color:green}+1 release audit{color}. The applied patch does not increase the total number of release audit warnings. {color:green}+1 core tests{color}. The patch passed unit tests in hadoop-hdfs-project/hadoop-hdfs. {color:green}+1 contrib tests{color}. The patch passed contrib unit tests. Test results: https://builds.apache.org/job/PreCommit-HDFS-Build/3433//testReport/ Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/3433//console This message is automatically generated. > Add an option to disable block-level state change logging > - > > Key: HDFS-4080 > URL: https://issues.apache.org/jira/browse/HDFS-4080 > Project: Hadoop HDFS > Issue Type: Bug > Components: name-node >Affects Versions: 3.0.0, 2.0.3-alpha >Reporter: Kihwal Lee >Assignee: Kihwal Lee > Attachments: hdfs-4080.1.patch, hdfs-4080.patch, hdfs-4080.patch > > > Although the block-level logging in namenode is useful for debugging, it can > add a significant overhead to busy hdfs clusters since they are done while > the namespace write lock is held. One example is shown in HDFS-4075. In this > example, the write lock was held for 5 minutes while logging 11 million log > messages for 5.5 million block invalidation events. > It will be useful if we have an option to disable these block-level log > messages and keep other state change messages going. If others feel that > they can turned into DEBUG (with addition of isDebugEnabled() checks), that > may also work too, but there might be people depending on the messages. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HDFS-4046) ChecksumTypeProto use NULL as enum value which is illegal in C/C++
[ https://issues.apache.org/jira/browse/HDFS-4046?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13487881#comment-13487881 ] Kihwal Lee commented on HDFS-4046: -- +1 (nb) Looks good to me. The namespace declaration was added in HADOOP-8985 and HDFS-4121 > ChecksumTypeProto use NULL as enum value which is illegal in C/C++ > -- > > Key: HDFS-4046 > URL: https://issues.apache.org/jira/browse/HDFS-4046 > Project: Hadoop HDFS > Issue Type: Bug >Reporter: Binglin Chang >Assignee: Binglin Chang >Priority: Minor > Attachments: HDFS-4046-ChecksumType-NULL-and-TestAuditLogs-bug.patch, > HDFS-4046-ChecksumType-NULL.patch, HDFS-4096-ChecksumTypeProto-NULL.patch > > > I tried to write a native hdfs client using protobuf based protocol, when I > generate c++ code using hdfs.proto, the generated file can not compile, > because NULL is an already defined macro. > I am thinking two solutions: > 1. refactor all DataChecksum.Type.NULL references to NONE, which should be > fine for all languages, but this may breaking compatibility. > 2. only change protobuf definition ChecksumTypeProto.NULL to NONE, and use > enum integer value(DataChecksum.Type.id) to convert between ChecksumTypeProto > and DataChecksum.Type, and make sure enum integer values are match(currently > already match). > I can make a patch for solution 2. > -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HDFS-4075) Reduce recommissioning overhead
[ https://issues.apache.org/jira/browse/HDFS-4075?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Kihwal Lee updated HDFS-4075: - Attachment: hdfs-4075.patch I excluded the logging change since HDFS-4080 will take care of it. The patch lets the recommissioning skip the over-replication check for dead nodes and logs the total number overreplicated blocks per node. I expect the future lock improvement will reduce the duration of namespace write locking. > Reduce recommissioning overhead > --- > > Key: HDFS-4075 > URL: https://issues.apache.org/jira/browse/HDFS-4075 > Project: Hadoop HDFS > Issue Type: Bug > Components: name-node >Affects Versions: 0.23.4, 2.0.2-alpha >Reporter: Kihwal Lee >Assignee: Kihwal Lee >Priority: Critical > Attachments: hdfs-4075.patch, hdfs-4075.patch > > > When datanodes are recommissioned, > {BlockManager#processOverReplicatedBlocksOnReCommission()} is called for each > rejoined node and excess blocks are added to the invalidate list. The problem > is this is done while the namesystem write lock is held. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HDFS-4080) Add an option to disable block-level state change logging
[ https://issues.apache.org/jira/browse/HDFS-4080?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Kihwal Lee updated HDFS-4080: - Attachment: hdfs-4080.patch The new patch addresses the review comment. The scope of the change is limited to the block state changes= messages from BlockManager & friends and the calls through DatanodeProtocol. > Add an option to disable block-level state change logging > - > > Key: HDFS-4080 > URL: https://issues.apache.org/jira/browse/HDFS-4080 > Project: Hadoop HDFS > Issue Type: Bug > Components: name-node >Affects Versions: 3.0.0, 2.0.3-alpha >Reporter: Kihwal Lee >Assignee: Kihwal Lee > Attachments: hdfs-4080.1.patch, hdfs-4080.patch, hdfs-4080.patch > > > Although the block-level logging in namenode is useful for debugging, it can > add a significant overhead to busy hdfs clusters since they are done while > the namespace write lock is held. One example is shown in HDFS-4075. In this > example, the write lock was held for 5 minutes while logging 11 million log > messages for 5.5 million block invalidation events. > It will be useful if we have an option to disable these block-level log > messages and keep other state change messages going. If others feel that > they can turned into DEBUG (with addition of isDebugEnabled() checks), that > may also work too, but there might be people depending on the messages. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HDFS-3809) Make BKJM use protobufs for all serialization with ZK
[ https://issues.apache.org/jira/browse/HDFS-3809?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13487794#comment-13487794 ] Hudson commented on HDFS-3809: -- Integrated in Hadoop-Mapreduce-trunk #1242 (See [https://builds.apache.org/job/Hadoop-Mapreduce-trunk/1242/]) Moved HDFS-3809 entry in CHANGES.txt from trunk to 2.0.3-alpha section (Revision 1403769) Result = FAILURE umamahesh : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1403769 Files : * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt > Make BKJM use protobufs for all serialization with ZK > - > > Key: HDFS-3809 > URL: https://issues.apache.org/jira/browse/HDFS-3809 > Project: Hadoop HDFS > Issue Type: Sub-task > Components: name-node >Affects Versions: 2.0.0-alpha, 3.0.0 >Reporter: Ivan Kelly >Assignee: Ivan Kelly > Fix For: 3.0.0, 2.0.3-alpha > > Attachments: > 0001-HDFS-3809.-Make-BKJM-use-protobufs-for-all-serializa.patch, > 0001-HDFS-3809.-Make-BKJM-use-protobufs-for-all-serializa.patch, > 0004-HDFS-3809-for-branch-2.patch, HDFS-3809.diff, HDFS-3809.diff, > HDFS-3809.diff > > > HDFS uses protobufs for serialization in many places. Protobufs allow fields > to be added without breaking bc or requiring new parsing code to be written. > For this reason, we should use them in BKJM also. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HDFS-3695) Genericize format() to non-file JournalManagers
[ https://issues.apache.org/jira/browse/HDFS-3695?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13487793#comment-13487793 ] Hudson commented on HDFS-3695: -- Integrated in Hadoop-Mapreduce-trunk #1242 (See [https://builds.apache.org/job/Hadoop-Mapreduce-trunk/1242/]) Moved HDFS-3695 entry in CHANGES.txt from trunk to 2.0.3-alpha section (Revision 1403748) Result = FAILURE umamahesh : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1403748 Files : * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt > Genericize format() to non-file JournalManagers > --- > > Key: HDFS-3695 > URL: https://issues.apache.org/jira/browse/HDFS-3695 > Project: Hadoop HDFS > Issue Type: Sub-task > Components: ha, name-node >Affects Versions: QuorumJournalManager (HDFS-3077) >Reporter: Todd Lipcon >Assignee: Todd Lipcon > Fix For: 3.0.0, 2.0.3-alpha > > Attachments: 0002-HDFS-3695-for-branch-2.patch, hdfs-3695.txt, > hdfs-3695.txt, hdfs-3695.txt > > > Currently, the "namenode -format" and "namenode -initializeSharedEdits" > commands do not understand how to do anything with non-file-based shared > storage. This affects both BookKeeperJournalManager and QuorumJournalManager. > This JIRA is to plumb through the formatting of edits directories using > pluggable journal manager implementations so that no separate step needs to > be taken to format them -- the same commands will work for NFS-based storage > or one of the alternate implementations. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HDFS-3916) libwebhdfs (C client) code cleanups
[ https://issues.apache.org/jira/browse/HDFS-3916?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13487792#comment-13487792 ] Hudson commented on HDFS-3916: -- Integrated in Hadoop-Mapreduce-trunk #1242 (See [https://builds.apache.org/job/Hadoop-Mapreduce-trunk/1242/]) HDFS-3916. libwebhdfs testing code cleanup. Contributed by Jing Zhao. (Revision 1403922) Result = FAILURE suresh : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1403922 Files : * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/contrib/libwebhdfs/src/test_libwebhdfs_multi_write.c * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/contrib/libwebhdfs/src/test_libwebhdfs_ops.c * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/contrib/libwebhdfs/src/test_libwebhdfs_read.c * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/contrib/libwebhdfs/src/test_libwebhdfs_threaded.c * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/contrib/libwebhdfs/src/test_libwebhdfs_write.c * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/contrib/libwebhdfs/src/test_read_bm.c * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/native/libhdfs/native_mini_dfs.c * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/native/libhdfs/native_mini_dfs.h > libwebhdfs (C client) code cleanups > --- > > Key: HDFS-3916 > URL: https://issues.apache.org/jira/browse/HDFS-3916 > Project: Hadoop HDFS > Issue Type: Bug > Components: webhdfs >Affects Versions: 2.0.2-alpha >Reporter: Colin Patrick McCabe >Assignee: Colin Patrick McCabe > Fix For: 2.0.3-alpha > > Attachments: 0002-fix.patch, HDFS-3916.003.patch, > HDFS-3916.004.patch, HDFS-3916.005.patch, HDFS-3916.006.patch > > > Code cleanups in libwebhdfs. > * don't duplicate exception.c, exception.h, expect.h, jni_helper.c. We have > one copy of these files; we don't need 2. > * remember to set errno in all public library functions (this is part of the > API) > * fix undefined symbols (if a function is not implemented, it should return > ENOTSUP, but still exist) > * don't expose private data structures in the (end-user visible) public > headers > * can't re-use hdfsBuilder as hdfsFS, because the strings in hdfsBuilder are > not dynamically allocated. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HDFS-4129) Add utility methods to dump NameNode in memory tree for testing
[ https://issues.apache.org/jira/browse/HDFS-4129?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13487787#comment-13487787 ] Hudson commented on HDFS-4129: -- Integrated in Hadoop-Mapreduce-trunk #1242 (See [https://builds.apache.org/job/Hadoop-Mapreduce-trunk/1242/]) HDFS-4129. Add utility methods to dump NameNode in memory tree for testing. Contributed by Tsz Wo (Nicholas), SZE. (Revision 1403956) Result = FAILURE suresh : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1403956 Files : * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSNamesystem.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/INode.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/INodeDirectory.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestFSDirectory.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestFsLimits.java > Add utility methods to dump NameNode in memory tree for testing > --- > > Key: HDFS-4129 > URL: https://issues.apache.org/jira/browse/HDFS-4129 > Project: Hadoop HDFS > Issue Type: Test > Components: name-node >Reporter: Tsz Wo (Nicholas), SZE >Assignee: Tsz Wo (Nicholas), SZE >Priority: Minor > Fix For: 3.0.0 > > Attachments: h4129_20121029b.patch, h4129_20121029.patch, > h4129_20121030.patch > > > The output of the utility methods looks like below. > {noformat} > \- foo (INodeDirectory) > \- sub1 (INodeDirectory) > +- file1 (INodeFile) > +- file2 (INodeFile) > +- sub11 (INodeDirectory) > \- file3 (INodeFile) > \- z_file4 (INodeFile) > {noformat} -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HDFS-3789) JournalManager#format() should be able to throw IOException
[ https://issues.apache.org/jira/browse/HDFS-3789?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13487790#comment-13487790 ] Hudson commented on HDFS-3789: -- Integrated in Hadoop-Mapreduce-trunk #1242 (See [https://builds.apache.org/job/Hadoop-Mapreduce-trunk/1242/]) Moved HDFS-3789 entry in CHANGES.txt from trunk to 2.0.3-alpha section (Revision 1403765) Result = FAILURE umamahesh : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1403765 Files : * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt > JournalManager#format() should be able to throw IOException > --- > > Key: HDFS-3789 > URL: https://issues.apache.org/jira/browse/HDFS-3789 > Project: Hadoop HDFS > Issue Type: Sub-task > Components: ha, name-node >Affects Versions: 3.0.0 >Reporter: Ivan Kelly >Assignee: Ivan Kelly > Fix For: 3.0.0, 2.0.3-alpha > > Attachments: 0003-HDFS-3789-for-branch-2.patch, HDFS-3789.diff > > > Currently JournalManager#format cannot throw any exception. As format can > fail, we should be able to propogate this failure upwards. Otherwise, format > will fail silently, and the admin will start using the cluster with a > failed/unusable journal manager. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HDFS-3573) Supply NamespaceInfo when instantiating JournalManagers
[ https://issues.apache.org/jira/browse/HDFS-3573?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13487789#comment-13487789 ] Hudson commented on HDFS-3573: -- Integrated in Hadoop-Mapreduce-trunk #1242 (See [https://builds.apache.org/job/Hadoop-Mapreduce-trunk/1242/]) Moved HDFS-3573 entry in CHANGES.txt from trunk to 2.0.3-alpha section (Revision 1403740) Result = FAILURE umamahesh : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1403740 Files : * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt > Supply NamespaceInfo when instantiating JournalManagers > --- > > Key: HDFS-3573 > URL: https://issues.apache.org/jira/browse/HDFS-3573 > Project: Hadoop HDFS > Issue Type: Sub-task > Components: name-node >Affects Versions: 3.0.0 >Reporter: Todd Lipcon >Assignee: Todd Lipcon >Priority: Minor > Fix For: 3.0.0, 2.0.3-alpha > > Attachments: 0001-HDFS-3573-for-branch-2.patch, hdfs-3573.txt, > hdfs-3573.txt, hdfs-3573.txt, hdfs-3573.txt > > > Currently, the JournalManagers are instantiated before the NamespaceInfo is > loaded from local storage directories. This is problematic since the JM may > want to verify that the storage info associated with the journal matches the > NN which is starting up (eg to prevent an operator accidentally configuring > two clusters against the same remote journal storage). This JIRA rejiggers > the initialization sequence so that the JMs receive NamespaceInfo as a > constructor argument. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HDFS-3804) TestHftpFileSystem fails intermittently with JDK7
[ https://issues.apache.org/jira/browse/HDFS-3804?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13487767#comment-13487767 ] Daryn Sharp commented on HDFS-3804: --- +1 For consistency, do you think {{hftpFs}} should also be non-static? > TestHftpFileSystem fails intermittently with JDK7 > - > > Key: HDFS-3804 > URL: https://issues.apache.org/jira/browse/HDFS-3804 > Project: Hadoop HDFS > Issue Type: Bug > Components: test > Environment: Apache Maven 3.0.4 > Maven home: /usr/share/maven > Java version: 1.7.0_04, vendor: Oracle Corporation > Java home: /usr/lib/jvm/jdk1.7.0_04/jre > Default locale: en_US, platform encoding: ISO-8859-1 > OS name: "linux", version: "3.2.0-25-generic", arch: "amd64", family: "unix" >Reporter: Trevor Robinson >Assignee: Trevor Robinson > Labels: java7 > Attachments: HDFS-3804-2.patch, HDFS-3804-3.patch, HDFS-3804.patch, > HDFS-3804.patch > > > For example: > testFileNameEncoding(org.apache.hadoop.hdfs.TestHftpFileSystem): Filesystem > closed > testDataNodeRedirect(org.apache.hadoop.hdfs.TestHftpFileSystem): Filesystem > closed > This test case sets up a filesystem that is used by the first half of the > test methods (in declaration order), but the second half of the tests start > by calling {{FileSystem.closeAll}}. With JDK7, test methods are run in an > arbitrary order, so if any first half methods run after any second half > methods, they fail. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HDFS-3809) Make BKJM use protobufs for all serialization with ZK
[ https://issues.apache.org/jira/browse/HDFS-3809?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13487729#comment-13487729 ] Hudson commented on HDFS-3809: -- Integrated in Hadoop-Hdfs-trunk #1212 (See [https://builds.apache.org/job/Hadoop-Hdfs-trunk/1212/]) Moved HDFS-3809 entry in CHANGES.txt from trunk to 2.0.3-alpha section (Revision 1403769) Result = FAILURE umamahesh : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1403769 Files : * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt > Make BKJM use protobufs for all serialization with ZK > - > > Key: HDFS-3809 > URL: https://issues.apache.org/jira/browse/HDFS-3809 > Project: Hadoop HDFS > Issue Type: Sub-task > Components: name-node >Affects Versions: 2.0.0-alpha, 3.0.0 >Reporter: Ivan Kelly >Assignee: Ivan Kelly > Fix For: 3.0.0, 2.0.3-alpha > > Attachments: > 0001-HDFS-3809.-Make-BKJM-use-protobufs-for-all-serializa.patch, > 0001-HDFS-3809.-Make-BKJM-use-protobufs-for-all-serializa.patch, > 0004-HDFS-3809-for-branch-2.patch, HDFS-3809.diff, HDFS-3809.diff, > HDFS-3809.diff > > > HDFS uses protobufs for serialization in many places. Protobufs allow fields > to be added without breaking bc or requiring new parsing code to be written. > For this reason, we should use them in BKJM also. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HDFS-3695) Genericize format() to non-file JournalManagers
[ https://issues.apache.org/jira/browse/HDFS-3695?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13487728#comment-13487728 ] Hudson commented on HDFS-3695: -- Integrated in Hadoop-Hdfs-trunk #1212 (See [https://builds.apache.org/job/Hadoop-Hdfs-trunk/1212/]) Moved HDFS-3695 entry in CHANGES.txt from trunk to 2.0.3-alpha section (Revision 1403748) Result = FAILURE umamahesh : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1403748 Files : * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt > Genericize format() to non-file JournalManagers > --- > > Key: HDFS-3695 > URL: https://issues.apache.org/jira/browse/HDFS-3695 > Project: Hadoop HDFS > Issue Type: Sub-task > Components: ha, name-node >Affects Versions: QuorumJournalManager (HDFS-3077) >Reporter: Todd Lipcon >Assignee: Todd Lipcon > Fix For: 3.0.0, 2.0.3-alpha > > Attachments: 0002-HDFS-3695-for-branch-2.patch, hdfs-3695.txt, > hdfs-3695.txt, hdfs-3695.txt > > > Currently, the "namenode -format" and "namenode -initializeSharedEdits" > commands do not understand how to do anything with non-file-based shared > storage. This affects both BookKeeperJournalManager and QuorumJournalManager. > This JIRA is to plumb through the formatting of edits directories using > pluggable journal manager implementations so that no separate step needs to > be taken to format them -- the same commands will work for NFS-based storage > or one of the alternate implementations. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HDFS-3916) libwebhdfs (C client) code cleanups
[ https://issues.apache.org/jira/browse/HDFS-3916?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13487727#comment-13487727 ] Hudson commented on HDFS-3916: -- Integrated in Hadoop-Hdfs-trunk #1212 (See [https://builds.apache.org/job/Hadoop-Hdfs-trunk/1212/]) HDFS-3916. libwebhdfs testing code cleanup. Contributed by Jing Zhao. (Revision 1403922) Result = FAILURE suresh : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1403922 Files : * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/contrib/libwebhdfs/src/test_libwebhdfs_multi_write.c * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/contrib/libwebhdfs/src/test_libwebhdfs_ops.c * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/contrib/libwebhdfs/src/test_libwebhdfs_read.c * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/contrib/libwebhdfs/src/test_libwebhdfs_threaded.c * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/contrib/libwebhdfs/src/test_libwebhdfs_write.c * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/contrib/libwebhdfs/src/test_read_bm.c * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/native/libhdfs/native_mini_dfs.c * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/native/libhdfs/native_mini_dfs.h > libwebhdfs (C client) code cleanups > --- > > Key: HDFS-3916 > URL: https://issues.apache.org/jira/browse/HDFS-3916 > Project: Hadoop HDFS > Issue Type: Bug > Components: webhdfs >Affects Versions: 2.0.2-alpha >Reporter: Colin Patrick McCabe >Assignee: Colin Patrick McCabe > Fix For: 2.0.3-alpha > > Attachments: 0002-fix.patch, HDFS-3916.003.patch, > HDFS-3916.004.patch, HDFS-3916.005.patch, HDFS-3916.006.patch > > > Code cleanups in libwebhdfs. > * don't duplicate exception.c, exception.h, expect.h, jni_helper.c. We have > one copy of these files; we don't need 2. > * remember to set errno in all public library functions (this is part of the > API) > * fix undefined symbols (if a function is not implemented, it should return > ENOTSUP, but still exist) > * don't expose private data structures in the (end-user visible) public > headers > * can't re-use hdfsBuilder as hdfsFS, because the strings in hdfsBuilder are > not dynamically allocated. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HDFS-3573) Supply NamespaceInfo when instantiating JournalManagers
[ https://issues.apache.org/jira/browse/HDFS-3573?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13487724#comment-13487724 ] Hudson commented on HDFS-3573: -- Integrated in Hadoop-Hdfs-trunk #1212 (See [https://builds.apache.org/job/Hadoop-Hdfs-trunk/1212/]) Moved HDFS-3573 entry in CHANGES.txt from trunk to 2.0.3-alpha section (Revision 1403740) Result = FAILURE umamahesh : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1403740 Files : * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt > Supply NamespaceInfo when instantiating JournalManagers > --- > > Key: HDFS-3573 > URL: https://issues.apache.org/jira/browse/HDFS-3573 > Project: Hadoop HDFS > Issue Type: Sub-task > Components: name-node >Affects Versions: 3.0.0 >Reporter: Todd Lipcon >Assignee: Todd Lipcon >Priority: Minor > Fix For: 3.0.0, 2.0.3-alpha > > Attachments: 0001-HDFS-3573-for-branch-2.patch, hdfs-3573.txt, > hdfs-3573.txt, hdfs-3573.txt, hdfs-3573.txt > > > Currently, the JournalManagers are instantiated before the NamespaceInfo is > loaded from local storage directories. This is problematic since the JM may > want to verify that the storage info associated with the journal matches the > NN which is starting up (eg to prevent an operator accidentally configuring > two clusters against the same remote journal storage). This JIRA rejiggers > the initialization sequence so that the JMs receive NamespaceInfo as a > constructor argument. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HDFS-3789) JournalManager#format() should be able to throw IOException
[ https://issues.apache.org/jira/browse/HDFS-3789?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13487725#comment-13487725 ] Hudson commented on HDFS-3789: -- Integrated in Hadoop-Hdfs-trunk #1212 (See [https://builds.apache.org/job/Hadoop-Hdfs-trunk/1212/]) Moved HDFS-3789 entry in CHANGES.txt from trunk to 2.0.3-alpha section (Revision 1403765) Result = FAILURE umamahesh : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1403765 Files : * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt > JournalManager#format() should be able to throw IOException > --- > > Key: HDFS-3789 > URL: https://issues.apache.org/jira/browse/HDFS-3789 > Project: Hadoop HDFS > Issue Type: Sub-task > Components: ha, name-node >Affects Versions: 3.0.0 >Reporter: Ivan Kelly >Assignee: Ivan Kelly > Fix For: 3.0.0, 2.0.3-alpha > > Attachments: 0003-HDFS-3789-for-branch-2.patch, HDFS-3789.diff > > > Currently JournalManager#format cannot throw any exception. As format can > fail, we should be able to propogate this failure upwards. Otherwise, format > will fail silently, and the admin will start using the cluster with a > failed/unusable journal manager. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HDFS-4129) Add utility methods to dump NameNode in memory tree for testing
[ https://issues.apache.org/jira/browse/HDFS-4129?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13487722#comment-13487722 ] Hudson commented on HDFS-4129: -- Integrated in Hadoop-Hdfs-trunk #1212 (See [https://builds.apache.org/job/Hadoop-Hdfs-trunk/1212/]) HDFS-4129. Add utility methods to dump NameNode in memory tree for testing. Contributed by Tsz Wo (Nicholas), SZE. (Revision 1403956) Result = FAILURE suresh : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1403956 Files : * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSNamesystem.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/INode.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/INodeDirectory.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestFSDirectory.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestFsLimits.java > Add utility methods to dump NameNode in memory tree for testing > --- > > Key: HDFS-4129 > URL: https://issues.apache.org/jira/browse/HDFS-4129 > Project: Hadoop HDFS > Issue Type: Test > Components: name-node >Reporter: Tsz Wo (Nicholas), SZE >Assignee: Tsz Wo (Nicholas), SZE >Priority: Minor > Fix For: 3.0.0 > > Attachments: h4129_20121029b.patch, h4129_20121029.patch, > h4129_20121030.patch > > > The output of the utility methods looks like below. > {noformat} > \- foo (INodeDirectory) > \- sub1 (INodeDirectory) > +- file1 (INodeFile) > +- file2 (INodeFile) > +- sub11 (INodeDirectory) > \- file3 (INodeFile) > \- z_file4 (INodeFile) > {noformat} -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HDFS-3695) Genericize format() to non-file JournalManagers
[ https://issues.apache.org/jira/browse/HDFS-3695?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13487672#comment-13487672 ] Hudson commented on HDFS-3695: -- Integrated in Hadoop-Yarn-trunk #22 (See [https://builds.apache.org/job/Hadoop-Yarn-trunk/22/]) Moved HDFS-3695 entry in CHANGES.txt from trunk to 2.0.3-alpha section (Revision 1403748) Result = SUCCESS umamahesh : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1403748 Files : * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt > Genericize format() to non-file JournalManagers > --- > > Key: HDFS-3695 > URL: https://issues.apache.org/jira/browse/HDFS-3695 > Project: Hadoop HDFS > Issue Type: Sub-task > Components: ha, name-node >Affects Versions: QuorumJournalManager (HDFS-3077) >Reporter: Todd Lipcon >Assignee: Todd Lipcon > Fix For: 3.0.0, 2.0.3-alpha > > Attachments: 0002-HDFS-3695-for-branch-2.patch, hdfs-3695.txt, > hdfs-3695.txt, hdfs-3695.txt > > > Currently, the "namenode -format" and "namenode -initializeSharedEdits" > commands do not understand how to do anything with non-file-based shared > storage. This affects both BookKeeperJournalManager and QuorumJournalManager. > This JIRA is to plumb through the formatting of edits directories using > pluggable journal manager implementations so that no separate step needs to > be taken to format them -- the same commands will work for NFS-based storage > or one of the alternate implementations. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HDFS-3809) Make BKJM use protobufs for all serialization with ZK
[ https://issues.apache.org/jira/browse/HDFS-3809?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13487673#comment-13487673 ] Hudson commented on HDFS-3809: -- Integrated in Hadoop-Yarn-trunk #22 (See [https://builds.apache.org/job/Hadoop-Yarn-trunk/22/]) Moved HDFS-3809 entry in CHANGES.txt from trunk to 2.0.3-alpha section (Revision 1403769) Result = SUCCESS umamahesh : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1403769 Files : * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt > Make BKJM use protobufs for all serialization with ZK > - > > Key: HDFS-3809 > URL: https://issues.apache.org/jira/browse/HDFS-3809 > Project: Hadoop HDFS > Issue Type: Sub-task > Components: name-node >Affects Versions: 2.0.0-alpha, 3.0.0 >Reporter: Ivan Kelly >Assignee: Ivan Kelly > Fix For: 3.0.0, 2.0.3-alpha > > Attachments: > 0001-HDFS-3809.-Make-BKJM-use-protobufs-for-all-serializa.patch, > 0001-HDFS-3809.-Make-BKJM-use-protobufs-for-all-serializa.patch, > 0004-HDFS-3809-for-branch-2.patch, HDFS-3809.diff, HDFS-3809.diff, > HDFS-3809.diff > > > HDFS uses protobufs for serialization in many places. Protobufs allow fields > to be added without breaking bc or requiring new parsing code to be written. > For this reason, we should use them in BKJM also. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HDFS-3916) libwebhdfs (C client) code cleanups
[ https://issues.apache.org/jira/browse/HDFS-3916?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13487671#comment-13487671 ] Hudson commented on HDFS-3916: -- Integrated in Hadoop-Yarn-trunk #22 (See [https://builds.apache.org/job/Hadoop-Yarn-trunk/22/]) HDFS-3916. libwebhdfs testing code cleanup. Contributed by Jing Zhao. (Revision 1403922) Result = SUCCESS suresh : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1403922 Files : * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/contrib/libwebhdfs/src/test_libwebhdfs_multi_write.c * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/contrib/libwebhdfs/src/test_libwebhdfs_ops.c * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/contrib/libwebhdfs/src/test_libwebhdfs_read.c * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/contrib/libwebhdfs/src/test_libwebhdfs_threaded.c * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/contrib/libwebhdfs/src/test_libwebhdfs_write.c * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/contrib/libwebhdfs/src/test_read_bm.c * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/native/libhdfs/native_mini_dfs.c * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/native/libhdfs/native_mini_dfs.h > libwebhdfs (C client) code cleanups > --- > > Key: HDFS-3916 > URL: https://issues.apache.org/jira/browse/HDFS-3916 > Project: Hadoop HDFS > Issue Type: Bug > Components: webhdfs >Affects Versions: 2.0.2-alpha >Reporter: Colin Patrick McCabe >Assignee: Colin Patrick McCabe > Fix For: 2.0.3-alpha > > Attachments: 0002-fix.patch, HDFS-3916.003.patch, > HDFS-3916.004.patch, HDFS-3916.005.patch, HDFS-3916.006.patch > > > Code cleanups in libwebhdfs. > * don't duplicate exception.c, exception.h, expect.h, jni_helper.c. We have > one copy of these files; we don't need 2. > * remember to set errno in all public library functions (this is part of the > API) > * fix undefined symbols (if a function is not implemented, it should return > ENOTSUP, but still exist) > * don't expose private data structures in the (end-user visible) public > headers > * can't re-use hdfsBuilder as hdfsFS, because the strings in hdfsBuilder are > not dynamically allocated. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HDFS-3573) Supply NamespaceInfo when instantiating JournalManagers
[ https://issues.apache.org/jira/browse/HDFS-3573?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13487668#comment-13487668 ] Hudson commented on HDFS-3573: -- Integrated in Hadoop-Yarn-trunk #22 (See [https://builds.apache.org/job/Hadoop-Yarn-trunk/22/]) Moved HDFS-3573 entry in CHANGES.txt from trunk to 2.0.3-alpha section (Revision 1403740) Result = SUCCESS umamahesh : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1403740 Files : * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt > Supply NamespaceInfo when instantiating JournalManagers > --- > > Key: HDFS-3573 > URL: https://issues.apache.org/jira/browse/HDFS-3573 > Project: Hadoop HDFS > Issue Type: Sub-task > Components: name-node >Affects Versions: 3.0.0 >Reporter: Todd Lipcon >Assignee: Todd Lipcon >Priority: Minor > Fix For: 3.0.0, 2.0.3-alpha > > Attachments: 0001-HDFS-3573-for-branch-2.patch, hdfs-3573.txt, > hdfs-3573.txt, hdfs-3573.txt, hdfs-3573.txt > > > Currently, the JournalManagers are instantiated before the NamespaceInfo is > loaded from local storage directories. This is problematic since the JM may > want to verify that the storage info associated with the journal matches the > NN which is starting up (eg to prevent an operator accidentally configuring > two clusters against the same remote journal storage). This JIRA rejiggers > the initialization sequence so that the JMs receive NamespaceInfo as a > constructor argument. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HDFS-3789) JournalManager#format() should be able to throw IOException
[ https://issues.apache.org/jira/browse/HDFS-3789?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13487669#comment-13487669 ] Hudson commented on HDFS-3789: -- Integrated in Hadoop-Yarn-trunk #22 (See [https://builds.apache.org/job/Hadoop-Yarn-trunk/22/]) Moved HDFS-3789 entry in CHANGES.txt from trunk to 2.0.3-alpha section (Revision 1403765) Result = SUCCESS umamahesh : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1403765 Files : * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt > JournalManager#format() should be able to throw IOException > --- > > Key: HDFS-3789 > URL: https://issues.apache.org/jira/browse/HDFS-3789 > Project: Hadoop HDFS > Issue Type: Sub-task > Components: ha, name-node >Affects Versions: 3.0.0 >Reporter: Ivan Kelly >Assignee: Ivan Kelly > Fix For: 3.0.0, 2.0.3-alpha > > Attachments: 0003-HDFS-3789-for-branch-2.patch, HDFS-3789.diff > > > Currently JournalManager#format cannot throw any exception. As format can > fail, we should be able to propogate this failure upwards. Otherwise, format > will fail silently, and the admin will start using the cluster with a > failed/unusable journal manager. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HDFS-4129) Add utility methods to dump NameNode in memory tree for testing
[ https://issues.apache.org/jira/browse/HDFS-4129?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13487666#comment-13487666 ] Hudson commented on HDFS-4129: -- Integrated in Hadoop-Yarn-trunk #22 (See [https://builds.apache.org/job/Hadoop-Yarn-trunk/22/]) HDFS-4129. Add utility methods to dump NameNode in memory tree for testing. Contributed by Tsz Wo (Nicholas), SZE. (Revision 1403956) Result = SUCCESS suresh : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1403956 Files : * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSNamesystem.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/INode.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/INodeDirectory.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestFSDirectory.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestFsLimits.java > Add utility methods to dump NameNode in memory tree for testing > --- > > Key: HDFS-4129 > URL: https://issues.apache.org/jira/browse/HDFS-4129 > Project: Hadoop HDFS > Issue Type: Test > Components: name-node >Reporter: Tsz Wo (Nicholas), SZE >Assignee: Tsz Wo (Nicholas), SZE >Priority: Minor > Fix For: 3.0.0 > > Attachments: h4129_20121029b.patch, h4129_20121029.patch, > h4129_20121030.patch > > > The output of the utility methods looks like below. > {noformat} > \- foo (INodeDirectory) > \- sub1 (INodeDirectory) > +- file1 (INodeFile) > +- file2 (INodeFile) > +- sub11 (INodeDirectory) > \- file3 (INodeFile) > \- z_file4 (INodeFile) > {noformat} -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HDFS-4047) BPServiceActor has nested shouldRun loops
[ https://issues.apache.org/jira/browse/HDFS-4047?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13487614#comment-13487614 ] Hadoop QA commented on HDFS-4047: - {color:red}-1 overall{color}. Here are the results of testing the latest attachment http://issues.apache.org/jira/secure/attachment/12551496/HADOOP-4047.patch against trunk revision . {color:green}+1 @author{color}. The patch does not contain any @author tags. {color:red}-1 tests included{color}. The patch doesn't appear to include any new or modified tests. Please justify why no new tests are needed for this patch. Also please list what manual steps were performed to verify this patch. {color:green}+1 javac{color}. The applied patch does not increase the total number of javac compiler warnings. {color:green}+1 javadoc{color}. The javadoc tool did not generate any warning messages. {color:green}+1 eclipse:eclipse{color}. The patch built with eclipse:eclipse. {color:green}+1 findbugs{color}. The patch does not introduce any new Findbugs (version 1.3.9) warnings. {color:green}+1 release audit{color}. The applied patch does not increase the total number of release audit warnings. {color:green}+1 core tests{color}. The patch passed unit tests in hadoop-hdfs-project/hadoop-hdfs. {color:green}+1 contrib tests{color}. The patch passed contrib unit tests. Test results: https://builds.apache.org/job/PreCommit-HDFS-Build/3432//testReport/ Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/3432//console This message is automatically generated. > BPServiceActor has nested shouldRun loops > - > > Key: HDFS-4047 > URL: https://issues.apache.org/jira/browse/HDFS-4047 > Project: Hadoop HDFS > Issue Type: Bug > Components: name-node >Affects Versions: 2.0.0-alpha >Reporter: Eli Collins >Priority: Minor > Attachments: HADOOP-4047.patch > > > BPServiceActor#run and offerService booth have while shouldRun loops. We only > need the outer one, ie we can hoist the info log from offerService out to run > and remove the while loop. > {code} > BPServiceActor#run: > while (shouldRun()) { > try { > offerService(); > } catch (Exception ex) { > ... > offerService: > while (shouldRun()) { > try { > {code} -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira