[jira] [Updated] (HDFS-6346) Optimize OP_SET_XATTRS by persisting single Xattr entry per setXattr/removeXattr api call

2014-05-13 Thread Yi Liu (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6346?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yi Liu updated HDFS-6346:
-

Attachment: HDFS-6346.1.patch

Thanks Uma for the review.

{quote}
Do you think, we need warn/log to the user that xattr attempting to remove does 
not exist?
{quote}
make sense, let's add a log when removing nonexistent xattr.

{quote}
Pls remove empty line between
{quote}
OK

{quote}
When existing and updated xattrs equal then we need not even call 
updateINodeXAttr. So code can be changed to 
{code}
if (existingXAttrs.size() != newXAttrs.size()) {
  XAttrStorage.updateINodeXAttrs(inode, newXAttrs, snapshotId);
  return xAttr;
}
return null; 
{code}
{quote}
Right, I will update it

> Optimize OP_SET_XATTRS by persisting single Xattr entry per 
> setXattr/removeXattr api call
> -
>
> Key: HDFS-6346
> URL: https://issues.apache.org/jira/browse/HDFS-6346
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: namenode
>Reporter: Uma Maheswara Rao G
>Assignee: Yi Liu
> Attachments: HDFS-6346.1.patch, HDFS-6346.2.patch, HDFS-6346.patch, 
> editsStored
>
>
> When a new xattrs set on an Inode, it may add this with OP_SET_XATTRS and 
> let's say [user.name1:value1]
> On a next call if we set another xattrs, then it may store along with older 
> existing xattrs again. It may be like [user.name1:value1, user.name2:value2]
> So, on adding more xattrs on same Inode, that list may grow and we keep store 
> the entries of already existing name, value fairs.
> Right now we defaulted the max Xattrs on an Inode to 32 and configured. If 
> user modified it to much larger value and start setting xattrs, then edits 
> loading may create issue like my above example.
> But I didn't refer any usecase of having large number of xattrs, this is just 
> the assumption to consider a case. My biggest doubt is whether we will have 
> such real usecases to have huge xattrs on a single INode.
> So, here is a thought on having OP_SET_XATTR for each setXAttr operation to 
> be logged, When we replay them we need to consolidate. This is some initial 
> thought we can think more if others also feel we need to consider this case 
> to handle.
> Otherwise we endup storing Xattrs entries in editlog file as n(n+1)/2 where n 
> is number xattrs for a file/dir. This may be issue only when we have large 
> number configured max xattrs for inode.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6381) Fix a typo in INodeReference.java

2014-05-13 Thread Jing Zhao (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6381?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jing Zhao updated HDFS-6381:


   Resolution: Fixed
Fix Version/s: 2.5.0
 Hadoop Flags: Reviewed
   Status: Resolved  (was: Patch Available)

I've committed this to trunk and branch-2. Thanks for the fix [~decster]!

> Fix a typo in INodeReference.java
> -
>
> Key: HDFS-6381
> URL: https://issues.apache.org/jira/browse/HDFS-6381
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: documentation
>Reporter: Binglin Chang
>Assignee: Binglin Chang
>Priority: Trivial
> Fix For: 2.5.0
>
> Attachments: HDFS-6381.v1.patch
>
>
> hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/INodeReference.java
> {code}
>   * For example,
> - * (1) Support we have /abc/foo, say the inode of foo is 
> inode(id=1000,name=foo)
> + * (1) Suppose we have /abc/foo, say the inode of foo is 
> inode(id=1000,name=foo)
> {code}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6283) Write end user documentation for xattrs.

2014-05-13 Thread Uma Maheswara Rao G (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6283?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13997300#comment-13997300
 ] 

Uma Maheswara Rao G commented on HDFS-6283:
---

{quote}
extended attributes are not interpreted by the system and are instead used by 
applications to store additional information about an inode.
{quote}
I saw this line. This is saying Xattrs can not be interpreted by System. But 
the xattrs with system/security name will be used only by system. So, we should 
not generically say that Xattrs can not be interpreted by system? Please 
correct me I interpreted this line wrongly :-)
Let me know if we need to change this, then I wil file a follow-up JIRA as it 
is already committed.


> Write end user documentation for xattrs.
> 
>
> Key: HDFS-6283
> URL: https://issues.apache.org/jira/browse/HDFS-6283
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: documentation
>Reporter: Chris Nauroth
>Assignee: Andrew Wang
> Fix For: HDFS XAttrs (HDFS-2006)
>
> Attachments: hdfs-6283-1.patch, hdfs-6283-2.patch, hdfs-6283-3.patch
>
>
> Update the File System Shell documentation to cover the new getfattr and 
> setfattr commands.  If warranted, consider adding a separate dedicated page 
> for fuller discussion of the xattrs model and how the feature works in more 
> detail.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6293) Issues with OIV processing PB-based fsimages

2014-05-13 Thread Kihwal Lee (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6293?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kihwal Lee updated HDFS-6293:
-

   Resolution: Fixed
Fix Version/s: 2.5.0
   3.0.0
 Release Note: Set "dfs.namenode.legacy-oiv-image.dir" to an appropriate 
directory to make standby name node or secondary name node save its file system 
state in the old fsimage format during checkpointing. This image can be used 
for offline analysis using the OfflineImageViewer.  Use the "hdfs oiv_legacy" 
command to process the old fsimage format.
 Hadoop Flags: Reviewed
   Status: Resolved  (was: Patch Available)

I've committed this to trunk and branch-2.  Thanks for discussions, code and 
reviews!

> Issues with OIV processing PB-based fsimages
> 
>
> Key: HDFS-6293
> URL: https://issues.apache.org/jira/browse/HDFS-6293
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 2.4.0
>Reporter: Kihwal Lee
>Assignee: Kihwal Lee
>Priority: Blocker
> Fix For: 3.0.0, 2.5.0
>
> Attachments: HDFS-6293.000.patch, HDFS-6293.001.patch, 
> HDFS-6293.002-save-deprecated-fsimage.patch, HDFS-6293.branch-2.patch, 
> HDFS-6293.trunk.patch, HDFS-6293.trunk.patch, HDFS-6293.v2.branch-2.patch, 
> HDFS-6293.v2.trunk.patch, HDFS-6293_sbn_ckpt_retention.patch, 
> HDFS-6293_sbn_ckpt_retention_oiv_legacy.patch, Heap Histogram.html
>
>
> There are issues with OIV when processing fsimages in protobuf. 
> Due to the internal layout changes introduced by the protobuf-based fsimage, 
> OIV consumes excessive amount of memory.  We have tested with a fsimage with 
> about 140M files/directories. The peak heap usage when processing this image 
> in pre-protobuf (i.e. pre-2.4.0) format was about 350MB.  After converting 
> the image to the protobuf format on 2.4.0, OIV would OOM even with 80GB of 
> heap (max new size was 1GB).  It should be possible to process any image with 
> the default heap size of 1.5GB.
> Another issue is the complete change of format/content in OIV's XML output.  
> I also noticed that the secret manager section has no tokens while there were 
> unexpired tokens in the original image (pre-2.4.0).  I did not check whether 
> they were also missing in the new pb fsimage.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (HDFS-6394) HDFS encryption documentation

2014-05-13 Thread Alejandro Abdelnur (JIRA)
Alejandro Abdelnur created HDFS-6394:


 Summary: HDFS encryption documentation
 Key: HDFS-6394
 URL: https://issues.apache.org/jira/browse/HDFS-6394
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: namenode, security
Reporter: Alejandro Abdelnur
Assignee: Charles Lamb


Documentation for HDFS encryption behavior and configuration




--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-2006) ability to support storing extended attributes per file

2014-05-13 Thread Uma Maheswara Rao G (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2006?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13997264#comment-13997264
 ] 

Uma Maheswara Rao G commented on HDFS-2006:
---

Thanks Nicholas, we will update the details in design doc about the size limit.
{code}
Also, I think we should have different limits for different XAttr namespaces 
(user, trusted, security and system) since we may want to put a smaller limit 
for the user visible namespaces. How does it sound?
{code}

This make sense to me. As in the initial thinking, we will have configuration 
parameter for controlling the size. Now as you pointed, if at all Kernel wants 
use it with a little larger value for some  purpose later, we need to modify 
the single parameter right now, from then it allows users also to write larger 
values. Storing larger values in editlog is not good with any namespace, but it 
will up to kernel implementation decision later.

I am +1 on this consideration.
Let me file a JIRA for it and discuss more about this point if others have any 
concerns.


> ability to support storing extended attributes per file
> ---
>
> Key: HDFS-2006
> URL: https://issues.apache.org/jira/browse/HDFS-2006
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: namenode
>Affects Versions: HDFS XAttrs (HDFS-2006)
>Reporter: dhruba borthakur
>Assignee: Yi Liu
> Attachments: HDFS-XAttrs-Design-1.pdf, HDFS-XAttrs-Design-2.pdf, 
> HDFS-XAttrs-Design-3.pdf, Test-Plan-for-Extended-Attributes-1.pdf, 
> xattrs.1.patch, xattrs.patch
>
>
> It would be nice if HDFS provides a feature to store extended attributes for 
> files, similar to the one described here: 
> http://en.wikipedia.org/wiki/Extended_file_attributes. 
> The challenge is that it has to be done in such a way that a site not using 
> this feature does not waste precious memory resources in the namenode.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6373) Remove support for extended attributes on symlinks

2014-05-13 Thread Andrew Wang (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6373?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13996856#comment-13996856
 ] 

Andrew Wang commented on HDFS-6373:
---

That stacktrace is probably just from how symlink resolution works. The client 
code catches the UnresolvedLinkException, attempts to resolve it, then retries 
with the new path. Maybe we should squish the trace since this is expected 
behavior.

> Remove support for extended attributes on symlinks
> --
>
> Key: HDFS-6373
> URL: https://issues.apache.org/jira/browse/HDFS-6373
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: namenode
>Reporter: Andrew Wang
>Assignee: Charles Lamb
>
> Looking in the Linux source code, we see the following:
> http://lxr.linux.no/linux+v3.14.3/fs/xattr.c
> {code}
>   60/*
>   61 * In the user.* namespace, only regular files and directories 
> can have
>   62 * extended attributes. For sticky directories, only the owner and
>   63 * privileged users can write attributes.
>   64 */
> {code}
> We should consider removing {{XAttrFeature}} from {{INodeSymlink}}.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6293) Issues with OIV processing PB-based fsimages

2014-05-13 Thread Kihwal Lee (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6293?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13996968#comment-13996968
 ] 

Kihwal Lee commented on HDFS-6293:
--

Thanks for the review, Suresh. I will commit this after precomit comes back.

> Issues with OIV processing PB-based fsimages
> 
>
> Key: HDFS-6293
> URL: https://issues.apache.org/jira/browse/HDFS-6293
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 2.4.0
>Reporter: Kihwal Lee
>Assignee: Kihwal Lee
>Priority: Blocker
> Attachments: HDFS-6293.000.patch, HDFS-6293.001.patch, 
> HDFS-6293.002-save-deprecated-fsimage.patch, HDFS-6293.branch-2.patch, 
> HDFS-6293.trunk.patch, HDFS-6293.trunk.patch, HDFS-6293.v2.branch-2.patch, 
> HDFS-6293.v2.trunk.patch, HDFS-6293_sbn_ckpt_retention.patch, 
> HDFS-6293_sbn_ckpt_retention_oiv_legacy.patch, Heap Histogram.html
>
>
> There are issues with OIV when processing fsimages in protobuf. 
> Due to the internal layout changes introduced by the protobuf-based fsimage, 
> OIV consumes excessive amount of memory.  We have tested with a fsimage with 
> about 140M files/directories. The peak heap usage when processing this image 
> in pre-protobuf (i.e. pre-2.4.0) format was about 350MB.  After converting 
> the image to the protobuf format on 2.4.0, OIV would OOM even with 80GB of 
> heap (max new size was 1GB).  It should be possible to process any image with 
> the default heap size of 1.5GB.
> Another issue is the complete change of format/content in OIV's XML output.  
> I also noticed that the secret manager section has no tokens while there were 
> unexpired tokens in the original image (pre-2.4.0).  I did not check whether 
> they were also missing in the new pb fsimage.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-2006) ability to support storing extended attributes per file

2014-05-13 Thread Yi Liu (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2006?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13997288#comment-13997288
 ] 

Yi Liu commented on HDFS-2006:
--

Thanks Nicholas, I agree we have different limits for different XAttr 
namespaces.

> ability to support storing extended attributes per file
> ---
>
> Key: HDFS-2006
> URL: https://issues.apache.org/jira/browse/HDFS-2006
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: namenode
>Affects Versions: HDFS XAttrs (HDFS-2006)
>Reporter: dhruba borthakur
>Assignee: Yi Liu
> Attachments: HDFS-XAttrs-Design-1.pdf, HDFS-XAttrs-Design-2.pdf, 
> HDFS-XAttrs-Design-3.pdf, Test-Plan-for-Extended-Attributes-1.pdf, 
> xattrs.1.patch, xattrs.patch
>
>
> It would be nice if HDFS provides a feature to store extended attributes for 
> files, similar to the one described here: 
> http://en.wikipedia.org/wiki/Extended_file_attributes. 
> The challenge is that it has to be done in such a way that a site not using 
> this feature does not waste precious memory resources in the namenode.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6366) FsImage loading failed with RemoveXattr op

2014-05-13 Thread Yi Liu (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6366?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13994182#comment-13994182
 ] 

Yi Liu commented on HDFS-6366:
--

+1 for the patch. Thanks Uma.

> FsImage loading failed with RemoveXattr op
> --
>
> Key: HDFS-6366
> URL: https://issues.apache.org/jira/browse/HDFS-6366
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: namenode
>Reporter: Uma Maheswara Rao G
>Assignee: Uma Maheswara Rao G
> Fix For: HDFS XAttrs (HDFS-2006)
>
> Attachments: HDFS-6366.patch
>
>
> Should break after the removeXattr op loading.
> Otherwise fsimage loading will fail:
> java.io.IOException: Invalid operation read OP_REMOVE_XATTR
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSEditLogLoader.applyEditLogOp(FSEditLogLoader.java:816)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSEditLogLoader.loadEditRecords(FSEditLogLoader.java:227)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSEditLogLoader.loadFSEdits(FSEditLogLoader.java:136)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.loadEdits(FSImage.java:805)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.loadFSImage(FSImage.java:665)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:272)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFSImage(FSNamesystem.java:902)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFromDisk(FSNamesystem.java:651)
>   at 
> org.apache.hadoop.hdfs.server.namenode.NameNode.loadNamesystem(NameNode.java:479)
>   at 
> org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:535)
>   at 
> org.apache.hadoop.hdfs.server.namenode.NameNode.(NameNode.java:694)
>   at 
> org.apache.hadoop.hdfs.server.namenode.NameNode.(NameNode.java:679)
>   at 
> org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1332)
>   at 
> org.apache.hadoop.hdfs.MiniDFSCluster.createNameNode(MiniDFSCluster.java:973)
>   at 
> org.apache.hadoop.hdfs.MiniDFSCluster.createNameNodesAndSetConf(MiniDFSCluster.java:854)
>   at 
> org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:700)
>   at org.apache.hadoop.hdfs.MiniDFSCluster.(MiniDFSCluster.java:373)
>   at 
> org.apache.hadoop.hdfs.MiniDFSCluster$Builder.build(MiniDFSCluster.java:354)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSXAttrBaseTest.initCluster(FSXAttrBaseTest.java:400)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSXAttrBaseTest.restart(FSXAttrBaseTest.java:418)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSXAttrBaseTest.testCreateXAttr(FSXAttrBaseTest.java:121)
>   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>   at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>   at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (HDFS-6367) Domain>#parse in EnumSetParam fails for parmater containing more than one enum.

2014-05-13 Thread Yi Liu (JIRA)
Yi Liu created HDFS-6367:


 Summary: Domain>#parse in EnumSetParam fails for 
parmater containing more than one enum.
 Key: HDFS-6367
 URL: https://issues.apache.org/jira/browse/HDFS-6367
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: webhdfs
Affects Versions: 3.0.0
Reporter: Yi Liu
 Fix For: 3.0.0
 Attachments: HDFS-6367.patch

Fails because additional "," 

java.lang.IllegalArgumentException: No enum const class 
org.apache.hadoop.fs.Options$Rename.,OVERWRITE
at java.lang.Enum.valueOf(Enum.java:196)
at 
org.apache.hadoop.hdfs.web.resources.EnumSetParam$Domain.parse(EnumSetParam.java:85)
at 
org.apache.hadoop.hdfs.web.resources.RenameOptionSetParam.(RenameOptionSetParam.java:45)
at 
org.apache.hadoop.hdfs.web.resources.TestParam.testRenameOptionSetParam(TestParam.java:355)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at 
org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:47)
at 
org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at 
org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:44)
at 
org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:271)
at 
org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:70)
at 
org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
at org.junit.runners.ParentRunner$3.run(ParentRunner.java:238)
at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:63)
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:236)
at org.junit.runners.ParentRunner.access$000(ParentRunner.java:53)
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:229)
at org.junit.runners.ParentRunner.run(ParentRunner.java:309)
at 
org.eclipse.jdt.internal.junit4.runner.JUnit4TestReference.run(JUnit4TestReference.java:50)
at 
org.eclipse.jdt.internal.junit.runner.TestExecution.run(TestExecution.java:38)
at 
org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:467)
at 
org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:683)
at 
org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.run(RemoteTestRunner.java:390)
at 
org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.main(RemoteTestRunner.java:197)





--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6376) Distcp data between two HA clusters requires another configuration

2014-05-13 Thread Dave Marion (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6376?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Dave Marion updated HDFS-6376:
--

Fix Version/s: 2.4.1

> Distcp data between two HA clusters requires another configuration
> --
>
> Key: HDFS-6376
> URL: https://issues.apache.org/jira/browse/HDFS-6376
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: datanode, federation, hdfs-client
>Affects Versions: 2.3.0, 2.4.0
> Environment: CDH 5.0 (Apache 2.3.0 + some patches)
>Reporter: Dave Marion
> Fix For: 2.4.1
>
> Attachments: HDFS-6376-2.patch, HDFS-6376-patch-1.patch
>
>
> User has to create a third set of configuration files for distcp when 
> transferring data between two HA clusters.
> Consider the scenario in [1]. You cannot put all of the required properties 
> in core-site.xml and hdfs-site.xml for the client to resolve the location of 
> both active namenodes. If you do, then the datanodes from cluster A may join 
> cluster B. I can not find a configuration option that tells the datanodes to 
> federate blocks for only one of the clusters in the configuration.
> [1] 
> http://mail-archives.apache.org/mod_mbox/hadoop-user/201404.mbox/%3CBAY172-W2133964E0C283968C161DD1520%40phx.gbl%3E



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6383) Upgrade S3n s3.fs.buffer.dir to suppoer multi directories

2014-05-13 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6383?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13997184#comment-13997184
 ] 

Hadoop QA commented on HDFS-6383:
-

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12644622/HDFS-6383.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:red}-1 tests included{color}.  The patch doesn't appear to include 
any new or modified tests.
Please justify why no new tests are needed for this 
patch.
Also please list what manual steps were performed to 
verify this patch.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  There were no new javadoc warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:red}-1 core tests{color}.  The following test timeouts occurred in 
hadoop-common-project/hadoop-common:

org.apache.hadoop.http.TestHttpServer

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-HDFS-Build/6897//testReport/
Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/6897//console

This message is automatically generated.

> Upgrade S3n s3.fs.buffer.dir to suppoer multi directories
> -
>
> Key: HDFS-6383
> URL: https://issues.apache.org/jira/browse/HDFS-6383
> Project: Hadoop HDFS
>  Issue Type: Improvement
>Affects Versions: 2.4.0
>Reporter: Ted Malaska
>Assignee: Ted Malaska
>Priority: Minor
> Attachments: HDFS-6383.patch
>
>
> s3.fs.buffer.dir defines the tmp folder where files will be written to before 
> getting sent to S3.  Right now this is limited to a single folder which 
> causes to major issues.
> 1. You need a drive with enough space to store all the tmp files at once
> 2. You are limited to the IO speeds of a single drive
> This solution will resolve both and has been tested to increase the S3 write 
> speed by 2.5x with 10 mappers on hs1.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Resolved] (HDFS-6283) Write end user documentation for xattrs.

2014-05-13 Thread Andrew Wang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6283?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Andrew Wang resolved HDFS-6283.
---

   Resolution: Fixed
Fix Version/s: HDFS XAttrs (HDFS-2006)

Committed to branch.

> Write end user documentation for xattrs.
> 
>
> Key: HDFS-6283
> URL: https://issues.apache.org/jira/browse/HDFS-6283
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: documentation
>Reporter: Chris Nauroth
>Assignee: Andrew Wang
> Fix For: HDFS XAttrs (HDFS-2006)
>
> Attachments: hdfs-6283-1.patch, hdfs-6283-2.patch, hdfs-6283-3.patch
>
>
> Update the File System Shell documentation to cover the new getfattr and 
> setfattr commands.  If warranted, consider adding a separate dedicated page 
> for fuller discussion of the xattrs model and how the feature works in more 
> detail.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-2006) ability to support storing extended attributes per file

2014-05-13 Thread Tsz Wo Nicholas Sze (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2006?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13997202#comment-13997202
 ] 

Tsz Wo Nicholas Sze commented on HDFS-2006:
---

Thanks [~cnauroth] for showing me the pointer.  I will look at the code in more 
details.

Although large value is not yet supported in the current phase, the "XAttr with 
large value" section in the design doc should mention that if  large value is 
supported, the large values should not be written to edit log.  Otherwise, edit 
log will grow very fast.

Also, I think we should have different limits for different XAttr namespaces 
(user, trusted, security and system) since we may want to put a smaller limit 
for the user visible namespaces.  How does it sound?

> ability to support storing extended attributes per file
> ---
>
> Key: HDFS-2006
> URL: https://issues.apache.org/jira/browse/HDFS-2006
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: namenode
>Affects Versions: HDFS XAttrs (HDFS-2006)
>Reporter: dhruba borthakur
>Assignee: Yi Liu
> Attachments: HDFS-XAttrs-Design-1.pdf, HDFS-XAttrs-Design-2.pdf, 
> HDFS-XAttrs-Design-3.pdf, Test-Plan-for-Extended-Attributes-1.pdf, 
> xattrs.1.patch, xattrs.patch
>
>
> It would be nice if HDFS provides a feature to store extended attributes for 
> files, similar to the one described here: 
> http://en.wikipedia.org/wiki/Extended_file_attributes. 
> The challenge is that it has to be done in such a way that a site not using 
> this feature does not waste precious memory resources in the namenode.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6293) Issues with OIV processing PB-based fsimages

2014-05-13 Thread Kihwal Lee (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6293?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kihwal Lee updated HDFS-6293:
-

Attachment: HDFS-6293.trunk.patch

The patch addresses review comments. I also added a simple check for OIV image 
generation in TestStandbyCheckpoints. Retention enforcement is tested in 
TestCheckpoint. This patch applies to trunk and branch-2.

> Issues with OIV processing PB-based fsimages
> 
>
> Key: HDFS-6293
> URL: https://issues.apache.org/jira/browse/HDFS-6293
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 2.4.0
>Reporter: Kihwal Lee
>Assignee: Kihwal Lee
>Priority: Blocker
> Attachments: HDFS-6293.000.patch, HDFS-6293.001.patch, 
> HDFS-6293.002-save-deprecated-fsimage.patch, HDFS-6293.trunk.patch, 
> HDFS-6293_sbn_ckpt_retention.patch, 
> HDFS-6293_sbn_ckpt_retention_oiv_legacy.patch, Heap Histogram.html
>
>
> There are issues with OIV when processing fsimages in protobuf. 
> Due to the internal layout changes introduced by the protobuf-based fsimage, 
> OIV consumes excessive amount of memory.  We have tested with a fsimage with 
> about 140M files/directories. The peak heap usage when processing this image 
> in pre-protobuf (i.e. pre-2.4.0) format was about 350MB.  After converting 
> the image to the protobuf format on 2.4.0, OIV would OOM even with 80GB of 
> heap (max new size was 1GB).  It should be possible to process any image with 
> the default heap size of 1.5GB.
> Another issue is the complete change of format/content in OIV's XML output.  
> I also noticed that the secret manager section has no tokens while there were 
> unexpired tokens in the original image (pre-2.4.0).  I did not check whether 
> they were also missing in the new pb fsimage.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6383) Upgrade S3n s3.fs.buffer.dir to suppoer multi directories

2014-05-13 Thread David S. Wang (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6383?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13997211#comment-13997211
 ] 

David S. Wang commented on HDFS-6383:
-

Thanks Ted for the patch.

We should probably use the LocalDirAllocator like what s3a uses.  That seems to 
be the proper way to do this in Hadoop.

> Upgrade S3n s3.fs.buffer.dir to suppoer multi directories
> -
>
> Key: HDFS-6383
> URL: https://issues.apache.org/jira/browse/HDFS-6383
> Project: Hadoop HDFS
>  Issue Type: Improvement
>Affects Versions: 2.4.0
>Reporter: Ted Malaska
>Assignee: Ted Malaska
>Priority: Minor
> Attachments: HDFS-6383.patch
>
>
> s3.fs.buffer.dir defines the tmp folder where files will be written to before 
> getting sent to S3.  Right now this is limited to a single folder which 
> causes to major issues.
> 1. You need a drive with enough space to store all the tmp files at once
> 2. You are limited to the IO speeds of a single drive
> This solution will resolve both and has been tested to increase the S3 write 
> speed by 2.5x with 10 mappers on hs1.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-2006) ability to support storing extended attributes per file

2014-05-13 Thread Tsz Wo Nicholas Sze (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2006?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13996507#comment-13996507
 ] 

Tsz Wo Nicholas Sze commented on HDFS-2006:
---

The latest design looks good.

For a XAttr with large value, writing it to editlog may be a problem.  Do we 
plan to have a size limit on the size of XAttr values?

> ability to support storing extended attributes per file
> ---
>
> Key: HDFS-2006
> URL: https://issues.apache.org/jira/browse/HDFS-2006
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: namenode
>Affects Versions: HDFS XAttrs (HDFS-2006)
>Reporter: dhruba borthakur
>Assignee: Yi Liu
> Attachments: HDFS-XAttrs-Design-1.pdf, HDFS-XAttrs-Design-2.pdf, 
> HDFS-XAttrs-Design-3.pdf, Test-Plan-for-Extended-Attributes-1.pdf, 
> xattrs.1.patch, xattrs.patch
>
>
> It would be nice if HDFS provides a feature to store extended attributes for 
> files, similar to the one described here: 
> http://en.wikipedia.org/wiki/Extended_file_attributes. 
> The challenge is that it has to be done in such a way that a site not using 
> this feature does not waste precious memory resources in the namenode.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (HDFS-6396) Remove support for ACL feature from INodeSymlink

2014-05-13 Thread Andrew Wang (JIRA)
Andrew Wang created HDFS-6396:
-

 Summary: Remove support for ACL feature from INodeSymlink
 Key: HDFS-6396
 URL: https://issues.apache.org/jira/browse/HDFS-6396
 Project: Hadoop HDFS
  Issue Type: Improvement
Affects Versions: 2.4.0
Reporter: Andrew Wang
Assignee: Charles Lamb
Priority: Minor


Symlinks cannot have ACLs, but we still have support for the ACL feature in 
INodeSymlink because of class inheritance. Let's remove this support for code 
consistency.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6375) Listing extended attributes with the search permission

2014-05-13 Thread Charles Lamb (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6375?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13997120#comment-13997120
 ] 

Charles Lamb commented on HDFS-6375:


Alternatively, we could just look at the permissions and only return the names 
(or names + values) based on the permissions that the caller has.


> Listing extended attributes with the search permission
> --
>
> Key: HDFS-6375
> URL: https://issues.apache.org/jira/browse/HDFS-6375
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: namenode
>Affects Versions: HDFS XAttrs (HDFS-2006)
>Reporter: Andrew Wang
>Assignee: Charles Lamb
>
> From the attr(5) manpage:
> {noformat}
>Users with search access to a file or directory may retrieve a list  of
>attribute names defined for that file or directory.
> {noformat}
> This is like doing {{getfattr}} without the {{-d}} flag, which we currently 
> don't support.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6376) Distcp data between two HA clusters requires another configuration

2014-05-13 Thread Dave Marion (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6376?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Dave Marion updated HDFS-6376:
--

Component/s: hdfs-client
 datanode

> Distcp data between two HA clusters requires another configuration
> --
>
> Key: HDFS-6376
> URL: https://issues.apache.org/jira/browse/HDFS-6376
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: datanode, federation, hdfs-client
>Affects Versions: 2.3.0, 2.4.0
> Environment: CDH 5.0 (Apache 2.3.0 + some patches)
>Reporter: Dave Marion
> Fix For: 2.4.1
>
> Attachments: HDFS-6376-2.patch, HDFS-6376-patch-1.patch
>
>
> User has to create a third set of configuration files for distcp when 
> transferring data between two HA clusters.
> Consider the scenario in [1]. You cannot put all of the required properties 
> in core-site.xml and hdfs-site.xml for the client to resolve the location of 
> both active namenodes. If you do, then the datanodes from cluster A may join 
> cluster B. I can not find a configuration option that tells the datanodes to 
> federate blocks for only one of the clusters in the configuration.
> [1] 
> http://mail-archives.apache.org/mod_mbox/hadoop-user/201404.mbox/%3CBAY172-W2133964E0C283968C161DD1520%40phx.gbl%3E



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6381) Fix a typo in INodeReference.java

2014-05-13 Thread Chen He (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6381?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chen He updated HDFS-6381:
--

Component/s: documentation

> Fix a typo in INodeReference.java
> -
>
> Key: HDFS-6381
> URL: https://issues.apache.org/jira/browse/HDFS-6381
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: documentation
>Reporter: Binglin Chang
>Assignee: Binglin Chang
>Priority: Trivial
> Attachments: HDFS-6381.v1.patch
>
>
> hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/INodeReference.java
> {code}
>   * For example,
> - * (1) Support we have /abc/foo, say the inode of foo is 
> inode(id=1000,name=foo)
> + * (1) Suppose we have /abc/foo, say the inode of foo is 
> inode(id=1000,name=foo)
> {code}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6293) Issues with OIV processing PB-based fsimages

2014-05-13 Thread Kihwal Lee (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6293?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13996873#comment-13996873
 ] 

Kihwal Lee commented on HDFS-6293:
--

I will fix all warnings except the two coming from FSImage due to the use of 
deprecated old image saving methods. 

> Issues with OIV processing PB-based fsimages
> 
>
> Key: HDFS-6293
> URL: https://issues.apache.org/jira/browse/HDFS-6293
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 2.4.0
>Reporter: Kihwal Lee
>Assignee: Kihwal Lee
>Priority: Blocker
> Attachments: HDFS-6293.000.patch, HDFS-6293.001.patch, 
> HDFS-6293.002-save-deprecated-fsimage.patch, HDFS-6293.branch-2.patch, 
> HDFS-6293.trunk.patch, HDFS-6293.trunk.patch, 
> HDFS-6293_sbn_ckpt_retention.patch, 
> HDFS-6293_sbn_ckpt_retention_oiv_legacy.patch, Heap Histogram.html
>
>
> There are issues with OIV when processing fsimages in protobuf. 
> Due to the internal layout changes introduced by the protobuf-based fsimage, 
> OIV consumes excessive amount of memory.  We have tested with a fsimage with 
> about 140M files/directories. The peak heap usage when processing this image 
> in pre-protobuf (i.e. pre-2.4.0) format was about 350MB.  After converting 
> the image to the protobuf format on 2.4.0, OIV would OOM even with 80GB of 
> heap (max new size was 1GB).  It should be possible to process any image with 
> the default heap size of 1.5GB.
> Another issue is the complete change of format/content in OIV's XML output.  
> I also noticed that the secret manager section has no tokens while there were 
> unexpired tokens in the original image (pre-2.4.0).  I did not check whether 
> they were also missing in the new pb fsimage.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (HDFS-6387) HDFS CLI admin tool for creating & deleting an encryption zone

2014-05-13 Thread Alejandro Abdelnur (JIRA)
Alejandro Abdelnur created HDFS-6387:


 Summary: HDFS CLI admin tool for creating & deleting an encryption 
zone
 Key: HDFS-6387
 URL: https://issues.apache.org/jira/browse/HDFS-6387
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: namenode, security
Reporter: Alejandro Abdelnur
Assignee: Charles Lamb


CLI admin tool to create/delete an encryption zone in HDFS.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6377) Unify xattr name and value limits into a single limit

2014-05-13 Thread Chris Nauroth (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6377?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13997065#comment-13997065
 ] 

Chris Nauroth commented on HDFS-6377:
-

Ah, thanks for the explanation of the {{cacheManager}} bit.  :-)

> Unify xattr name and value limits into a single limit
> -
>
> Key: HDFS-6377
> URL: https://issues.apache.org/jira/browse/HDFS-6377
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: namenode
>Affects Versions: HDFS XAttrs (HDFS-2006)
>Reporter: Andrew Wang
>Assignee: Andrew Wang
> Fix For: HDFS XAttrs (HDFS-2006)
>
> Attachments: hdfs-6377-1.patch, hdfs-6377-2.patch
>
>
> Instead of having separate limits and config options for the size of an 
> xattr's name and value, let's use a single limit.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6373) Remove support for extended attributes on symlinks

2014-05-13 Thread Andrew Wang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6373?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Andrew Wang updated HDFS-6373:
--

   Resolution: Fixed
Fix Version/s: HDFS XAttrs (HDFS-2006)
   Status: Resolved  (was: Patch Available)

Committed to branch, thanks all.

> Remove support for extended attributes on symlinks
> --
>
> Key: HDFS-6373
> URL: https://issues.apache.org/jira/browse/HDFS-6373
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: namenode
>Reporter: Andrew Wang
>Assignee: Charles Lamb
> Fix For: HDFS XAttrs (HDFS-2006)
>
> Attachments: HDFS-6373.1.patch, HDFS-6373.2.patch
>
>
> Looking in the Linux source code, we see the following:
> http://lxr.linux.no/linux+v3.14.3/fs/xattr.c
> {code}
>   60/*
>   61 * In the user.* namespace, only regular files and directories 
> can have
>   62 * extended attributes. For sticky directories, only the owner and
>   63 * privileged users can write attributes.
>   64 */
> {code}
> We should consider removing {{XAttrFeature}} from {{INodeSymlink}}.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-2576) Namenode should have a favored nodes hint to enable clients to have control over block placement.

2014-05-13 Thread samira elhami (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2576?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13996841#comment-13996841
 ] 

samira elhami commented on HDFS-2576:
-

This patch is really helpful. I know it is added to hadoop-2.1.0-Beta. I've 
install this version of hadoop but I don't know how to set the block placement. 
I've searched everywhere. please help me.
Thanks

> Namenode should have a favored nodes hint to enable clients to have control 
> over block placement.
> -
>
> Key: HDFS-2576
> URL: https://issues.apache.org/jira/browse/HDFS-2576
> Project: Hadoop HDFS
>  Issue Type: New Feature
>  Components: hdfs-client, namenode
>Reporter: Pritam Damania
>Assignee: Devaraj Das
> Fix For: 2.1.0-beta
>
> Attachments: hdfs-2576-1.txt, hdfs-2576-trunk-1.patch, 
> hdfs-2576-trunk-2.patch, hdfs-2576-trunk-7.1.patch, hdfs-2576-trunk-7.patch, 
> hdfs-2576-trunk-8.1.patch, hdfs-2576-trunk-8.2.patch, 
> hdfs-2576-trunk-8.3.patch, hdfs-2576-trunk-8.patch
>
>
> Sometimes Clients like HBase are required to dynamically compute the 
> datanodes it wishes to place the blocks for a file for higher level of 
> locality. For this purpose there is a need of a way to give the Namenode a 
> hint in terms of a favoredNodes parameter about the locations where the 
> client wants to put each block. The proposed solution is a favored nodes 
> parameter in the addBlock() method and in the create() file method to enable 
> the clients to give the hints to the NameNode about the locations of each 
> replica of the block. Note that this would be just a hint and finally the 
> NameNode would look at disk usage, datanode load etc. and decide whether it 
> can respect the hints or not.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6293) Issues with OIV processing PB-based fsimages

2014-05-13 Thread Kihwal Lee (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6293?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13997142#comment-13997142
 ] 

Kihwal Lee commented on HDFS-6293:
--

The failed test is unrelated to this JIRA. It will be fixed in HDFS-6257.
The two warnings (deprecation) are expected.

> Issues with OIV processing PB-based fsimages
> 
>
> Key: HDFS-6293
> URL: https://issues.apache.org/jira/browse/HDFS-6293
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 2.4.0
>Reporter: Kihwal Lee
>Assignee: Kihwal Lee
>Priority: Blocker
> Attachments: HDFS-6293.000.patch, HDFS-6293.001.patch, 
> HDFS-6293.002-save-deprecated-fsimage.patch, HDFS-6293.branch-2.patch, 
> HDFS-6293.trunk.patch, HDFS-6293.trunk.patch, HDFS-6293.v2.branch-2.patch, 
> HDFS-6293.v2.trunk.patch, HDFS-6293_sbn_ckpt_retention.patch, 
> HDFS-6293_sbn_ckpt_retention_oiv_legacy.patch, Heap Histogram.html
>
>
> There are issues with OIV when processing fsimages in protobuf. 
> Due to the internal layout changes introduced by the protobuf-based fsimage, 
> OIV consumes excessive amount of memory.  We have tested with a fsimage with 
> about 140M files/directories. The peak heap usage when processing this image 
> in pre-protobuf (i.e. pre-2.4.0) format was about 350MB.  After converting 
> the image to the protobuf format on 2.4.0, OIV would OOM even with 80GB of 
> heap (max new size was 1GB).  It should be possible to process any image with 
> the default heap size of 1.5GB.
> Another issue is the complete change of format/content in OIV's XML output.  
> I also noticed that the secret manager section has no tokens while there were 
> unexpired tokens in the original image (pre-2.4.0).  I did not check whether 
> they were also missing in the new pb fsimage.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6283) Write end user documentation for xattrs.

2014-05-13 Thread Andrew Wang (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6283?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13997053#comment-13997053
 ] 

Andrew Wang commented on HDFS-6283:
---

New patch attached, thanks for the reviews all. I think this addresses all the 
comments except as follows:

- I didn't split out the common changes, seems like we could move this to a 
HADOOP subtask if people prefer, but it seems nicer to do the FsShell updates 
with this same patch as Chris requested. Plus, this way the typo is 
consistently fixed in all three places.
- However, for wording changes, I'd prefer to punt that to another JIRA. I used 
the exactly text from the FsShell help, so we should update all three places 
(FsShell, ExtendedAttributes.apt.vm, FileSystemShell.apt.vm) if we want to 
change the wording anywhere.

> Write end user documentation for xattrs.
> 
>
> Key: HDFS-6283
> URL: https://issues.apache.org/jira/browse/HDFS-6283
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: documentation
>Reporter: Chris Nauroth
>Assignee: Andrew Wang
> Attachments: hdfs-6283-1.patch, hdfs-6283-2.patch
>
>
> Update the File System Shell documentation to cover the new getfattr and 
> setfattr commands.  If warranted, consider adding a separate dedicated page 
> for fuller discussion of the xattrs model and how the feature works in more 
> detail.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (HDFS-6395) Assorted improvements to xattr limit checking

2014-05-13 Thread Andrew Wang (JIRA)
Andrew Wang created HDFS-6395:
-

 Summary: Assorted improvements to xattr limit checking
 Key: HDFS-6395
 URL: https://issues.apache.org/jira/browse/HDFS-6395
 Project: Hadoop HDFS
  Issue Type: Sub-task
Affects Versions: HDFS XAttrs (HDFS-2006)
Reporter: Andrew Wang


It'd be nice to print messages during fsimage and editlog loading if we hit 
either the # of xattrs per inode or the xattr size limits.

We should also consider making the # of xattrs limit only apply to the user 
namespace, or to each namespace separately, to prevent users from locking out 
access to other namespaces.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6377) Unify xattr name and value limits into a single limit

2014-05-13 Thread Andrew Wang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6377?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Andrew Wang updated HDFS-6377:
--

Attachment: hdfs-6377-2.patch

> Unify xattr name and value limits into a single limit
> -
>
> Key: HDFS-6377
> URL: https://issues.apache.org/jira/browse/HDFS-6377
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: namenode
>Affects Versions: HDFS XAttrs (HDFS-2006)
>Reporter: Andrew Wang
>Assignee: Andrew Wang
> Attachments: hdfs-6377-1.patch, hdfs-6377-2.patch
>
>
> Instead of having separate limits and config options for the size of an 
> xattr's name and value, let's use a single limit.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6056) Clean up NFS config settings

2014-05-13 Thread Aaron T. Myers (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6056?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13997056#comment-13997056
 ] 

Aaron T. Myers commented on HDFS-6056:
--

Why not add proper config deprecations of the old config names to the new 
config names so that this won't be an incompatible change? That seems vastly 
preferable to me.

> Clean up NFS config settings
> 
>
> Key: HDFS-6056
> URL: https://issues.apache.org/jira/browse/HDFS-6056
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: nfs
>Affects Versions: 2.3.0
>Reporter: Aaron T. Myers
>Assignee: Brandon Li
> Attachments: HDFS-6056.001.patch, HDFS-6056.002.patch, 
> HDFS-6056.003.patch, HDFS-6056.004.patch
>
>
> As discussed on HDFS-6050, there's a few opportunities to improve the config 
> settings related to NFS. This JIRA is to implement those changes, which 
> include: moving hdfs-nfs related properties into hadoop-hdfs-nfs project, and 
> replacing 'nfs' with 'nfs' in the property names.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6293) Issues with OIV processing PB-based fsimages

2014-05-13 Thread Kihwal Lee (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6293?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kihwal Lee updated HDFS-6293:
-

Status: Patch Available  (was: In Progress)

> Issues with OIV processing PB-based fsimages
> 
>
> Key: HDFS-6293
> URL: https://issues.apache.org/jira/browse/HDFS-6293
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 2.4.0
>Reporter: Kihwal Lee
>Assignee: Kihwal Lee
>Priority: Blocker
> Attachments: HDFS-6293.000.patch, HDFS-6293.001.patch, 
> HDFS-6293.002-save-deprecated-fsimage.patch, HDFS-6293.trunk.patch, 
> HDFS-6293_sbn_ckpt_retention.patch, 
> HDFS-6293_sbn_ckpt_retention_oiv_legacy.patch, Heap Histogram.html
>
>
> There are issues with OIV when processing fsimages in protobuf. 
> Due to the internal layout changes introduced by the protobuf-based fsimage, 
> OIV consumes excessive amount of memory.  We have tested with a fsimage with 
> about 140M files/directories. The peak heap usage when processing this image 
> in pre-protobuf (i.e. pre-2.4.0) format was about 350MB.  After converting 
> the image to the protobuf format on 2.4.0, OIV would OOM even with 80GB of 
> heap (max new size was 1GB).  It should be possible to process any image with 
> the default heap size of 1.5GB.
> Another issue is the complete change of format/content in OIV's XML output.  
> I also noticed that the secret manager section has no tokens while there were 
> unexpired tokens in the original image (pre-2.4.0).  I did not check whether 
> they were also missing in the new pb fsimage.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6293) Issues with OIV processing PB-based fsimages

2014-05-13 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6293?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13997051#comment-13997051
 ] 

Hadoop QA commented on HDFS-6293:
-

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  
http://issues.apache.org/jira/secure/attachment/12644696/HDFS-6293.v2.trunk.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 2 new 
or modified test files.

  {color:red}-1 javac{color}.  The applied patch generated 1277 javac 
compiler warnings (more than the trunk's current 1275 warnings).

{color:green}+1 javadoc{color}.  There were no new javadoc warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:red}-1 core tests{color}.  The patch failed these unit tests in 
hadoop-hdfs-project/hadoop-hdfs:

  org.apache.hadoop.hdfs.server.namenode.TestCacheDirectives

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-HDFS-Build/6896//testReport/
Javac warnings: 
https://builds.apache.org/job/PreCommit-HDFS-Build/6896//artifact/trunk/patchprocess/diffJavacWarnings.txt
Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/6896//console

This message is automatically generated.

> Issues with OIV processing PB-based fsimages
> 
>
> Key: HDFS-6293
> URL: https://issues.apache.org/jira/browse/HDFS-6293
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 2.4.0
>Reporter: Kihwal Lee
>Assignee: Kihwal Lee
>Priority: Blocker
> Attachments: HDFS-6293.000.patch, HDFS-6293.001.patch, 
> HDFS-6293.002-save-deprecated-fsimage.patch, HDFS-6293.branch-2.patch, 
> HDFS-6293.trunk.patch, HDFS-6293.trunk.patch, HDFS-6293.v2.branch-2.patch, 
> HDFS-6293.v2.trunk.patch, HDFS-6293_sbn_ckpt_retention.patch, 
> HDFS-6293_sbn_ckpt_retention_oiv_legacy.patch, Heap Histogram.html
>
>
> There are issues with OIV when processing fsimages in protobuf. 
> Due to the internal layout changes introduced by the protobuf-based fsimage, 
> OIV consumes excessive amount of memory.  We have tested with a fsimage with 
> about 140M files/directories. The peak heap usage when processing this image 
> in pre-protobuf (i.e. pre-2.4.0) format was about 350MB.  After converting 
> the image to the protobuf format on 2.4.0, OIV would OOM even with 80GB of 
> heap (max new size was 1GB).  It should be possible to process any image with 
> the default heap size of 1.5GB.
> Another issue is the complete change of format/content in OIV's XML output.  
> I also noticed that the secret manager section has no tokens while there were 
> unexpired tokens in the original image (pre-2.4.0).  I did not check whether 
> they were also missing in the new pb fsimage.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (HDFS-6385) Show when block deletion will start after NameNode startup in WebUI

2014-05-13 Thread Jing Zhao (JIRA)
Jing Zhao created HDFS-6385:
---

 Summary: Show when block deletion will start after NameNode 
startup in WebUI
 Key: HDFS-6385
 URL: https://issues.apache.org/jira/browse/HDFS-6385
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Jing Zhao


HDFS-6186 provides functionality to delay block deletion for a period of time 
after NameNode startup. Currently we only show the number of pending block 
deletions in WebUI. We should also show when the block deletion will start in 
WebUI.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6293) Issues with OIV processing PB-based fsimages

2014-05-13 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6293?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13996582#comment-13996582
 ] 

Hadoop QA commented on HDFS-6293:
-

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12644631/HDFS-6293.trunk.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 2 new 
or modified test files.

{color:red}-1 javac{color:red}.  The patch appears to cause the build to 
fail.

Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/6892//console

This message is automatically generated.

> Issues with OIV processing PB-based fsimages
> 
>
> Key: HDFS-6293
> URL: https://issues.apache.org/jira/browse/HDFS-6293
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 2.4.0
>Reporter: Kihwal Lee
>Assignee: Kihwal Lee
>Priority: Blocker
> Attachments: HDFS-6293.000.patch, HDFS-6293.001.patch, 
> HDFS-6293.002-save-deprecated-fsimage.patch, HDFS-6293.trunk.patch, 
> HDFS-6293_sbn_ckpt_retention.patch, 
> HDFS-6293_sbn_ckpt_retention_oiv_legacy.patch, Heap Histogram.html
>
>
> There are issues with OIV when processing fsimages in protobuf. 
> Due to the internal layout changes introduced by the protobuf-based fsimage, 
> OIV consumes excessive amount of memory.  We have tested with a fsimage with 
> about 140M files/directories. The peak heap usage when processing this image 
> in pre-protobuf (i.e. pre-2.4.0) format was about 350MB.  After converting 
> the image to the protobuf format on 2.4.0, OIV would OOM even with 80GB of 
> heap (max new size was 1GB).  It should be possible to process any image with 
> the default heap size of 1.5GB.
> Another issue is the complete change of format/content in OIV's XML output.  
> I also noticed that the secret manager section has no tokens while there were 
> unexpired tokens in the original image (pre-2.4.0).  I did not check whether 
> they were also missing in the new pb fsimage.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6283) Write end user documentation for xattrs.

2014-05-13 Thread Andrew Wang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6283?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Andrew Wang updated HDFS-6283:
--

Attachment: hdfs-6283-2.patch

> Write end user documentation for xattrs.
> 
>
> Key: HDFS-6283
> URL: https://issues.apache.org/jira/browse/HDFS-6283
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: documentation
>Reporter: Chris Nauroth
>Assignee: Andrew Wang
> Attachments: hdfs-6283-1.patch, hdfs-6283-2.patch
>
>
> Update the File System Shell documentation to cover the new getfattr and 
> setfattr commands.  If warranted, consider adding a separate dedicated page 
> for fuller discussion of the xattrs model and how the feature works in more 
> detail.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6334) Client failover proxy provider for IP failover based NN HA

2014-05-13 Thread Kihwal Lee (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6334?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kihwal Lee updated HDFS-6334:
-

   Resolution: Fixed
Fix Version/s: 2.5.0
   3.0.0
 Hadoop Flags: Reviewed
   Status: Resolved  (was: Patch Available)

Thanks for the review, Aaron. I've committed this to trunk and branch-2.

> Client failover proxy provider for IP failover based NN HA
> --
>
> Key: HDFS-6334
> URL: https://issues.apache.org/jira/browse/HDFS-6334
> Project: Hadoop HDFS
>  Issue Type: Improvement
>Reporter: Kihwal Lee
>Assignee: Kihwal Lee
> Fix For: 3.0.0, 2.5.0
>
> Attachments: HDFS-6334.patch, HDFS-6334.v2.patch, HDFS-6334.v3.patch
>
>
> With RPCv9 and improvements in the SPNEGO auth handling, it is possible to 
> set up a pair of HA namenodes utilizing IP failover as client-request fencing 
> mechanism.
> This jira will make it possible for HA to be configured without requiring use 
> of logical URI and provide a simple IP failover proxy provider.  The change 
> will allow any old implementation of {{FailoverProxyProvider}} to continue to 
> work.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6373) Remove support for extended attributes on symlinks

2014-05-13 Thread Chris Nauroth (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6373?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13996975#comment-13996975
 ] 

Chris Nauroth commented on HDFS-6373:
-

{quote}
This makes sense to me, since symlink permissions are always ignored. Linux 
actually doesn't even let you change symlink permissions, though it is allowed 
by the Unix spec. Chris Nauroth, could you comment on whether this was 
intentional, or just an oversight? It seems like we might want to remove 
ACLs-on-symlinks, which is compatible since symlinks are still hard-disabled.
{quote}

It is impossible to set an ACL on a symlink.  Any ACL operation run on a 
symlink automatically dereferences the symlink and operates on its target, i.e. 
by passing {{true}} to {{FSDirectory#getINodesInPath4Write}}.  We wrote tests 
asserting this behavior, so I don't think there is any way to store an 
{{AclFeature}} on an {{INodeSymlink}}.

I quickly skimmed the patch, and it looks like you're going for some additional 
prevention inside {{INodeSymlink}} to prevent mistakes.  That seems like a good 
idea, and we could do similar for ACLs.

> Remove support for extended attributes on symlinks
> --
>
> Key: HDFS-6373
> URL: https://issues.apache.org/jira/browse/HDFS-6373
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: namenode
>Reporter: Andrew Wang
>Assignee: Charles Lamb
> Attachments: HDFS-6373.1.patch, HDFS-6373.2.patch
>
>
> Looking in the Linux source code, we see the following:
> http://lxr.linux.no/linux+v3.14.3/fs/xattr.c
> {code}
>   60/*
>   61 * In the user.* namespace, only regular files and directories 
> can have
>   62 * extended attributes. For sticky directories, only the owner and
>   63 * privileged users can write attributes.
>   64 */
> {code}
> We should consider removing {{XAttrFeature}} from {{INodeSymlink}}.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6313) WebHdfs may use the wrong NN when configured for multiple HA NNs

2014-05-13 Thread Kihwal Lee (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6313?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13993042#comment-13993042
 ] 

Kihwal Lee commented on HDFS-6313:
--

MiniDFSCluster or NN does not alter the config after its up. It was a easy way 
to get HA config set up for the test.

> WebHdfs may use the wrong NN when configured for multiple HA NNs
> 
>
> Key: HDFS-6313
> URL: https://issues.apache.org/jira/browse/HDFS-6313
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: webhdfs
>Affects Versions: 3.0.0, 2.4.0
>Reporter: Daryn Sharp
>Assignee: Kihwal Lee
>Priority: Blocker
> Fix For: 3.0.0, 2.4.1
>
> Attachments: HDFS-6313.branch-2.4.patch, HDFS-6313.patch
>
>
> WebHdfs resolveNNAddr will return a union of addresses for all HA configured 
> NNs.  The client may access the wrong NN.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6373) Remove support for extended attributes on symlinks

2014-05-13 Thread Charles Lamb (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6373?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13996912#comment-13996912
 ] 

Charles Lamb commented on HDFS-6373:


Yeah, good idea. I've uploaded a new patch.

> Remove support for extended attributes on symlinks
> --
>
> Key: HDFS-6373
> URL: https://issues.apache.org/jira/browse/HDFS-6373
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: namenode
>Reporter: Andrew Wang
>Assignee: Charles Lamb
> Attachments: HDFS-6373.1.patch, HDFS-6373.2.patch
>
>
> Looking in the Linux source code, we see the following:
> http://lxr.linux.no/linux+v3.14.3/fs/xattr.c
> {code}
>   60/*
>   61 * In the user.* namespace, only regular files and directories 
> can have
>   62 * extended attributes. For sticky directories, only the owner and
>   63 * privileged users can write attributes.
>   64 */
> {code}
> We should consider removing {{XAttrFeature}} from {{INodeSymlink}}.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6370) Web UI fails to display in intranet under IE

2014-05-13 Thread Chris Nauroth (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6370?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chris Nauroth updated HDFS-6370:


   Resolution: Fixed
Fix Version/s: 2.5.0
   3.0.0
   Status: Resolved  (was: Patch Available)

I committed this to trunk and branch-2.  Haohui, thank you for contributing the 
fix.

> Web UI fails to display in intranet under IE
> 
>
> Key: HDFS-6370
> URL: https://issues.apache.org/jira/browse/HDFS-6370
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: datanode, journal-node, namenode
>Affects Versions: 3.0.0, 2.4.0
>Reporter: Haohui Mai
>Assignee: Haohui Mai
> Fix For: 3.0.0, 2.5.0
>
> Attachments: HDFS-6370.000.patch
>
>
> When IE renders the web UI of a cluster than runs in the intranet, it forces 
> the compatibility mode to be turned on which causes the UI fails to render 
> correctly.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Work started] (HDFS-6346) Optimize OP_SET_XATTRS by persisting single Xattr entry per setXattr/removeXattr api call

2014-05-13 Thread Yi Liu (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6346?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Work on HDFS-6346 started by Yi Liu.

> Optimize OP_SET_XATTRS by persisting single Xattr entry per 
> setXattr/removeXattr api call
> -
>
> Key: HDFS-6346
> URL: https://issues.apache.org/jira/browse/HDFS-6346
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: namenode
>Reporter: Uma Maheswara Rao G
>Assignee: Yi Liu
> Attachments: HDFS-6346.patch, editsStored
>
>
> When a new xattrs set on an Inode, it may add this with OP_SET_XATTRS and 
> let's say [user.name1:value1]
> On a next call if we set another xattrs, then it may store along with older 
> existing xattrs again. It may be like [user.name1:value1, user.name2:value2]
> So, on adding more xattrs on same Inode, that list may grow and we keep store 
> the entries of already existing name, value fairs.
> Right now we defaulted the max Xattrs on an Inode to 32 and configured. If 
> user modified it to much larger value and start setting xattrs, then edits 
> loading may create issue like my above example.
> But I didn't refer any usecase of having large number of xattrs, this is just 
> the assumption to consider a case. My biggest doubt is whether we will have 
> such real usecases to have huge xattrs on a single INode.
> So, here is a thought on having OP_SET_XATTR for each setXAttr operation to 
> be logged, When we replay them we need to consolidate. This is some initial 
> thought we can think more if others also feel we need to consider this case 
> to handle.
> Otherwise we endup storing Xattrs entries in editlog file as n(n+1)/2 where n 
> is number xattrs for a file/dir. This may be issue only when we have large 
> number configured max xattrs for inode.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6373) Remove support for extended attributes on symlinks

2014-05-13 Thread Charles Lamb (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6373?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Charles Lamb updated HDFS-6373:
---

Attachment: HDFS-6373.1.patch

Andrew, attached is a patch which overrides the appropriate xattr methods in 
INodeWithAdditionalFields.

> Remove support for extended attributes on symlinks
> --
>
> Key: HDFS-6373
> URL: https://issues.apache.org/jira/browse/HDFS-6373
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: namenode
>Reporter: Andrew Wang
>Assignee: Charles Lamb
> Attachments: HDFS-6373.1.patch
>
>
> Looking in the Linux source code, we see the following:
> http://lxr.linux.no/linux+v3.14.3/fs/xattr.c
> {code}
>   60/*
>   61 * In the user.* namespace, only regular files and directories 
> can have
>   62 * extended attributes. For sticky directories, only the owner and
>   63 * privileged users can write attributes.
>   64 */
> {code}
> We should consider removing {{XAttrFeature}} from {{INodeSymlink}}.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Comment Edited] (HDFS-6293) Issues with OIV processing PB-based fsimages

2014-05-13 Thread Suresh Srinivas (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6293?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13996922#comment-13996922
 ] 

Suresh Srinivas edited comment on HDFS-6293 at 5/13/14 9:36 PM:


+1 for the patch. Thanks [~wheat9] and [~kihwal] for making these changes.


was (Author: sureshms):
+1 for the patch. Thanks [~hmai] and [~kihwal] for making these changes.

> Issues with OIV processing PB-based fsimages
> 
>
> Key: HDFS-6293
> URL: https://issues.apache.org/jira/browse/HDFS-6293
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 2.4.0
>Reporter: Kihwal Lee
>Assignee: Kihwal Lee
>Priority: Blocker
> Attachments: HDFS-6293.000.patch, HDFS-6293.001.patch, 
> HDFS-6293.002-save-deprecated-fsimage.patch, HDFS-6293.branch-2.patch, 
> HDFS-6293.trunk.patch, HDFS-6293.trunk.patch, HDFS-6293.v2.branch-2.patch, 
> HDFS-6293.v2.trunk.patch, HDFS-6293_sbn_ckpt_retention.patch, 
> HDFS-6293_sbn_ckpt_retention_oiv_legacy.patch, Heap Histogram.html
>
>
> There are issues with OIV when processing fsimages in protobuf. 
> Due to the internal layout changes introduced by the protobuf-based fsimage, 
> OIV consumes excessive amount of memory.  We have tested with a fsimage with 
> about 140M files/directories. The peak heap usage when processing this image 
> in pre-protobuf (i.e. pre-2.4.0) format was about 350MB.  After converting 
> the image to the protobuf format on 2.4.0, OIV would OOM even with 80GB of 
> heap (max new size was 1GB).  It should be possible to process any image with 
> the default heap size of 1.5GB.
> Another issue is the complete change of format/content in OIV's XML output.  
> I also noticed that the secret manager section has no tokens while there were 
> unexpired tokens in the original image (pre-2.4.0).  I did not check whether 
> they were also missing in the new pb fsimage.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6395) Assorted improvements to xattr limit checking

2014-05-13 Thread Andrew Wang (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6395?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13996989#comment-13996989
 ] 

Andrew Wang commented on HDFS-6395:
---

Let's also remove the extra import in FSNamesystem that Chris noticed from 
HDFS-6377:

{code}
+import java.io.UnsupportedEncodingException;
{code}

> Assorted improvements to xattr limit checking
> -
>
> Key: HDFS-6395
> URL: https://issues.apache.org/jira/browse/HDFS-6395
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: namenode
>Affects Versions: HDFS XAttrs (HDFS-2006)
>Reporter: Andrew Wang
>
> It'd be nice to print messages during fsimage and editlog loading if we hit 
> either the # of xattrs per inode or the xattr size limits.
> We should also consider making the # of xattrs limit only apply to the user 
> namespace, or to each namespace separately, to prevent users from locking out 
> access to other namespaces.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6373) Remove support for extended attributes on symlinks

2014-05-13 Thread Charles Lamb (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6373?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Charles Lamb updated HDFS-6373:
---

Attachment: HDFS-6373.2.patch

> Remove support for extended attributes on symlinks
> --
>
> Key: HDFS-6373
> URL: https://issues.apache.org/jira/browse/HDFS-6373
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: namenode
>Reporter: Andrew Wang
>Assignee: Charles Lamb
> Attachments: HDFS-6373.1.patch, HDFS-6373.2.patch
>
>
> Looking in the Linux source code, we see the following:
> http://lxr.linux.no/linux+v3.14.3/fs/xattr.c
> {code}
>   60/*
>   61 * In the user.* namespace, only regular files and directories 
> can have
>   62 * extended attributes. For sticky directories, only the owner and
>   63 * privileged users can write attributes.
>   64 */
> {code}
> We should consider removing {{XAttrFeature}} from {{INodeSymlink}}.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6370) Web UI fails to display in intranet under IE

2014-05-13 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6370?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13996809#comment-13996809
 ] 

Hadoop QA commented on HDFS-6370:
-

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12644321/HDFS-6370.000.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:red}-1 tests included{color}.  The patch doesn't appear to include 
any new or modified tests.
Please justify why no new tests are needed for this 
patch.
Also please list what manual steps were performed to 
verify this patch.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  There were no new javadoc warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:green}+1 core tests{color}.  The patch passed unit tests in 
hadoop-hdfs-project/hadoop-hdfs.

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-HDFS-Build/6893//testReport/
Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/6893//console

This message is automatically generated.

> Web UI fails to display in intranet under IE
> 
>
> Key: HDFS-6370
> URL: https://issues.apache.org/jira/browse/HDFS-6370
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: datanode, journal-node, namenode
>Affects Versions: 3.0.0, 2.4.0
>Reporter: Haohui Mai
>Assignee: Haohui Mai
> Attachments: HDFS-6370.000.patch
>
>
> When IE renders the web UI of a cluster than runs in the intranet, it forces 
> the compatibility mode to be turned on which causes the UI fails to render 
> correctly.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6293) Issues with OIV processing PB-based fsimages

2014-05-13 Thread Suresh Srinivas (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6293?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13996922#comment-13996922
 ] 

Suresh Srinivas commented on HDFS-6293:
---

+1 for the patch. Thanks [~hmai] and [~kihwal] for making these changes.

> Issues with OIV processing PB-based fsimages
> 
>
> Key: HDFS-6293
> URL: https://issues.apache.org/jira/browse/HDFS-6293
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 2.4.0
>Reporter: Kihwal Lee
>Assignee: Kihwal Lee
>Priority: Blocker
> Attachments: HDFS-6293.000.patch, HDFS-6293.001.patch, 
> HDFS-6293.002-save-deprecated-fsimage.patch, HDFS-6293.branch-2.patch, 
> HDFS-6293.trunk.patch, HDFS-6293.trunk.patch, HDFS-6293.v2.branch-2.patch, 
> HDFS-6293.v2.trunk.patch, HDFS-6293_sbn_ckpt_retention.patch, 
> HDFS-6293_sbn_ckpt_retention_oiv_legacy.patch, Heap Histogram.html
>
>
> There are issues with OIV when processing fsimages in protobuf. 
> Due to the internal layout changes introduced by the protobuf-based fsimage, 
> OIV consumes excessive amount of memory.  We have tested with a fsimage with 
> about 140M files/directories. The peak heap usage when processing this image 
> in pre-protobuf (i.e. pre-2.4.0) format was about 350MB.  After converting 
> the image to the protobuf format on 2.4.0, OIV would OOM even with 80GB of 
> heap (max new size was 1GB).  It should be possible to process any image with 
> the default heap size of 1.5GB.
> Another issue is the complete change of format/content in OIV's XML output.  
> I also noticed that the secret manager section has no tokens while there were 
> unexpired tokens in the original image (pre-2.4.0).  I did not check whether 
> they were also missing in the new pb fsimage.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6378) NFS: when portmap/rpcbind is not available, NFS registration should timeout instead of hanging

2014-05-13 Thread Brandon Li (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6378?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Brandon Li updated HDFS-6378:
-

Description: NFS gateway should  

> NFS: when portmap/rpcbind is not available, NFS registration should timeout 
> instead of hanging 
> ---
>
> Key: HDFS-6378
> URL: https://issues.apache.org/jira/browse/HDFS-6378
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: nfs
>Reporter: Brandon Li
>
> NFS gateway should  



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6134) Transparent data at rest encryption

2014-05-13 Thread Alejandro Abdelnur (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6134?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13995608#comment-13995608
 ] 

Alejandro Abdelnur commented on HDFS-6134:
--

Owen, I think the docs posted in HADOOP-10150 will address most if not all your 
questions/concerns:

https://issues.apache.org/jira/secure/attachment/12640388/HDFSDataAtRestEncryptionAlternatives.pdf

https://issues.apache.org/jira/secure/attachment/12640389/HDFSDataatRestEncryptionProposal.pdf

https://issues.apache.org/jira/secure/attachment/12640390/HDFSDataatRestEncryptionAttackVectors.pdf

> Transparent data at rest encryption
> ---
>
> Key: HDFS-6134
> URL: https://issues.apache.org/jira/browse/HDFS-6134
> Project: Hadoop HDFS
>  Issue Type: New Feature
>  Components: security
>Affects Versions: 2.3.0
>Reporter: Alejandro Abdelnur
>Assignee: Alejandro Abdelnur
> Attachments: HDFSDataAtRestEncryption.pdf
>
>
> Because of privacy and security regulations, for many industries, sensitive 
> data at rest must be in encrypted form. For example: the health­care industry 
> (HIPAA regulations), the card payment industry (PCI DSS regulations) or the 
> US government (FISMA regulations).
> This JIRA aims to provide a mechanism to encrypt HDFS data at rest that can 
> be used transparently by any application accessing HDFS via Hadoop Filesystem 
> Java API, Hadoop libhdfs C library, or WebHDFS REST API.
> The resulting implementation should be able to be used in compliance with 
> different regulation requirements.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6373) Remove support for extended attributes on symlinks

2014-05-13 Thread Andrew Wang (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6373?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13996905#comment-13996905
 ] 

Andrew Wang commented on HDFS-6373:
---

This is basically what I had in mind, but how do you feel about throwing some 
kind of RuntimeException, e.g. UnsupportedOperationException? It should prevent 
anyone from trying this in the future.

> Remove support for extended attributes on symlinks
> --
>
> Key: HDFS-6373
> URL: https://issues.apache.org/jira/browse/HDFS-6373
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: namenode
>Reporter: Andrew Wang
>Assignee: Charles Lamb
> Attachments: HDFS-6373.1.patch
>
>
> Looking in the Linux source code, we see the following:
> http://lxr.linux.no/linux+v3.14.3/fs/xattr.c
> {code}
>   60/*
>   61 * In the user.* namespace, only regular files and directories 
> can have
>   62 * extended attributes. For sticky directories, only the owner and
>   63 * privileged users can write attributes.
>   64 */
> {code}
> We should consider removing {{XAttrFeature}} from {{INodeSymlink}}.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6345) DFS.listCacheDirectives() should allow filtering based on cache directive ID

2014-05-13 Thread Colin Patrick McCabe (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6345?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13996783#comment-13996783
 ] 

Colin Patrick McCabe commented on HDFS-6345:


{code}
+// Querying for a single ID
+final Long id = filter.getId();
+if (id != null) {
+  if (!directivesById.containsKey(id)) {
+throw new InvalidRequestException("Did not find requested id " + id);
+  }
+  prevId = id - 1;
 }
{code}

This is a bit tricky, so I think we should add a comment to the code explaining 
why we're changing prevId.

{code}
+
 ArrayList replies =
 new ArrayList(NUM_PRE_ALLOCATED_ENTRIES);
 int numReplies = 0;
@@ -710,6 +718,11 @@ public void removeDirective(long id, FSPermissionChecker 
pc)
   }
   CacheDirective curDirective = cur.getValue();
   CacheDirectiveInfo info = cur.getValue().toInfo();
+  if (id != null &&
+  !(info.getId().equals(id))) {
+// If we didn't find the requested ID, we're done
+break;
+  }
{code}

Maybe change the comment to "if we didn't find, or filtered out the requested 
ID".  Filtering by ID is only one kind of filtering... if an additional filter 
is set which eliminates the requested item, nothing will show up.

> DFS.listCacheDirectives() should allow filtering based on cache directive ID
> 
>
> Key: HDFS-6345
> URL: https://issues.apache.org/jira/browse/HDFS-6345
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: caching
>Affects Versions: 2.4.0
>Reporter: Lenni Kuff
>Assignee: Andrew Wang
> Attachments: hdfs-6345-1.patch
>
>
> DFS.listCacheDirectives() should allow filtering based on cache directive ID. 
> Currently it throws an exception.
> For example:
> {code}
> long directiveId = ;
> CacheDirectiveInfo filter = new CacheDirectiveInfo.Builder()  
> .setId(directiveId)
> .build();
> RemoteIterator itr = dfs.listCacheDirectives(filter);
> {code}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6283) Write end user documentation for xattrs.

2014-05-13 Thread Chris Nauroth (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6283?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13996767#comment-13996767
 ] 

Chris Nauroth commented on HDFS-6283:
-

Hi, Andrew.  The new content looks great.  Here are a few comments:

# Would you please add a hyperlink to the new doc in 
hadoop-project/src/site/site.xml?
# Would you please also update FileSystemShell.apt.vm to add getfattr and 
setfattr?
# The "Go Back" link is broken.  It looks like some existing Go Back links are 
broken too in the published documentation.  Let's fix this one here, and I 
filed HADOOP-10602 to clean up the others.
# Let's mention that setting a value for an xattr is optional.  (Some might 
find it helpful to set a valueless xattr to use as a kind of tag.)

> Write end user documentation for xattrs.
> 
>
> Key: HDFS-6283
> URL: https://issues.apache.org/jira/browse/HDFS-6283
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: documentation
>Reporter: Chris Nauroth
>Assignee: Andrew Wang
> Attachments: hdfs-6283-1.patch
>
>
> Update the File System Shell documentation to cover the new getfattr and 
> setfattr commands.  If warranted, consider adding a separate dedicated page 
> for fuller discussion of the xattrs model and how the feature works in more 
> detail.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6056) Clean up NFS config settings

2014-05-13 Thread Brandon Li (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6056?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Brandon Li updated HDFS-6056:
-

Attachment: HDFS-6056.004.patch

> Clean up NFS config settings
> 
>
> Key: HDFS-6056
> URL: https://issues.apache.org/jira/browse/HDFS-6056
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: nfs
>Affects Versions: 2.3.0
>Reporter: Aaron T. Myers
>Assignee: Brandon Li
> Attachments: HDFS-6056.001.patch, HDFS-6056.002.patch, 
> HDFS-6056.003.patch, HDFS-6056.004.patch
>
>
> As discussed on HDFS-6050, there's a few opportunities to improve the config 
> settings related to NFS. This JIRA is to implement those changes.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6370) Web UI fails to display in intranet under IE

2014-05-13 Thread Chris Nauroth (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6370?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13996834#comment-13996834
 ] 

Chris Nauroth commented on HDFS-6370:
-

bq. -1 tests included. The patch doesn't appear to include any new or modified 
tests.

This patch changes HTML only.  There are no new automated tests.  It has been 
tested manually by browsing with IE.

I'm going to commit this.

> Web UI fails to display in intranet under IE
> 
>
> Key: HDFS-6370
> URL: https://issues.apache.org/jira/browse/HDFS-6370
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: datanode, journal-node, namenode
>Affects Versions: 3.0.0, 2.4.0
>Reporter: Haohui Mai
>Assignee: Haohui Mai
> Attachments: HDFS-6370.000.patch
>
>
> When IE renders the web UI of a cluster than runs in the intranet, it forces 
> the compatibility mode to be turned on which causes the UI fails to render 
> correctly.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6293) Issues with OIV processing PB-based fsimages

2014-05-13 Thread Kihwal Lee (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6293?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13996649#comment-13996649
 ] 

Kihwal Lee commented on HDFS-6293:
--

Ouch. The patch applies to trunk but it introduces a duplicate import in the 
test case. I will upload separate patches.

> Issues with OIV processing PB-based fsimages
> 
>
> Key: HDFS-6293
> URL: https://issues.apache.org/jira/browse/HDFS-6293
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 2.4.0
>Reporter: Kihwal Lee
>Assignee: Kihwal Lee
>Priority: Blocker
> Attachments: HDFS-6293.000.patch, HDFS-6293.001.patch, 
> HDFS-6293.002-save-deprecated-fsimage.patch, HDFS-6293.trunk.patch, 
> HDFS-6293_sbn_ckpt_retention.patch, 
> HDFS-6293_sbn_ckpt_retention_oiv_legacy.patch, Heap Histogram.html
>
>
> There are issues with OIV when processing fsimages in protobuf. 
> Due to the internal layout changes introduced by the protobuf-based fsimage, 
> OIV consumes excessive amount of memory.  We have tested with a fsimage with 
> about 140M files/directories. The peak heap usage when processing this image 
> in pre-protobuf (i.e. pre-2.4.0) format was about 350MB.  After converting 
> the image to the protobuf format on 2.4.0, OIV would OOM even with 80GB of 
> heap (max new size was 1GB).  It should be possible to process any image with 
> the default heap size of 1.5GB.
> Another issue is the complete change of format/content in OIV's XML output.  
> I also noticed that the secret manager section has no tokens while there were 
> unexpired tokens in the original image (pre-2.4.0).  I did not check whether 
> they were also missing in the new pb fsimage.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (HDFS-6384) Provide an alternative offline file system analysis tool

2014-05-13 Thread Kihwal Lee (JIRA)
Kihwal Lee created HDFS-6384:


 Summary: Provide an alternative offline file system analysis tool
 Key: HDFS-6384
 URL: https://issues.apache.org/jira/browse/HDFS-6384
 Project: Hadoop HDFS
  Issue Type: Improvement
Affects Versions: 2.5.0
Reporter: Kihwal Lee
Priority: Critical


Per discussion in HDFS-6293, we need an alternative way of analyzing file 
system offline.  Some use cases can be covered by post-processing the new 
PB-based fsimage without consuming much resources, but others may need a way to 
efficiently generating fully constructed name space.

Once we have a better solution, we should deprecate NN's ability to save its 
partial state in the old fsimage format.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6383) Upgrade S3n s3.fs.buffer.dir to suppoer multi directories

2014-05-13 Thread Ted Malaska (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6383?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ted Malaska updated HDFS-6383:
--

Attachment: HDFS-6383.patch

This solves the problem.

Thanks to:
Joe Prosser, Dave Wang, Andrei Savu
and 
Govind Kamat for testing

> Upgrade S3n s3.fs.buffer.dir to suppoer multi directories
> -
>
> Key: HDFS-6383
> URL: https://issues.apache.org/jira/browse/HDFS-6383
> Project: Hadoop HDFS
>  Issue Type: Improvement
>Reporter: Ted Malaska
>Priority: Minor
> Attachments: HDFS-6383.patch
>
>
> s3.fs.buffer.dir defines the tmp folder where files will be written to before 
> getting sent to S3.  Right now this is limited to a single folder which 
> causes to major issues.
> 1. You need a drive with enough space to store all the tmp files at once
> 2. You are limited to the IO speeds of a single drive
> This solution will resolve both and has been tested to increase the S3 write 
> speed by 2.5x with 10 mappers on hs1.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6367) EnumSetParam$Domain#parse fails for parameter containing more than one enum.

2014-05-13 Thread Uma Maheswara Rao G (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6367?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Uma Maheswara Rao G updated HDFS-6367:
--

Description: 
Fails because additional "," 

java.lang.IllegalArgumentException: No enum const class 
org.apache.hadoop.fs.Options$Rename.,OVERWRITE
at java.lang.Enum.valueOf(Enum.java:196)
at 
org.apache.hadoop.hdfs.web.resources.EnumSetParam$Domain.parse(EnumSetParam.java:85)
at 
org.apache.hadoop.hdfs.web.resources.RenameOptionSetParam.(RenameOptionSetParam.java:45)
at 
org.apache.hadoop.hdfs.web.resources.TestParam.testRenameOptionSetParam(TestParam.java:355)


  was:
Fails because additional "," 

java.lang.IllegalArgumentException: No enum const class 
org.apache.hadoop.fs.Options$Rename.,OVERWRITE
at java.lang.Enum.valueOf(Enum.java:196)
at 
org.apache.hadoop.hdfs.web.resources.EnumSetParam$Domain.parse(EnumSetParam.java:85)
at 
org.apache.hadoop.hdfs.web.resources.RenameOptionSetParam.(RenameOptionSetParam.java:45)
at 
org.apache.hadoop.hdfs.web.resources.TestParam.testRenameOptionSetParam(TestParam.java:355)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at 
org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:47)
at 
org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at 
org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:44)
at 
org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:271)
at 
org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:70)
at 
org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
at org.junit.runners.ParentRunner$3.run(ParentRunner.java:238)
at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:63)
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:236)
at org.junit.runners.ParentRunner.access$000(ParentRunner.java:53)
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:229)
at org.junit.runners.ParentRunner.run(ParentRunner.java:309)
at 
org.eclipse.jdt.internal.junit4.runner.JUnit4TestReference.run(JUnit4TestReference.java:50)
at 
org.eclipse.jdt.internal.junit.runner.TestExecution.run(TestExecution.java:38)
at 
org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:467)
at 
org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:683)
at 
org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.run(RemoteTestRunner.java:390)
at 
org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.main(RemoteTestRunner.java:197)




> EnumSetParam$Domain#parse fails for parameter containing more than one enum.
> 
>
> Key: HDFS-6367
> URL: https://issues.apache.org/jira/browse/HDFS-6367
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: webhdfs
>Affects Versions: 3.0.0
>Reporter: Yi Liu
>Assignee: Yi Liu
> Fix For: 3.0.0
>
> Attachments: HDFS-6367.patch
>
>
> Fails because additional "," 
> 
> java.lang.IllegalArgumentException: No enum const class 
> org.apache.hadoop.fs.Options$Rename.,OVERWRITE
>   at java.lang.Enum.valueOf(Enum.java:196)
>   at 
> org.apache.hadoop.hdfs.web.resources.EnumSetParam$Domain.parse(EnumSetParam.java:85)
>   at 
> org.apache.hadoop.hdfs.web.resources.RenameOptionSetParam.(RenameOptionSetParam.java:45)
>   at 
> org.apache.hadoop.hdfs.web.resources.TestParam.testRenameOptionSetParam(TestParam.java:355)
>   



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6355) Fix divide-by-zero, improper use of wall-clock time in BlockPoolSliceScanner

2014-05-13 Thread Colin Patrick McCabe (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6355?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Colin Patrick McCabe updated HDFS-6355:
---

   Resolution: Fixed
Fix Version/s: 2.5.0
   Status: Resolved  (was: Patch Available)

committed, thanks

> Fix divide-by-zero, improper use of wall-clock time in BlockPoolSliceScanner
> 
>
> Key: HDFS-6355
> URL: https://issues.apache.org/jira/browse/HDFS-6355
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 2.4.0
>Reporter: Colin Patrick McCabe
>Assignee: Colin Patrick McCabe
> Fix For: 2.5.0
>
> Attachments: HDFS-6355.001.patch
>
>
> BlockPoolSliceScanner uses {{Time.now}} to calculate an interval.  But this 
> is incorrect, since if the wall-clock time changes, we will end up setting 
> the scan periods to a shorter or longer time than we configured.
> There is also a case where we may divide by zero if we get unlucky, because 
> we calculate an interval and divide by it, without checking whether the 
> interval is 0 milliseconds.  This would produce an {{ArithmeticException}} 
> since we are using longs.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6056) Clean up NFS config settings

2014-05-13 Thread Brandon Li (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6056?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Brandon Li updated HDFS-6056:
-

Description: As discussed on HDFS-6050, there's a few opportunities to 
improve the config settings related to NFS. This JIRA is to implement those 
changes, which include: moving hdfs-nfs related properties into hadoop-hdfs-nfs 
project, and replacing 'nfs' with 'nfs' in the property names.  (was: As 
discussed on HDFS-6050, there's a few opportunities to improve the config 
settings related to NFS. This JIRA is to implement those changes.)

> Clean up NFS config settings
> 
>
> Key: HDFS-6056
> URL: https://issues.apache.org/jira/browse/HDFS-6056
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: nfs
>Affects Versions: 2.3.0
>Reporter: Aaron T. Myers
>Assignee: Brandon Li
> Attachments: HDFS-6056.001.patch, HDFS-6056.002.patch, 
> HDFS-6056.003.patch, HDFS-6056.004.patch
>
>
> As discussed on HDFS-6050, there's a few opportunities to improve the config 
> settings related to NFS. This JIRA is to implement those changes, which 
> include: moving hdfs-nfs related properties into hadoop-hdfs-nfs project, and 
> replacing 'nfs' with 'nfs' in the property names.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6377) Unify xattr name and value limits into a single limit

2014-05-13 Thread Uma Maheswara Rao G (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6377?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13995997#comment-13995997
 ] 

Uma Maheswara Rao G commented on HDFS-6377:
---


Thanks a lot, Andrew for filing this JIRA and for the patch.
Adding to Liu's  and Chris comments, I have following comments:
- . {code}
 "XAttr is too big, maximum size = " + nnConf.xattrMaxSize
+  + ", but the size is = " + xAttr.getName().length());
{code}

Seems like now max size is combination of name and value size together, but 
here you considered on name#length?

- . {code}
 public static void init() throws Exception {
+conf = new Configuration();
{code}
Lets use HdfsCOnfiguration? [ but no harm in using either :-) ]

- . {code}
 fs.setXAttr(path, name1, longValue);
+  fs.setXAttr(path, "user.a", longValue);
{code}
you can use name1 for it?

- 
Do you think we need to have minimum configuration limit?  Lets say user 
configured size as 3, then this is always invalid size as Namespace itself 
occupy this space?  [ I am not insisting, just to discuss this point ]
Atleast we can warn on startup saying, you can not add xattrs with configured 
limit? [ I think 7 is the minimum size for adding xattr with user name space 
where name and value can have one char each at least]

- 
Thanks for the other cleanup.
Yes, I agree with Chris, having fs-limits for consistency make sense to me.

> Unify xattr name and value limits into a single limit
> -
>
> Key: HDFS-6377
> URL: https://issues.apache.org/jira/browse/HDFS-6377
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: namenode
>Affects Versions: HDFS XAttrs (HDFS-2006)
>Reporter: Andrew Wang
>Assignee: Andrew Wang
> Attachments: hdfs-6377-1.patch
>
>
> Instead of having separate limits and config options for the size of an 
> xattr's name and value, let's use a single limit.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Resolved] (HDFS-6380) TestBalancerWithNodeGroup is failing in trunk

2014-05-13 Thread Uma Maheswara Rao G (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6380?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Uma Maheswara Rao G resolved HDFS-6380.
---

Resolution: Duplicate

> TestBalancerWithNodeGroup is failing in trunk
> -
>
> Key: HDFS-6380
> URL: https://issues.apache.org/jira/browse/HDFS-6380
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: namenode
>Affects Versions: 3.0.0
>Reporter: Uma Maheswara Rao G
>
> Error Message
>  expected:<1800> but was:<1814>
> Stacktrace
> {noformat}
> java.lang.AssertionError: expected:<1800> but was:<1814>
>   at org.junit.Assert.fail(Assert.java:88)
>   at org.junit.Assert.failNotEquals(Assert.java:743)
>   at org.junit.Assert.assertEquals(Assert.java:118)
>   at org.junit.Assert.assertEquals(Assert.java:144)
>   at 
> org.apache.hadoop.hdfs.server.balancer.TestBalancerWithNodeGroup.testBalancerWithRackLocality(TestBalancerWithNodeGroup.java:253)
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6186) Pause deletion of blocks when the namenode starts up

2014-05-13 Thread Suresh Srinivas (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6186?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13996688#comment-13996688
 ] 

Suresh Srinivas commented on HDFS-6186:
---

+1 for the patch.

We should create another jira to print along with current pending deletion 
count displayed on namenode webUI, something like:
{{Pending Deletion Blocks:  (Block deletion will resume after xxx seconds)}}

> Pause deletion of blocks when the namenode starts up
> 
>
> Key: HDFS-6186
> URL: https://issues.apache.org/jira/browse/HDFS-6186
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: namenode
>Reporter: Suresh Srinivas
>Assignee: Jing Zhao
> Attachments: HDFS-6186.000.patch, HDFS-6186.002.patch, 
> HDFS-6186.003.patch
>
>
> HDFS namenode can delete blocks very quickly, given the deletion happens as a 
> parallel operation spread across many datanodes. One of the frequent 
> anxieties I see is that a lot of data can be deleted very quickly, when a 
> cluster is brought up, especially when one of the storage directories has 
> failed and namenode metadata was copied from another storage. Copying wrong 
> metadata would results in some of the newer files (if old metadata was 
> copied) being deleted along with their blocks. 
> HDFS-5986 now captures the number of pending deletion block on namenode webUI 
> and JMX. I propose pausing deletion of blocks for a configured period of time 
> (default 1 hour?) after namenode comes out of safemode. This will give enough 
> time for the administrator to notice large number of pending deletion blocks 
> and take corrective action.
> Thoughts?



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6346) Optimize OP_SET_XATTRS by persisting single Xattr entry per setXattr/removeXattr api call

2014-05-13 Thread Yi Liu (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6346?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yi Liu updated HDFS-6346:
-

Attachment: HDFS-6346.2.patch

Uma, thanks for review and collecting the improvement data. That's great. The 
new patch includes update for your comment.

> Optimize OP_SET_XATTRS by persisting single Xattr entry per 
> setXattr/removeXattr api call
> -
>
> Key: HDFS-6346
> URL: https://issues.apache.org/jira/browse/HDFS-6346
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: namenode
>Reporter: Uma Maheswara Rao G
>Assignee: Yi Liu
> Attachments: HDFS-6346.1.patch, HDFS-6346.2.patch, HDFS-6346.patch, 
> editsStored
>
>
> When a new xattrs set on an Inode, it may add this with OP_SET_XATTRS and 
> let's say [user.name1:value1]
> On a next call if we set another xattrs, then it may store along with older 
> existing xattrs again. It may be like [user.name1:value1, user.name2:value2]
> So, on adding more xattrs on same Inode, that list may grow and we keep store 
> the entries of already existing name, value fairs.
> Right now we defaulted the max Xattrs on an Inode to 32 and configured. If 
> user modified it to much larger value and start setting xattrs, then edits 
> loading may create issue like my above example.
> But I didn't refer any usecase of having large number of xattrs, this is just 
> the assumption to consider a case. My biggest doubt is whether we will have 
> such real usecases to have huge xattrs on a single INode.
> So, here is a thought on having OP_SET_XATTR for each setXAttr operation to 
> be logged, When we replay them we need to consolidate. This is some initial 
> thought we can think more if others also feel we need to consider this case 
> to handle.
> Otherwise we endup storing Xattrs entries in editlog file as n(n+1)/2 where n 
> is number xattrs for a file/dir. This may be issue only when we have large 
> number configured max xattrs for inode.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6369) RemoteBlockReader#available() should call FSInputChecker.available()

2014-05-13 Thread Colin Patrick McCabe (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6369?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13995294#comment-13995294
 ] 

Colin Patrick McCabe commented on HDFS-6369:


The calling code in {{DFSInputStream}} already takes into account where the end 
of the block is.  See this code in {{DFSInputStream}}:
{code}
if (pos <= targetPos && targetPos <= blockEnd) {
  //
  // If this seek is to a positive position in the current
  // block, and this piece of data might already be lying in
  // the TCP buffer, then just eat up the intervening data.
  //
  int diff = (int)(targetPos - pos);
  if (diff <= blockReader.available()) {
  ...
{code}

Notice that we don't call {{available}} if {{targetPos}} is beyond {{blockEnd}}

Also, {{RemoteBlockReader}} is deprecated.  See the comment at the top:

{code}
/**
 * \@deprecated this is an old implementation that is being left around
 * in case any issues spring up with the new {\@link RemoteBlockReader2} 
implementation.
 * It will be removed in the next release.
 */
{code}

We should really fix {{RemoteBlockReader2}} to support SOCKS sockets (those are 
the only ones without channels, as I recall) and remove {{RemoteBlockReader}} 
as the comment promised we'd do.

> RemoteBlockReader#available() should call FSInputChecker.available()
> 
>
> Key: HDFS-6369
> URL: https://issues.apache.org/jira/browse/HDFS-6369
> Project: Hadoop HDFS
>  Issue Type: Bug
>Reporter: Ted Yu
>Assignee: Chen He
>Priority: Trivial
> Attachments: HDFS-6369.patch
>
>
> Currently DFSClient.TCP_WINDOW_SIZE is directly returned.
> However, FSInputChecker.available(), in the superclass, may return value 
> lower than the constant.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6230) Expose upgrade status through NameNode web UI

2014-05-13 Thread Mit Desai (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6230?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mit Desai updated HDFS-6230:


Attachment: HDFS-6230.patch

Thanks for looking at the patch [~wheat9]. Posting the updated patch.

> Expose upgrade status through NameNode web UI
> -
>
> Key: HDFS-6230
> URL: https://issues.apache.org/jira/browse/HDFS-6230
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: namenode
>Affects Versions: 2.4.0
>Reporter: Arpit Agarwal
>Assignee: Mit Desai
> Attachments: HDFS-6230-NoUpgradesInProgress.png, 
> HDFS-6230-UpgradeInProgress.jpg, HDFS-6230.patch, HDFS-6230.patch
>
>
> The NameNode web UI does not show upgrade information anymore. Hadoop 2.0 
> also does not have the _hadoop dfsadmin -upgradeProgress_ command to check 
> the upgrade status.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6326) WebHdfs ACL compatibility is broken

2014-05-13 Thread Chris Nauroth (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6326?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chris Nauroth updated HDFS-6326:


Attachment: HDFS-6326.5.patch

Daryn, that looks like a good suggestion to me.  Here is patch v5 incorporating 
those ideas.  Mostly, this is just folding your code right into the v4 patch 
with some additional comments.  Note however that I kept the 
{{FsAclPermission}} class as {{Private}} inside the HDFS package, not Common.  
I'd prefer not to publish this into the public API footprint.

I was hesitant to add a boolean member to track true/false for ACL presence 
because of the risk of increasing memory footprint.  However, I confirmed with 
jmap -histo that it makes no difference after your suggested change:

{code}
375: 1 32  org.apache.hadoop.fs.permission.FsPermission
382: 1 32  
org.apache.hadoop.hdfs.protocol.FsAclPermission
{code}

Memory utilization of {{FsPermission}} and {{FsAclPermission}} are very much 
dominated by Java object overhead/padding, so adding the extra boolean didn't 
make a difference.  Haohui and I have had some offline discussion about 
optimizations here, but that's not directly relevant to this issue.

[~daryn], can you please let me know if you're still +1 for v5?  Thanks!

> WebHdfs ACL compatibility is broken
> ---
>
> Key: HDFS-6326
> URL: https://issues.apache.org/jira/browse/HDFS-6326
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: webhdfs
>Affects Versions: 3.0.0, 2.4.0
>Reporter: Daryn Sharp
>Assignee: Chris Nauroth
>Priority: Blocker
> Attachments: HDFS-6326.1.patch, HDFS-6326.2.patch, HDFS-6326.3.patch, 
> HDFS-6326.4.patch, HDFS-6326.5.patch, aclfsperm.example
>
>
> 2.4 ACL support is completely incompatible with <2.4 webhdfs servers.  The NN 
> throws an {{IllegalArgumentException}} exception.
> {code}
> hadoop fs -ls webhdfs://nn/
> Found 21 items
> ls: Invalid value for webhdfs parameter "op": No enum constant 
> org.apache.hadoop.hdfs.web.resources.GetOpParam.Op.GETACLSTATUS
> [... 20 more times...]
> {code}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6367) Domain>#parse in EnumSetParam fails for parmater containing more than one enum.

2014-05-13 Thread Yi Liu (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6367?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yi Liu updated HDFS-6367:
-

Attachment: HDFS-6367.patch

The "," should not be contained.

> Domain>#parse in EnumSetParam fails for parmater containing 
> more than one enum.
> -
>
> Key: HDFS-6367
> URL: https://issues.apache.org/jira/browse/HDFS-6367
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: webhdfs
>Affects Versions: 3.0.0
>Reporter: Yi Liu
> Fix For: 3.0.0
>
> Attachments: HDFS-6367.patch
>
>
> Fails because additional "," 
> 
> java.lang.IllegalArgumentException: No enum const class 
> org.apache.hadoop.fs.Options$Rename.,OVERWRITE
>   at java.lang.Enum.valueOf(Enum.java:196)
>   at 
> org.apache.hadoop.hdfs.web.resources.EnumSetParam$Domain.parse(EnumSetParam.java:85)
>   at 
> org.apache.hadoop.hdfs.web.resources.RenameOptionSetParam.(RenameOptionSetParam.java:45)
>   at 
> org.apache.hadoop.hdfs.web.resources.TestParam.testRenameOptionSetParam(TestParam.java:355)
>   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>   at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
>   at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>   at java.lang.reflect.Method.invoke(Method.java:597)
>   at 
> org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:47)
>   at 
> org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
>   at 
> org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:44)
>   at 
> org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
>   at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:271)
>   at 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:70)
>   at 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
>   at org.junit.runners.ParentRunner$3.run(ParentRunner.java:238)
>   at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:63)
>   at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:236)
>   at org.junit.runners.ParentRunner.access$000(ParentRunner.java:53)
>   at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:229)
>   at org.junit.runners.ParentRunner.run(ParentRunner.java:309)
>   at 
> org.eclipse.jdt.internal.junit4.runner.JUnit4TestReference.run(JUnit4TestReference.java:50)
>   at 
> org.eclipse.jdt.internal.junit.runner.TestExecution.run(TestExecution.java:38)
>   at 
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:467)
>   at 
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:683)
>   at 
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.run(RemoteTestRunner.java:390)
>   at 
> org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.main(RemoteTestRunner.java:197)



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6370) Web UI fails to display in intranet under IE

2014-05-13 Thread Chris Nauroth (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6370?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chris Nauroth updated HDFS-6370:


  Component/s: namenode
   journal-node
   datanode
 Target Version/s: 3.0.0, 2.5.0
Affects Version/s: 3.0.0
   2.4.0
 Hadoop Flags: Reviewed

> Web UI fails to display in intranet under IE
> 
>
> Key: HDFS-6370
> URL: https://issues.apache.org/jira/browse/HDFS-6370
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: datanode, journal-node, namenode
>Affects Versions: 3.0.0, 2.4.0
>Reporter: Haohui Mai
>Assignee: Haohui Mai
> Attachments: HDFS-6370.000.patch
>
>
> When IE renders the web UI of a cluster than runs in the intranet, it forces 
> the compatibility mode to be turned on which causes the UI fails to render 
> correctly.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6056) Clean up NFS config settings

2014-05-13 Thread Brandon Li (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6056?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13996698#comment-13996698
 ] 

Brandon Li commented on HDFS-6056:
--

[~jingzhao], thanks for the review. I've update the jira description and marked 
it as incompatible change.
I've uploaded a new patch to address your comments.

> Clean up NFS config settings
> 
>
> Key: HDFS-6056
> URL: https://issues.apache.org/jira/browse/HDFS-6056
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: nfs
>Affects Versions: 2.3.0
>Reporter: Aaron T. Myers
>Assignee: Brandon Li
> Attachments: HDFS-6056.001.patch, HDFS-6056.002.patch, 
> HDFS-6056.003.patch, HDFS-6056.004.patch
>
>
> As discussed on HDFS-6050, there's a few opportunities to improve the config 
> settings related to NFS. This JIRA is to implement those changes, which 
> include: moving hdfs-nfs related properties into hadoop-hdfs-nfs project, and 
> replacing 'nfs' with 'nfs' in the property names.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6361) TestIdUserGroup.testUserUpdateSetting failed due to out of range nfsnobody Id

2014-05-13 Thread Yongjun Zhang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6361?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yongjun Zhang updated HDFS-6361:


Attachment: HDFS-6361.003.patch

> TestIdUserGroup.testUserUpdateSetting failed due to out of range nfsnobody Id
> -
>
> Key: HDFS-6361
> URL: https://issues.apache.org/jira/browse/HDFS-6361
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: nfs
>Affects Versions: 2.4.0
>Reporter: Yongjun Zhang
>Assignee: Yongjun Zhang
> Attachments: HDFS-6361.001.patch, HDFS-6361.002.patch, 
> HDFS-6361.003.patch
>
>
> The following error happens pretty often:
> org.apache.hadoop.nfs.nfs3.TestIdUserGroup.testUserUpdateSetting
> Failing for the past 1 build (Since Unstable#61 )
> Took 0.1 sec.
> add description
> Error Message
> For input string: "4294967294"
> Stacktrace
> java.lang.NumberFormatException: For input string: "4294967294"
>   at 
> java.lang.NumberFormatException.forInputString(NumberFormatException.java:65)
>   at java.lang.Integer.parseInt(Integer.java:495)
>   at java.lang.Integer.valueOf(Integer.java:582)
>   at 
> org.apache.hadoop.nfs.nfs3.IdUserGroup.updateMapInternal(IdUserGroup.java:137)
>   at 
> org.apache.hadoop.nfs.nfs3.IdUserGroup.updateMaps(IdUserGroup.java:188)
>   at org.apache.hadoop.nfs.nfs3.IdUserGroup.(IdUserGroup.java:60)
>   at 
> org.apache.hadoop.nfs.nfs3.TestIdUserGroup.testUserUpdateSetting(TestIdUserGroup.java:71)
> Standard Output
> log4j:WARN No appenders could be found for logger 
> (org.apache.hadoop.nfs.nfs3.IdUserGroup).
> log4j:WARN Please initialize the log4j system properly.
> log4j:WARN See http://logging.apache.org/log4j/1.2/faq.html#noconfig for more 
> info.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6340) DN can't finalize upgarde

2014-05-13 Thread Arpit Agarwal (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6340?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13992964#comment-13992964
 ] 

Arpit Agarwal commented on HDFS-6340:
-

It looks like the previous build aborted. I kicked off a new one.

> DN can't finalize upgarde
> -
>
> Key: HDFS-6340
> URL: https://issues.apache.org/jira/browse/HDFS-6340
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: datanode
>Affects Versions: 2.4.0
>Reporter: Rahul Singhal
>Priority: Blocker
> Fix For: 3.0.0, 2.4.1
>
> Attachments: HDFS-6340-branch-2.4.0.patch, HDFS-6340.02.patch, 
> HDFS-6340.patch
>
>
> I upgraded a (NN) HA cluster from 2.2.0 to 2.4.0. After I issued the 
> '-finalizeUpgarde' command, NN was able to finalize the upgrade but DN 
> couldn't (I waited for the next block report).
> I think I have found the problem to be due to HDFS-5153. I will attach a 
> proposed fix.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6357) SetXattr should persist rpcIDs for handling retrycache with Namenode restart and HA

2014-05-13 Thread Uma Maheswara Rao G (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6357?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13994293#comment-13994293
 ] 

Uma Maheswara Rao G commented on HDFS-6357:
---

Thanks a lot, for the reviews Yi and Andrew!
I have just committed this to branch!

> SetXattr should persist rpcIDs for handling retrycache with Namenode restart 
> and HA
> ---
>
> Key: HDFS-6357
> URL: https://issues.apache.org/jira/browse/HDFS-6357
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: namenode
>Affects Versions: HDFS XAttrs (HDFS-2006)
>Reporter: Uma Maheswara Rao G
>Assignee: Uma Maheswara Rao G
> Fix For: HDFS XAttrs (HDFS-2006)
>
> Attachments: HDFS-6357.patch, HDFS-6357.patch, HDFS-6357.patch
>
>
> SetXattr should persist rpcIDs for handling restart Namenode and HA scenarios



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Comment Edited] (HDFS-6293) Issues with OIV processing PB-based fsimages

2014-05-13 Thread Kihwal Lee (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6293?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13996649#comment-13996649
 ] 

Kihwal Lee edited comment on HDFS-6293 at 5/13/14 5:19 PM:
---

Oops. The patch applies to trunk but it introduces a duplicate import in the 
test case. I will upload separate patches.


was (Author: kihwal):
Ouch. The patch applies to trunk but it introduces a duplicate import in the 
test case. I will upload separate patches.

> Issues with OIV processing PB-based fsimages
> 
>
> Key: HDFS-6293
> URL: https://issues.apache.org/jira/browse/HDFS-6293
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 2.4.0
>Reporter: Kihwal Lee
>Assignee: Kihwal Lee
>Priority: Blocker
> Attachments: HDFS-6293.000.patch, HDFS-6293.001.patch, 
> HDFS-6293.002-save-deprecated-fsimage.patch, HDFS-6293.branch-2.patch, 
> HDFS-6293.trunk.patch, HDFS-6293.trunk.patch, 
> HDFS-6293_sbn_ckpt_retention.patch, 
> HDFS-6293_sbn_ckpt_retention_oiv_legacy.patch, Heap Histogram.html
>
>
> There are issues with OIV when processing fsimages in protobuf. 
> Due to the internal layout changes introduced by the protobuf-based fsimage, 
> OIV consumes excessive amount of memory.  We have tested with a fsimage with 
> about 140M files/directories. The peak heap usage when processing this image 
> in pre-protobuf (i.e. pre-2.4.0) format was about 350MB.  After converting 
> the image to the protobuf format on 2.4.0, OIV would OOM even with 80GB of 
> heap (max new size was 1GB).  It should be possible to process any image with 
> the default heap size of 1.5GB.
> Another issue is the complete change of format/content in OIV's XML output.  
> I also noticed that the secret manager section has no tokens while there were 
> unexpired tokens in the original image (pre-2.4.0).  I did not check whether 
> they were also missing in the new pb fsimage.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-5682) Heterogeneous Storage phase 2 - APIs to expose Storage Types

2014-05-13 Thread Zesheng Wu (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5682?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13996072#comment-13996072
 ] 

Zesheng Wu commented on HDFS-5682:
--

Thanks [~ndimiduk], here are some details of our SLA requirements:
The background is that we have a structured data storage system which is very 
similar with amazon DynamoDB and is built on HBase. This system is intended to 
serve multi-tenants, in order to ensure the SLA, such as 99% read/write 
latency, we need to place the data on SSD. One straightforward way is to set up 
a hdfs cluster with all SSD data disks, but this is cost too much. So we need 
to reduce the cost. Another way is to set up a hdfs cluster with both SSD and 
HDD data disks, we can put 1 replica on SSD, and the other 2 on HDD. This is 
really a typical scenario of hdfs heterogeneous storage.

> Heterogeneous Storage phase 2 - APIs to expose Storage Types
> 
>
> Key: HDFS-5682
> URL: https://issues.apache.org/jira/browse/HDFS-5682
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 3.0.0
>Reporter: Arpit Agarwal
>Assignee: Arpit Agarwal
>
> Phase 1 (HDFS-2832) added support to present the DataNode as a collection of 
> discrete storages of different types.
> This Jira is to track phase 2 of the Heterogeneous Storage work which 
> involves exposing Storage Types to applications and adding Quota Management 
> support for administrators.
> This phase will also include tools support for administrators/users.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6250) TestBalancerWithNodeGroup.testBalancerWithRackLocality fails

2014-05-13 Thread Junping Du (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6250?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13996207#comment-13996207
 ] 

Junping Du commented on HDFS-6250:
--

Hi [~decster], patch LGTM. Would you run 50 iterations and update us your 
results here? Thx!

> TestBalancerWithNodeGroup.testBalancerWithRackLocality fails
> 
>
> Key: HDFS-6250
> URL: https://issues.apache.org/jira/browse/HDFS-6250
> Project: Hadoop HDFS
>  Issue Type: Bug
>Reporter: Kihwal Lee
>Assignee: Chen He
> Attachments: HDFS-6250-v2.patch, HDFS-6250-v3.patch, HDFS-6250.patch, 
> test_log.txt
>
>
> It was seen in https://builds.apache.org/job/PreCommit-HDFS-Build/6669/
> {panel}
> java.lang.AssertionError: expected:<1800> but was:<1810>
>   at org.junit.Assert.fail(Assert.java:93)
>   at org.junit.Assert.failNotEquals(Assert.java:647)
>   at org.junit.Assert.assertEquals(Assert.java:128)
>   at org.junit.Assert.assertEquals(Assert.java:147)
>   at org.apache.hadoop.hdfs.server.balancer.TestBalancerWithNodeGroup
>  .testBalancerWithRackLocality(TestBalancerWithNodeGroup.java:253)
> {panel}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6326) WebHdfs ACL compatibility is broken

2014-05-13 Thread Chris Nauroth (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6326?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13995821#comment-13995821
 ] 

Chris Nauroth commented on HDFS-6326:
-

I've repeated my compatibility testing on patch v4 and confirmed that it still 
works.

> WebHdfs ACL compatibility is broken
> ---
>
> Key: HDFS-6326
> URL: https://issues.apache.org/jira/browse/HDFS-6326
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: webhdfs
>Affects Versions: 3.0.0, 2.4.0
>Reporter: Daryn Sharp
>Assignee: Chris Nauroth
>Priority: Blocker
> Attachments: HDFS-6326.1.patch, HDFS-6326.2.patch, HDFS-6326.3.patch, 
> HDFS-6326.4.patch
>
>
> 2.4 ACL support is completely incompatible with <2.4 webhdfs servers.  The NN 
> throws an {{IllegalArgumentException}} exception.
> {code}
> hadoop fs -ls webhdfs://nn/
> Found 21 items
> ls: Invalid value for webhdfs parameter "op": No enum constant 
> org.apache.hadoop.hdfs.web.resources.GetOpParam.Op.GETACLSTATUS
> [... 20 more times...]
> {code}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Assigned] (HDFS-6293) Issues with OIV processing PB-based fsimages

2014-05-13 Thread Kihwal Lee (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6293?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kihwal Lee reassigned HDFS-6293:


Assignee: Kihwal Lee  (was: Haohui Mai)

> Issues with OIV processing PB-based fsimages
> 
>
> Key: HDFS-6293
> URL: https://issues.apache.org/jira/browse/HDFS-6293
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 2.4.0
>Reporter: Kihwal Lee
>Assignee: Kihwal Lee
>Priority: Blocker
> Attachments: HDFS-6293.000.patch, HDFS-6293.001.patch, 
> HDFS-6293.002-save-deprecated-fsimage.patch, 
> HDFS-6293_sbn_ckpt_retention.patch, 
> HDFS-6293_sbn_ckpt_retention_oiv_legacy.patch, Heap Histogram.html
>
>
> There are issues with OIV when processing fsimages in protobuf. 
> Due to the internal layout changes introduced by the protobuf-based fsimage, 
> OIV consumes excessive amount of memory.  We have tested with a fsimage with 
> about 140M files/directories. The peak heap usage when processing this image 
> in pre-protobuf (i.e. pre-2.4.0) format was about 350MB.  After converting 
> the image to the protobuf format on 2.4.0, OIV would OOM even with 80GB of 
> heap (max new size was 1GB).  It should be possible to process any image with 
> the default heap size of 1.5GB.
> Another issue is the complete change of format/content in OIV's XML output.  
> I also noticed that the secret manager section has no tokens while there were 
> unexpired tokens in the original image (pre-2.4.0).  I did not check whether 
> they were also missing in the new pb fsimage.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6371) In HA setup, the standby NN should redirect WebHDFS write requests to the active NN

2014-05-13 Thread Daryn Sharp (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6371?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13996394#comment-13996394
 ] 

Daryn Sharp commented on HDFS-6371:
---

Agreed on client side retries.  With IP-based failover, a standby NN that 
attempts to redirect to the other NN will create a redirect loop to itself 
until the IP failover to the active occurs.

> In HA setup, the standby NN should redirect WebHDFS write requests to the 
> active NN
> ---
>
> Key: HDFS-6371
> URL: https://issues.apache.org/jira/browse/HDFS-6371
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: namenode, webhdfs
>Reporter: Tsz Wo Nicholas Sze
>Assignee: Tsz Wo Nicholas Sze
>
> The current WebHDFS implementation in namenode does not check its HA state -- 
> it does the same thing no matter it is active or standby.
> Suppose a http client talk to the standby NN via WebHDFS.  For the read 
> operations, there is no problem.  For the write operations, if the operation 
> requires http redirect (e.g. creating a file), it will work since the standby 
> NN will also redirect the client to a DN.  When the client connect to the DN, 
> the DN will fulfill the request with the active NN.  However, for the write 
> operations not requiring http redirect (e.g. mkdir), the operation will fail 
> with StandbyException since it will be executed with the standby NN.
> There are two solutions:
> # The http client could catch StandbyException and then retries with the 
> other NN in this case.
> # The standby NN redirects the request to the active NN.
> The second solution seems better since the client does not need to know both 
> active NN and standby NN.
> Note that WebHdfsFileSystem is already able to handle HA failover.  The JIRA 
> is for other http clients.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6381) Fix a typo in INodeReference.java

2014-05-13 Thread Junping Du (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6381?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13996213#comment-13996213
 ] 

Junping Du commented on HDFS-6381:
--

Kick off Jenkins test manually.

> Fix a typo in INodeReference.java
> -
>
> Key: HDFS-6381
> URL: https://issues.apache.org/jira/browse/HDFS-6381
> Project: Hadoop HDFS
>  Issue Type: Bug
>Reporter: Binglin Chang
>Assignee: Binglin Chang
>Priority: Trivial
> Attachments: HDFS-6381.v1.patch
>
>
> hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/INodeReference.java
> {code}
>   * For example,
> - * (1) Support we have /abc/foo, say the inode of foo is 
> inode(id=1000,name=foo)
> + * (1) Suppose we have /abc/foo, say the inode of foo is 
> inode(id=1000,name=foo)
> {code}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6293) Issues with OIV processing PB-based fsimages

2014-05-13 Thread Kihwal Lee (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6293?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13996442#comment-13996442
 ] 

Kihwal Lee commented on HDFS-6293:
--

Here is recap.
* oiv_legacy for debugging and inspecting old fsimages. As long as NN is 
capable of reading and upgrading from it, we should keep it.
* NN's ability to save in old fsimage format:  This should go away once we have 
a more sustainable solution. Postprocessing ([~wheat9] 's proposal) will cover 
many use cases.  We may need something else. This is to be determined.
* Keep adding features to the old fsimage format: Please don't!
* Changing the layout of the new PB-based fsimage back to the old way: Heck no!

We will deprecate NN's ability to save in the old fsimage format, once we agree 
on what the long term solution is. We should create a separate JIRA and 
continue discussion.

As [~sureshms] suggested, I will prepare the patch for trunk as well with 
[~ajisakaa]'s  comment incorporated.

> Issues with OIV processing PB-based fsimages
> 
>
> Key: HDFS-6293
> URL: https://issues.apache.org/jira/browse/HDFS-6293
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 2.4.0
>Reporter: Kihwal Lee
>Assignee: Haohui Mai
>Priority: Blocker
> Attachments: HDFS-6293.000.patch, HDFS-6293.001.patch, 
> HDFS-6293.002-save-deprecated-fsimage.patch, 
> HDFS-6293_sbn_ckpt_retention.patch, 
> HDFS-6293_sbn_ckpt_retention_oiv_legacy.patch, Heap Histogram.html
>
>
> There are issues with OIV when processing fsimages in protobuf. 
> Due to the internal layout changes introduced by the protobuf-based fsimage, 
> OIV consumes excessive amount of memory.  We have tested with a fsimage with 
> about 140M files/directories. The peak heap usage when processing this image 
> in pre-protobuf (i.e. pre-2.4.0) format was about 350MB.  After converting 
> the image to the protobuf format on 2.4.0, OIV would OOM even with 80GB of 
> heap (max new size was 1GB).  It should be possible to process any image with 
> the default heap size of 1.5GB.
> Another issue is the complete change of format/content in OIV's XML output.  
> I also noticed that the secret manager section has no tokens while there were 
> unexpired tokens in the original image (pre-2.4.0).  I did not check whether 
> they were also missing in the new pb fsimage.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6383) Upgrade S3n s3.fs.buffer.dir to suppoer multi directories

2014-05-13 Thread Ted Malaska (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6383?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ted Malaska updated HDFS-6383:
--

Status: Patch Available  (was: Open)

Patch is ready

> Upgrade S3n s3.fs.buffer.dir to suppoer multi directories
> -
>
> Key: HDFS-6383
> URL: https://issues.apache.org/jira/browse/HDFS-6383
> Project: Hadoop HDFS
>  Issue Type: Improvement
>Reporter: Ted Malaska
>Priority: Minor
> Attachments: HDFS-6383.patch
>
>
> s3.fs.buffer.dir defines the tmp folder where files will be written to before 
> getting sent to S3.  Right now this is limited to a single folder which 
> causes to major issues.
> 1. You need a drive with enough space to store all the tmp files at once
> 2. You are limited to the IO speeds of a single drive
> This solution will resolve both and has been tested to increase the S3 write 
> speed by 2.5x with 10 mappers on hs1.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6373) Remove support for extended attributes on symlinks

2014-05-13 Thread Charles Lamb (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6373?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13996425#comment-13996425
 ] 

Charles Lamb commented on HDFS-6373:


While investigating this, I noticed this in the TestNameNodeXAttr output. I 
suspect it's not intended, but I'll investigate further:

2014-05-13 10:04:33,584 INFO  hdfs.StateChange 
(FSNamesystem.java:completeFile(3028)) - DIR* completeFile: /symdir2/target is 
closed by DFSClient_NONMAPREDUCE_239063403_1
2014-05-13 10:04:33,590 INFO  FSNamesystem.audit 
(FSNamesystem.java:logAuditMessage(7980)) - allowed=true   ugi=cwl 
(auth:SIMPLE)   ip=/127.0.0.1   cmd=createSymlink   src=/symdir1/link   
dst=/symdir2/target perm=cwl:supergroup:rwxrwxrwx   proto=rpc
2014-05-13 10:04:33,601 INFO  FSNamesystem.audit 
(FSNamesystem.java:logAuditMessage(7980)) - allowed=true   ugi=cwl 
(auth:SIMPLE)   ip=/127.0.0.1   cmd=setXAttrsrc=/symdir2/target 
dst=nullperm=cwl:supergroup:rw-r--r--   proto=rpc
2014-05-13 10:04:33,605 INFO  FSNamesystem.audit 
(FSNamesystem.java:logAuditMessage(7980)) - allowed=true   ugi=cwl 
(auth:SIMPLE)   ip=/127.0.0.1   cmd=setXAttrsrc=/symdir2/target 
dst=nullperm=cwl:supergroup:rw-r--r--   proto=rpc
2014-05-13 10:04:33,608 INFO  ipc.Server (Server.java:run(2120)) - IPC Server 
handler 9 on 38622, call 
org.apache.hadoop.hdfs.protocol.ClientProtocol.getXAttrs from 127.0.0.1:41722 
Call#18 Retry#0
org.apache.hadoop.hdfs.protocol.UnresolvedPathException: /symdir2/target
at 
org.apache.hadoop.hdfs.server.namenode.INodesInPath.resolve(INodesInPath.java:203)
at 
org.apache.hadoop.hdfs.server.namenode.FSDirectory.getLastINodeInPath(FSDirectory.java:3135)
at 
org.apache.hadoop.hdfs.server.namenode.FSDirectory.getXAttrs(FSDirectory.java:2952)
at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getXAttrs(FSNamesystem.java:7858)
at 
org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.getXAttrs(NameNodeRpcServer.java:1396)
at 
org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.getXAttrs(ClientNamenodeProtocolServerSideTranslatorPB.java:1294)
at 
org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
at 
org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:605)
at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:932)
at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2099)
at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2095)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:415)
at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1608)
at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2093)

> Remove support for extended attributes on symlinks
> --
>
> Key: HDFS-6373
> URL: https://issues.apache.org/jira/browse/HDFS-6373
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: namenode
>Reporter: Andrew Wang
>Assignee: Charles Lamb
>
> Looking in the Linux source code, we see the following:
> http://lxr.linux.no/linux+v3.14.3/fs/xattr.c
> {code}
>   60/*
>   61 * In the user.* namespace, only regular files and directories 
> can have
>   62 * extended attributes. For sticky directories, only the owner and
>   63 * privileged users can write attributes.
>   64 */
> {code}
> We should consider removing {{XAttrFeature}} from {{INodeSymlink}}.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6186) Pause deletion of blocks when the namenode starts up

2014-05-13 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6186?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13996194#comment-13996194
 ] 

Hadoop QA commented on HDFS-6186:
-

{color:green}+1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12644511/HDFS-6186.003.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 1 new 
or modified test files.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  There were no new javadoc warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:green}+1 core tests{color}.  The patch passed unit tests in 
hadoop-hdfs-project/hadoop-hdfs.

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-HDFS-Build/6890//testReport/
Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/6890//console

This message is automatically generated.

> Pause deletion of blocks when the namenode starts up
> 
>
> Key: HDFS-6186
> URL: https://issues.apache.org/jira/browse/HDFS-6186
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: namenode
>Reporter: Suresh Srinivas
>Assignee: Jing Zhao
> Attachments: HDFS-6186.000.patch, HDFS-6186.002.patch, 
> HDFS-6186.003.patch
>
>
> HDFS namenode can delete blocks very quickly, given the deletion happens as a 
> parallel operation spread across many datanodes. One of the frequent 
> anxieties I see is that a lot of data can be deleted very quickly, when a 
> cluster is brought up, especially when one of the storage directories has 
> failed and namenode metadata was copied from another storage. Copying wrong 
> metadata would results in some of the newer files (if old metadata was 
> copied) being deleted along with their blocks. 
> HDFS-5986 now captures the number of pending deletion block on namenode webUI 
> and JMX. I propose pausing deletion of blocks for a configured period of time 
> (default 1 hour?) after namenode comes out of safemode. This will give enough 
> time for the administrator to notice large number of pending deletion blocks 
> and take corrective action.
> Thoughts?



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6345) DFS.listCacheDirectives() should allow filtering based on cache directive ID

2014-05-13 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6345?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13996149#comment-13996149
 ] 

Hadoop QA commented on HDFS-6345:
-

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12644544/hdfs-6345-1.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 2 new 
or modified test files.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  There were no new javadoc warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:red}-1 core tests{color}.  The patch failed these unit tests in 
hadoop-hdfs-project/hadoop-hdfs:

  org.apache.hadoop.hdfs.TestDistributedFileSystem

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-HDFS-Build/6888//testReport/
Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/6888//console

This message is automatically generated.

> DFS.listCacheDirectives() should allow filtering based on cache directive ID
> 
>
> Key: HDFS-6345
> URL: https://issues.apache.org/jira/browse/HDFS-6345
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: caching
>Affects Versions: 2.4.0
>Reporter: Lenni Kuff
>Assignee: Andrew Wang
> Attachments: hdfs-6345-1.patch
>
>
> DFS.listCacheDirectives() should allow filtering based on cache directive ID. 
> Currently it throws an exception.
> For example:
> {code}
> long directiveId = ;
> CacheDirectiveInfo filter = new CacheDirectiveInfo.Builder()  
> .setId(directiveId)
> .build();
> RemoteIterator itr = dfs.listCacheDirectives(filter);
> {code}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6377) Unify xattr name and value limits into a single limit

2014-05-13 Thread Charles Lamb (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6377?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13996396#comment-13996396
 ] 

Charles Lamb commented on HDFS-6377:


In general, LGTM. One nit (piling on Uma's comment):

+  "XAttr is too big, maximum size = " + nnConf.xattrMaxSize
+  + ", but the size is = " + xAttr.getName().length());

Could we change the wording to

"The XAttr is too big. The maximum size of the name + value is 
, but the total size is  Unify xattr name and value limits into a single limit
> -
>
> Key: HDFS-6377
> URL: https://issues.apache.org/jira/browse/HDFS-6377
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: namenode
>Affects Versions: HDFS XAttrs (HDFS-2006)
>Reporter: Andrew Wang
>Assignee: Andrew Wang
> Attachments: hdfs-6377-1.patch
>
>
> Instead of having separate limits and config options for the size of an 
> xattr's name and value, let's use a single limit.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6325) Append should fail if the last block has insufficient number of replicas

2014-05-13 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6325?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13995837#comment-13995837
 ] 

Hadoop QA commented on HDFS-6325:
-

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12644461/HDFS-6325.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 1 new 
or modified test files.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  There were no new javadoc warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:red}-1 core tests{color}.  The patch failed these unit tests in 
hadoop-hdfs-project/hadoop-hdfs:

  
org.apache.hadoop.hdfs.server.balancer.TestBalancerWithNodeGroup
  org.apache.hadoop.hdfs.TestFileAppend4

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-HDFS-Build/6879//testReport/
Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/6879//console

This message is automatically generated.

> Append should fail if the last block has insufficient number of replicas
> 
>
> Key: HDFS-6325
> URL: https://issues.apache.org/jira/browse/HDFS-6325
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: namenode
>Affects Versions: 2.2.0
>Reporter: Konstantin Shvachko
>Assignee: Keith Pak
> Attachments: HDFS-6325.patch, HDFS-6325.patch, HDFS-6325_test.patch, 
> appendTest.patch
>
>
> Currently append() succeeds on a file with the last block that has no 
> replicas. But the subsequent updatePipeline() fails as there are no replicas 
> with the exception "Unable to retrieve blocks locations for last block". This 
> leaves the file unclosed, and others can not do anything with it until its 
> lease expires.
> The solution is to check replicas of the last block on the NameNode and fail 
> during append() rather than during updatePipeline().
> How many replicas should be present before NN allows to append? I see two 
> options:
> # min-replication: allow append if the last block is minimally replicated (1 
> by default)
> # full-replication: allow append if the last block is fully replicated (3 by 
> default)



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6293) Issues with OIV processing PB-based fsimages

2014-05-13 Thread Kihwal Lee (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6293?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13995444#comment-13995444
 ] 

Kihwal Lee commented on HDFS-6293:
--

bq. can you please mark the class that are brought back deprecated?
I am not sure deprecating it now is a good idea.  There are two apsects in this.
* Ability to read the old fsimage format in OIV. This has always been the case 
for OIV. Now it will be the combination of oiv and oiv_legacy that will 
continue this.   As long as NN can read/upgrade old images, there is no reason 
to hurry and remove this capability from oiv.
* Support for the use case built around the old oiv.  Users won't care much 
about the image format as long as they can obtain what they want, using 
reasonable and comparable amount of resources as before.  If we only think 
about this, we can deprecate OIV and NN's ability to save the old format, once 
we have a long term alternative solution.


> Issues with OIV processing PB-based fsimages
> 
>
> Key: HDFS-6293
> URL: https://issues.apache.org/jira/browse/HDFS-6293
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 2.4.0
>Reporter: Kihwal Lee
>Assignee: Haohui Mai
>Priority: Blocker
> Attachments: HDFS-6293.000.patch, HDFS-6293.001.patch, 
> HDFS-6293.002-save-deprecated-fsimage.patch, 
> HDFS-6293_sbn_ckpt_retention.patch, 
> HDFS-6293_sbn_ckpt_retention_oiv_legacy.patch, Heap Histogram.html
>
>
> There are issues with OIV when processing fsimages in protobuf. 
> Due to the internal layout changes introduced by the protobuf-based fsimage, 
> OIV consumes excessive amount of memory.  We have tested with a fsimage with 
> about 140M files/directories. The peak heap usage when processing this image 
> in pre-protobuf (i.e. pre-2.4.0) format was about 350MB.  After converting 
> the image to the protobuf format on 2.4.0, OIV would OOM even with 80GB of 
> heap (max new size was 1GB).  It should be possible to process any image with 
> the default heap size of 1.5GB.
> Another issue is the complete change of format/content in OIV's XML output.  
> I also noticed that the secret manager section has no tokens while there were 
> unexpired tokens in the original image (pre-2.4.0).  I did not check whether 
> they were also missing in the new pb fsimage.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6328) Clean up dead code in FSDirectory

2014-05-13 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6328?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13995409#comment-13995409
 ] 

Hudson commented on HDFS-6328:
--

SUCCESS: Integrated in Hadoop-trunk-Commit #5603 (See 
[https://builds.apache.org/job/Hadoop-trunk-Commit/5603/])
HDFS-6328. Clean up dead code in FSDirectory. Contributed by Haohui Mai. 
(wheat9: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1593755)
* /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSDirectory.java


> Clean up dead code in FSDirectory
> -
>
> Key: HDFS-6328
> URL: https://issues.apache.org/jira/browse/HDFS-6328
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: namenode
>Reporter: Haohui Mai
>Assignee: Haohui Mai
> Attachments: HDFS-6328.000.patch, HDFS-6328.001.patch, 
> HDFS-6328.002.patch, HDFS-6328.003.patch, HDFS-6328.004.patch
>
>
> This jira proposes:
> # Cleaning up dead code in FSDirectory.
> # Simplify the control flows that IntelliJ flags as warnings.
> # Move functions related to resolving paths into one place.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (HDFS-6382) HDFS File/Directory TTL

2014-05-13 Thread Zesheng Wu (JIRA)
Zesheng Wu created HDFS-6382:


 Summary: HDFS File/Directory TTL
 Key: HDFS-6382
 URL: https://issues.apache.org/jira/browse/HDFS-6382
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: hdfs-client, namenode
Affects Versions: 2.4.0
Reporter: Zesheng Wu


In production environment, we always have scenario like this, we want to backup 
files on hdfs for some time and then hope to delete theses files automatically. 
For example, we keep only 1 day's logs on local disk due to limited disk space, 
but we need to keep about 1 month's logs in order to debug program bugs, so we 
keep all the logs on hdfs and delete logs which are older than 1 month. This is 
a typical scenario of HDFS TTL. So here we propose that hdfs can support TTL.

Following are some details of this proposal:
1. HDFS can support TTL on a specified file or directory
2. If a TTL is set on a file, the file will be deleted automatically after the 
TTL is expired
3. If a TTL is set on a directory, the child files and directories will be 
deleted automatically after the TTL is expired
4. The child file/directory's TTL configuration should override its parent 
directory's
5. A global configuration is needed to configure that whether the deleted 
files/directories should go to the trash or not
6. A global configuration is needed to configure that whether a directory with 
TTL should be deleted when it is emptied by TTL mechanism or not.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6230) Expose upgrade status through NameNode web UI

2014-05-13 Thread Haohui Mai (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6230?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13995340#comment-13995340
 ] 

Haohui Mai commented on HDFS-6230:
--

+1. I'll commit it shortly.

> Expose upgrade status through NameNode web UI
> -
>
> Key: HDFS-6230
> URL: https://issues.apache.org/jira/browse/HDFS-6230
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: namenode
>Affects Versions: 2.4.0
>Reporter: Arpit Agarwal
>Assignee: Mit Desai
> Attachments: HDFS-6230-NoUpgradesInProgress.png, 
> HDFS-6230-UpgradeInProgress.jpg, HDFS-6230.patch, HDFS-6230.patch
>
>
> The NameNode web UI does not show upgrade information anymore. Hadoop 2.0 
> also does not have the _hadoop dfsadmin -upgradeProgress_ command to check 
> the upgrade status.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6305) WebHdfs response decoding may throw RuntimeExceptions

2014-05-13 Thread Jonathan Eagles (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6305?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13995737#comment-13995737
 ] 

Jonathan Eagles commented on HDFS-6305:
---

Kicked jenkins. there are 3 or 4 builds ahead of this one. It may be late 
tonight to see the results.

Jon

> WebHdfs response decoding may throw RuntimeExceptions
> -
>
> Key: HDFS-6305
> URL: https://issues.apache.org/jira/browse/HDFS-6305
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: webhdfs
>Affects Versions: 2.0.0-alpha, 3.0.0
>Reporter: Daryn Sharp
>Assignee: Daryn Sharp
>Priority: Critical
> Attachments: HDFS-6305.patch
>
>
> WebHdfs does not guard against exceptions while decoding the response 
> payload.  The json parser will throw RunTime exceptions on malformed 
> responses.  The json decoding routines do not validate the expected fields 
> are present which may cause NPEs.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (HDFS-6379) HTTPFS - Implement ACLs support

2014-05-13 Thread Alejandro Abdelnur (JIRA)
Alejandro Abdelnur created HDFS-6379:


 Summary: HTTPFS - Implement ACLs support
 Key: HDFS-6379
 URL: https://issues.apache.org/jira/browse/HDFS-6379
 Project: Hadoop HDFS
  Issue Type: Bug
Reporter: Alejandro Abdelnur
Assignee: Alejandro Abdelnur
 Fix For: 2.4.0


HDFS-4685 added ACLs support to WebHDFS but missed adding them to HttpFS.

This JIRA is for such.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6345) DFS.listCacheDirectives() should allow filtering based on cache directive ID

2014-05-13 Thread Andrew Wang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6345?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Andrew Wang updated HDFS-6345:
--

Attachment: hdfs-6345-1.patch

Patch attached. I did manual testing with a branch-2 NN and the new client, 
verified that it still filters correctly, and that we get the same exception 
thrown on error.

> DFS.listCacheDirectives() should allow filtering based on cache directive ID
> 
>
> Key: HDFS-6345
> URL: https://issues.apache.org/jira/browse/HDFS-6345
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: caching
>Affects Versions: 2.4.0
>Reporter: Lenni Kuff
>Assignee: Andrew Wang
> Attachments: hdfs-6345-1.patch
>
>
> DFS.listCacheDirectives() should allow filtering based on cache directive ID. 
> Currently it throws an exception.
> For example:
> {code}
> long directiveId = ;
> CacheDirectiveInfo filter = new CacheDirectiveInfo.Builder()  
> .setId(directiveId)
> .build();
> RemoteIterator itr = dfs.listCacheDirectives(filter);
> {code}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6134) Transparent data at rest encryption

2014-05-13 Thread Steve Loughran (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6134?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13996245#comment-13996245
 ] 

Steve Loughran commented on HDFS-6134:
--

Some side comments
# The ASF don't want to be distributing cryptography libraries; the design has 
to address that distribution model.
# HADOOP-9565 proposes a {{BlobStore}} extending FileSystem to let users know 
that the FS doesn't have the semantics of a real filesystem -for blobstore 
security we'd need the same marker, so apps using a wrapped blobstore will know 
that some operations are non-atomic and potentially O(files)

> Transparent data at rest encryption
> ---
>
> Key: HDFS-6134
> URL: https://issues.apache.org/jira/browse/HDFS-6134
> Project: Hadoop HDFS
>  Issue Type: New Feature
>  Components: security
>Affects Versions: 2.3.0
>Reporter: Alejandro Abdelnur
>Assignee: Alejandro Abdelnur
> Attachments: HDFSDataAtRestEncryption.pdf
>
>
> Because of privacy and security regulations, for many industries, sensitive 
> data at rest must be in encrypted form. For example: the health­care industry 
> (HIPAA regulations), the card payment industry (PCI DSS regulations) or the 
> US government (FISMA regulations).
> This JIRA aims to provide a mechanism to encrypt HDFS data at rest that can 
> be used transparently by any application accessing HDFS via Hadoop Filesystem 
> Java API, Hadoop libhdfs C library, or WebHDFS REST API.
> The resulting implementation should be able to be used in compliance with 
> different regulation requirements.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6374) setXAttr should require the user to be the owner of the file or directory

2014-05-13 Thread Yi Liu (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6374?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13995988#comment-13995988
 ] 

Yi Liu commented on HDFS-6374:
--

setXAttr/removeXAttr are only for owner, and getXAttr are *not* only for Owner 
and by "r" permission.

> setXAttr should require the user to be the owner of the file or directory
> -
>
> Key: HDFS-6374
> URL: https://issues.apache.org/jira/browse/HDFS-6374
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: namenode
>Affects Versions: HDFS XAttrs (HDFS-2006)
>Reporter: Andrew Wang
>Assignee: Charles Lamb
>
> From the attr(5) manpage:
> {noformat}
>For  this reason, extended user attributes are only allowed for regular
>files and directories,  and  access  to  extended  user  attributes  is
>restricted  to the owner and to users with appropriate capabilities for
>directories with the sticky bit set (see the chmod(1) manual  page  for
>an explanation of Sticky Directories).
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6382) HDFS File/Directory TTL

2014-05-13 Thread Zesheng Wu (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6382?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zesheng Wu updated HDFS-6382:
-

Description: 
In production environment, we always have scenario like this, we want to backup 
files on hdfs for some time and then hope to delete these files automatically. 
For example, we keep only 1 day's logs on local disk due to limited disk space, 
but we need to keep about 1 month's logs in order to debug program bugs, so we 
keep all the logs on hdfs and delete logs which are older than 1 month. This is 
a typical scenario of HDFS TTL. So here we propose that hdfs can support TTL.

Following are some details of this proposal:
1. HDFS can support TTL on a specified file or directory
2. If a TTL is set on a file, the file will be deleted automatically after the 
TTL is expired
3. If a TTL is set on a directory, the child files and directories will be 
deleted automatically after the TTL is expired
4. The child file/directory's TTL configuration should override its parent 
directory's
5. A global configuration is needed to configure that whether the deleted 
files/directories should go to the trash or not
6. A global configuration is needed to configure that whether a directory with 
TTL should be deleted when it is emptied by TTL mechanism or not.

  was:
In production environment, we always have scenario like this, we want to backup 
files on hdfs for some time and then hope to delete theses files automatically. 
For example, we keep only 1 day's logs on local disk due to limited disk space, 
but we need to keep about 1 month's logs in order to debug program bugs, so we 
keep all the logs on hdfs and delete logs which are older than 1 month. This is 
a typical scenario of HDFS TTL. So here we propose that hdfs can support TTL.

Following are some details of this proposal:
1. HDFS can support TTL on a specified file or directory
2. If a TTL is set on a file, the file will be deleted automatically after the 
TTL is expired
3. If a TTL is set on a directory, the child files and directories will be 
deleted automatically after the TTL is expired
4. The child file/directory's TTL configuration should override its parent 
directory's
5. A global configuration is needed to configure that whether the deleted 
files/directories should go to the trash or not
6. A global configuration is needed to configure that whether a directory with 
TTL should be deleted when it is emptied by TTL mechanism or not.


> HDFS File/Directory TTL
> ---
>
> Key: HDFS-6382
> URL: https://issues.apache.org/jira/browse/HDFS-6382
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: hdfs-client, namenode
>Affects Versions: 2.4.0
>Reporter: Zesheng Wu
>
> In production environment, we always have scenario like this, we want to 
> backup files on hdfs for some time and then hope to delete these files 
> automatically. For example, we keep only 1 day's logs on local disk due to 
> limited disk space, but we need to keep about 1 month's logs in order to 
> debug program bugs, so we keep all the logs on hdfs and delete logs which are 
> older than 1 month. This is a typical scenario of HDFS TTL. So here we 
> propose that hdfs can support TTL.
> Following are some details of this proposal:
> 1. HDFS can support TTL on a specified file or directory
> 2. If a TTL is set on a file, the file will be deleted automatically after 
> the TTL is expired
> 3. If a TTL is set on a directory, the child files and directories will be 
> deleted automatically after the TTL is expired
> 4. The child file/directory's TTL configuration should override its parent 
> directory's
> 5. A global configuration is needed to configure that whether the deleted 
> files/directories should go to the trash or not
> 6. A global configuration is needed to configure that whether a directory 
> with TTL should be deleted when it is emptied by TTL mechanism or not.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Assigned] (HDFS-6377) Unify xattr name and value limits into a single limit

2014-05-13 Thread Andrew Wang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6377?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Andrew Wang reassigned HDFS-6377:
-

Assignee: Andrew Wang

> Unify xattr name and value limits into a single limit
> -
>
> Key: HDFS-6377
> URL: https://issues.apache.org/jira/browse/HDFS-6377
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: namenode
>Affects Versions: HDFS XAttrs (HDFS-2006)
>Reporter: Andrew Wang
>Assignee: Andrew Wang
>
> Instead of having separate limits and config options for the size of an 
> xattr's name and value, let's use a single limit.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


  1   2   >