[jira] [Updated] (HDFS-5840) Follow-up to HDFS-5138 to improve error handling during partial upgrade failures

2014-03-24 Thread Jing Zhao (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5840?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jing Zhao updated HDFS-5840:


   Resolution: Fixed
Fix Version/s: 2.4.0
 Hadoop Flags: Reviewed
   Status: Resolved  (was: Patch Available)

I've committed this.

> Follow-up to HDFS-5138 to improve error handling during partial upgrade 
> failures
> 
>
> Key: HDFS-5840
> URL: https://issues.apache.org/jira/browse/HDFS-5840
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: ha, journal-node, namenode
>Affects Versions: 3.0.0, 2.4.0
>Reporter: Aaron T. Myers
>Assignee: Jing Zhao
>Priority: Blocker
> Fix For: 2.4.0
>
> Attachments: HDFS-5840.001.patch, HDFS-5840.002.patch, 
> HDFS-5840.003.patch, HDFS-5840.patch
>
>
> Suresh posted some good comment in HDFS-5138 after that patch had already 
> been committed to trunk. This JIRA is to address those. See the first comment 
> of this JIRA for the full content of the review.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6153) Document "fileId" and "childrenNum" fields in the FileStatus Json schema

2014-03-24 Thread Akira AJISAKA (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6153?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Akira AJISAKA updated HDFS-6153:


Status: Patch Available  (was: Open)

> Document "fileId" and "childrenNum" fields in the FileStatus Json schema
> 
>
> Key: HDFS-6153
> URL: https://issues.apache.org/jira/browse/HDFS-6153
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: documentation, webhdfs
>Affects Versions: 2.3.0
>Reporter: Akira AJISAKA
>Assignee: Akira AJISAKA
>Priority: Minor
>  Labels: newbie
> Attachments: HDFS-6153.patch
>
>
> Now WebHDFS returns FileStatus Json objects include "fileId" and 
> "childrenNum" fields but these fields are not documented.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6153) Document "fileId" and "childrenNum" fields in the FileStatus Json schema

2014-03-24 Thread Akira AJISAKA (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6153?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Akira AJISAKA updated HDFS-6153:


Attachment: HDFS-6153.patch

Attaching a patch.

> Document "fileId" and "childrenNum" fields in the FileStatus Json schema
> 
>
> Key: HDFS-6153
> URL: https://issues.apache.org/jira/browse/HDFS-6153
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: documentation, webhdfs
>Affects Versions: 2.3.0
>Reporter: Akira AJISAKA
>Assignee: Akira AJISAKA
>Priority: Minor
>  Labels: newbie
> Attachments: HDFS-6153.patch
>
>
> Now WebHDFS returns FileStatus Json objects include "fileId" and 
> "childrenNum" fields but these fields are not documented.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6153) Document "fileId" and "childrenNum" fields in the FileStatus Json schema

2014-03-24 Thread Akira AJISAKA (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6153?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Akira AJISAKA updated HDFS-6153:


Summary: Document "fileId" and "childrenNum" fields in the FileStatus Json 
schema  (was: Add "fileId" and "childrenNum" fields to Json schema in the 
WebHDFS document)

> Document "fileId" and "childrenNum" fields in the FileStatus Json schema
> 
>
> Key: HDFS-6153
> URL: https://issues.apache.org/jira/browse/HDFS-6153
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: documentation, webhdfs
>Affects Versions: 2.3.0
>Reporter: Akira AJISAKA
>Assignee: Akira AJISAKA
>Priority: Minor
>  Labels: newbie
>
> Now WebHDFS returns FileStatus Json objects include "fileId" and 
> "childrenNum" fields but these fields are not documented.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-5840) Follow-up to HDFS-5138 to improve error handling during partial upgrade failures

2014-03-24 Thread Jing Zhao (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5840?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13946237#comment-13946237
 ] 

Jing Zhao commented on HDFS-5840:
-

Thanks for the review, Suresh and Aaron! +1 for the 003 patch. I will commit it 
soon.

> Follow-up to HDFS-5138 to improve error handling during partial upgrade 
> failures
> 
>
> Key: HDFS-5840
> URL: https://issues.apache.org/jira/browse/HDFS-5840
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: ha, journal-node, namenode
>Affects Versions: 3.0.0, 2.4.0
>Reporter: Aaron T. Myers
>Assignee: Jing Zhao
>Priority: Blocker
> Attachments: HDFS-5840.001.patch, HDFS-5840.002.patch, 
> HDFS-5840.003.patch, HDFS-5840.patch
>
>
> Suresh posted some good comment in HDFS-5138 after that patch had already 
> been committed to trunk. This JIRA is to address those. See the first comment 
> of this JIRA for the full content of the review.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-5978) Create a tool to take fsimage and expose read-only WebHDFS API

2014-03-24 Thread Akira AJISAKA (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5978?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Akira AJISAKA updated HDFS-5978:


Attachment: HDFS-5978.7.patch

Thanks Haohui for the comment. Updated the patch.

> Create a tool to take fsimage and expose read-only WebHDFS API
> --
>
> Key: HDFS-5978
> URL: https://issues.apache.org/jira/browse/HDFS-5978
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: tools
>Reporter: Akira AJISAKA
>Assignee: Akira AJISAKA
>  Labels: newbie
> Attachments: HDFS-5978.2.patch, HDFS-5978.3.patch, HDFS-5978.4.patch, 
> HDFS-5978.5.patch, HDFS-5978.6.patch, HDFS-5978.7.patch, HDFS-5978.patch
>
>
> Suggested in HDFS-5975.
> Add an option to OfflineImageViewer to launch HTTP server that takes fsimage 
> and exposes the read-only version of WebHDFS API for OfflineImageViewer. The 
> server only supports {{LISTSTATUS}} operation and other operations such as 
> {{GETFILESTATUS}} will be added by other JIRAs.
> That way we can allow the operator to use the existing command-line tool, or 
> even the web UI to debug the fsimage. It also allows the operator to 
> interactively browsing the file system, figuring out what goes wrong.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6134) Transparent data at rest encryption

2014-03-24 Thread Alejandro Abdelnur (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6134?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13946222#comment-13946222
 ] 

Alejandro Abdelnur commented on HDFS-6134:
--

[~lmccay], I've just opened the following JIRAs for the KeyProvider API 
improvements and a KeyProvider server: HADOOP-10427, HADOOP-10428, 
HADOOP-10429, HADOOP-10430, HADOOP-10431, HADOOP-10432 & HADOOP-10433. I've 
already posted patches for all but the last one (I'll try to get it ready by 
EOW). 

> Transparent data at rest encryption
> ---
>
> Key: HDFS-6134
> URL: https://issues.apache.org/jira/browse/HDFS-6134
> Project: Hadoop HDFS
>  Issue Type: New Feature
>  Components: security
>Affects Versions: 2.3.0
>Reporter: Alejandro Abdelnur
>Assignee: Alejandro Abdelnur
> Attachments: HDFSDataAtRestEncryption.pdf
>
>
> Because of privacy and security regulations, for many industries, sensitive 
> data at rest must be in encrypted form. For example: the health­care industry 
> (HIPAA regulations), the card payment industry (PCI DSS regulations) or the 
> US government (FISMA regulations).
> This JIRA aims to provide a mechanism to encrypt HDFS data at rest that can 
> be used transparently by any application accessing HDFS via Hadoop Filesystem 
> Java API, Hadoop libhdfs C library, or WebHDFS REST API.
> The resulting implementation should be able to be used in compliance with 
> different regulation requirements.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-5840) Follow-up to HDFS-5138 to improve error handling during partial upgrade failures

2014-03-24 Thread Suresh Srinivas (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5840?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Suresh Srinivas updated HDFS-5840:
--

Attachment: HDFS-5840.003.patch

+1 for the patch. I imported DFSUtil and minor edit to fix the javadoc link 
error to DFSUtil#getRpcAddressesForNameserviceId.

> Follow-up to HDFS-5138 to improve error handling during partial upgrade 
> failures
> 
>
> Key: HDFS-5840
> URL: https://issues.apache.org/jira/browse/HDFS-5840
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: ha, journal-node, namenode
>Affects Versions: 3.0.0, 2.4.0
>Reporter: Aaron T. Myers
>Assignee: Jing Zhao
>Priority: Blocker
> Attachments: HDFS-5840.001.patch, HDFS-5840.002.patch, 
> HDFS-5840.003.patch, HDFS-5840.patch
>
>
> Suresh posted some good comment in HDFS-5138 after that patch had already 
> been committed to trunk. This JIRA is to address those. See the first comment 
> of this JIRA for the full content of the review.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6133) Make Balancer support exclude specified path

2014-03-24 Thread zhaoyunjiong (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6133?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

zhaoyunjiong updated HDFS-6133:
---

Status: Patch Available  (was: Open)

> Make Balancer support exclude specified path
> 
>
> Key: HDFS-6133
> URL: https://issues.apache.org/jira/browse/HDFS-6133
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: balancer, namenode
>Reporter: zhaoyunjiong
>Assignee: zhaoyunjiong
> Attachments: HDFS-6133.patch
>
>
> Currently, run Balancer will destroying Regionserver's data locality.
> If getBlocks could exclude blocks belongs to files which have specific path 
> prefix, like "/hbase", then we can run Balancer without destroying 
> Regionserver's data locality.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-5978) Create a tool to take fsimage and expose read-only WebHDFS API

2014-03-24 Thread Haohui Mai (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5978?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13946214#comment-13946214
 ] 

Haohui Mai commented on HDFS-5978:
--

You'll need to add all client-side channels into the {{allChannels}} objects, 
by putting a new pipeline stage at the very beginning of the pipeline:

{code}
@Override
public void channelOpen(ChannelHandlerContext ctx, ChannelStateEvent e) {
allChannels.add(e.getChannel());
}
{code}

> Create a tool to take fsimage and expose read-only WebHDFS API
> --
>
> Key: HDFS-5978
> URL: https://issues.apache.org/jira/browse/HDFS-5978
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: tools
>Reporter: Akira AJISAKA
>Assignee: Akira AJISAKA
>  Labels: newbie
> Attachments: HDFS-5978.2.patch, HDFS-5978.3.patch, HDFS-5978.4.patch, 
> HDFS-5978.5.patch, HDFS-5978.6.patch, HDFS-5978.patch
>
>
> Suggested in HDFS-5975.
> Add an option to OfflineImageViewer to launch HTTP server that takes fsimage 
> and exposes the read-only version of WebHDFS API for OfflineImageViewer. The 
> server only supports {{LISTSTATUS}} operation and other operations such as 
> {{GETFILESTATUS}} will be added by other JIRAs.
> That way we can allow the operator to use the existing command-line tool, or 
> even the web UI to debug the fsimage. It also allows the operator to 
> interactively browsing the file system, figuring out what goes wrong.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6125) Cleanup unnecessary cast in HDFS code base

2014-03-24 Thread Suresh Srinivas (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6125?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Suresh Srinivas updated HDFS-6125:
--

   Resolution: Fixed
Fix Version/s: 2.5.0
   Status: Resolved  (was: Patch Available)

I've committed the patch to trunk, branch-2. Thanks [~cnauroth] for the review 
and patch rebase.

> Cleanup unnecessary cast in HDFS code base
> --
>
> Key: HDFS-6125
> URL: https://issues.apache.org/jira/browse/HDFS-6125
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: test
>Affects Versions: 2.4.0
>Reporter: Suresh Srinivas
>Assignee: Suresh Srinivas
> Fix For: 2.5.0
>
> Attachments: HDFS-6125.2.patch, HDFS-6125.patch
>
>




--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6150) Add inode id information in the logs to make debugging easier

2014-03-24 Thread Suresh Srinivas (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6150?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Suresh Srinivas updated HDFS-6150:
--

Attachment: HDFS-6150.3.patch

Updated patch to fix the test failure.

> Add inode id information in the logs to make debugging easier
> -
>
> Key: HDFS-6150
> URL: https://issues.apache.org/jira/browse/HDFS-6150
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: namenode
>Reporter: Suresh Srinivas
>Assignee: Suresh Srinivas
>Priority: Minor
> Attachments: HDFS-6150.1.patch, HDFS-6150.2.patch, HDFS-6150.3.patch, 
> HDFS-6150.patch
>
>
> Inode information and path information are missing in the logs and 
> exceptions. Adding this will help debug multi threading issues related to 
> using incode INode ID information.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-5978) Create a tool to take fsimage and expose read-only WebHDFS API

2014-03-24 Thread Akira AJISAKA (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5978?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Akira AJISAKA updated HDFS-5978:


Attachment: HDFS-5978.6.patch

Updated the patch.

> Create a tool to take fsimage and expose read-only WebHDFS API
> --
>
> Key: HDFS-5978
> URL: https://issues.apache.org/jira/browse/HDFS-5978
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: tools
>Reporter: Akira AJISAKA
>Assignee: Akira AJISAKA
>  Labels: newbie
> Attachments: HDFS-5978.2.patch, HDFS-5978.3.patch, HDFS-5978.4.patch, 
> HDFS-5978.5.patch, HDFS-5978.6.patch, HDFS-5978.patch
>
>
> Suggested in HDFS-5975.
> Add an option to OfflineImageViewer to launch HTTP server that takes fsimage 
> and exposes the read-only version of WebHDFS API for OfflineImageViewer. The 
> server only supports {{LISTSTATUS}} operation and other operations such as 
> {{GETFILESTATUS}} will be added by other JIRAs.
> That way we can allow the operator to use the existing command-line tool, or 
> even the web UI to debug the fsimage. It also allows the operator to 
> interactively browsing the file system, figuring out what goes wrong.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6130) NPE during namenode upgrade from old release

2014-03-24 Thread Fengdong Yu (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6130?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13946201#comment-13946201
 ] 

Fengdong Yu commented on HDFS-6130:
---

[~wheat9], 

fsimage was uploaded.
please read my following steps carefully before fix the bug.

1)There is no HA enabled during these steps.
2)all test files are all less than one block size

a. start hadoop-1.0.4 hdfs
b. put  one files on the hdfs
c. stop hdfs.
d. start dfs with upgrade option to the lastest trunk
e. put more than ten files on the hdfs
f. stop hdfs
g. start hdfs  (NPE here)

NOTE. if put a few files(such as one file) at step e, there is no NPE at step g.



> NPE during namenode upgrade from old release
> 
>
> Key: HDFS-6130
> URL: https://issues.apache.org/jira/browse/HDFS-6130
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: namenode
>Affects Versions: 2.4.0
>Reporter: Fengdong Yu
> Attachments: fsimage.tar.gz
>
>
> I want upgrade an old cluster(0.20.2-cdh3u1) to trunk instance, 
> I can upgrade successfully if I don't configurage HA, but if HA enabled,
> there is NPE when I run ' hdfs namenode -initializeSharedEdits'
> {code}
> 14/03/20 15:06:41 INFO namenode.FSNamesystem: Retry cache on namenode is 
> enabled
> 14/03/20 15:06:41 INFO namenode.FSNamesystem: Retry cache will use 0.03 of 
> total heap and retry cache entry expiry time is 60 millis
> 14/03/20 15:06:41 INFO util.GSet: Computing capacity for map 
> NameNodeRetryCache
> 14/03/20 15:06:41 INFO util.GSet: VM type   = 64-bit
> 14/03/20 15:06:41 INFO util.GSet: 0.02999329447746% max memory 896 MB = 
> 275.3 KB
> 14/03/20 15:06:41 INFO util.GSet: capacity  = 2^15 = 32768 entries
> 14/03/20 15:06:41 INFO namenode.AclConfigFlag: ACLs enabled? false
> 14/03/20 15:06:41 INFO common.Storage: Lock on 
> /data/hadoop/data1/dfs/name/in_use.lock acquired by nodename 
> 7326@10-150-170-176
> 14/03/20 15:06:42 INFO common.Storage: Lock on 
> /data/hadoop/data2/dfs/name/in_use.lock acquired by nodename 
> 7326@10-150-170-176
> 14/03/20 15:06:42 INFO namenode.FSImage: No edit log streams selected.
> 14/03/20 15:06:42 INFO namenode.FSImageFormatPBINode: Loading 1 INodes.
> 14/03/20 15:06:42 FATAL namenode.NameNode: Exception in namenode join
> java.lang.NullPointerException
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSDirectory.isReservedName(FSDirectory.java:2984)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImageFormatPBINode$Loader.addToParent(FSImageFormatPBINode.java:205)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImageFormatPBINode$Loader.loadINodeDirectorySection(FSImageFormatPBINode.java:162)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImageFormatProtobuf$Loader.loadInternal(FSImageFormatProtobuf.java:243)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImageFormatProtobuf$Loader.load(FSImageFormatProtobuf.java:168)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImageFormat$LoaderDelegator.load(FSImageFormat.java:120)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.loadFSImage(FSImage.java:895)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.loadFSImage(FSImage.java:881)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.loadFSImageFile(FSImage.java:704)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.loadFSImage(FSImage.java:642)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:271)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFSImage(FSNamesystem.java:894)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFromDisk(FSNamesystem.java:653)
>   at 
> org.apache.hadoop.hdfs.server.namenode.NameNode.initializeSharedEdits(NameNode.java:912)
>   at 
> org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1276)
>   at 
> org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1360)
> 14/03/20 15:06:42 INFO util.ExitUtil: Exiting with status 1
> 14/03/20 15:06:42 INFO namenode.NameNode: SHUTDOWN_MSG: 
> /
> SHUTDOWN_MSG: Shutting down NameNode at 10-150-170-176/10.150.170.176
> /
> {code}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6130) NPE during namenode upgrade from old release

2014-03-24 Thread Fengdong Yu (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6130?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Fengdong Yu updated HDFS-6130:
--

Attachment: fsimage.tar.gz

> NPE during namenode upgrade from old release
> 
>
> Key: HDFS-6130
> URL: https://issues.apache.org/jira/browse/HDFS-6130
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: namenode
>Affects Versions: 2.4.0
>Reporter: Fengdong Yu
> Attachments: fsimage.tar.gz
>
>
> I want upgrade an old cluster(0.20.2-cdh3u1) to trunk instance, 
> I can upgrade successfully if I don't configurage HA, but if HA enabled,
> there is NPE when I run ' hdfs namenode -initializeSharedEdits'
> {code}
> 14/03/20 15:06:41 INFO namenode.FSNamesystem: Retry cache on namenode is 
> enabled
> 14/03/20 15:06:41 INFO namenode.FSNamesystem: Retry cache will use 0.03 of 
> total heap and retry cache entry expiry time is 60 millis
> 14/03/20 15:06:41 INFO util.GSet: Computing capacity for map 
> NameNodeRetryCache
> 14/03/20 15:06:41 INFO util.GSet: VM type   = 64-bit
> 14/03/20 15:06:41 INFO util.GSet: 0.02999329447746% max memory 896 MB = 
> 275.3 KB
> 14/03/20 15:06:41 INFO util.GSet: capacity  = 2^15 = 32768 entries
> 14/03/20 15:06:41 INFO namenode.AclConfigFlag: ACLs enabled? false
> 14/03/20 15:06:41 INFO common.Storage: Lock on 
> /data/hadoop/data1/dfs/name/in_use.lock acquired by nodename 
> 7326@10-150-170-176
> 14/03/20 15:06:42 INFO common.Storage: Lock on 
> /data/hadoop/data2/dfs/name/in_use.lock acquired by nodename 
> 7326@10-150-170-176
> 14/03/20 15:06:42 INFO namenode.FSImage: No edit log streams selected.
> 14/03/20 15:06:42 INFO namenode.FSImageFormatPBINode: Loading 1 INodes.
> 14/03/20 15:06:42 FATAL namenode.NameNode: Exception in namenode join
> java.lang.NullPointerException
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSDirectory.isReservedName(FSDirectory.java:2984)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImageFormatPBINode$Loader.addToParent(FSImageFormatPBINode.java:205)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImageFormatPBINode$Loader.loadINodeDirectorySection(FSImageFormatPBINode.java:162)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImageFormatProtobuf$Loader.loadInternal(FSImageFormatProtobuf.java:243)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImageFormatProtobuf$Loader.load(FSImageFormatProtobuf.java:168)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImageFormat$LoaderDelegator.load(FSImageFormat.java:120)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.loadFSImage(FSImage.java:895)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.loadFSImage(FSImage.java:881)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.loadFSImageFile(FSImage.java:704)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.loadFSImage(FSImage.java:642)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:271)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFSImage(FSNamesystem.java:894)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFromDisk(FSNamesystem.java:653)
>   at 
> org.apache.hadoop.hdfs.server.namenode.NameNode.initializeSharedEdits(NameNode.java:912)
>   at 
> org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1276)
>   at 
> org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1360)
> 14/03/20 15:06:42 INFO util.ExitUtil: Exiting with status 1
> 14/03/20 15:06:42 INFO namenode.NameNode: SHUTDOWN_MSG: 
> /
> SHUTDOWN_MSG: Shutting down NameNode at 10-150-170-176/10.150.170.176
> /
> {code}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6150) Add inode id information in the logs to make debugging easier

2014-03-24 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6150?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13946180#comment-13946180
 ] 

Hadoop QA commented on HDFS-6150:
-

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12636483/HDFS-6150.2.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 1 new 
or modified test files.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  There were no new javadoc warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:red}-1 core tests{color}.  The patch failed these unit tests in 
hadoop-hdfs-project/hadoop-hdfs:

  org.apache.hadoop.hdfs.TestLeaseRecovery2

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-HDFS-Build/6483//testReport/
Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/6483//console

This message is automatically generated.

> Add inode id information in the logs to make debugging easier
> -
>
> Key: HDFS-6150
> URL: https://issues.apache.org/jira/browse/HDFS-6150
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: namenode
>Reporter: Suresh Srinivas
>Assignee: Suresh Srinivas
>Priority: Minor
> Attachments: HDFS-6150.1.patch, HDFS-6150.2.patch, HDFS-6150.patch
>
>
> Inode information and path information are missing in the logs and 
> exceptions. Adding this will help debug multi threading issues related to 
> using incode INode ID information.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-5978) Create a tool to take fsimage and expose read-only WebHDFS API

2014-03-24 Thread Haohui Mai (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5978?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13946160#comment-13946160
 ] 

Haohui Mai commented on HDFS-5978:
--

{code}
+  public void shutdown() {
+bootstrap.shutdown();
+  }
{code}

Calling {{bootstrap.shutdown()}} only will have resource leaks. You'll need to 
add all channels into a {{ChannelGroup}} and calls {{ChannelGroup.close()}} and 
{{ChannelFactory.releaseExternalResources()}} to guarantee proper clean ups. It 
is important to clean up the resources as other tests might reuse the JVM.

Please see http://docs.jboss.org/netty/3.2/guide/html_single for code examples. 
The portmap daemon in the nfs project contains the example as well.

for {{initServerAndWait()}}, you replace it with a single line of code that 
waits on the close future of the server-side channel:

{code}
bootstrap.bind(address).getCloseFuture().await();
{code}

Other than that the patch looks good to me.

> Create a tool to take fsimage and expose read-only WebHDFS API
> --
>
> Key: HDFS-5978
> URL: https://issues.apache.org/jira/browse/HDFS-5978
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: tools
>Reporter: Akira AJISAKA
>Assignee: Akira AJISAKA
>  Labels: newbie
> Attachments: HDFS-5978.2.patch, HDFS-5978.3.patch, HDFS-5978.4.patch, 
> HDFS-5978.5.patch, HDFS-5978.patch
>
>
> Suggested in HDFS-5975.
> Add an option to OfflineImageViewer to launch HTTP server that takes fsimage 
> and exposes the read-only version of WebHDFS API for OfflineImageViewer. The 
> server only supports {{LISTSTATUS}} operation and other operations such as 
> {{GETFILESTATUS}} will be added by other JIRAs.
> That way we can allow the operator to use the existing command-line tool, or 
> even the web UI to debug the fsimage. It also allows the operator to 
> interactively browsing the file system, figuring out what goes wrong.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (HDFS-6153) Add "fileId" and "childrenNum" fields to Json schema in the WebHDFS document

2014-03-24 Thread Akira AJISAKA (JIRA)
Akira AJISAKA created HDFS-6153:
---

 Summary: Add "fileId" and "childrenNum" fields to Json schema in 
the WebHDFS document
 Key: HDFS-6153
 URL: https://issues.apache.org/jira/browse/HDFS-6153
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: documentation, webhdfs
Affects Versions: 2.3.0
Reporter: Akira AJISAKA
Assignee: Akira AJISAKA
Priority: Minor


Now WebHDFS returns FileStatus Json objects include "fileId" and "childrenNum" 
fields but these fields are not documented.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-5978) Create a tool to take fsimage and expose read-only WebHDFS API

2014-03-24 Thread Akira AJISAKA (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5978?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Akira AJISAKA updated HDFS-5978:


Description: 
Suggested in HDFS-5975.

Add an option to OfflineImageViewer to launch HTTP server that takes fsimage 
and exposes the read-only version of WebHDFS API for OfflineImageViewer. The 
server only supports {{LISTSTATUS}} operation and other operations such as 
{{GETFILESTATUS}} will be added by other JIRAs.

That way we can allow the operator to use the existing command-line tool, or 
even the web UI to debug the fsimage. It also allows the operator to 
interactively browsing the file system, figuring out what goes wrong.


  was:
Suggested in HDFS-5975.

Add an option to exposes the read-only version of WebHDFS API for 
OfflineImageViewer. You can imagine it looks very similar to jhat.

That way we can allow the operator to use the existing command-line tool, or 
even the web UI to debug the fsimage. It also allows the operator to 
interactively browsing the file system, figuring out what goes wrong.



> Create a tool to take fsimage and expose read-only WebHDFS API
> --
>
> Key: HDFS-5978
> URL: https://issues.apache.org/jira/browse/HDFS-5978
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: tools
>Reporter: Akira AJISAKA
>Assignee: Akira AJISAKA
>  Labels: newbie
> Attachments: HDFS-5978.2.patch, HDFS-5978.3.patch, HDFS-5978.4.patch, 
> HDFS-5978.5.patch, HDFS-5978.patch
>
>
> Suggested in HDFS-5975.
> Add an option to OfflineImageViewer to launch HTTP server that takes fsimage 
> and exposes the read-only version of WebHDFS API for OfflineImageViewer. The 
> server only supports {{LISTSTATUS}} operation and other operations such as 
> {{GETFILESTATUS}} will be added by other JIRAs.
> That way we can allow the operator to use the existing command-line tool, or 
> even the web UI to debug the fsimage. It also allows the operator to 
> interactively browsing the file system, figuring out what goes wrong.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-5978) Create a tool to take fsimage and expose read-only WebHDFS API

2014-03-24 Thread Akira AJISAKA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5978?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13946147#comment-13946147
 ] 

Akira AJISAKA commented on HDFS-5978:
-

Thanks Haohui for the comments. I renewed the patch.
bq. You can pick a default port if you want.
I choose 5978 (the number of this issue). If you come up with better number, 
please tell me.
{quote}
{code}
public void initServerAndWait(String fsimage) throws IOException {
...
{code}
This is unnecessary.
{quote}
It is necessary to wait for Ctrl+C because the server shutdowns automatically 
when OfflineImageViewer exits.

> Create a tool to take fsimage and expose read-only WebHDFS API
> --
>
> Key: HDFS-5978
> URL: https://issues.apache.org/jira/browse/HDFS-5978
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: tools
>Reporter: Akira AJISAKA
>Assignee: Akira AJISAKA
>  Labels: newbie
> Attachments: HDFS-5978.2.patch, HDFS-5978.3.patch, HDFS-5978.4.patch, 
> HDFS-5978.5.patch, HDFS-5978.patch
>
>
> Suggested in HDFS-5975.
> Add an option to exposes the read-only version of WebHDFS API for 
> OfflineImageViewer. You can imagine it looks very similar to jhat.
> That way we can allow the operator to use the existing command-line tool, or 
> even the web UI to debug the fsimage. It also allows the operator to 
> interactively browsing the file system, figuring out what goes wrong.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6152) distcp V2 doesn't preserve root dir's attributes when -p is specified

2014-03-24 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6152?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13946146#comment-13946146
 ] 

Hadoop QA commented on HDFS-6152:
-

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12636501/HDFS-6152.001.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 9 new 
or modified test files.

  {color:red}-1 javac{color}.  The applied patch generated 1493 javac 
compiler warnings (more than the trunk's current 1492 warnings).

{color:green}+1 javadoc{color}.  There were no new javadoc warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:red}-1 core tests{color}.  The patch failed these unit tests in 
hadoop-tools/hadoop-distcp:

  org.apache.hadoop.tools.mapred.TestUniformSizeInputFormat

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-HDFS-Build/6484//testReport/
Javac warnings: 
https://builds.apache.org/job/PreCommit-HDFS-Build/6484//artifact/trunk/patchprocess/diffJavacWarnings.txt
Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/6484//console

This message is automatically generated.

> distcp V2 doesn't preserve root dir's attributes when -p is specified
> -
>
> Key: HDFS-6152
> URL: https://issues.apache.org/jira/browse/HDFS-6152
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: hdfs-client
>Affects Versions: 2.3.0
>Reporter: Yongjun Zhang
>Assignee: Yongjun Zhang
> Attachments: HDFS-6152.001.patch
>
>
> Two issues were observed with distcpV2
> ISSUE 1. when copying a source dir to target dir with "-pu" option using 
> command 
>   "distcp -pu source-dir target-dir"
>  
> The source dir's owner is not preserved at target dir. Simiarly other 
> attributes of source dir are not preserved.  Supposedly they should be 
> preserved when no -update and no -overwrite specified. 
> There are two scenarios with the above command:
> a. when target-dir already exists. Issuing the above command will  result in 
> target-dir/source-dir (source-dir here refers to the last component of the 
> source-dir path in the command line) at target file system, with all contents 
> in source-dir copied to under target-dir/src-dir. The issue in this case is, 
> the attributes of src-dir is not preserved.
> b. when target-dir doesn't exist. It will result in target-dir with all 
> contents of source-dir copied to under target-dir. This issue in this  case 
> is, the attributes of source-dir is not carried over to target-dir.
> For multiple source cases, e.g., command 
>   "distcp -pu source-dir1 source-dir2 target-dir"
> No matter whether the target-dir exists or not, the multiple sources are 
> copied to under the target dir (target-dir is created if it didn't exist). 
> And their attributes are preserved. 
> ISSUE 2. with the following command:
>   "distcp source-dir target-dir"
> when source-dir is an empty directory, and when target-dir doesn't exist, 
> source-dir is not copied, actually the command behaves like a no-op. However, 
> when the source-dir is not empty, it would be copied and results in 
> target-dir at the target file system containing a copy of source-dir's 
> children.
> To be consistent, empty source dir should be copied too. Basically the  above 
> distcp command should cause target-dir get created at target file system, and 
> the source-dir's attributes are preserved at target-dir when -p is passed.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-5978) Create a tool to take fsimage and expose read-only WebHDFS API

2014-03-24 Thread Akira AJISAKA (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5978?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Akira AJISAKA updated HDFS-5978:


Attachment: HDFS-5978.5.patch

> Create a tool to take fsimage and expose read-only WebHDFS API
> --
>
> Key: HDFS-5978
> URL: https://issues.apache.org/jira/browse/HDFS-5978
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: tools
>Reporter: Akira AJISAKA
>Assignee: Akira AJISAKA
>  Labels: newbie
> Attachments: HDFS-5978.2.patch, HDFS-5978.3.patch, HDFS-5978.4.patch, 
> HDFS-5978.5.patch, HDFS-5978.patch
>
>
> Suggested in HDFS-5975.
> Add an option to exposes the read-only version of WebHDFS API for 
> OfflineImageViewer. You can imagine it looks very similar to jhat.
> That way we can allow the operator to use the existing command-line tool, or 
> even the web UI to debug the fsimage. It also allows the operator to 
> interactively browsing the file system, figuring out what goes wrong.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-5840) Follow-up to HDFS-5138 to improve error handling during partial upgrade failures

2014-03-24 Thread Aaron T. Myers (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5840?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13946099#comment-13946099
 ] 

Aaron T. Myers commented on HDFS-5840:
--

Thanks a lot for taking this across the finish line, Jing. The updated patch 
you posted looks pretty good to me.

Suresh, does it look OK to you as well? If so, +1 from my perspective.

> Follow-up to HDFS-5138 to improve error handling during partial upgrade 
> failures
> 
>
> Key: HDFS-5840
> URL: https://issues.apache.org/jira/browse/HDFS-5840
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: ha, journal-node, namenode
>Affects Versions: 3.0.0, 2.4.0
>Reporter: Aaron T. Myers
>Assignee: Jing Zhao
>Priority: Blocker
> Attachments: HDFS-5840.001.patch, HDFS-5840.002.patch, HDFS-5840.patch
>
>
> Suresh posted some good comment in HDFS-5138 after that patch had already 
> been committed to trunk. This JIRA is to address those. See the first comment 
> of this JIRA for the full content of the review.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6130) NPE during namenode upgrade from old release

2014-03-24 Thread Fengdong Yu (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6130?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13946089#comment-13946089
 ] 

Fengdong Yu commented on HDFS-6130:
---

please ignore my create check point method, that's wrong.

> NPE during namenode upgrade from old release
> 
>
> Key: HDFS-6130
> URL: https://issues.apache.org/jira/browse/HDFS-6130
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: namenode
>Affects Versions: 2.4.0
>Reporter: Fengdong Yu
>
> I want upgrade an old cluster(0.20.2-cdh3u1) to trunk instance, 
> I can upgrade successfully if I don't configurage HA, but if HA enabled,
> there is NPE when I run ' hdfs namenode -initializeSharedEdits'
> {code}
> 14/03/20 15:06:41 INFO namenode.FSNamesystem: Retry cache on namenode is 
> enabled
> 14/03/20 15:06:41 INFO namenode.FSNamesystem: Retry cache will use 0.03 of 
> total heap and retry cache entry expiry time is 60 millis
> 14/03/20 15:06:41 INFO util.GSet: Computing capacity for map 
> NameNodeRetryCache
> 14/03/20 15:06:41 INFO util.GSet: VM type   = 64-bit
> 14/03/20 15:06:41 INFO util.GSet: 0.02999329447746% max memory 896 MB = 
> 275.3 KB
> 14/03/20 15:06:41 INFO util.GSet: capacity  = 2^15 = 32768 entries
> 14/03/20 15:06:41 INFO namenode.AclConfigFlag: ACLs enabled? false
> 14/03/20 15:06:41 INFO common.Storage: Lock on 
> /data/hadoop/data1/dfs/name/in_use.lock acquired by nodename 
> 7326@10-150-170-176
> 14/03/20 15:06:42 INFO common.Storage: Lock on 
> /data/hadoop/data2/dfs/name/in_use.lock acquired by nodename 
> 7326@10-150-170-176
> 14/03/20 15:06:42 INFO namenode.FSImage: No edit log streams selected.
> 14/03/20 15:06:42 INFO namenode.FSImageFormatPBINode: Loading 1 INodes.
> 14/03/20 15:06:42 FATAL namenode.NameNode: Exception in namenode join
> java.lang.NullPointerException
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSDirectory.isReservedName(FSDirectory.java:2984)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImageFormatPBINode$Loader.addToParent(FSImageFormatPBINode.java:205)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImageFormatPBINode$Loader.loadINodeDirectorySection(FSImageFormatPBINode.java:162)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImageFormatProtobuf$Loader.loadInternal(FSImageFormatProtobuf.java:243)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImageFormatProtobuf$Loader.load(FSImageFormatProtobuf.java:168)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImageFormat$LoaderDelegator.load(FSImageFormat.java:120)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.loadFSImage(FSImage.java:895)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.loadFSImage(FSImage.java:881)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.loadFSImageFile(FSImage.java:704)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.loadFSImage(FSImage.java:642)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:271)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFSImage(FSNamesystem.java:894)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFromDisk(FSNamesystem.java:653)
>   at 
> org.apache.hadoop.hdfs.server.namenode.NameNode.initializeSharedEdits(NameNode.java:912)
>   at 
> org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1276)
>   at 
> org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1360)
> 14/03/20 15:06:42 INFO util.ExitUtil: Exiting with status 1
> 14/03/20 15:06:42 INFO namenode.NameNode: SHUTDOWN_MSG: 
> /
> SHUTDOWN_MSG: Shutting down NameNode at 10-150-170-176/10.150.170.176
> /
> {code}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-5840) Follow-up to HDFS-5138 to improve error handling during partial upgrade failures

2014-03-24 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5840?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13946080#comment-13946080
 ] 

Hadoop QA commented on HDFS-5840:
-

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12636462/HDFS-5840.002.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 2 new 
or modified test files.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  There were no new javadoc warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:red}-1 core tests{color}.  The patch failed these unit tests in 
hadoop-hdfs-project/hadoop-hdfs:

  org.apache.hadoop.hdfs.server.namenode.TestCacheDirectives

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-HDFS-Build/6481//testReport/
Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/6481//console

This message is automatically generated.

> Follow-up to HDFS-5138 to improve error handling during partial upgrade 
> failures
> 
>
> Key: HDFS-5840
> URL: https://issues.apache.org/jira/browse/HDFS-5840
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: ha, journal-node, namenode
>Affects Versions: 3.0.0, 2.4.0
>Reporter: Aaron T. Myers
>Assignee: Jing Zhao
>Priority: Blocker
> Attachments: HDFS-5840.001.patch, HDFS-5840.002.patch, HDFS-5840.patch
>
>
> Suresh posted some good comment in HDFS-5138 after that patch had already 
> been committed to trunk. This JIRA is to address those. See the first comment 
> of this JIRA for the full content of the review.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-5910) Enhance DataTransferProtocol to allow per-connection choice of encryption/plain-text

2014-03-24 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5910?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13946075#comment-13946075
 ] 

Hadoop QA commented on HDFS-5910:
-

{color:green}+1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12636464/HDFS-5910.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 1 new 
or modified test files.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  There were no new javadoc warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:green}+1 core tests{color}.  The patch passed unit tests in 
hadoop-hdfs-project/hadoop-hdfs.

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-HDFS-Build/6482//testReport/
Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/6482//console

This message is automatically generated.

> Enhance DataTransferProtocol to allow per-connection choice of 
> encryption/plain-text
> 
>
> Key: HDFS-5910
> URL: https://issues.apache.org/jira/browse/HDFS-5910
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: security
>Affects Versions: 2.2.0
>Reporter: Benoy Antony
>Assignee: Benoy Antony
> Attachments: HDFS-5910.patch, HDFS-5910.patch, HDFS-5910.patch
>
>
> It is possible to enable encryption of DataTransferProtocol. 
> In some use cases, it is required to encrypt data transfer with some clients 
> , but communicate in plain text with some other clients and data nodes.
> A sample use case will be that any data transfer inside a firewall can be in 
> plain text whereas any data transfer from clients  outside the firewall needs 
> to be encrypted.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6130) NPE during namenode upgrade from old release

2014-03-24 Thread Fengdong Yu (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6130?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13946073#comment-13946073
 ] 

Fengdong Yu commented on HDFS-6130:
---

hi [~szetszwo], where you get 1.3.0? 

> NPE during namenode upgrade from old release
> 
>
> Key: HDFS-6130
> URL: https://issues.apache.org/jira/browse/HDFS-6130
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: namenode
>Affects Versions: 2.4.0
>Reporter: Fengdong Yu
>
> I want upgrade an old cluster(0.20.2-cdh3u1) to trunk instance, 
> I can upgrade successfully if I don't configurage HA, but if HA enabled,
> there is NPE when I run ' hdfs namenode -initializeSharedEdits'
> {code}
> 14/03/20 15:06:41 INFO namenode.FSNamesystem: Retry cache on namenode is 
> enabled
> 14/03/20 15:06:41 INFO namenode.FSNamesystem: Retry cache will use 0.03 of 
> total heap and retry cache entry expiry time is 60 millis
> 14/03/20 15:06:41 INFO util.GSet: Computing capacity for map 
> NameNodeRetryCache
> 14/03/20 15:06:41 INFO util.GSet: VM type   = 64-bit
> 14/03/20 15:06:41 INFO util.GSet: 0.02999329447746% max memory 896 MB = 
> 275.3 KB
> 14/03/20 15:06:41 INFO util.GSet: capacity  = 2^15 = 32768 entries
> 14/03/20 15:06:41 INFO namenode.AclConfigFlag: ACLs enabled? false
> 14/03/20 15:06:41 INFO common.Storage: Lock on 
> /data/hadoop/data1/dfs/name/in_use.lock acquired by nodename 
> 7326@10-150-170-176
> 14/03/20 15:06:42 INFO common.Storage: Lock on 
> /data/hadoop/data2/dfs/name/in_use.lock acquired by nodename 
> 7326@10-150-170-176
> 14/03/20 15:06:42 INFO namenode.FSImage: No edit log streams selected.
> 14/03/20 15:06:42 INFO namenode.FSImageFormatPBINode: Loading 1 INodes.
> 14/03/20 15:06:42 FATAL namenode.NameNode: Exception in namenode join
> java.lang.NullPointerException
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSDirectory.isReservedName(FSDirectory.java:2984)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImageFormatPBINode$Loader.addToParent(FSImageFormatPBINode.java:205)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImageFormatPBINode$Loader.loadINodeDirectorySection(FSImageFormatPBINode.java:162)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImageFormatProtobuf$Loader.loadInternal(FSImageFormatProtobuf.java:243)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImageFormatProtobuf$Loader.load(FSImageFormatProtobuf.java:168)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImageFormat$LoaderDelegator.load(FSImageFormat.java:120)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.loadFSImage(FSImage.java:895)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.loadFSImage(FSImage.java:881)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.loadFSImageFile(FSImage.java:704)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.loadFSImage(FSImage.java:642)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:271)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFSImage(FSNamesystem.java:894)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFromDisk(FSNamesystem.java:653)
>   at 
> org.apache.hadoop.hdfs.server.namenode.NameNode.initializeSharedEdits(NameNode.java:912)
>   at 
> org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1276)
>   at 
> org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1360)
> 14/03/20 15:06:42 INFO util.ExitUtil: Exiting with status 1
> 14/03/20 15:06:42 INFO namenode.NameNode: SHUTDOWN_MSG: 
> /
> SHUTDOWN_MSG: Shutting down NameNode at 10-150-170-176/10.150.170.176
> /
> {code}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6152) distcp V2 doesn't preserve root dir's attributes when -p is specified

2014-03-24 Thread Yongjun Zhang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6152?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yongjun Zhang updated HDFS-6152:


Status: Patch Available  (was: Open)

The submitted patch tries to address the two issues reported.

Some notable changes:

1.  A new boolean field "targetPathExists" is introduced to DistCpOptions 
class. The value of this field is a derived by checking whether the target path 
exists or not in the beginning of distcp.  (Arguably, this information could be 
put somewhere else, but I found DistCpOption is the most suitable place based 
on the current DistCp implementation). 

A new corresponding jobconf property CONF_LABEL_TARGET_PATH_EXISTS is 
introduced, and it's initialized at the same time as the targetPathExists field.

The reason is that the result of class SimpleCopyListing's method 
computeSourceRootPath depends on DistCpOption. E.g., whether the distcp target 
exists or not, whether -update or -overwrite switches are passed. And Item 3 
below needs this info (via the new jobconf property). 

Unit tests that use DistCpOptions need to be aware of the need to set this 
filed according to the test-case's setting.

2. For the issues reported in this JIRA, an entry that  was skipped by 
writeToFileListing method with the following code:
{code}
 if (fileStatus.getPath().equals(sourcePathRoot) && fileStatus.isDirectory())
return; // Skip the root-paths.
{code}
is now added to the filelisting when no -update/-overwrite is specified.

This entry is recognized by both the CopyMapper and CopyCommitter.
Using this entry, the CopyMapper will create dir accordingly (for ISSUE 2), and 
the CopyCommitter will update attributes when specified (for ISSUE 1).

E.g., distcp a/b xyz, where a/b is the source dir, 
a. if xyz doesn't exist, then "a/b" is written to the copyListing with empty 
relative path "". 
b. if xyz exists, then "a/b" is written to the copyListing with relative path 
"b".

3. 
class CopyCommitter's method deleteMissing creates a DistCpOption object with 
default setting, and collect listing from prior-to-committing result of distcp. 
This is not sufficient for the above mentioned reason  (The result of class 
SimpleCopyListing's method computeSourceRootPath depends on DistCpOption). The 
problem is revealed with the change I added to fix this JIRA, and the patch I 
submitted addressed it.

Thanks for reviewing.


> distcp V2 doesn't preserve root dir's attributes when -p is specified
> -
>
> Key: HDFS-6152
> URL: https://issues.apache.org/jira/browse/HDFS-6152
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: hdfs-client
>Affects Versions: 2.3.0
>Reporter: Yongjun Zhang
>Assignee: Yongjun Zhang
> Attachments: HDFS-6152.001.patch
>
>
> Two issues were observed with distcpV2
> ISSUE 1. when copying a source dir to target dir with "-pu" option using 
> command 
>   "distcp -pu source-dir target-dir"
>  
> The source dir's owner is not preserved at target dir. Simiarly other 
> attributes of source dir are not preserved.  Supposedly they should be 
> preserved when no -update and no -overwrite specified. 
> There are two scenarios with the above command:
> a. when target-dir already exists. Issuing the above command will  result in 
> target-dir/source-dir (source-dir here refers to the last component of the 
> source-dir path in the command line) at target file system, with all contents 
> in source-dir copied to under target-dir/src-dir. The issue in this case is, 
> the attributes of src-dir is not preserved.
> b. when target-dir doesn't exist. It will result in target-dir with all 
> contents of source-dir copied to under target-dir. This issue in this  case 
> is, the attributes of source-dir is not carried over to target-dir.
> For multiple source cases, e.g., command 
>   "distcp -pu source-dir1 source-dir2 target-dir"
> No matter whether the target-dir exists or not, the multiple sources are 
> copied to under the target dir (target-dir is created if it didn't exist). 
> And their attributes are preserved. 
> ISSUE 2. with the following command:
>   "distcp source-dir target-dir"
> when source-dir is an empty directory, and when target-dir doesn't exist, 
> source-dir is not copied, actually the command behaves like a no-op. However, 
> when the source-dir is not empty, it would be copied and results in 
> target-dir at the target file system containing a copy of source-dir's 
> children.
> To be consistent, empty source dir should be copied too. Basically the  above 
> distcp command should cause target-dir get created at target file system, and 
> the source-dir's attributes are preserved at target-dir when -p is passed.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6152) distcp V2 doesn't preserve root dir's attributes when -p is specified

2014-03-24 Thread Yongjun Zhang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6152?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yongjun Zhang updated HDFS-6152:


Attachment: HDFS-6152.001.patch

> distcp V2 doesn't preserve root dir's attributes when -p is specified
> -
>
> Key: HDFS-6152
> URL: https://issues.apache.org/jira/browse/HDFS-6152
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: hdfs-client
>Affects Versions: 2.3.0
>Reporter: Yongjun Zhang
>Assignee: Yongjun Zhang
> Attachments: HDFS-6152.001.patch
>
>
> Two issues were observed with distcpV2
> ISSUE 1. when copying a source dir to target dir with "-pu" option using 
> command 
>   "distcp -pu source-dir target-dir"
>  
> The source dir's owner is not preserved at target dir. Simiarly other 
> attributes of source dir are not preserved.  Supposedly they should be 
> preserved when no -update and no -overwrite specified. 
> There are two scenarios with the above command:
> a. when target-dir already exists. Issuing the above command will  result in 
> target-dir/source-dir (source-dir here refers to the last component of the 
> source-dir path in the command line) at target file system, with all contents 
> in source-dir copied to under target-dir/src-dir. The issue in this case is, 
> the attributes of src-dir is not preserved.
> b. when target-dir doesn't exist. It will result in target-dir with all 
> contents of source-dir copied to under target-dir. This issue in this  case 
> is, the attributes of source-dir is not carried over to target-dir.
> For multiple source cases, e.g., command 
>   "distcp -pu source-dir1 source-dir2 target-dir"
> No matter whether the target-dir exists or not, the multiple sources are 
> copied to under the target dir (target-dir is created if it didn't exist). 
> And their attributes are preserved. 
> ISSUE 2. with the following command:
>   "distcp source-dir target-dir"
> when source-dir is an empty directory, and when target-dir doesn't exist, 
> source-dir is not copied, actually the command behaves like a no-op. However, 
> when the source-dir is not empty, it would be copied and results in 
> target-dir at the target file system containing a copy of source-dir's 
> children.
> To be consistent, empty source dir should be copied too. Basically the  above 
> distcp command should cause target-dir get created at target file system, and 
> the source-dir's attributes are preserved at target-dir when -p is passed.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6152) distcp V2 doesn't preserve root dir's attributes when -p is specified

2014-03-24 Thread Yongjun Zhang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6152?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yongjun Zhang updated HDFS-6152:


Attachment: (was: HDFS-6152.001.patch)

> distcp V2 doesn't preserve root dir's attributes when -p is specified
> -
>
> Key: HDFS-6152
> URL: https://issues.apache.org/jira/browse/HDFS-6152
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: hdfs-client
>Affects Versions: 2.3.0
>Reporter: Yongjun Zhang
>Assignee: Yongjun Zhang
>
> Two issues were observed with distcpV2
> ISSUE 1. when copying a source dir to target dir with "-pu" option using 
> command 
>   "distcp -pu source-dir target-dir"
>  
> The source dir's owner is not preserved at target dir. Simiarly other 
> attributes of source dir are not preserved.  Supposedly they should be 
> preserved when no -update and no -overwrite specified. 
> There are two scenarios with the above command:
> a. when target-dir already exists. Issuing the above command will  result in 
> target-dir/source-dir (source-dir here refers to the last component of the 
> source-dir path in the command line) at target file system, with all contents 
> in source-dir copied to under target-dir/src-dir. The issue in this case is, 
> the attributes of src-dir is not preserved.
> b. when target-dir doesn't exist. It will result in target-dir with all 
> contents of source-dir copied to under target-dir. This issue in this  case 
> is, the attributes of source-dir is not carried over to target-dir.
> For multiple source cases, e.g., command 
>   "distcp -pu source-dir1 source-dir2 target-dir"
> No matter whether the target-dir exists or not, the multiple sources are 
> copied to under the target dir (target-dir is created if it didn't exist). 
> And their attributes are preserved. 
> ISSUE 2. with the following command:
>   "distcp source-dir target-dir"
> when source-dir is an empty directory, and when target-dir doesn't exist, 
> source-dir is not copied, actually the command behaves like a no-op. However, 
> when the source-dir is not empty, it would be copied and results in 
> target-dir at the target file system containing a copy of source-dir's 
> children.
> To be consistent, empty source dir should be copied too. Basically the  above 
> distcp command should cause target-dir get created at target file system, and 
> the source-dir's attributes are preserved at target-dir when -p is passed.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6136) PacketReceiver#doRead calls setFieldsFromData with wrong argument

2014-03-24 Thread Todd Lipcon (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6136?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13946022#comment-13946022
 ] 

Todd Lipcon commented on HDFS-6136:
---

Hey Nick. This is a surprising bug fix since data transfer apparently works 
fine without it. Is there any effect of this change? If not, why not? Seems 
like getting this wrong should cause problems :)

> PacketReceiver#doRead calls setFieldsFromData with wrong argument
> -
>
> Key: HDFS-6136
> URL: https://issues.apache.org/jira/browse/HDFS-6136
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: datanode
>Affects Versions: 2.3.0
>Reporter: Nick White
>  Labels: patch
> Attachments: HDFS-6136.patch
>
>
> PacketHeader#setFieldsFromData takes the packet length as the first argument, 
> but PacketReceiver#doRead passes the dataPlusChecksumLen (which is 4 less).



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6125) Cleanup unnecessary cast in HDFS code base

2014-03-24 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6125?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13946015#comment-13946015
 ] 

Hadoop QA commented on HDFS-6125:
-

{color:green}+1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12636455/HDFS-6125.2.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 43 new 
or modified test files.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  There were no new javadoc warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:green}+1 core tests{color}.  The patch passed unit tests in 
hadoop-hdfs-project/hadoop-hdfs 
hadoop-hdfs-project/hadoop-hdfs/src/contrib/bkjournal.

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-HDFS-Build/6480//testReport/
Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/6480//console

This message is automatically generated.

> Cleanup unnecessary cast in HDFS code base
> --
>
> Key: HDFS-6125
> URL: https://issues.apache.org/jira/browse/HDFS-6125
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: test
>Affects Versions: 2.4.0
>Reporter: Suresh Srinivas
>Assignee: Suresh Srinivas
> Attachments: HDFS-6125.2.patch, HDFS-6125.patch
>
>




--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6152) distcp V2 doesn't preserve root dir's attributes when -p is specified

2014-03-24 Thread Yongjun Zhang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6152?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yongjun Zhang updated HDFS-6152:


Description: 
Two issues were observed with distcpV2

ISSUE 1. when copying a source dir to target dir with "-pu" option using 
command 

  "distcp -pu source-dir target-dir"
 
The source dir's owner is not preserved at target dir. Simiarly other 
attributes of source dir are not preserved.  Supposedly they should be 
preserved when no -update and no -overwrite specified. 

There are two scenarios with the above command:

a. when target-dir already exists. Issuing the above command will  result in 
target-dir/source-dir (source-dir here refers to the last component of the 
source-dir path in the command line) at target file system, with all contents 
in source-dir copied to under target-dir/src-dir. The issue in this case is, 
the attributes of src-dir is not preserved.

b. when target-dir doesn't exist. It will result in target-dir with all 
contents of source-dir copied to under target-dir. This issue in this  case is, 
the attributes of source-dir is not carried over to target-dir.

For multiple source cases, e.g., command 

  "distcp -pu source-dir1 source-dir2 target-dir"

No matter whether the target-dir exists or not, the multiple sources are copied 
to under the target dir (target-dir is created if it didn't exist). And their 
attributes are preserved. 

ISSUE 2. with the following command:

  "distcp source-dir target-dir"

when source-dir is an empty directory, and when target-dir doesn't exist, 
source-dir is not copied, actually the command behaves like a no-op. However, 
when the source-dir is not empty, it would be copied and results in target-dir 
at the target file system containing a copy of source-dir's children.

To be consistent, empty source dir should be copied too. Basically the  above 
distcp command should cause target-dir get created at target file system, and 
the source-dir's attributes are preserved at target-dir when -p is passed.


  was:
Two issues were observed with distcpV2

ISSUE 1. when copying a source dir to target dir with "-pu" option using 
command 

  "distcp -pu source-dir target-dir"
 
The source dir's owner is not preserved at target dir. Simiarly other 
attributes of source dir are not preserved.  Supposedly they should be 
preserved when no -update and no -overwrite specified. 

There are two scenarios with the above command:

a. when target-dir already exists. Issuing the above command will  result in 
target-dir/source-dir (source-dir here refers to the last component of the 
source-dir path in the command line) at target file system, with all contents 
in source-dir copied to under target-dir/src-dir. The issue in this case is, 
the attributes of src-dir is not preserved.

b. when target-dir doesn't exist. It will result in target-dir with all 
contents of source-dir copied to under target-dir. This issue in this  case is, 
the attributes of source-dir is not carried over to target-dir.

For multiple source cases, e.g., command 

  "distcp -pu source-dir1 source-dir2 target-dir"

No matter whether the target-dir exists or not, the multiple sources are copied 
to under the target dir (target-dir is created if it didn't exist). And their 
attributes are preserved. 

ISSUE 2. with the following command:

distcp source-dir target-dir

when source-dir is an empty directory, and when target-dir doesn't exist, 
source-dir is not copied, actually the command behaves like a no-op. 

However, when the source-dir is not empty, it would be copied and results in 
target-dir at the target file system containing a copy of source-dir's children.

To be consistent, empty source dir should be copied too. Basically the  above 
distcp command should cause target-dir get created at target file 
system, and the source-dir's attributes are preserved at target-dir when 
-p is passed.



> distcp V2 doesn't preserve root dir's attributes when -p is specified
> -
>
> Key: HDFS-6152
> URL: https://issues.apache.org/jira/browse/HDFS-6152
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: hdfs-client
>Affects Versions: 2.3.0
>Reporter: Yongjun Zhang
>Assignee: Yongjun Zhang
> Attachments: HDFS-6152.001.patch
>
>
> Two issues were observed with distcpV2
> ISSUE 1. when copying a source dir to target dir with "-pu" option using 
> command 
>   "distcp -pu source-dir target-dir"
>  
> The source dir's owner is not preserved at target dir. Simiarly other 
> attributes of source dir are not preserved.  Supposedly they should be 
> preserved when no -update and no -overwrite specified. 
> There are two scenarios with the above command:
> a. when target-dir already exists. Issuing the above command will  result in 
> target-dir/sou

[jira] [Commented] (HDFS-6130) NPE during namenode upgrade from old release

2014-03-24 Thread Fengdong Yu (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6130?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13945992#comment-13945992
 ] 

Fengdong Yu commented on HDFS-6130:
---

OK, no problem, I can using  rollingUpgrade -prepare to create check point.

> NPE during namenode upgrade from old release
> 
>
> Key: HDFS-6130
> URL: https://issues.apache.org/jira/browse/HDFS-6130
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: namenode
>Affects Versions: 2.4.0
>Reporter: Fengdong Yu
>
> I want upgrade an old cluster(0.20.2-cdh3u1) to trunk instance, 
> I can upgrade successfully if I don't configurage HA, but if HA enabled,
> there is NPE when I run ' hdfs namenode -initializeSharedEdits'
> {code}
> 14/03/20 15:06:41 INFO namenode.FSNamesystem: Retry cache on namenode is 
> enabled
> 14/03/20 15:06:41 INFO namenode.FSNamesystem: Retry cache will use 0.03 of 
> total heap and retry cache entry expiry time is 60 millis
> 14/03/20 15:06:41 INFO util.GSet: Computing capacity for map 
> NameNodeRetryCache
> 14/03/20 15:06:41 INFO util.GSet: VM type   = 64-bit
> 14/03/20 15:06:41 INFO util.GSet: 0.02999329447746% max memory 896 MB = 
> 275.3 KB
> 14/03/20 15:06:41 INFO util.GSet: capacity  = 2^15 = 32768 entries
> 14/03/20 15:06:41 INFO namenode.AclConfigFlag: ACLs enabled? false
> 14/03/20 15:06:41 INFO common.Storage: Lock on 
> /data/hadoop/data1/dfs/name/in_use.lock acquired by nodename 
> 7326@10-150-170-176
> 14/03/20 15:06:42 INFO common.Storage: Lock on 
> /data/hadoop/data2/dfs/name/in_use.lock acquired by nodename 
> 7326@10-150-170-176
> 14/03/20 15:06:42 INFO namenode.FSImage: No edit log streams selected.
> 14/03/20 15:06:42 INFO namenode.FSImageFormatPBINode: Loading 1 INodes.
> 14/03/20 15:06:42 FATAL namenode.NameNode: Exception in namenode join
> java.lang.NullPointerException
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSDirectory.isReservedName(FSDirectory.java:2984)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImageFormatPBINode$Loader.addToParent(FSImageFormatPBINode.java:205)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImageFormatPBINode$Loader.loadINodeDirectorySection(FSImageFormatPBINode.java:162)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImageFormatProtobuf$Loader.loadInternal(FSImageFormatProtobuf.java:243)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImageFormatProtobuf$Loader.load(FSImageFormatProtobuf.java:168)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImageFormat$LoaderDelegator.load(FSImageFormat.java:120)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.loadFSImage(FSImage.java:895)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.loadFSImage(FSImage.java:881)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.loadFSImageFile(FSImage.java:704)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.loadFSImage(FSImage.java:642)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:271)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFSImage(FSNamesystem.java:894)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFromDisk(FSNamesystem.java:653)
>   at 
> org.apache.hadoop.hdfs.server.namenode.NameNode.initializeSharedEdits(NameNode.java:912)
>   at 
> org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1276)
>   at 
> org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1360)
> 14/03/20 15:06:42 INFO util.ExitUtil: Exiting with status 1
> 14/03/20 15:06:42 INFO namenode.NameNode: SHUTDOWN_MSG: 
> /
> SHUTDOWN_MSG: Shutting down NameNode at 10-150-170-176/10.150.170.176
> /
> {code}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6130) NPE during namenode upgrade from old release

2014-03-24 Thread Haohui Mai (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6130?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13945988#comment-13945988
 ] 

Haohui Mai commented on HDFS-6130:
--

Can you create a checkpoint so that upgrading from the checkpointed fsimage 
will triggered the bug?

> NPE during namenode upgrade from old release
> 
>
> Key: HDFS-6130
> URL: https://issues.apache.org/jira/browse/HDFS-6130
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: namenode
>Affects Versions: 2.4.0
>Reporter: Fengdong Yu
>
> I want upgrade an old cluster(0.20.2-cdh3u1) to trunk instance, 
> I can upgrade successfully if I don't configurage HA, but if HA enabled,
> there is NPE when I run ' hdfs namenode -initializeSharedEdits'
> {code}
> 14/03/20 15:06:41 INFO namenode.FSNamesystem: Retry cache on namenode is 
> enabled
> 14/03/20 15:06:41 INFO namenode.FSNamesystem: Retry cache will use 0.03 of 
> total heap and retry cache entry expiry time is 60 millis
> 14/03/20 15:06:41 INFO util.GSet: Computing capacity for map 
> NameNodeRetryCache
> 14/03/20 15:06:41 INFO util.GSet: VM type   = 64-bit
> 14/03/20 15:06:41 INFO util.GSet: 0.02999329447746% max memory 896 MB = 
> 275.3 KB
> 14/03/20 15:06:41 INFO util.GSet: capacity  = 2^15 = 32768 entries
> 14/03/20 15:06:41 INFO namenode.AclConfigFlag: ACLs enabled? false
> 14/03/20 15:06:41 INFO common.Storage: Lock on 
> /data/hadoop/data1/dfs/name/in_use.lock acquired by nodename 
> 7326@10-150-170-176
> 14/03/20 15:06:42 INFO common.Storage: Lock on 
> /data/hadoop/data2/dfs/name/in_use.lock acquired by nodename 
> 7326@10-150-170-176
> 14/03/20 15:06:42 INFO namenode.FSImage: No edit log streams selected.
> 14/03/20 15:06:42 INFO namenode.FSImageFormatPBINode: Loading 1 INodes.
> 14/03/20 15:06:42 FATAL namenode.NameNode: Exception in namenode join
> java.lang.NullPointerException
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSDirectory.isReservedName(FSDirectory.java:2984)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImageFormatPBINode$Loader.addToParent(FSImageFormatPBINode.java:205)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImageFormatPBINode$Loader.loadINodeDirectorySection(FSImageFormatPBINode.java:162)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImageFormatProtobuf$Loader.loadInternal(FSImageFormatProtobuf.java:243)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImageFormatProtobuf$Loader.load(FSImageFormatProtobuf.java:168)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImageFormat$LoaderDelegator.load(FSImageFormat.java:120)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.loadFSImage(FSImage.java:895)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.loadFSImage(FSImage.java:881)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.loadFSImageFile(FSImage.java:704)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.loadFSImage(FSImage.java:642)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:271)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFSImage(FSNamesystem.java:894)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFromDisk(FSNamesystem.java:653)
>   at 
> org.apache.hadoop.hdfs.server.namenode.NameNode.initializeSharedEdits(NameNode.java:912)
>   at 
> org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1276)
>   at 
> org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1360)
> 14/03/20 15:06:42 INFO util.ExitUtil: Exiting with status 1
> 14/03/20 15:06:42 INFO namenode.NameNode: SHUTDOWN_MSG: 
> /
> SHUTDOWN_MSG: Shutting down NameNode at 10-150-170-176/10.150.170.176
> /
> {code}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6150) Add inode id information in the logs to make debugging easier

2014-03-24 Thread Tsz Wo Nicholas Sze (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6150?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tsz Wo Nicholas Sze updated HDFS-6150:
--

Priority: Minor  (was: Major)
Assignee: Suresh Srinivas
Hadoop Flags: Reviewed

+1  HDFS-6150.2.patch looks good.

> Add inode id information in the logs to make debugging easier
> -
>
> Key: HDFS-6150
> URL: https://issues.apache.org/jira/browse/HDFS-6150
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: namenode
>Reporter: Suresh Srinivas
>Assignee: Suresh Srinivas
>Priority: Minor
> Attachments: HDFS-6150.1.patch, HDFS-6150.2.patch, HDFS-6150.patch
>
>
> Inode information and path information are missing in the logs and 
> exceptions. Adding this will help debug multi threading issues related to 
> using incode INode ID information.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6150) Add inode id information in the logs to make debugging easier

2014-03-24 Thread Suresh Srinivas (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6150?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Suresh Srinivas updated HDFS-6150:
--

Attachment: HDFS-6150.2.patch

New patch to fix NPE.

> Add inode id information in the logs to make debugging easier
> -
>
> Key: HDFS-6150
> URL: https://issues.apache.org/jira/browse/HDFS-6150
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: namenode
>Reporter: Suresh Srinivas
> Attachments: HDFS-6150.1.patch, HDFS-6150.2.patch, HDFS-6150.patch
>
>
> Inode information and path information are missing in the logs and 
> exceptions. Adding this will help debug multi threading issues related to 
> using incode INode ID information.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6130) NPE during namenode upgrade from old release

2014-03-24 Thread Fengdong Yu (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6130?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13945943#comment-13945943
 ] 

Fengdong Yu commented on HDFS-6130:
---

Thanks [~szetszwo]!

[~wheat9], do you want only fsimage or both image and edit log? I'll reproduce 
today using 1.3.0 and the latest trunk, then I'll keep the corresponding 
fsimage and edit logs.

> NPE during namenode upgrade from old release
> 
>
> Key: HDFS-6130
> URL: https://issues.apache.org/jira/browse/HDFS-6130
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: namenode
>Affects Versions: 2.4.0
>Reporter: Fengdong Yu
>
> I want upgrade an old cluster(0.20.2-cdh3u1) to trunk instance, 
> I can upgrade successfully if I don't configurage HA, but if HA enabled,
> there is NPE when I run ' hdfs namenode -initializeSharedEdits'
> {code}
> 14/03/20 15:06:41 INFO namenode.FSNamesystem: Retry cache on namenode is 
> enabled
> 14/03/20 15:06:41 INFO namenode.FSNamesystem: Retry cache will use 0.03 of 
> total heap and retry cache entry expiry time is 60 millis
> 14/03/20 15:06:41 INFO util.GSet: Computing capacity for map 
> NameNodeRetryCache
> 14/03/20 15:06:41 INFO util.GSet: VM type   = 64-bit
> 14/03/20 15:06:41 INFO util.GSet: 0.02999329447746% max memory 896 MB = 
> 275.3 KB
> 14/03/20 15:06:41 INFO util.GSet: capacity  = 2^15 = 32768 entries
> 14/03/20 15:06:41 INFO namenode.AclConfigFlag: ACLs enabled? false
> 14/03/20 15:06:41 INFO common.Storage: Lock on 
> /data/hadoop/data1/dfs/name/in_use.lock acquired by nodename 
> 7326@10-150-170-176
> 14/03/20 15:06:42 INFO common.Storage: Lock on 
> /data/hadoop/data2/dfs/name/in_use.lock acquired by nodename 
> 7326@10-150-170-176
> 14/03/20 15:06:42 INFO namenode.FSImage: No edit log streams selected.
> 14/03/20 15:06:42 INFO namenode.FSImageFormatPBINode: Loading 1 INodes.
> 14/03/20 15:06:42 FATAL namenode.NameNode: Exception in namenode join
> java.lang.NullPointerException
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSDirectory.isReservedName(FSDirectory.java:2984)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImageFormatPBINode$Loader.addToParent(FSImageFormatPBINode.java:205)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImageFormatPBINode$Loader.loadINodeDirectorySection(FSImageFormatPBINode.java:162)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImageFormatProtobuf$Loader.loadInternal(FSImageFormatProtobuf.java:243)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImageFormatProtobuf$Loader.load(FSImageFormatProtobuf.java:168)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImageFormat$LoaderDelegator.load(FSImageFormat.java:120)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.loadFSImage(FSImage.java:895)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.loadFSImage(FSImage.java:881)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.loadFSImageFile(FSImage.java:704)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.loadFSImage(FSImage.java:642)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:271)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFSImage(FSNamesystem.java:894)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFromDisk(FSNamesystem.java:653)
>   at 
> org.apache.hadoop.hdfs.server.namenode.NameNode.initializeSharedEdits(NameNode.java:912)
>   at 
> org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1276)
>   at 
> org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1360)
> 14/03/20 15:06:42 INFO util.ExitUtil: Exiting with status 1
> 14/03/20 15:06:42 INFO namenode.NameNode: SHUTDOWN_MSG: 
> /
> SHUTDOWN_MSG: Shutting down NameNode at 10-150-170-176/10.150.170.176
> /
> {code}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6150) Add inode id information in the logs to make debugging easier

2014-03-24 Thread Suresh Srinivas (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6150?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13945938#comment-13945938
 ] 

Suresh Srinivas commented on HDFS-6150:
---

New patch fixes the test failures and address the comments.

> Add inode id information in the logs to make debugging easier
> -
>
> Key: HDFS-6150
> URL: https://issues.apache.org/jira/browse/HDFS-6150
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: namenode
>Reporter: Suresh Srinivas
> Attachments: HDFS-6150.1.patch, HDFS-6150.patch
>
>
> Inode information and path information are missing in the logs and 
> exceptions. Adding this will help debug multi threading issues related to 
> using incode INode ID information.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6150) Add inode id information in the logs to make debugging easier

2014-03-24 Thread Suresh Srinivas (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6150?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Suresh Srinivas updated HDFS-6150:
--

Attachment: HDFS-6150.1.patch

> Add inode id information in the logs to make debugging easier
> -
>
> Key: HDFS-6150
> URL: https://issues.apache.org/jira/browse/HDFS-6150
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: namenode
>Reporter: Suresh Srinivas
> Attachments: HDFS-6150.1.patch, HDFS-6150.patch
>
>
> Inode information and path information are missing in the logs and 
> exceptions. Adding this will help debug multi threading issues related to 
> using incode INode ID information.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-5910) Enhance DataTransferProtocol to allow per-connection choice of encryption/plain-text

2014-03-24 Thread Arpit Agarwal (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5910?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13945925#comment-13945925
 ] 

Arpit Agarwal commented on HDFS-5910:
-

Thanks for making the changes Benoy.

{quote}
The ip address of namenode or datanode is not available at some of the client 
invocations. Please let me know if there is a way to get an ip address..
{quote}
Just for my understanding - lacking the peer's IP address is it your intention 
to use configuration to decide the client's behavior?

I looked through the usages of {{isTrusted}} and some of them already have the 
connected socket available, so it is fairly easy to query the remote's socket 
address and pass it to {{isTrusted}}.

For the usage in getDataEncryptionKey(), we can refactor to pass a functor as 
the encryption key to e.g. {{getFileChecksum}}. However I am okay with doing 
the refactoring in a separate change. We can leave the parameter-less overload 
of {{isTrusted}} for now and just use it from {{getEcnryptionKey}} and file a 
separate Jira to fix it.

{quote}
 I wanted to use InetAddress as the argument to TrustedChannelResolver than a 
string-ip-address to maintain parity with SaslPropertiesResolver. To convert a 
string ip, I use InetAddress.getByName
{quote}
Thanks for the explanation. Will 
[InetAddresses#forString|http://docs.guava-libraries.googlecode.com/git/javadoc/com/google/common/net/InetAddresses.html#forString%28java.lang.String%29]
 from Guava work for you? I just checked and it's available in our build.

> Enhance DataTransferProtocol to allow per-connection choice of 
> encryption/plain-text
> 
>
> Key: HDFS-5910
> URL: https://issues.apache.org/jira/browse/HDFS-5910
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: security
>Affects Versions: 2.2.0
>Reporter: Benoy Antony
>Assignee: Benoy Antony
> Attachments: HDFS-5910.patch, HDFS-5910.patch, HDFS-5910.patch
>
>
> It is possible to enable encryption of DataTransferProtocol. 
> In some use cases, it is required to encrypt data transfer with some clients 
> , but communicate in plain text with some other clients and data nodes.
> A sample use case will be that any data transfer inside a firewall can be in 
> plain text whereas any data transfer from clients  outside the firewall needs 
> to be encrypted.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6124) Add final modifier to class members

2014-03-24 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6124?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13945912#comment-13945912
 ] 

Hudson commented on HDFS-6124:
--

SUCCESS: Integrated in Hadoop-trunk-Commit #5395 (See 
[https://builds.apache.org/job/Hadoop-trunk-Commit/5395/])
HDFS-6124. Add final modifier to class members. (Contributed by Suresh 
Srinivas) (arp: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1581124)
* /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/BlockMissingException.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/BlockReaderLocal.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/BlockReaderLocalLegacy.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/CorruptFileBlockIterator.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/DFSClient.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/DFSHedgedReadMetrics.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/DFSInputStream.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/DFSOutputStream.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/DFSUtil.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/DomainSocketFactory.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/RemoteBlockReader2.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/StorageType.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/client/ShortCircuitCache.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/net/DomainPeerServer.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/net/TcpPeerServer.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/protocol/BlockListAsLongs.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/protocol/BlockLocalPathInfo.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/protocol/CorruptFileBlocks.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/protocol/DatanodeLocalInfo.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/protocol/HdfsConstants.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/protocol/HdfsFileStatus.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/protocol/HdfsLocatedFileStatus.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/protocol/LocatedBlock.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/protocol/LocatedBlocks.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/protocol/RollingUpgradeStatus.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/protocol/SnapshotDiffReport.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/protocol/SnapshottableDirectoryStatus.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/protocol/datatransfer/DataTransferEncryptor.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/protocol/datatransfer/PacketReceiver.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/protocolPB/ClientNamenodeProtocolTranslatorPB.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/qjournal/client/IPCLoggerChannel.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/qjournal/protocol/RequestInfo.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/qjournal/server/JournalMetrics.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/qjournal/server/JournalNode.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/qjournal/server/JournalNodeHttpServer.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/qjournal/s

[jira] [Commented] (HDFS-5840) Follow-up to HDFS-5138 to improve error handling during partial upgrade failures

2014-03-24 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5840?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13945899#comment-13945899
 ] 

Hadoop QA commented on HDFS-5840:
-

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12636424/HDFS-5840.001.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 2 new 
or modified test files.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  There were no new javadoc warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:red}-1 core tests{color}.  The patch failed these unit tests in 
hadoop-hdfs-project/hadoop-hdfs:

  org.apache.hadoop.hdfs.TestSafeMode
  org.apache.hadoop.hdfs.qjournal.TestNNWithQJM

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-HDFS-Build/6479//testReport/
Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/6479//console

This message is automatically generated.

> Follow-up to HDFS-5138 to improve error handling during partial upgrade 
> failures
> 
>
> Key: HDFS-5840
> URL: https://issues.apache.org/jira/browse/HDFS-5840
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: ha, journal-node, namenode
>Affects Versions: 3.0.0, 2.4.0
>Reporter: Aaron T. Myers
>Assignee: Jing Zhao
>Priority: Blocker
> Attachments: HDFS-5840.001.patch, HDFS-5840.002.patch, HDFS-5840.patch
>
>
> Suresh posted some good comment in HDFS-5138 after that patch had already 
> been committed to trunk. This JIRA is to address those. See the first comment 
> of this JIRA for the full content of the review.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6124) Add final modifier to class members

2014-03-24 Thread Arpit Agarwal (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6124?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arpit Agarwal updated HDFS-6124:


  Resolution: Fixed
   Fix Version/s: 2.4.0
  3.0.0
Target Version/s: 2.4.0
Hadoop Flags: Reviewed
  Status: Resolved  (was: Patch Available)

I committed the patch to trunk, branch-2 and branch-2.4.

Thanks for the code cleanup Suresh!

> Add final modifier to class members
> ---
>
> Key: HDFS-6124
> URL: https://issues.apache.org/jira/browse/HDFS-6124
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>Affects Versions: 2.4.0
>Reporter: Suresh Srinivas
>Assignee: Suresh Srinivas
> Fix For: 3.0.0, 2.4.0
>
> Attachments: HDFS-6124.1.patch, HDFS-6124.patch
>
>
> Many of the member variables declaration for classes are missing final 
> modifier in HDFS. This jira adds final modifier where possible.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-5840) Follow-up to HDFS-5138 to improve error handling during partial upgrade failures

2014-03-24 Thread Jing Zhao (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5840?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jing Zhao updated HDFS-5840:


Component/s: journal-node
 ha

> Follow-up to HDFS-5138 to improve error handling during partial upgrade 
> failures
> 
>
> Key: HDFS-5840
> URL: https://issues.apache.org/jira/browse/HDFS-5840
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: ha, journal-node, namenode
>Affects Versions: 3.0.0, 2.4.0
>Reporter: Aaron T. Myers
>Assignee: Jing Zhao
>Priority: Blocker
> Attachments: HDFS-5840.001.patch, HDFS-5840.002.patch, HDFS-5840.patch
>
>
> Suresh posted some good comment in HDFS-5138 after that patch had already 
> been committed to trunk. This JIRA is to address those. See the first comment 
> of this JIRA for the full content of the review.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-5840) Follow-up to HDFS-5138 to improve error handling during partial upgrade failures

2014-03-24 Thread Jing Zhao (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5840?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jing Zhao updated HDFS-5840:


Status: Patch Available  (was: In Progress)

> Follow-up to HDFS-5138 to improve error handling during partial upgrade 
> failures
> 
>
> Key: HDFS-5840
> URL: https://issues.apache.org/jira/browse/HDFS-5840
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: namenode
>Affects Versions: 3.0.0, 2.4.0
>Reporter: Aaron T. Myers
>Assignee: Jing Zhao
>Priority: Blocker
> Attachments: HDFS-5840.001.patch, HDFS-5840.002.patch, HDFS-5840.patch
>
>
> Suresh posted some good comment in HDFS-5138 after that patch had already 
> been committed to trunk. This JIRA is to address those. See the first comment 
> of this JIRA for the full content of the review.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-5910) Enhance DataTransferProtocol to allow per-connection choice of encryption/plain-text

2014-03-24 Thread Benoy Antony (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5910?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Benoy Antony updated HDFS-5910:
---

Attachment: HDFS-5910.patch

Thanks [~arpitagarwal]  for the comments.

I am attaching a new based on some of your comments. I request guidance on #1 
and #4.
{quote}
1.  isSecureOnClient may also want to use the peer's address to 
make a decision. e.g. intra-cluster transfer vs. distcp to remote cluster.
{quote}


The ip address of namenode or datanode is not available at some of the client 
invocations. Please let me know if there is a way to get an ip address..

{quote}
2.  Related to #1, isSecureOnClient and isSecureOnServer look awkward. How 
about replacing both with isTrustedChannel that takes the peer's IP address? We 
should probably avoid overloading the term secure in this context since there 
is a related concept ofPeer#hasSecureChannel().
{quote}

I have renamed the class to TrustedChannelResolver and function to isTrusted() 
. 

{quote}
3. Could you please update the documentation
{quote}

done

{quote}
4. Is the InetAddress.getByName call in DataXceiver#getClientAddress necessary? 
If it were necessary it would have been a security hole since DNS resolution 
may yield a different IP address than the one used by the client. It turns out 
for the kinds of Peers we are interested in this will be an IP address, so 
let's just remove the call.
{quote}

 I wanted to use InetAddress as the argument to TrustedChannelResolver than a 
string-ip-address to maintain parity with _SaslPropertiesResolver_. To convert 
a string ip, I use InetAddress.getByName

>From the documentation of InetAddress.getByName(String host):
The host name can either be a machine name, such as "java.sun.com", or a 
textual representation of its IP address. If a literal IP address is supplied, 
only the validity of the address format is checked.
So basically , if the argument is ip address, getByName doesn't do a DNS check.
If there is a different way to get the InetAddress , we can definitely use 
that. 

Other option is to not care about the parity with _SaslPropertiesResolver_  and 
pass the string ip address. 
Yet another option will be to pass the Peer itself to TrustedChannelResolver  
so that the custom implementation can take care of getting the ip address etc.  
 Will be great to get your opinion on this. 
 


> Enhance DataTransferProtocol to allow per-connection choice of 
> encryption/plain-text
> 
>
> Key: HDFS-5910
> URL: https://issues.apache.org/jira/browse/HDFS-5910
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: security
>Affects Versions: 2.2.0
>Reporter: Benoy Antony
>Assignee: Benoy Antony
> Attachments: HDFS-5910.patch, HDFS-5910.patch, HDFS-5910.patch
>
>
> It is possible to enable encryption of DataTransferProtocol. 
> In some use cases, it is required to encrypt data transfer with some clients 
> , but communicate in plain text with some other clients and data nodes.
> A sample use case will be that any data transfer inside a firewall can be in 
> plain text whereas any data transfer from clients  outside the firewall needs 
> to be encrypted.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-5840) Follow-up to HDFS-5138 to improve error handling during partial upgrade failures

2014-03-24 Thread Jing Zhao (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5840?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jing Zhao updated HDFS-5840:


Fix Version/s: (was: 3.0.0)
 Target Version/s: 2.4.0  (was: 3.0.0)
Affects Version/s: 2.4.0
   Status: In Progress  (was: Patch Available)

> Follow-up to HDFS-5138 to improve error handling during partial upgrade 
> failures
> 
>
> Key: HDFS-5840
> URL: https://issues.apache.org/jira/browse/HDFS-5840
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: namenode
>Affects Versions: 3.0.0, 2.4.0
>Reporter: Aaron T. Myers
>Assignee: Aaron T. Myers
>Priority: Blocker
> Attachments: HDFS-5840.001.patch, HDFS-5840.002.patch, HDFS-5840.patch
>
>
> Suresh posted some good comment in HDFS-5138 after that patch had already 
> been committed to trunk. This JIRA is to address those. See the first comment 
> of this JIRA for the full content of the review.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Assigned] (HDFS-5840) Follow-up to HDFS-5138 to improve error handling during partial upgrade failures

2014-03-24 Thread Jing Zhao (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5840?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jing Zhao reassigned HDFS-5840:
---

Assignee: Jing Zhao  (was: Aaron T. Myers)

> Follow-up to HDFS-5138 to improve error handling during partial upgrade 
> failures
> 
>
> Key: HDFS-5840
> URL: https://issues.apache.org/jira/browse/HDFS-5840
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: namenode
>Affects Versions: 3.0.0, 2.4.0
>Reporter: Aaron T. Myers
>Assignee: Jing Zhao
>Priority: Blocker
> Attachments: HDFS-5840.001.patch, HDFS-5840.002.patch, HDFS-5840.patch
>
>
> Suresh posted some good comment in HDFS-5138 after that patch had already 
> been committed to trunk. This JIRA is to address those. See the first comment 
> of this JIRA for the full content of the review.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-5840) Follow-up to HDFS-5138 to improve error handling during partial upgrade failures

2014-03-24 Thread Jing Zhao (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5840?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jing Zhao updated HDFS-5840:


Attachment: HDFS-5840.002.patch

Update the patch to address Suresh's comments. The patch also switch the 
sequence of the doUpgrade calls on shared edits and on local storage.

Now with the patch we have the following:
# If the doPreUpgrade call on JNs fail (i.e., not all the JNs succeed), we can 
restart NN and JNs for recovery, and NN and JNs will go back to the status 
before upgrade.
# If the doUpgrade call on JNs fail, some JNs may have both previous and 
current directories. Restarting JN cannot solve the issue. The issue has to be 
manually fixed. But the probability of this kind of failure is relatively low 
considering the doPreUpgrade call succeeds on all the JNs.

> Follow-up to HDFS-5138 to improve error handling during partial upgrade 
> failures
> 
>
> Key: HDFS-5840
> URL: https://issues.apache.org/jira/browse/HDFS-5840
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: namenode
>Affects Versions: 3.0.0
>Reporter: Aaron T. Myers
>Assignee: Aaron T. Myers
>Priority: Blocker
> Fix For: 3.0.0
>
> Attachments: HDFS-5840.001.patch, HDFS-5840.002.patch, HDFS-5840.patch
>
>
> Suresh posted some good comment in HDFS-5138 after that patch had already 
> been committed to trunk. This JIRA is to address those. See the first comment 
> of this JIRA for the full content of the review.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6152) distcp V2 doesn't preserve root dir's attributes when -p is specified

2014-03-24 Thread Yongjun Zhang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6152?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yongjun Zhang updated HDFS-6152:


Attachment: HDFS-6152.001.patch

> distcp V2 doesn't preserve root dir's attributes when -p is specified
> -
>
> Key: HDFS-6152
> URL: https://issues.apache.org/jira/browse/HDFS-6152
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: hdfs-client
>Affects Versions: 2.3.0
>Reporter: Yongjun Zhang
>Assignee: Yongjun Zhang
> Attachments: HDFS-6152.001.patch
>
>
> Two issues were observed with distcpV2
> ISSUE 1. when copying a source dir to target dir with "-pu" option using 
> command 
>   "distcp -pu source-dir target-dir"
>  
> The source dir's owner is not preserved at target dir. Simiarly other 
> attributes of source dir are not preserved.  Supposedly they should be 
> preserved when no -update and no -overwrite specified. 
> There are two scenarios with the above command:
> a. when target-dir already exists. Issuing the above command will  result in 
> target-dir/source-dir (source-dir here refers to the last component of the 
> source-dir path in the command line) at target file system, with all contents 
> in source-dir copied to under target-dir/src-dir. The issue in this case is, 
> the attributes of src-dir is not preserved.
> b. when target-dir doesn't exist. It will result in target-dir with all 
> contents of source-dir copied to under target-dir. This issue in this  case 
> is, the attributes of source-dir is not carried over to target-dir.
> For multiple source cases, e.g., command 
>   "distcp -pu source-dir1 source-dir2 target-dir"
> No matter whether the target-dir exists or not, the multiple sources are 
> copied to under the target dir (target-dir is created if it didn't exist). 
> And their attributes are preserved. 
> ISSUE 2. with the following command:
> distcp source-dir target-dir
> when source-dir is an empty directory, and when target-dir doesn't exist, 
> source-dir is not copied, actually the command behaves like a no-op. 
> However, when the source-dir is not empty, it would be copied and results in 
> target-dir at the target file system containing a copy of source-dir's 
> children.
> To be consistent, empty source dir should be copied too. Basically the  above 
> distcp command should cause target-dir get created at target file 
> system, and the source-dir's attributes are preserved at target-dir when 
> -p is passed.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Assigned] (HDFS-6152) distcp V2 doesn't preserve root dir's attributes when -p is specified

2014-03-24 Thread Yongjun Zhang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6152?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yongjun Zhang reassigned HDFS-6152:
---

Assignee: Yongjun Zhang

> distcp V2 doesn't preserve root dir's attributes when -p is specified
> -
>
> Key: HDFS-6152
> URL: https://issues.apache.org/jira/browse/HDFS-6152
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: hdfs-client
>Affects Versions: 2.3.0
>Reporter: Yongjun Zhang
>Assignee: Yongjun Zhang
>
> Two issues were observed with distcpV2
> ISSUE 1. when copying a source dir to target dir with "-pu" option using 
> command 
>   "distcp -pu source-dir target-dir"
>  
> The source dir's owner is not preserved at target dir. Simiarly other 
> attributes of source dir are not preserved.  Supposedly they should be 
> preserved when no -update and no -overwrite specified. 
> There are two scenarios with the above command:
> a. when target-dir already exists. Issuing the above command will  result in 
> target-dir/source-dir (source-dir here refers to the last component of the 
> source-dir path in the command line) at target file system, with all contents 
> in source-dir copied to under target-dir/src-dir. The issue in this case is, 
> the attributes of src-dir is not preserved.
> b. when target-dir doesn't exist. It will result in target-dir with all 
> contents of source-dir copied to under target-dir. This issue in this  case 
> is, the attributes of source-dir is not carried over to target-dir.
> For multiple source cases, e.g., command 
>   "distcp -pu source-dir1 source-dir2 target-dir"
> No matter whether the target-dir exists or not, the multiple sources are 
> copied to under the target dir (target-dir is created if it didn't exist). 
> And their attributes are preserved. 
> ISSUE 2. with the following command:
> distcp source-dir target-dir
> when source-dir is an empty directory, and when target-dir doesn't exist, 
> source-dir is not copied, actually the command behaves like a no-op. 
> However, when the source-dir is not empty, it would be copied and results in 
> target-dir at the target file system containing a copy of source-dir's 
> children.
> To be consistent, empty source dir should be copied too. Basically the  above 
> distcp command should cause target-dir get created at target file 
> system, and the source-dir's attributes are preserved at target-dir when 
> -p is passed.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (HDFS-6152) distcp V2 doesn't preserve root dir's attributes when -p is specified

2014-03-24 Thread Yongjun Zhang (JIRA)
Yongjun Zhang created HDFS-6152:
---

 Summary: distcp V2 doesn't preserve root dir's attributes when -p 
is specified
 Key: HDFS-6152
 URL: https://issues.apache.org/jira/browse/HDFS-6152
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: hdfs-client
Affects Versions: 2.3.0
Reporter: Yongjun Zhang


Two issues were observed with distcpV2

ISSUE 1. when copying a source dir to target dir with "-pu" option using 
command 

  "distcp -pu source-dir target-dir"
 
The source dir's owner is not preserved at target dir. Simiarly other 
attributes of source dir are not preserved.  Supposedly they should be 
preserved when no -update and no -overwrite specified. 

There are two scenarios with the above command:

a. when target-dir already exists. Issuing the above command will  result in 
target-dir/source-dir (source-dir here refers to the last component of the 
source-dir path in the command line) at target file system, with all contents 
in source-dir copied to under target-dir/src-dir. The issue in this case is, 
the attributes of src-dir is not preserved.

b. when target-dir doesn't exist. It will result in target-dir with all 
contents of source-dir copied to under target-dir. This issue in this  case is, 
the attributes of source-dir is not carried over to target-dir.

For multiple source cases, e.g., command 

  "distcp -pu source-dir1 source-dir2 target-dir"

No matter whether the target-dir exists or not, the multiple sources are copied 
to under the target dir (target-dir is created if it didn't exist). And their 
attributes are preserved. 

ISSUE 2. with the following command:

distcp source-dir target-dir

when source-dir is an empty directory, and when target-dir doesn't exist, 
source-dir is not copied, actually the command behaves like a no-op. 

However, when the source-dir is not empty, it would be copied and results in 
target-dir at the target file system containing a copy of source-dir's children.

To be consistent, empty source dir should be copied too. Basically the  above 
distcp command should cause target-dir get created at target file 
system, and the source-dir's attributes are preserved at target-dir when 
-p is passed.




--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6124) Add final modifier to class members

2014-03-24 Thread Arpit Agarwal (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6124?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13945865#comment-13945865
 ] 

Arpit Agarwal commented on HDFS-6124:
-

+1 for the patch. I will commit it shortly.

> Add final modifier to class members
> ---
>
> Key: HDFS-6124
> URL: https://issues.apache.org/jira/browse/HDFS-6124
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>Affects Versions: 2.4.0
>Reporter: Suresh Srinivas
>Assignee: Suresh Srinivas
> Attachments: HDFS-6124.1.patch, HDFS-6124.patch
>
>
> Many of the member variables declaration for classes are missing final 
> modifier in HDFS. This jira adds final modifier where possible.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6150) Add inode id information in the logs to make debugging easier

2014-03-24 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6150?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13945840#comment-13945840
 ] 

Hadoop QA commented on HDFS-6150:
-

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12636419/HDFS-6150.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:red}-1 tests included{color}.  The patch doesn't appear to include 
any new or modified tests.
Please justify why no new tests are needed for this 
patch.
Also please list what manual steps were performed to 
verify this patch.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  There were no new javadoc warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:red}-1 findbugs{color}.  The patch appears to introduce 2 new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:red}-1 core tests{color}.  The patch failed these unit tests in 
hadoop-hdfs-project/hadoop-hdfs:

  org.apache.hadoop.hdfs.TestFileCreation
  org.apache.hadoop.hdfs.TestLeaseRecovery2

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-HDFS-Build/6478//testReport/
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HDFS-Build/6478//artifact/trunk/patchprocess/newPatchFindbugsWarningshadoop-hdfs.html
Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/6478//console

This message is automatically generated.

> Add inode id information in the logs to make debugging easier
> -
>
> Key: HDFS-6150
> URL: https://issues.apache.org/jira/browse/HDFS-6150
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: namenode
>Reporter: Suresh Srinivas
> Attachments: HDFS-6150.patch
>
>
> Inode information and path information are missing in the logs and 
> exceptions. Adding this will help debug multi threading issues related to 
> using incode INode ID information.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6125) Cleanup unnecessary cast in HDFS code base

2014-03-24 Thread Chris Nauroth (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6125?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chris Nauroth updated HDFS-6125:


  Component/s: test
 Target Version/s: 3.0.0, 2.5.0
Affects Version/s: 2.4.0
 Hadoop Flags: Reviewed

> Cleanup unnecessary cast in HDFS code base
> --
>
> Key: HDFS-6125
> URL: https://issues.apache.org/jira/browse/HDFS-6125
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: test
>Affects Versions: 2.4.0
>Reporter: Suresh Srinivas
>Assignee: Suresh Srinivas
> Attachments: HDFS-6125.2.patch, HDFS-6125.patch
>
>




--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6134) Transparent data at rest encryption

2014-03-24 Thread Alejandro Abdelnur (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6134?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13945828#comment-13945828
 ] 

Alejandro Abdelnur commented on HDFS-6134:
--

[~lmccay], gr8, I have done some work already on this area while prototyping, 
I'll create a few JIRAs later tonight and put up patches for the stuff I 
already have.

> Transparent data at rest encryption
> ---
>
> Key: HDFS-6134
> URL: https://issues.apache.org/jira/browse/HDFS-6134
> Project: Hadoop HDFS
>  Issue Type: New Feature
>  Components: security
>Affects Versions: 2.3.0
>Reporter: Alejandro Abdelnur
>Assignee: Alejandro Abdelnur
> Attachments: HDFSDataAtRestEncryption.pdf
>
>
> Because of privacy and security regulations, for many industries, sensitive 
> data at rest must be in encrypted form. For example: the health­care industry 
> (HIPAA regulations), the card payment industry (PCI DSS regulations) or the 
> US government (FISMA regulations).
> This JIRA aims to provide a mechanism to encrypt HDFS data at rest that can 
> be used transparently by any application accessing HDFS via Hadoop Filesystem 
> Java API, Hadoop libhdfs C library, or WebHDFS REST API.
> The resulting implementation should be able to be used in compliance with 
> different regulation requirements.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6125) Cleanup unnecessary cast in HDFS code base

2014-03-24 Thread Chris Nauroth (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6125?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chris Nauroth updated HDFS-6125:


Attachment: HDFS-6125.2.patch

{{FSEditLogOp}} needed a very minor rebase to apply on current trunk.  Instead 
of going back and forth with comments, I just made the change locally, and I'm 
uploading the new patch.

I'm +1 for patch v2, pending Jenkins.  Thanks again, Suresh.

> Cleanup unnecessary cast in HDFS code base
> --
>
> Key: HDFS-6125
> URL: https://issues.apache.org/jira/browse/HDFS-6125
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>Reporter: Suresh Srinivas
>Assignee: Suresh Srinivas
> Attachments: HDFS-6125.2.patch, HDFS-6125.patch
>
>




--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6150) Add inode id information in the logs to make debugging easier

2014-03-24 Thread Tsz Wo Nicholas Sze (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6150?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13945823#comment-13945823
 ] 

Tsz Wo Nicholas Sze commented on HDFS-6150:
---

In the existing code, there are two places below that it says "Failed to ... " 
+ src + " on client " + clientMachine.  It seems saying the src is "on" the 
client machine.  It should be "for" the client.  Could you also fix it as well?
{code}
@@ -2292,8 +2292,8 @@ private void startFileInternal(FSPermissionChecker pc, 
String src,
 try {
   if (myFile == null) {
 if (!create) {
-  throw new FileNotFoundException("failed to overwrite non-existent 
file "
-+ src + " on client " + clientMachine);
+  throw new FileNotFoundException("Can't overwrite non-existent " +
+  src + " on client " + clientMachine);
 }
   } else {
 if (overwrite) {
@@ -2306,8 +2306,8 @@ private void startFileInternal(FSPermissionChecker pc, 
String src,
 } else {
   // If lease soft limit time is expired, recover the lease
   recoverLeaseInternal(myFile, src, holder, clientMachine, false);
-  throw new FileAlreadyExistsException("failed to create file " + src
-  + " on client " + clientMachine + " because the file exists");
+  throw new FileAlreadyExistsException(src + " on client " +
+  clientMachine + " already exists");
 }
   }
{code}


> Add inode id information in the logs to make debugging easier
> -
>
> Key: HDFS-6150
> URL: https://issues.apache.org/jira/browse/HDFS-6150
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: namenode
>Reporter: Suresh Srinivas
> Attachments: HDFS-6150.patch
>
>
> Inode information and path information are missing in the logs and 
> exceptions. Adding this will help debug multi threading issues related to 
> using incode INode ID information.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6125) Cleanup unnecessary cast in HDFS code base

2014-03-24 Thread Chris Nauroth (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6125?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13945811#comment-13945811
 ] 

Chris Nauroth commented on HDFS-6125:
-

Thanks for doing these clean-ups, Suresh.  I can start code reviewing this.

> Cleanup unnecessary cast in HDFS code base
> --
>
> Key: HDFS-6125
> URL: https://issues.apache.org/jira/browse/HDFS-6125
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>Reporter: Suresh Srinivas
>Assignee: Suresh Srinivas
> Attachments: HDFS-6125.patch
>
>




--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-5846) Assigning DEFAULT_RACK in resolveNetworkLocation method can break data resiliency

2014-03-24 Thread Chris Nauroth (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5846?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chris Nauroth updated HDFS-5846:


   Resolution: Fixed
Fix Version/s: 2.4.0
   3.0.0
   Status: Resolved  (was: Patch Available)

I committed this to trunk, branch-2 and branch-2.4.  Nikola, thank you for 
contributing this code.

> Assigning DEFAULT_RACK in resolveNetworkLocation method can break data 
> resiliency
> -
>
> Key: HDFS-5846
> URL: https://issues.apache.org/jira/browse/HDFS-5846
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: namenode
>Affects Versions: 3.0.0, 2.3.0
>Reporter: Nikola Vujic
>Assignee: Nikola Vujic
> Fix For: 3.0.0, 2.4.0
>
> Attachments: hdfs-5846.patch, hdfs-5846.patch, hdfs-5846.patch
>
>
> Medhod CachedDNSToSwitchMapping::resolve() can return NULL which requires 
> careful handling. Null can be returned in two cases:
> • An error occurred with topology script execution (script crashes).
> • Script returns wrong number of values (other than expected)
> Critical handling is in the DN registration code. DN registration code is 
> responsible for assigning proper topology paths to all registered datanodes. 
> Existing code handles this NULL pointer on the following way 
> ({{resolveNetworkLocation}} method):
> {code}
> / /resolve its network location
> List rName = dnsToSwitchMapping.resolve(names);
> String networkLocation;
> if (rName == null) {
>   LOG.error("The resolve call returned null! Using " + 
>   NetworkTopology.DEFAULT_RACK + " for host " + names);
>   networkLocation = NetworkTopology.DEFAULT_RACK;
> } else {
>   networkLocation = rName.get(0);
> }
> return networkLocation;
> {code}
> The line of code that is assigning default rack:
> {code} networkLocation = NetworkTopology.DEFAULT_RACK; {code} 
> can cause a serious problem. This means if somehow we got NULL, then the 
> default rack will be assigned as a DN's network location and DN's 
> registration will finish successfully. Under this circumstances, we will be 
> able to load data into cluster which is working with a wrong topology. Wrong  
> topology means that fault domains are not honored. 
> For the end user, it means that two data replicas can end up in the same 
> fault domain and a single failure can cause loss of two, or more, replicas. 
> Cluster would be in the inconsistent state but it would not be aware of that 
> and the whole thing would work as if everything was fine. We can notice that 
> something wrong happened almost only by looking in the log for the error:
> {code}
> LOG.error("The resolve call returned null! Using " + 
> NetworkTopology.DEFAULT_RACK + " for host " + names);
> {code}
>  



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6134) Transparent data at rest encryption

2014-03-24 Thread Larry McCay (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6134?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13945790#comment-13945790
 ] 

Larry McCay commented on HDFS-6134:
---

Hi [~tucu00] - I like what I see here. We should file jira's for the 
KeyProvider API work that you mention in your document and discuss some of 
those aspects there. We have a number of common interests in that area.

> Transparent data at rest encryption
> ---
>
> Key: HDFS-6134
> URL: https://issues.apache.org/jira/browse/HDFS-6134
> Project: Hadoop HDFS
>  Issue Type: New Feature
>  Components: security
>Affects Versions: 2.3.0
>Reporter: Alejandro Abdelnur
>Assignee: Alejandro Abdelnur
> Attachments: HDFSDataAtRestEncryption.pdf
>
>
> Because of privacy and security regulations, for many industries, sensitive 
> data at rest must be in encrypted form. For example: the health­care industry 
> (HIPAA regulations), the card payment industry (PCI DSS regulations) or the 
> US government (FISMA regulations).
> This JIRA aims to provide a mechanism to encrypt HDFS data at rest that can 
> be used transparently by any application accessing HDFS via Hadoop Filesystem 
> Java API, Hadoop libhdfs C library, or WebHDFS REST API.
> The resulting implementation should be able to be used in compliance with 
> different regulation requirements.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-5807) TestBalancerWithNodeGroup.testBalancerWithNodeGroup fails intermittently on Branch-2

2014-03-24 Thread Chen He (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5807?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13945788#comment-13945788
 ] 

Chen He commented on HDFS-5807:
---

Working on it.

> TestBalancerWithNodeGroup.testBalancerWithNodeGroup fails intermittently on 
> Branch-2
> 
>
> Key: HDFS-5807
> URL: https://issues.apache.org/jira/browse/HDFS-5807
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: test
>Affects Versions: 2.3.0
>Reporter: Mit Desai
>Assignee: Chen He
> Fix For: 3.0.0, 2.4.0
>
> Attachments: HDFS-5807.patch
>
>
> The test times out after some time.
> {noformat}
> java.util.concurrent.TimeoutException: Rebalancing expected avg utilization 
> to become 0.16, but on datanode 127.0.0.1:42451 it remains at 0.3 after more 
> than 2 msec.
>   at 
> org.apache.hadoop.hdfs.server.balancer.TestBalancerWithNodeGroup.waitForBalancer(TestBalancerWithNodeGroup.java:151)
>   at 
> org.apache.hadoop.hdfs.server.balancer.TestBalancerWithNodeGroup.runBalancer(TestBalancerWithNodeGroup.java:178)
>   at 
> org.apache.hadoop.hdfs.server.balancer.TestBalancerWithNodeGroup.testBalancerWithNodeGroup(TestBalancerWithNodeGroup.java:302)
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-5846) Assigning DEFAULT_RACK in resolveNetworkLocation method can break data resiliency

2014-03-24 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5846?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13945782#comment-13945782
 ] 

Hudson commented on HDFS-5846:
--

SUCCESS: Integrated in Hadoop-trunk-Commit #5393 (See 
[https://builds.apache.org/job/Hadoop-trunk-Commit/5393/])
HDFS-5846. Shuffle phase is slow in Windows - FadviseFileRegion::transferTo 
does not read disks efficiently. Contributed by Nikola Vujic. (cnauroth: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1581091)
* /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/DFSConfigKeys.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/DatanodeManager.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/UnresolvedTopologyException.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/resources/hdfs-default.xml
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/blockmanagement/TestDatanodeManager.java


> Assigning DEFAULT_RACK in resolveNetworkLocation method can break data 
> resiliency
> -
>
> Key: HDFS-5846
> URL: https://issues.apache.org/jira/browse/HDFS-5846
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: namenode
>Affects Versions: 3.0.0, 2.3.0
>Reporter: Nikola Vujic
>Assignee: Nikola Vujic
> Attachments: hdfs-5846.patch, hdfs-5846.patch, hdfs-5846.patch
>
>
> Medhod CachedDNSToSwitchMapping::resolve() can return NULL which requires 
> careful handling. Null can be returned in two cases:
> • An error occurred with topology script execution (script crashes).
> • Script returns wrong number of values (other than expected)
> Critical handling is in the DN registration code. DN registration code is 
> responsible for assigning proper topology paths to all registered datanodes. 
> Existing code handles this NULL pointer on the following way 
> ({{resolveNetworkLocation}} method):
> {code}
> / /resolve its network location
> List rName = dnsToSwitchMapping.resolve(names);
> String networkLocation;
> if (rName == null) {
>   LOG.error("The resolve call returned null! Using " + 
>   NetworkTopology.DEFAULT_RACK + " for host " + names);
>   networkLocation = NetworkTopology.DEFAULT_RACK;
> } else {
>   networkLocation = rName.get(0);
> }
> return networkLocation;
> {code}
> The line of code that is assigning default rack:
> {code} networkLocation = NetworkTopology.DEFAULT_RACK; {code} 
> can cause a serious problem. This means if somehow we got NULL, then the 
> default rack will be assigned as a DN's network location and DN's 
> registration will finish successfully. Under this circumstances, we will be 
> able to load data into cluster which is working with a wrong topology. Wrong  
> topology means that fault domains are not honored. 
> For the end user, it means that two data replicas can end up in the same 
> fault domain and a single failure can cause loss of two, or more, replicas. 
> Cluster would be in the inconsistent state but it would not be aware of that 
> and the whole thing would work as if everything was fine. We can notice that 
> something wrong happened almost only by looking in the log for the error:
> {code}
> LOG.error("The resolve call returned null! Using " + 
> NetworkTopology.DEFAULT_RACK + " for host " + names);
> {code}
>  



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-5138) Support HDFS upgrade in HA

2014-03-24 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5138?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13945748#comment-13945748
 ] 

Hudson commented on HDFS-5138:
--

SUCCESS: Integrated in Hadoop-trunk-Commit #5392 (See 
[https://builds.apache.org/job/Hadoop-trunk-Commit/5392/])
Move HDFS-5138 to 2.4.0 section in CHANGES.txt (jing9: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1581074)
* /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt


> Support HDFS upgrade in HA
> --
>
> Key: HDFS-5138
> URL: https://issues.apache.org/jira/browse/HDFS-5138
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 2.1.1-beta
>Reporter: Kihwal Lee
>Assignee: Aaron T. Myers
>Priority: Blocker
> Fix For: 3.0.0, 2.4.0
>
> Attachments: HDFS-5138.branch-2.001.patch, HDFS-5138.patch, 
> HDFS-5138.patch, HDFS-5138.patch, HDFS-5138.patch, HDFS-5138.patch, 
> HDFS-5138.patch, HDFS-5138.patch, HDFS-5138.patch, HDFS-5138.patch, 
> HDFS-5138.patch, HDFS-5138.patch, hdfs-5138-branch-2.txt
>
>
> With HA enabled, NN wo't start with "-upgrade". Since there has been a layout 
> version change between 2.0.x and 2.1.x, starting NN in upgrade mode was 
> necessary when deploying 2.1.x to an existing 2.0.x cluster. But the only way 
> to get around this was to disable HA and upgrade. 
> The NN and the cluster cannot be flipped back to HA until the upgrade is 
> finalized. If HA is disabled only on NN for layout upgrade and HA is turned 
> back on without involving DNs, things will work, but finaliizeUpgrade won't 
> work (the NN is in HA and it cannot be in upgrade mode) and DN's upgrade 
> snapshots won't get removed.
> We will need a different ways of doing layout upgrade and upgrade snapshot.  
> I am marking this as a 2.1.1-beta blocker based on feedback from others.  If 
> there is a reasonable workaround that does not increase maintenance window 
> greatly, we can lower its priority from blocker to critical.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6135) In HDFS upgrade with HA setup, JournalNode cannot handle layout version bump when rolling back

2014-03-24 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6135?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13945749#comment-13945749
 ] 

Hudson commented on HDFS-6135:
--

SUCCESS: Integrated in Hadoop-trunk-Commit #5392 (See 
[https://builds.apache.org/job/Hadoop-trunk-Commit/5392/])
HDFS-6135. In HDFS upgrade with HA setup, JournalNode cannot handle layout 
version bump when rolling back. Contributed by Jing Zhao. (jing9: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1581070)
* /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/qjournal/server/JNStorage.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/common/StorageInfo.java


> In HDFS upgrade with HA setup, JournalNode cannot handle layout version bump 
> when rolling back
> --
>
> Key: HDFS-6135
> URL: https://issues.apache.org/jira/browse/HDFS-6135
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: journal-node
>Affects Versions: 3.0.0
>Reporter: Jing Zhao
>Assignee: Jing Zhao
>Priority: Blocker
> Fix For: 2.4.0
>
> Attachments: HDFS-6135.000.patch, HDFS-6135.001.patch, 
> HDFS-6135.002.patch, HDFS-6135.test.txt
>
>
> While doing HDFS upgrade with HA setup, if the layoutversion gets changed in 
> the upgrade, the rollback may trigger the following exception in JournalNodes 
> (suppose the new software bumped the layoutversion from -55 to -56):
> {code}
> 14/03/21 01:01:53 FATAL namenode.NameNode: Exception in namenode join
> org.apache.hadoop.hdfs.qjournal.client.QuorumException: Could not check if 
> roll back possible for one or more JournalNodes. 1 exceptions thrown:
> Unexpected version of storage directory /grid/1/tmp/journal/mycluster. 
> Reported: -56. Expecting = -55.
>   at 
> org.apache.hadoop.hdfs.server.common.StorageInfo.setLayoutVersion(StorageInfo.java:203)
>   at 
> org.apache.hadoop.hdfs.server.common.StorageInfo.setFieldsFromProperties(StorageInfo.java:156)
>   at 
> org.apache.hadoop.hdfs.server.common.StorageInfo.readProperties(StorageInfo.java:135)
>   at 
> org.apache.hadoop.hdfs.qjournal.server.JNStorage.analyzeStorage(JNStorage.java:202)
>   at 
> org.apache.hadoop.hdfs.qjournal.server.JNStorage.(JNStorage.java:73)
>   at 
> org.apache.hadoop.hdfs.qjournal.server.Journal.(Journal.java:142)
>   at 
> org.apache.hadoop.hdfs.qjournal.server.JournalNode.getOrCreateJournal(JournalNode.java:87)
>   at 
> org.apache.hadoop.hdfs.qjournal.server.JournalNode.canRollBack(JournalNode.java:304)
>   at 
> org.apache.hadoop.hdfs.qjournal.server.JournalNodeRpcServer.canRollBack(JournalNodeRpcServer.java:228)
> {code}
> Looks like for rollback JN with old software cannot handle future 
> layoutversion brought by new software.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Reopened] (HDFS-5807) TestBalancerWithNodeGroup.testBalancerWithNodeGroup fails intermittently on Branch-2

2014-03-24 Thread Mit Desai (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5807?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mit Desai reopened HDFS-5807:
-


[~airbots], I found this test failing again in our nightly builds, Can you take 
a look into it again? 

{noformat}
Error Message

Rebalancing expected avg utilization to become 0.16, but on datanode 
X.X.X.X: it remains at 0.3 after more than 4 msec.

Stacktrace

java.util.concurrent.TimeoutException: Rebalancing expected avg utilization to 
become 0.16, but on datanode X.X.X.X: it remains at 0.3 after more than 
4 msec.
at 
org.apache.hadoop.hdfs.server.balancer.TestBalancerWithNodeGroup.waitForBalancer(TestBalancerWithNodeGroup.java:151)
at 
org.apache.hadoop.hdfs.server.balancer.TestBalancerWithNodeGroup.runBalancer(TestBalancerWithNodeGroup.java:178)
at 
org.apache.hadoop.hdfs.server.balancer.TestBalancerWithNodeGroup.testBalancerWithNodeGroup(TestBalancerWithNodeGroup.java:302)

{noformat}

> TestBalancerWithNodeGroup.testBalancerWithNodeGroup fails intermittently on 
> Branch-2
> 
>
> Key: HDFS-5807
> URL: https://issues.apache.org/jira/browse/HDFS-5807
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: test
>Affects Versions: 2.3.0
>Reporter: Mit Desai
>Assignee: Chen He
> Fix For: 3.0.0, 2.4.0
>
> Attachments: HDFS-5807.patch
>
>
> The test times out after some time.
> {noformat}
> java.util.concurrent.TimeoutException: Rebalancing expected avg utilization 
> to become 0.16, but on datanode 127.0.0.1:42451 it remains at 0.3 after more 
> than 2 msec.
>   at 
> org.apache.hadoop.hdfs.server.balancer.TestBalancerWithNodeGroup.waitForBalancer(TestBalancerWithNodeGroup.java:151)
>   at 
> org.apache.hadoop.hdfs.server.balancer.TestBalancerWithNodeGroup.runBalancer(TestBalancerWithNodeGroup.java:178)
>   at 
> org.apache.hadoop.hdfs.server.balancer.TestBalancerWithNodeGroup.testBalancerWithNodeGroup(TestBalancerWithNodeGroup.java:302)
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-5840) Follow-up to HDFS-5138 to improve error handling during partial upgrade failures

2014-03-24 Thread Jing Zhao (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5840?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13945728#comment-13945728
 ] 

Jing Zhao commented on HDFS-5840:
-

Thanks for the comments, Suresh! Will update the patch to address the comments.

> Follow-up to HDFS-5138 to improve error handling during partial upgrade 
> failures
> 
>
> Key: HDFS-5840
> URL: https://issues.apache.org/jira/browse/HDFS-5840
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: namenode
>Affects Versions: 3.0.0
>Reporter: Aaron T. Myers
>Assignee: Aaron T. Myers
>Priority: Blocker
> Fix For: 3.0.0
>
> Attachments: HDFS-5840.001.patch, HDFS-5840.patch
>
>
> Suresh posted some good comment in HDFS-5138 after that patch had already 
> been committed to trunk. This JIRA is to address those. See the first comment 
> of this JIRA for the full content of the review.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6135) In HDFS upgrade with HA setup, JournalNode cannot handle layout version bump when rolling back

2014-03-24 Thread Jing Zhao (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6135?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jing Zhao updated HDFS-6135:


   Resolution: Fixed
Fix Version/s: 2.4.0
   Status: Resolved  (was: Patch Available)

Thanks for the review, Nicholas! I've committed this to trunk, branch-2, and 
branch-2.4.0.

> In HDFS upgrade with HA setup, JournalNode cannot handle layout version bump 
> when rolling back
> --
>
> Key: HDFS-6135
> URL: https://issues.apache.org/jira/browse/HDFS-6135
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: journal-node
>Affects Versions: 3.0.0
>Reporter: Jing Zhao
>Assignee: Jing Zhao
>Priority: Blocker
> Fix For: 2.4.0
>
> Attachments: HDFS-6135.000.patch, HDFS-6135.001.patch, 
> HDFS-6135.002.patch, HDFS-6135.test.txt
>
>
> While doing HDFS upgrade with HA setup, if the layoutversion gets changed in 
> the upgrade, the rollback may trigger the following exception in JournalNodes 
> (suppose the new software bumped the layoutversion from -55 to -56):
> {code}
> 14/03/21 01:01:53 FATAL namenode.NameNode: Exception in namenode join
> org.apache.hadoop.hdfs.qjournal.client.QuorumException: Could not check if 
> roll back possible for one or more JournalNodes. 1 exceptions thrown:
> Unexpected version of storage directory /grid/1/tmp/journal/mycluster. 
> Reported: -56. Expecting = -55.
>   at 
> org.apache.hadoop.hdfs.server.common.StorageInfo.setLayoutVersion(StorageInfo.java:203)
>   at 
> org.apache.hadoop.hdfs.server.common.StorageInfo.setFieldsFromProperties(StorageInfo.java:156)
>   at 
> org.apache.hadoop.hdfs.server.common.StorageInfo.readProperties(StorageInfo.java:135)
>   at 
> org.apache.hadoop.hdfs.qjournal.server.JNStorage.analyzeStorage(JNStorage.java:202)
>   at 
> org.apache.hadoop.hdfs.qjournal.server.JNStorage.(JNStorage.java:73)
>   at 
> org.apache.hadoop.hdfs.qjournal.server.Journal.(Journal.java:142)
>   at 
> org.apache.hadoop.hdfs.qjournal.server.JournalNode.getOrCreateJournal(JournalNode.java:87)
>   at 
> org.apache.hadoop.hdfs.qjournal.server.JournalNode.canRollBack(JournalNode.java:304)
>   at 
> org.apache.hadoop.hdfs.qjournal.server.JournalNodeRpcServer.canRollBack(JournalNodeRpcServer.java:228)
> {code}
> Looks like for rollback JN with old software cannot handle future 
> layoutversion brought by new software.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-5846) Assigning DEFAULT_RACK in resolveNetworkLocation method can break data resiliency

2014-03-24 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5846?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13945710#comment-13945710
 ] 

Hadoop QA commented on HDFS-5846:
-

{color:green}+1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12636402/hdfs-5846.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 1 new 
or modified test files.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  There were no new javadoc warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:green}+1 core tests{color}.  The patch passed unit tests in 
hadoop-hdfs-project/hadoop-hdfs.

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-HDFS-Build/6477//testReport/
Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/6477//console

This message is automatically generated.

> Assigning DEFAULT_RACK in resolveNetworkLocation method can break data 
> resiliency
> -
>
> Key: HDFS-5846
> URL: https://issues.apache.org/jira/browse/HDFS-5846
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: namenode
>Affects Versions: 3.0.0, 2.3.0
>Reporter: Nikola Vujic
>Assignee: Nikola Vujic
> Attachments: hdfs-5846.patch, hdfs-5846.patch, hdfs-5846.patch
>
>
> Medhod CachedDNSToSwitchMapping::resolve() can return NULL which requires 
> careful handling. Null can be returned in two cases:
> • An error occurred with topology script execution (script crashes).
> • Script returns wrong number of values (other than expected)
> Critical handling is in the DN registration code. DN registration code is 
> responsible for assigning proper topology paths to all registered datanodes. 
> Existing code handles this NULL pointer on the following way 
> ({{resolveNetworkLocation}} method):
> {code}
> / /resolve its network location
> List rName = dnsToSwitchMapping.resolve(names);
> String networkLocation;
> if (rName == null) {
>   LOG.error("The resolve call returned null! Using " + 
>   NetworkTopology.DEFAULT_RACK + " for host " + names);
>   networkLocation = NetworkTopology.DEFAULT_RACK;
> } else {
>   networkLocation = rName.get(0);
> }
> return networkLocation;
> {code}
> The line of code that is assigning default rack:
> {code} networkLocation = NetworkTopology.DEFAULT_RACK; {code} 
> can cause a serious problem. This means if somehow we got NULL, then the 
> default rack will be assigned as a DN's network location and DN's 
> registration will finish successfully. Under this circumstances, we will be 
> able to load data into cluster which is working with a wrong topology. Wrong  
> topology means that fault domains are not honored. 
> For the end user, it means that two data replicas can end up in the same 
> fault domain and a single failure can cause loss of two, or more, replicas. 
> Cluster would be in the inconsistent state but it would not be aware of that 
> and the whole thing would work as if everything was fine. We can notice that 
> something wrong happened almost only by looking in the log for the error:
> {code}
> LOG.error("The resolve call returned null! Using " + 
> NetworkTopology.DEFAULT_RACK + " for host " + names);
> {code}
>  



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Resolved] (HDFS-5138) Support HDFS upgrade in HA

2014-03-24 Thread Jing Zhao (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5138?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jing Zhao resolved HDFS-5138.
-

   Resolution: Fixed
Fix Version/s: 2.4.0

I've merged this to branch-2 and branch-2.4.0.

> Support HDFS upgrade in HA
> --
>
> Key: HDFS-5138
> URL: https://issues.apache.org/jira/browse/HDFS-5138
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 2.1.1-beta
>Reporter: Kihwal Lee
>Assignee: Aaron T. Myers
>Priority: Blocker
> Fix For: 3.0.0, 2.4.0
>
> Attachments: HDFS-5138.branch-2.001.patch, HDFS-5138.patch, 
> HDFS-5138.patch, HDFS-5138.patch, HDFS-5138.patch, HDFS-5138.patch, 
> HDFS-5138.patch, HDFS-5138.patch, HDFS-5138.patch, HDFS-5138.patch, 
> HDFS-5138.patch, HDFS-5138.patch, hdfs-5138-branch-2.txt
>
>
> With HA enabled, NN wo't start with "-upgrade". Since there has been a layout 
> version change between 2.0.x and 2.1.x, starting NN in upgrade mode was 
> necessary when deploying 2.1.x to an existing 2.0.x cluster. But the only way 
> to get around this was to disable HA and upgrade. 
> The NN and the cluster cannot be flipped back to HA until the upgrade is 
> finalized. If HA is disabled only on NN for layout upgrade and HA is turned 
> back on without involving DNs, things will work, but finaliizeUpgrade won't 
> work (the NN is in HA and it cannot be in upgrade mode) and DN's upgrade 
> snapshots won't get removed.
> We will need a different ways of doing layout upgrade and upgrade snapshot.  
> I am marking this as a 2.1.1-beta blocker based on feedback from others.  If 
> there is a reasonable workaround that does not increase maintenance window 
> greatly, we can lower its priority from blocker to critical.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6149) Running Httpfs UTs with testKerberos profile has failures.

2014-03-24 Thread Jinghui Wang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6149?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jinghui Wang updated HDFS-6149:
---

Summary: Running Httpfs UTs with testKerberos profile has failures.  (was: 
Running UTs with testKerberos profile has failures.)

> Running Httpfs UTs with testKerberos profile has failures.
> --
>
> Key: HDFS-6149
> URL: https://issues.apache.org/jira/browse/HDFS-6149
> Project: Hadoop HDFS
>  Issue Type: Test
>  Components: test
>Affects Versions: 2.2.0
>Reporter: Jinghui Wang
>Assignee: Jinghui Wang
> Fix For: 2.3.0
>
> Attachments: HDFS-6149.patch
>
>
> UT failures in TestHttpFSWithKerberos.
> Tests using testDelegationTokenWithinDoAs fail because of the statically set 
> keytab file.
> Test testDelegationTokenHttpFSAccess also fails due the incorrect assumption 
> that CANCELDELEGATIONTOKEN does not require credentials.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-5138) Support HDFS upgrade in HA

2014-03-24 Thread Jing Zhao (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5138?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13945688#comment-13945688
 ] 

Jing Zhao commented on HDFS-5138:
-

Thanks for the quick review, Nicholas! I will commit the backport patch to 
branch-2 and 2.4. We can continue to fix remaining issues in HDFS-5840 and 
HDFS-6135.

> Support HDFS upgrade in HA
> --
>
> Key: HDFS-5138
> URL: https://issues.apache.org/jira/browse/HDFS-5138
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 2.1.1-beta
>Reporter: Kihwal Lee
>Assignee: Aaron T. Myers
>Priority: Blocker
> Fix For: 3.0.0
>
> Attachments: HDFS-5138.branch-2.001.patch, HDFS-5138.patch, 
> HDFS-5138.patch, HDFS-5138.patch, HDFS-5138.patch, HDFS-5138.patch, 
> HDFS-5138.patch, HDFS-5138.patch, HDFS-5138.patch, HDFS-5138.patch, 
> HDFS-5138.patch, HDFS-5138.patch, hdfs-5138-branch-2.txt
>
>
> With HA enabled, NN wo't start with "-upgrade". Since there has been a layout 
> version change between 2.0.x and 2.1.x, starting NN in upgrade mode was 
> necessary when deploying 2.1.x to an existing 2.0.x cluster. But the only way 
> to get around this was to disable HA and upgrade. 
> The NN and the cluster cannot be flipped back to HA until the upgrade is 
> finalized. If HA is disabled only on NN for layout upgrade and HA is turned 
> back on without involving DNs, things will work, but finaliizeUpgrade won't 
> work (the NN is in HA and it cannot be in upgrade mode) and DN's upgrade 
> snapshots won't get removed.
> We will need a different ways of doing layout upgrade and upgrade snapshot.  
> I am marking this as a 2.1.1-beta blocker based on feedback from others.  If 
> there is a reasonable workaround that does not increase maintenance window 
> greatly, we can lower its priority from blocker to critical.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-5840) Follow-up to HDFS-5138 to improve error handling during partial upgrade failures

2014-03-24 Thread Suresh Srinivas (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5840?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13945683#comment-13945683
 ] 

Suresh Srinivas commented on HDFS-5840:
---

[~jingzhao], thanks for doing the patch.

Should JN throw an exception, if pre-upgrade or upgrade is retried, that 
upgrade is already in progress, restart the journal node before attempting 
namenode upgrade again?

> Follow-up to HDFS-5138 to improve error handling during partial upgrade 
> failures
> 
>
> Key: HDFS-5840
> URL: https://issues.apache.org/jira/browse/HDFS-5840
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: namenode
>Affects Versions: 3.0.0
>Reporter: Aaron T. Myers
>Assignee: Aaron T. Myers
>Priority: Blocker
> Fix For: 3.0.0
>
> Attachments: HDFS-5840.001.patch, HDFS-5840.patch
>
>
> Suresh posted some good comment in HDFS-5138 after that patch had already 
> been committed to trunk. This JIRA is to address those. See the first comment 
> of this JIRA for the full content of the review.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6050) NFS does not handle exceptions correctly in a few places

2014-03-24 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6050?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13945680#comment-13945680
 ] 

Hudson commented on HDFS-6050:
--

SUCCESS: Integrated in Hadoop-trunk-Commit #5391 (See 
[https://builds.apache.org/job/Hadoop-trunk-Commit/5391/])
HDFS-6050. NFS does not handle exceptions correctly in a few places. 
Contributed by Brandon Li (brandonli: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1581055)
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-nfs/src/main/java/org/apache/hadoop/nfs/nfs3/IdUserGroup.java
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-nfs/src/main/java/org/apache/hadoop/nfs/nfs3/Nfs3FileAttributes.java
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-nfs/src/main/java/org/apache/hadoop/portmap/Portmap.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs-nfs/src/main/java/org/apache/hadoop/hdfs/nfs/mount/RpcProgramMountd.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs-nfs/src/main/java/org/apache/hadoop/hdfs/nfs/nfs3/AsyncDataService.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs-nfs/src/main/java/org/apache/hadoop/hdfs/nfs/nfs3/Nfs3Utils.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs-nfs/src/main/java/org/apache/hadoop/hdfs/nfs/nfs3/OpenFileCtx.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs-nfs/src/main/java/org/apache/hadoop/hdfs/nfs/nfs3/RpcProgramNfs3.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs-nfs/src/main/java/org/apache/hadoop/hdfs/nfs/nfs3/WriteManager.java
* /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt


> NFS does not handle exceptions correctly in a few places
> 
>
> Key: HDFS-6050
> URL: https://issues.apache.org/jira/browse/HDFS-6050
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: nfs
>Reporter: Brock Noland
>Assignee: Brandon Li
> Attachments: HDFS-6050.002.patch, HDFS-6050.patch
>
>
> I noticed this file does not log exceptions appropriately in multiple 
> locations.
> Not logging the stack of Throwable:
> https://github.com/apache/hadoop-common/blob/f567f09091368fe800f3f70605da38a69c953fe3/hadoop-hdfs-project/hadoop-hdfs-nfs/src/main/java/org/apache/hadoop/hdfs/nfs/nfs3/OpenFileCtx.java#L364
> Printing exceptions to stderr:
> https://github.com/apache/hadoop-common/blob/f567f09091368fe800f3f70605da38a69c953fe3/hadoop-hdfs-project/hadoop-hdfs-nfs/src/main/java/org/apache/hadoop/hdfs/nfs/nfs3/OpenFileCtx.java#L1160
> https://github.com/apache/hadoop-common/blob/f567f09091368fe800f3f70605da38a69c953fe3/hadoop-hdfs-project/hadoop-hdfs-nfs/src/main/java/org/apache/hadoop/hdfs/nfs/nfs3/OpenFileCtx.java#L1149
> Not logging the stack trace:
> https://github.com/apache/hadoop-common/blob/f567f09091368fe800f3f70605da38a69c953fe3/hadoop-hdfs-project/hadoop-hdfs-nfs/src/main/java/org/apache/hadoop/hdfs/nfs/nfs3/OpenFileCtx.java#L1062
> https://github.com/apache/hadoop-common/blob/f567f09091368fe800f3f70605da38a69c953fe3/hadoop-hdfs-project/hadoop-hdfs-nfs/src/main/java/org/apache/hadoop/hdfs/nfs/nfs3/OpenFileCtx.java#L966
> https://github.com/apache/hadoop-common/blob/f567f09091368fe800f3f70605da38a69c953fe3/hadoop-hdfs-project/hadoop-hdfs-nfs/src/main/java/org/apache/hadoop/hdfs/nfs/nfs3/OpenFileCtx.java#L961
> https://github.com/apache/hadoop-common/blob/f567f09091368fe800f3f70605da38a69c953fe3/hadoop-hdfs-project/hadoop-hdfs-nfs/src/main/java/org/apache/hadoop/hdfs/nfs/nfs3/OpenFileCtx.java#L680



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6050) NFS does not handle exceptions correctly in a few places

2014-03-24 Thread Brandon Li (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6050?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Brandon Li updated HDFS-6050:
-

  Resolution: Fixed
Hadoop Flags: Reviewed
  Status: Resolved  (was: Patch Available)

> NFS does not handle exceptions correctly in a few places
> 
>
> Key: HDFS-6050
> URL: https://issues.apache.org/jira/browse/HDFS-6050
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: nfs
>Reporter: Brock Noland
>Assignee: Brandon Li
> Attachments: HDFS-6050.002.patch, HDFS-6050.patch
>
>
> I noticed this file does not log exceptions appropriately in multiple 
> locations.
> Not logging the stack of Throwable:
> https://github.com/apache/hadoop-common/blob/f567f09091368fe800f3f70605da38a69c953fe3/hadoop-hdfs-project/hadoop-hdfs-nfs/src/main/java/org/apache/hadoop/hdfs/nfs/nfs3/OpenFileCtx.java#L364
> Printing exceptions to stderr:
> https://github.com/apache/hadoop-common/blob/f567f09091368fe800f3f70605da38a69c953fe3/hadoop-hdfs-project/hadoop-hdfs-nfs/src/main/java/org/apache/hadoop/hdfs/nfs/nfs3/OpenFileCtx.java#L1160
> https://github.com/apache/hadoop-common/blob/f567f09091368fe800f3f70605da38a69c953fe3/hadoop-hdfs-project/hadoop-hdfs-nfs/src/main/java/org/apache/hadoop/hdfs/nfs/nfs3/OpenFileCtx.java#L1149
> Not logging the stack trace:
> https://github.com/apache/hadoop-common/blob/f567f09091368fe800f3f70605da38a69c953fe3/hadoop-hdfs-project/hadoop-hdfs-nfs/src/main/java/org/apache/hadoop/hdfs/nfs/nfs3/OpenFileCtx.java#L1062
> https://github.com/apache/hadoop-common/blob/f567f09091368fe800f3f70605da38a69c953fe3/hadoop-hdfs-project/hadoop-hdfs-nfs/src/main/java/org/apache/hadoop/hdfs/nfs/nfs3/OpenFileCtx.java#L966
> https://github.com/apache/hadoop-common/blob/f567f09091368fe800f3f70605da38a69c953fe3/hadoop-hdfs-project/hadoop-hdfs-nfs/src/main/java/org/apache/hadoop/hdfs/nfs/nfs3/OpenFileCtx.java#L961
> https://github.com/apache/hadoop-common/blob/f567f09091368fe800f3f70605da38a69c953fe3/hadoop-hdfs-project/hadoop-hdfs-nfs/src/main/java/org/apache/hadoop/hdfs/nfs/nfs3/OpenFileCtx.java#L680



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-5840) Follow-up to HDFS-5138 to improve error handling during partial upgrade failures

2014-03-24 Thread Jing Zhao (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5840?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jing Zhao updated HDFS-5840:


Attachment: HDFS-5840.001.patch

Rebase [~atm]'s patch. Also add storage recovery logic for JN based on the 
discussion in the comments.

> Follow-up to HDFS-5138 to improve error handling during partial upgrade 
> failures
> 
>
> Key: HDFS-5840
> URL: https://issues.apache.org/jira/browse/HDFS-5840
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: namenode
>Affects Versions: 3.0.0
>Reporter: Aaron T. Myers
>Assignee: Aaron T. Myers
>Priority: Blocker
> Fix For: 3.0.0
>
> Attachments: HDFS-5840.001.patch, HDFS-5840.patch
>
>
> Suresh posted some good comment in HDFS-5138 after that patch had already 
> been committed to trunk. This JIRA is to address those. See the first comment 
> of this JIRA for the full content of the review.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6050) NFS does not handle exceptions correctly in a few places

2014-03-24 Thread Brandon Li (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6050?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Brandon Li updated HDFS-6050:
-

Issue Type: Improvement  (was: Bug)

> NFS does not handle exceptions correctly in a few places
> 
>
> Key: HDFS-6050
> URL: https://issues.apache.org/jira/browse/HDFS-6050
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: nfs
>Reporter: Brock Noland
>Assignee: Brandon Li
> Attachments: HDFS-6050.002.patch, HDFS-6050.patch
>
>
> I noticed this file does not log exceptions appropriately in multiple 
> locations.
> Not logging the stack of Throwable:
> https://github.com/apache/hadoop-common/blob/f567f09091368fe800f3f70605da38a69c953fe3/hadoop-hdfs-project/hadoop-hdfs-nfs/src/main/java/org/apache/hadoop/hdfs/nfs/nfs3/OpenFileCtx.java#L364
> Printing exceptions to stderr:
> https://github.com/apache/hadoop-common/blob/f567f09091368fe800f3f70605da38a69c953fe3/hadoop-hdfs-project/hadoop-hdfs-nfs/src/main/java/org/apache/hadoop/hdfs/nfs/nfs3/OpenFileCtx.java#L1160
> https://github.com/apache/hadoop-common/blob/f567f09091368fe800f3f70605da38a69c953fe3/hadoop-hdfs-project/hadoop-hdfs-nfs/src/main/java/org/apache/hadoop/hdfs/nfs/nfs3/OpenFileCtx.java#L1149
> Not logging the stack trace:
> https://github.com/apache/hadoop-common/blob/f567f09091368fe800f3f70605da38a69c953fe3/hadoop-hdfs-project/hadoop-hdfs-nfs/src/main/java/org/apache/hadoop/hdfs/nfs/nfs3/OpenFileCtx.java#L1062
> https://github.com/apache/hadoop-common/blob/f567f09091368fe800f3f70605da38a69c953fe3/hadoop-hdfs-project/hadoop-hdfs-nfs/src/main/java/org/apache/hadoop/hdfs/nfs/nfs3/OpenFileCtx.java#L966
> https://github.com/apache/hadoop-common/blob/f567f09091368fe800f3f70605da38a69c953fe3/hadoop-hdfs-project/hadoop-hdfs-nfs/src/main/java/org/apache/hadoop/hdfs/nfs/nfs3/OpenFileCtx.java#L961
> https://github.com/apache/hadoop-common/blob/f567f09091368fe800f3f70605da38a69c953fe3/hadoop-hdfs-project/hadoop-hdfs-nfs/src/main/java/org/apache/hadoop/hdfs/nfs/nfs3/OpenFileCtx.java#L680



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6050) NFS does not handle exceptions correctly in a few places

2014-03-24 Thread Brandon Li (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6050?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13945673#comment-13945673
 ] 

Brandon Li commented on HDFS-6050:
--

{quote}
preOpAttr is always null here
{quote}
This may not be true later so I think it's better to keep preOpAttr.

> NFS does not handle exceptions correctly in a few places
> 
>
> Key: HDFS-6050
> URL: https://issues.apache.org/jira/browse/HDFS-6050
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: nfs
>Reporter: Brock Noland
>Assignee: Brandon Li
> Attachments: HDFS-6050.002.patch, HDFS-6050.patch
>
>
> I noticed this file does not log exceptions appropriately in multiple 
> locations.
> Not logging the stack of Throwable:
> https://github.com/apache/hadoop-common/blob/f567f09091368fe800f3f70605da38a69c953fe3/hadoop-hdfs-project/hadoop-hdfs-nfs/src/main/java/org/apache/hadoop/hdfs/nfs/nfs3/OpenFileCtx.java#L364
> Printing exceptions to stderr:
> https://github.com/apache/hadoop-common/blob/f567f09091368fe800f3f70605da38a69c953fe3/hadoop-hdfs-project/hadoop-hdfs-nfs/src/main/java/org/apache/hadoop/hdfs/nfs/nfs3/OpenFileCtx.java#L1160
> https://github.com/apache/hadoop-common/blob/f567f09091368fe800f3f70605da38a69c953fe3/hadoop-hdfs-project/hadoop-hdfs-nfs/src/main/java/org/apache/hadoop/hdfs/nfs/nfs3/OpenFileCtx.java#L1149
> Not logging the stack trace:
> https://github.com/apache/hadoop-common/blob/f567f09091368fe800f3f70605da38a69c953fe3/hadoop-hdfs-project/hadoop-hdfs-nfs/src/main/java/org/apache/hadoop/hdfs/nfs/nfs3/OpenFileCtx.java#L1062
> https://github.com/apache/hadoop-common/blob/f567f09091368fe800f3f70605da38a69c953fe3/hadoop-hdfs-project/hadoop-hdfs-nfs/src/main/java/org/apache/hadoop/hdfs/nfs/nfs3/OpenFileCtx.java#L966
> https://github.com/apache/hadoop-common/blob/f567f09091368fe800f3f70605da38a69c953fe3/hadoop-hdfs-project/hadoop-hdfs-nfs/src/main/java/org/apache/hadoop/hdfs/nfs/nfs3/OpenFileCtx.java#L961
> https://github.com/apache/hadoop-common/blob/f567f09091368fe800f3f70605da38a69c953fe3/hadoop-hdfs-project/hadoop-hdfs-nfs/src/main/java/org/apache/hadoop/hdfs/nfs/nfs3/OpenFileCtx.java#L680



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6050) NFS does not handle exceptions correctly in a few places

2014-03-24 Thread Brandon Li (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6050?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13945674#comment-13945674
 ] 

Brandon Li commented on HDFS-6050:
--

Thank you, Haohui, for the review. I've committed the patch.

> NFS does not handle exceptions correctly in a few places
> 
>
> Key: HDFS-6050
> URL: https://issues.apache.org/jira/browse/HDFS-6050
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: nfs
>Reporter: Brock Noland
>Assignee: Brandon Li
> Attachments: HDFS-6050.002.patch, HDFS-6050.patch
>
>
> I noticed this file does not log exceptions appropriately in multiple 
> locations.
> Not logging the stack of Throwable:
> https://github.com/apache/hadoop-common/blob/f567f09091368fe800f3f70605da38a69c953fe3/hadoop-hdfs-project/hadoop-hdfs-nfs/src/main/java/org/apache/hadoop/hdfs/nfs/nfs3/OpenFileCtx.java#L364
> Printing exceptions to stderr:
> https://github.com/apache/hadoop-common/blob/f567f09091368fe800f3f70605da38a69c953fe3/hadoop-hdfs-project/hadoop-hdfs-nfs/src/main/java/org/apache/hadoop/hdfs/nfs/nfs3/OpenFileCtx.java#L1160
> https://github.com/apache/hadoop-common/blob/f567f09091368fe800f3f70605da38a69c953fe3/hadoop-hdfs-project/hadoop-hdfs-nfs/src/main/java/org/apache/hadoop/hdfs/nfs/nfs3/OpenFileCtx.java#L1149
> Not logging the stack trace:
> https://github.com/apache/hadoop-common/blob/f567f09091368fe800f3f70605da38a69c953fe3/hadoop-hdfs-project/hadoop-hdfs-nfs/src/main/java/org/apache/hadoop/hdfs/nfs/nfs3/OpenFileCtx.java#L1062
> https://github.com/apache/hadoop-common/blob/f567f09091368fe800f3f70605da38a69c953fe3/hadoop-hdfs-project/hadoop-hdfs-nfs/src/main/java/org/apache/hadoop/hdfs/nfs/nfs3/OpenFileCtx.java#L966
> https://github.com/apache/hadoop-common/blob/f567f09091368fe800f3f70605da38a69c953fe3/hadoop-hdfs-project/hadoop-hdfs-nfs/src/main/java/org/apache/hadoop/hdfs/nfs/nfs3/OpenFileCtx.java#L961
> https://github.com/apache/hadoop-common/blob/f567f09091368fe800f3f70605da38a69c953fe3/hadoop-hdfs-project/hadoop-hdfs-nfs/src/main/java/org/apache/hadoop/hdfs/nfs/nfs3/OpenFileCtx.java#L680



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (HDFS-6151) HDFS should refuse to cache blocks >=2GB

2014-03-24 Thread Andrew Wang (JIRA)
Andrew Wang created HDFS-6151:
-

 Summary: HDFS should refuse to cache blocks >=2GB
 Key: HDFS-6151
 URL: https://issues.apache.org/jira/browse/HDFS-6151
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: caching, datanode
Affects Versions: 2.4.0
Reporter: Andrew Wang
Assignee: Andrew Wang


If you try to cache a block that's >=2GB, the DN will silently fail to cache it 
since {{MappedByteBuffer}} uses a signed int to represent size. Blocks this 
large are rare, but we should log or alert the user somehow.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6050) NFS does not handle exceptions correctly in a few places

2014-03-24 Thread Brandon Li (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6050?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Brandon Li updated HDFS-6050:
-

Summary: NFS does not handle exceptions correctly in a few places  (was: 
NFS OpenFileCtx does not handle exceptions correctly)

> NFS does not handle exceptions correctly in a few places
> 
>
> Key: HDFS-6050
> URL: https://issues.apache.org/jira/browse/HDFS-6050
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: nfs
>Reporter: Brock Noland
>Assignee: Brandon Li
> Attachments: HDFS-6050.002.patch, HDFS-6050.patch
>
>
> I noticed this file does not log exceptions appropriately in multiple 
> locations.
> Not logging the stack of Throwable:
> https://github.com/apache/hadoop-common/blob/f567f09091368fe800f3f70605da38a69c953fe3/hadoop-hdfs-project/hadoop-hdfs-nfs/src/main/java/org/apache/hadoop/hdfs/nfs/nfs3/OpenFileCtx.java#L364
> Printing exceptions to stderr:
> https://github.com/apache/hadoop-common/blob/f567f09091368fe800f3f70605da38a69c953fe3/hadoop-hdfs-project/hadoop-hdfs-nfs/src/main/java/org/apache/hadoop/hdfs/nfs/nfs3/OpenFileCtx.java#L1160
> https://github.com/apache/hadoop-common/blob/f567f09091368fe800f3f70605da38a69c953fe3/hadoop-hdfs-project/hadoop-hdfs-nfs/src/main/java/org/apache/hadoop/hdfs/nfs/nfs3/OpenFileCtx.java#L1149
> Not logging the stack trace:
> https://github.com/apache/hadoop-common/blob/f567f09091368fe800f3f70605da38a69c953fe3/hadoop-hdfs-project/hadoop-hdfs-nfs/src/main/java/org/apache/hadoop/hdfs/nfs/nfs3/OpenFileCtx.java#L1062
> https://github.com/apache/hadoop-common/blob/f567f09091368fe800f3f70605da38a69c953fe3/hadoop-hdfs-project/hadoop-hdfs-nfs/src/main/java/org/apache/hadoop/hdfs/nfs/nfs3/OpenFileCtx.java#L966
> https://github.com/apache/hadoop-common/blob/f567f09091368fe800f3f70605da38a69c953fe3/hadoop-hdfs-project/hadoop-hdfs-nfs/src/main/java/org/apache/hadoop/hdfs/nfs/nfs3/OpenFileCtx.java#L961
> https://github.com/apache/hadoop-common/blob/f567f09091368fe800f3f70605da38a69c953fe3/hadoop-hdfs-project/hadoop-hdfs-nfs/src/main/java/org/apache/hadoop/hdfs/nfs/nfs3/OpenFileCtx.java#L680



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-5910) Enhance DataTransferProtocol to allow per-connection choice of encryption/plain-text

2014-03-24 Thread Suresh Srinivas (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5910?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13945654#comment-13945654
 ] 

Suresh Srinivas commented on HDFS-5910:
---

[~benoyantony], any updates on this?

> Enhance DataTransferProtocol to allow per-connection choice of 
> encryption/plain-text
> 
>
> Key: HDFS-5910
> URL: https://issues.apache.org/jira/browse/HDFS-5910
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: security
>Affects Versions: 2.2.0
>Reporter: Benoy Antony
>Assignee: Benoy Antony
> Attachments: HDFS-5910.patch, HDFS-5910.patch
>
>
> It is possible to enable encryption of DataTransferProtocol. 
> In some use cases, it is required to encrypt data transfer with some clients 
> , but communicate in plain text with some other clients and data nodes.
> A sample use case will be that any data transfer inside a firewall can be in 
> plain text whereas any data transfer from clients  outside the firewall needs 
> to be encrypted.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6130) NPE during namenode upgrade from old release

2014-03-24 Thread Haohui Mai (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6130?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13945653#comment-13945653
 ] 

Haohui Mai commented on HDFS-6130:
--

It would be very helpful if the corresponding fsimage is available.

> NPE during namenode upgrade from old release
> 
>
> Key: HDFS-6130
> URL: https://issues.apache.org/jira/browse/HDFS-6130
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: namenode
>Affects Versions: 2.4.0
>Reporter: Fengdong Yu
>
> I want upgrade an old cluster(0.20.2-cdh3u1) to trunk instance, 
> I can upgrade successfully if I don't configurage HA, but if HA enabled,
> there is NPE when I run ' hdfs namenode -initializeSharedEdits'
> {code}
> 14/03/20 15:06:41 INFO namenode.FSNamesystem: Retry cache on namenode is 
> enabled
> 14/03/20 15:06:41 INFO namenode.FSNamesystem: Retry cache will use 0.03 of 
> total heap and retry cache entry expiry time is 60 millis
> 14/03/20 15:06:41 INFO util.GSet: Computing capacity for map 
> NameNodeRetryCache
> 14/03/20 15:06:41 INFO util.GSet: VM type   = 64-bit
> 14/03/20 15:06:41 INFO util.GSet: 0.02999329447746% max memory 896 MB = 
> 275.3 KB
> 14/03/20 15:06:41 INFO util.GSet: capacity  = 2^15 = 32768 entries
> 14/03/20 15:06:41 INFO namenode.AclConfigFlag: ACLs enabled? false
> 14/03/20 15:06:41 INFO common.Storage: Lock on 
> /data/hadoop/data1/dfs/name/in_use.lock acquired by nodename 
> 7326@10-150-170-176
> 14/03/20 15:06:42 INFO common.Storage: Lock on 
> /data/hadoop/data2/dfs/name/in_use.lock acquired by nodename 
> 7326@10-150-170-176
> 14/03/20 15:06:42 INFO namenode.FSImage: No edit log streams selected.
> 14/03/20 15:06:42 INFO namenode.FSImageFormatPBINode: Loading 1 INodes.
> 14/03/20 15:06:42 FATAL namenode.NameNode: Exception in namenode join
> java.lang.NullPointerException
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSDirectory.isReservedName(FSDirectory.java:2984)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImageFormatPBINode$Loader.addToParent(FSImageFormatPBINode.java:205)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImageFormatPBINode$Loader.loadINodeDirectorySection(FSImageFormatPBINode.java:162)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImageFormatProtobuf$Loader.loadInternal(FSImageFormatProtobuf.java:243)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImageFormatProtobuf$Loader.load(FSImageFormatProtobuf.java:168)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImageFormat$LoaderDelegator.load(FSImageFormat.java:120)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.loadFSImage(FSImage.java:895)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.loadFSImage(FSImage.java:881)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.loadFSImageFile(FSImage.java:704)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.loadFSImage(FSImage.java:642)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:271)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFSImage(FSNamesystem.java:894)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFromDisk(FSNamesystem.java:653)
>   at 
> org.apache.hadoop.hdfs.server.namenode.NameNode.initializeSharedEdits(NameNode.java:912)
>   at 
> org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1276)
>   at 
> org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1360)
> 14/03/20 15:06:42 INFO util.ExitUtil: Exiting with status 1
> 14/03/20 15:06:42 INFO namenode.NameNode: SHUTDOWN_MSG: 
> /
> SHUTDOWN_MSG: Shutting down NameNode at 10-150-170-176/10.150.170.176
> /
> {code}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6135) In HDFS upgrade with HA setup, JournalNode cannot handle layout version bump when rolling back

2014-03-24 Thread Tsz Wo Nicholas Sze (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6135?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tsz Wo Nicholas Sze updated HDFS-6135:
--

 Component/s: journal-node
Hadoop Flags: Reviewed

+1 patch looks good.

> In HDFS upgrade with HA setup, JournalNode cannot handle layout version bump 
> when rolling back
> --
>
> Key: HDFS-6135
> URL: https://issues.apache.org/jira/browse/HDFS-6135
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: journal-node
>Affects Versions: 3.0.0
>Reporter: Jing Zhao
>Assignee: Jing Zhao
>Priority: Blocker
> Attachments: HDFS-6135.000.patch, HDFS-6135.001.patch, 
> HDFS-6135.002.patch, HDFS-6135.test.txt
>
>
> While doing HDFS upgrade with HA setup, if the layoutversion gets changed in 
> the upgrade, the rollback may trigger the following exception in JournalNodes 
> (suppose the new software bumped the layoutversion from -55 to -56):
> {code}
> 14/03/21 01:01:53 FATAL namenode.NameNode: Exception in namenode join
> org.apache.hadoop.hdfs.qjournal.client.QuorumException: Could not check if 
> roll back possible for one or more JournalNodes. 1 exceptions thrown:
> Unexpected version of storage directory /grid/1/tmp/journal/mycluster. 
> Reported: -56. Expecting = -55.
>   at 
> org.apache.hadoop.hdfs.server.common.StorageInfo.setLayoutVersion(StorageInfo.java:203)
>   at 
> org.apache.hadoop.hdfs.server.common.StorageInfo.setFieldsFromProperties(StorageInfo.java:156)
>   at 
> org.apache.hadoop.hdfs.server.common.StorageInfo.readProperties(StorageInfo.java:135)
>   at 
> org.apache.hadoop.hdfs.qjournal.server.JNStorage.analyzeStorage(JNStorage.java:202)
>   at 
> org.apache.hadoop.hdfs.qjournal.server.JNStorage.(JNStorage.java:73)
>   at 
> org.apache.hadoop.hdfs.qjournal.server.Journal.(Journal.java:142)
>   at 
> org.apache.hadoop.hdfs.qjournal.server.JournalNode.getOrCreateJournal(JournalNode.java:87)
>   at 
> org.apache.hadoop.hdfs.qjournal.server.JournalNode.canRollBack(JournalNode.java:304)
>   at 
> org.apache.hadoop.hdfs.qjournal.server.JournalNodeRpcServer.canRollBack(JournalNodeRpcServer.java:228)
> {code}
> Looks like for rollback JN with old software cannot handle future 
> layoutversion brought by new software.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6150) Add inode id information in the logs to make debugging easier

2014-03-24 Thread Suresh Srinivas (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6150?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Suresh Srinivas updated HDFS-6150:
--

Attachment: HDFS-6150.patch

> Add inode id information in the logs to make debugging easier
> -
>
> Key: HDFS-6150
> URL: https://issues.apache.org/jira/browse/HDFS-6150
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: namenode
>Reporter: Suresh Srinivas
> Attachments: HDFS-6150.patch
>
>
> Inode information and path information are missing in the logs and 
> exceptions. Adding this will help debug multi threading issues related to 
> using incode INode ID information.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6150) Add inode id information in the logs to make debugging easier

2014-03-24 Thread Suresh Srinivas (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6150?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Suresh Srinivas updated HDFS-6150:
--

Status: Patch Available  (was: Open)

> Add inode id information in the logs to make debugging easier
> -
>
> Key: HDFS-6150
> URL: https://issues.apache.org/jira/browse/HDFS-6150
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: namenode
>Reporter: Suresh Srinivas
> Attachments: HDFS-6150.patch
>
>
> Inode information and path information are missing in the logs and 
> exceptions. Adding this will help debug multi threading issues related to 
> using incode INode ID information.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (HDFS-6150) Add inode id information in the logs to make debugging easier

2014-03-24 Thread Suresh Srinivas (JIRA)
Suresh Srinivas created HDFS-6150:
-

 Summary: Add inode id information in the logs to make debugging 
easier
 Key: HDFS-6150
 URL: https://issues.apache.org/jira/browse/HDFS-6150
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: namenode
Reporter: Suresh Srinivas
 Attachments: HDFS-6150.patch

Inode information and path information are missing in the logs and exceptions. 
Adding this will help debug multi threading issues related to using incode 
INode ID information.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6130) NPE during namenode upgrade from old release

2014-03-24 Thread Tsz Wo Nicholas Sze (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6130?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13945642#comment-13945642
 ] 

Tsz Wo Nicholas Sze commented on HDFS-6130:
---

> I believe that this is a duplicate of HDFS-5988.

Hi [~wheat9], the stack trace posted here is indeed different from the one 
posted in HDFS-6021 (a dup of HDFS-5988).  So it seems that this is a different 
issue.  In this bug, FSImageFormatPBINode somehow passes a null inode to 
FSDirectory.  Could you take a look?

- Stack trace posted here 

{noformat}
14/03/20 15:06:42 FATAL namenode.NameNode: Exception in namenode join
java.lang.NullPointerException
at 
org.apache.hadoop.hdfs.server.namenode.FSDirectory.isReservedName(FSDirectory.java:2984)
at 
org.apache.hadoop.hdfs.server.namenode.FSImageFormatPBINode$Loader.addToParent(FSImageFormatPBINode.java:205)
at 
org.apache.hadoop.hdfs.server.namenode.FSImageFormatPBINode$Loader.loadINodeDirectorySection(FSImageFormatPBINode.java:162)
at 
org.apache.hadoop.hdfs.server.namenode.FSImageFormatProtobuf$Loader.loadInternal(FSImageFormatProtobuf.java:243)
at 
org.apache.hadoop.hdfs.server.namenode.FSImageFormatProtobuf$Loader.load(FSImageFormatProtobuf.java:168)
at 
org.apache.hadoop.hdfs.server.namenode.FSImageFormat$LoaderDelegator.load(FSImageFormat.java:120)
at 
org.apache.hadoop.hdfs.server.namenode.FSImage.loadFSImage(FSImage.java:895)
at 
org.apache.hadoop.hdfs.server.namenode.FSImage.loadFSImage(FSImage.java:881)
at 
org.apache.hadoop.hdfs.server.namenode.FSImage.loadFSImageFile(FSImage.java:704)
at 
org.apache.hadoop.hdfs.server.namenode.FSImage.loadFSImage(FSImage.java:642)
at 
org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:271)
at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFSImage(FSNamesystem.java:894)
at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFromDisk(FSNamesystem.java:653)
at 
org.apache.hadoop.hdfs.server.namenode.NameNode.initializeSharedEdits(NameNode.java:912)
at 
org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1276)
...
{noformat}

- Stack trace posted in HDFS-6021 (a dup of HDFS-5988)

{noformat}
2014-02-26 17:03:11,755 FATAL [main] namenode.NameNode 
(NameNode.java:main(1351)) - Exception in namenode join
java.lang.NullPointerException
at 
org.apache.hadoop.hdfs.server.namenode.FSImageFormatProtobuf$Loader.loadInternal(FSImageFormatProtobuf.java:227)
at 
org.apache.hadoop.hdfs.server.namenode.FSImageFormatProtobuf$Loader.load(FSImageFormatProtobuf.java:169)
at 
org.apache.hadoop.hdfs.server.namenode.FSImageFormat$LoaderDelegator.load(FSImageFormat.java:225)
at 
org.apache.hadoop.hdfs.server.namenode.FSImage.loadFSImage(FSImage.java:802)
at 
org.apache.hadoop.hdfs.server.namenode.FSImage.loadFSImage(FSImage.java:792)
at 
org.apache.hadoop.hdfs.server.namenode.FSImage.loadFSImageFile(FSImage.java:624)
at 
org.apache.hadoop.hdfs.server.namenode.FSImage.loadFSImage(FSImage.java:593)
at 
org.apache.hadoop.hdfs.server.namenode.FSImage.doUpgrade(FSImage.java:331)
at 
org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:251)
at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFSImage(FSNamesystem.java:882)
at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFromDisk(FSNamesystem.java:641)
at 
org.apache.hadoop.hdfs.server.namenode.NameNode.loadNamesystem(NameNode.java:435)
at 
org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:491)
at 
org.apache.hadoop.hdfs.server.namenode.NameNode.(NameNode.java:647)
at 
org.apache.hadoop.hdfs.server.namenode.NameNode.(NameNode.java:632)
at 
org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1280)
...
{noformat}


> NPE during namenode upgrade from old release
> 
>
> Key: HDFS-6130
> URL: https://issues.apache.org/jira/browse/HDFS-6130
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: namenode
>Affects Versions: 2.4.0
>Reporter: Fengdong Yu
>
> I want upgrade an old cluster(0.20.2-cdh3u1) to trunk instance, 
> I can upgrade successfully if I don't configurage HA, but if HA enabled,
> there is NPE when I run ' hdfs namenode -initializeSharedEdits'
> {code}
> 14/03/20 15:06:41 INFO namenode.FSNamesystem: Retry cache on namenode is 
> enabled
> 14/03/20 15:06:41 INFO namenode.FSNamesystem: Retry cache will use 0.03 of 
> total heap and retry cache entry expiry time is 60 millis
> 14/03/20 15:06:41 INFO util.GSet: Computing capacity for map 
> NameNodeRetryCache
> 14/03/20 15:06:41 INFO 

[jira] [Commented] (HDFS-6130) NPE during namenode upgrade from old release

2014-03-24 Thread Tsz Wo Nicholas Sze (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6130?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13945628#comment-13945628
 ] 

Tsz Wo Nicholas Sze commented on HDFS-6130:
---

> Apache release also has this issue. Apache 1.0.4 upgrade to the trunk, you 
> can reproduce this issue.

Hi Fengdong, I just have tried it but cannot reproduce the NPE.  There were a 
log of changes since Apache 1.0.4.  I was using 1.3.0 in my test.  Could you 
also try it?

> NPE during namenode upgrade from old release
> 
>
> Key: HDFS-6130
> URL: https://issues.apache.org/jira/browse/HDFS-6130
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: namenode
>Affects Versions: 2.4.0
>Reporter: Fengdong Yu
>
> I want upgrade an old cluster(0.20.2-cdh3u1) to trunk instance, 
> I can upgrade successfully if I don't configurage HA, but if HA enabled,
> there is NPE when I run ' hdfs namenode -initializeSharedEdits'
> {code}
> 14/03/20 15:06:41 INFO namenode.FSNamesystem: Retry cache on namenode is 
> enabled
> 14/03/20 15:06:41 INFO namenode.FSNamesystem: Retry cache will use 0.03 of 
> total heap and retry cache entry expiry time is 60 millis
> 14/03/20 15:06:41 INFO util.GSet: Computing capacity for map 
> NameNodeRetryCache
> 14/03/20 15:06:41 INFO util.GSet: VM type   = 64-bit
> 14/03/20 15:06:41 INFO util.GSet: 0.02999329447746% max memory 896 MB = 
> 275.3 KB
> 14/03/20 15:06:41 INFO util.GSet: capacity  = 2^15 = 32768 entries
> 14/03/20 15:06:41 INFO namenode.AclConfigFlag: ACLs enabled? false
> 14/03/20 15:06:41 INFO common.Storage: Lock on 
> /data/hadoop/data1/dfs/name/in_use.lock acquired by nodename 
> 7326@10-150-170-176
> 14/03/20 15:06:42 INFO common.Storage: Lock on 
> /data/hadoop/data2/dfs/name/in_use.lock acquired by nodename 
> 7326@10-150-170-176
> 14/03/20 15:06:42 INFO namenode.FSImage: No edit log streams selected.
> 14/03/20 15:06:42 INFO namenode.FSImageFormatPBINode: Loading 1 INodes.
> 14/03/20 15:06:42 FATAL namenode.NameNode: Exception in namenode join
> java.lang.NullPointerException
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSDirectory.isReservedName(FSDirectory.java:2984)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImageFormatPBINode$Loader.addToParent(FSImageFormatPBINode.java:205)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImageFormatPBINode$Loader.loadINodeDirectorySection(FSImageFormatPBINode.java:162)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImageFormatProtobuf$Loader.loadInternal(FSImageFormatProtobuf.java:243)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImageFormatProtobuf$Loader.load(FSImageFormatProtobuf.java:168)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImageFormat$LoaderDelegator.load(FSImageFormat.java:120)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.loadFSImage(FSImage.java:895)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.loadFSImage(FSImage.java:881)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.loadFSImageFile(FSImage.java:704)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.loadFSImage(FSImage.java:642)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:271)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFSImage(FSNamesystem.java:894)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFromDisk(FSNamesystem.java:653)
>   at 
> org.apache.hadoop.hdfs.server.namenode.NameNode.initializeSharedEdits(NameNode.java:912)
>   at 
> org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1276)
>   at 
> org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1360)
> 14/03/20 15:06:42 INFO util.ExitUtil: Exiting with status 1
> 14/03/20 15:06:42 INFO namenode.NameNode: SHUTDOWN_MSG: 
> /
> SHUTDOWN_MSG: Shutting down NameNode at 10-150-170-176/10.150.170.176
> /
> {code}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6135) In HDFS upgrade with HA setup, JournalNode cannot handle layout version bump when rolling back

2014-03-24 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6135?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13945614#comment-13945614
 ] 

Hadoop QA commented on HDFS-6135:
-

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12636293/HDFS-6135.002.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:red}-1 tests included{color}.  The patch doesn't appear to include 
any new or modified tests.
Please justify why no new tests are needed for this 
patch.
Also please list what manual steps were performed to 
verify this patch.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  There were no new javadoc warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:green}+1 core tests{color}.  The patch passed unit tests in 
hadoop-hdfs-project/hadoop-hdfs.

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-HDFS-Build/6476//testReport/
Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/6476//console

This message is automatically generated.

> In HDFS upgrade with HA setup, JournalNode cannot handle layout version bump 
> when rolling back
> --
>
> Key: HDFS-6135
> URL: https://issues.apache.org/jira/browse/HDFS-6135
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 3.0.0
>Reporter: Jing Zhao
>Assignee: Jing Zhao
>Priority: Blocker
> Attachments: HDFS-6135.000.patch, HDFS-6135.001.patch, 
> HDFS-6135.002.patch, HDFS-6135.test.txt
>
>
> While doing HDFS upgrade with HA setup, if the layoutversion gets changed in 
> the upgrade, the rollback may trigger the following exception in JournalNodes 
> (suppose the new software bumped the layoutversion from -55 to -56):
> {code}
> 14/03/21 01:01:53 FATAL namenode.NameNode: Exception in namenode join
> org.apache.hadoop.hdfs.qjournal.client.QuorumException: Could not check if 
> roll back possible for one or more JournalNodes. 1 exceptions thrown:
> Unexpected version of storage directory /grid/1/tmp/journal/mycluster. 
> Reported: -56. Expecting = -55.
>   at 
> org.apache.hadoop.hdfs.server.common.StorageInfo.setLayoutVersion(StorageInfo.java:203)
>   at 
> org.apache.hadoop.hdfs.server.common.StorageInfo.setFieldsFromProperties(StorageInfo.java:156)
>   at 
> org.apache.hadoop.hdfs.server.common.StorageInfo.readProperties(StorageInfo.java:135)
>   at 
> org.apache.hadoop.hdfs.qjournal.server.JNStorage.analyzeStorage(JNStorage.java:202)
>   at 
> org.apache.hadoop.hdfs.qjournal.server.JNStorage.(JNStorage.java:73)
>   at 
> org.apache.hadoop.hdfs.qjournal.server.Journal.(Journal.java:142)
>   at 
> org.apache.hadoop.hdfs.qjournal.server.JournalNode.getOrCreateJournal(JournalNode.java:87)
>   at 
> org.apache.hadoop.hdfs.qjournal.server.JournalNode.canRollBack(JournalNode.java:304)
>   at 
> org.apache.hadoop.hdfs.qjournal.server.JournalNodeRpcServer.canRollBack(JournalNodeRpcServer.java:228)
> {code}
> Looks like for rollback JN with old software cannot handle future 
> layoutversion brought by new software.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6134) Transparent data at rest encryption

2014-03-24 Thread Alejandro Abdelnur (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6134?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13945576#comment-13945576
 ] 

Alejandro Abdelnur commented on HDFS-6134:
--

(Cross-posting HADOOP-10150 & HDFS-6134]

[~avik_...@yahoo.com], I’ve just looked at the MAR/21 proposal in HADOOP-10150 
(the patches uploaded on MAR/21 do not apply on trunk cleanly, so I cannot look 
at them easily. It seems to have missing pieces, like getXAttrs() and wiring to 
KeyProvider API. Would be possible to rebased them so they apply to trunk?)

bq. do we need a new proposal for the work already being done on HADOOP-10150?

HADOOP-10150 aims to provide encryption for any filesystem implementation as a 
decorator filesystem. While HDFS-6134 aims to provide encryption for HDFS. 

The 2 approaches differ on the level of transparency you get. The comparison 
table in the "HDFS Data at Rest Encryption" attachment 
(https://issues.apache.org/jira/secure/attachment/12635964/HDFSDataAtRestEncryption.pdf)
 highlights the differences.

Particularly, the things I’m concerned the most with HADOOP-10150 are:

* All clients (doing encryption/decryption) must have access the key management 
service.
* Secure key propagation to tasks running in the cluster (i.e. mapper and 
reducer tasks)
* Use of AES-CTR (instead of an authenticated encryption mode such as AES-GCM)
* Not clear how hflush()

bq. are there design choices in this proposal that are superior to the patch 
already provided on HADOOP-10150?

IMO, a consolidated access/distribution of keys by the NN (as opposed to every 
client) improves the security of the system.

bq. do you have additional requirement listed in this JIRA that could be 
incorporated in to HADOOP-10150, 

They are enumerated in the "HDFS Data at Rest Encryption" attachment. The ones 
I don’t see them address in HADOOP-10150 are: #6, #8.A. And it is not clear how 
 #4 & #5 can be achieved.

bq. so we can collaborate and not duplicate?

Definitely, I want to work together with you guys to leverage as much as 
posible. Either by unifying the 2 proposal or by sharing common code if we 
think both approaches have merits and we decide to move forward with both.

Happy to jump on a call to discuss things and the report back to the community 
if you think that will speed up the discussion.

--
By looking at the latest design doc of HADOOP-10150 I can see that things have 
been modified a bit (from the original design doc) bringing it a bit closer to 
some of the HDFS-6134 requirements.

Still, it is not clear how transparency will be achieved for existing 
applications: HDFS URI changes, clients must connect to the Key store to 
retrieve the encryption key (clients will need key store principals). The 
encryption key must be propagated to jobs tasks (i.e. Mapper/Reducer processes)

Requirement #4 "Can decorate HDFS and all other file systems in Hadoop, and 
will not modify existing structure of file system, such as namenode and 
datanode structure if the wrapped file system is HDFS." This is contradicted by 
the design, in the "Storage of IV and data key" is stated  "So we implement 
extended information based on INode feature, and use it to store data key and 
IV. "

Requirement #5 "Admin can configure encryption policies, such as which 
directory will be encrypted.", this seems driven by HDFS client configuration 
file (hdfs-site.xml). This is not really admin driven as clients could break 
this by configuring their hdfs-site.xml file)

Restrictions of move operations for files within an encrypted directory. The 
original design had something about it (not entirely correct), now is gone.

(Mentioned before), how thing flush() operations will be handled as the 
encryption block will be cut short? How this is handled on writes? How this is 
handled on reads?

Explicit auditing on encrypted files access does not seem handled.



> Transparent data at rest encryption
> ---
>
> Key: HDFS-6134
> URL: https://issues.apache.org/jira/browse/HDFS-6134
> Project: Hadoop HDFS
>  Issue Type: New Feature
>  Components: security
>Affects Versions: 2.3.0
>Reporter: Alejandro Abdelnur
>Assignee: Alejandro Abdelnur
> Attachments: HDFSDataAtRestEncryption.pdf
>
>
> Because of privacy and security regulations, for many industries, sensitive 
> data at rest must be in encrypted form. For example: the health­care industry 
> (HIPAA regulations), the card payment industry (PCI DSS regulations) or the 
> US government (FISMA regulations).
> This JIRA aims to provide a mechanism to encrypt HDFS data at rest that can 
> be used transparently by any application accessing HDFS via Hadoop Filesystem 
> Java API, Hadoop libhdfs C library, or WebHDFS REST API.
> The resulting implementation should be able to be used in complianc

[jira] [Commented] (HDFS-5138) Support HDFS upgrade in HA

2014-03-24 Thread Tsz Wo Nicholas Sze (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5138?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13945560#comment-13945560
 ] 

Tsz Wo Nicholas Sze commented on HDFS-5138:
---

+1 the branch-2 patch looks good.

> Support HDFS upgrade in HA
> --
>
> Key: HDFS-5138
> URL: https://issues.apache.org/jira/browse/HDFS-5138
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 2.1.1-beta
>Reporter: Kihwal Lee
>Assignee: Aaron T. Myers
>Priority: Blocker
> Fix For: 3.0.0
>
> Attachments: HDFS-5138.branch-2.001.patch, HDFS-5138.patch, 
> HDFS-5138.patch, HDFS-5138.patch, HDFS-5138.patch, HDFS-5138.patch, 
> HDFS-5138.patch, HDFS-5138.patch, HDFS-5138.patch, HDFS-5138.patch, 
> HDFS-5138.patch, HDFS-5138.patch, hdfs-5138-branch-2.txt
>
>
> With HA enabled, NN wo't start with "-upgrade". Since there has been a layout 
> version change between 2.0.x and 2.1.x, starting NN in upgrade mode was 
> necessary when deploying 2.1.x to an existing 2.0.x cluster. But the only way 
> to get around this was to disable HA and upgrade. 
> The NN and the cluster cannot be flipped back to HA until the upgrade is 
> finalized. If HA is disabled only on NN for layout upgrade and HA is turned 
> back on without involving DNs, things will work, but finaliizeUpgrade won't 
> work (the NN is in HA and it cannot be in upgrade mode) and DN's upgrade 
> snapshots won't get removed.
> We will need a different ways of doing layout upgrade and upgrade snapshot.  
> I am marking this as a 2.1.1-beta blocker based on feedback from others.  If 
> there is a reasonable workaround that does not increase maintenance window 
> greatly, we can lower its priority from blocker to critical.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6148) LeaseManager crashes while initiating block recovery

2014-03-24 Thread Kihwal Lee (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6148?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kihwal Lee updated HDFS-6148:
-

Affects Version/s: (was: 2.4.0)
   2.3.0

> LeaseManager crashes while initiating block recovery
> 
>
> Key: HDFS-6148
> URL: https://issues.apache.org/jira/browse/HDFS-6148
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 2.3.0
>Reporter: Kihwal Lee
>Priority: Blocker
>
> While running branch-2.4, the LeaseManager crashed with an NPE. This does not 
> always happen on block recovery.
> {panel}
> Exception in thread
> "org.apache.hadoop.hdfs.server.namenode.LeaseManager$Monitor@5d66b728"
> java.lang.NullPointerException
> at 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockInfoUnderConstruction$
> 
> ReplicaUnderConstruction.isAlive(BlockInfoUnderConstruction.java:121)
> at 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockInfoUnderConstruction.
> initializeBlockRecovery(BlockInfoUnderConstruction.java:286)
> at 
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.internalReleaseLease(FSNamesystem.java:3746)
> at 
> org.apache.hadoop.hdfs.server.namenode.LeaseManager.checkLeases(LeaseManager.java:474)
> at 
> org.apache.hadoop.hdfs.server.namenode.LeaseManager.access$900(LeaseManager.java:68)
> at 
> org.apache.hadoop.hdfs.server.namenode.LeaseManager$Monitor.run(LeaseManager.java:411)
> at java.lang.Thread.run(Thread.java:722)
> {panel}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6148) LeaseManager crashes while initiating block recovery

2014-03-24 Thread Kihwal Lee (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6148?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13945530#comment-13945530
 ] 

Kihwal Lee commented on HDFS-6148:
--

Sorry, it was seen on a 2.3 cluster. I will verify wether we still have this 
bug in 2.4

> LeaseManager crashes while initiating block recovery
> 
>
> Key: HDFS-6148
> URL: https://issues.apache.org/jira/browse/HDFS-6148
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 2.4.0
>Reporter: Kihwal Lee
>Priority: Blocker
>
> While running branch-2.4, the LeaseManager crashed with an NPE. This does not 
> always happen on block recovery.
> {panel}
> Exception in thread
> "org.apache.hadoop.hdfs.server.namenode.LeaseManager$Monitor@5d66b728"
> java.lang.NullPointerException
> at 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockInfoUnderConstruction$
> 
> ReplicaUnderConstruction.isAlive(BlockInfoUnderConstruction.java:121)
> at 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockInfoUnderConstruction.
> initializeBlockRecovery(BlockInfoUnderConstruction.java:286)
> at 
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.internalReleaseLease(FSNamesystem.java:3746)
> at 
> org.apache.hadoop.hdfs.server.namenode.LeaseManager.checkLeases(LeaseManager.java:474)
> at 
> org.apache.hadoop.hdfs.server.namenode.LeaseManager.access$900(LeaseManager.java:68)
> at 
> org.apache.hadoop.hdfs.server.namenode.LeaseManager$Monitor.run(LeaseManager.java:411)
> at java.lang.Thread.run(Thread.java:722)
> {panel}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-5846) Assigning DEFAULT_RACK in resolveNetworkLocation method can break data resiliency

2014-03-24 Thread Chris Nauroth (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5846?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chris Nauroth updated HDFS-5846:


  Component/s: namenode
 Target Version/s: 3.0.0, 2.4.0
Affects Version/s: 3.0.0
   2.3.0
 Hadoop Flags: Reviewed

> Assigning DEFAULT_RACK in resolveNetworkLocation method can break data 
> resiliency
> -
>
> Key: HDFS-5846
> URL: https://issues.apache.org/jira/browse/HDFS-5846
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: namenode
>Affects Versions: 3.0.0, 2.3.0
>Reporter: Nikola Vujic
>Assignee: Nikola Vujic
> Attachments: hdfs-5846.patch, hdfs-5846.patch, hdfs-5846.patch
>
>
> Medhod CachedDNSToSwitchMapping::resolve() can return NULL which requires 
> careful handling. Null can be returned in two cases:
> • An error occurred with topology script execution (script crashes).
> • Script returns wrong number of values (other than expected)
> Critical handling is in the DN registration code. DN registration code is 
> responsible for assigning proper topology paths to all registered datanodes. 
> Existing code handles this NULL pointer on the following way 
> ({{resolveNetworkLocation}} method):
> {code}
> / /resolve its network location
> List rName = dnsToSwitchMapping.resolve(names);
> String networkLocation;
> if (rName == null) {
>   LOG.error("The resolve call returned null! Using " + 
>   NetworkTopology.DEFAULT_RACK + " for host " + names);
>   networkLocation = NetworkTopology.DEFAULT_RACK;
> } else {
>   networkLocation = rName.get(0);
> }
> return networkLocation;
> {code}
> The line of code that is assigning default rack:
> {code} networkLocation = NetworkTopology.DEFAULT_RACK; {code} 
> can cause a serious problem. This means if somehow we got NULL, then the 
> default rack will be assigned as a DN's network location and DN's 
> registration will finish successfully. Under this circumstances, we will be 
> able to load data into cluster which is working with a wrong topology. Wrong  
> topology means that fault domains are not honored. 
> For the end user, it means that two data replicas can end up in the same 
> fault domain and a single failure can cause loss of two, or more, replicas. 
> Cluster would be in the inconsistent state but it would not be aware of that 
> and the whole thing would work as if everything was fine. We can notice that 
> something wrong happened almost only by looking in the log for the error:
> {code}
> LOG.error("The resolve call returned null! Using " + 
> NetworkTopology.DEFAULT_RACK + " for host " + names);
> {code}
>  



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-5846) Assigning DEFAULT_RACK in resolveNetworkLocation method can break data resiliency

2014-03-24 Thread Chris Nauroth (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5846?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chris Nauroth updated HDFS-5846:


Attachment: hdfs-5846.patch

+1 for the patch.

The test failure looks unrelated to me.  I couldn't reproduce it locally.  I'm 
re-uploading the same patch just to kick off another Jenkins run to confirm.

> Assigning DEFAULT_RACK in resolveNetworkLocation method can break data 
> resiliency
> -
>
> Key: HDFS-5846
> URL: https://issues.apache.org/jira/browse/HDFS-5846
> Project: Hadoop HDFS
>  Issue Type: Bug
>Reporter: Nikola Vujic
>Assignee: Nikola Vujic
> Attachments: hdfs-5846.patch, hdfs-5846.patch, hdfs-5846.patch
>
>
> Medhod CachedDNSToSwitchMapping::resolve() can return NULL which requires 
> careful handling. Null can be returned in two cases:
> • An error occurred with topology script execution (script crashes).
> • Script returns wrong number of values (other than expected)
> Critical handling is in the DN registration code. DN registration code is 
> responsible for assigning proper topology paths to all registered datanodes. 
> Existing code handles this NULL pointer on the following way 
> ({{resolveNetworkLocation}} method):
> {code}
> / /resolve its network location
> List rName = dnsToSwitchMapping.resolve(names);
> String networkLocation;
> if (rName == null) {
>   LOG.error("The resolve call returned null! Using " + 
>   NetworkTopology.DEFAULT_RACK + " for host " + names);
>   networkLocation = NetworkTopology.DEFAULT_RACK;
> } else {
>   networkLocation = rName.get(0);
> }
> return networkLocation;
> {code}
> The line of code that is assigning default rack:
> {code} networkLocation = NetworkTopology.DEFAULT_RACK; {code} 
> can cause a serious problem. This means if somehow we got NULL, then the 
> default rack will be assigned as a DN's network location and DN's 
> registration will finish successfully. Under this circumstances, we will be 
> able to load data into cluster which is working with a wrong topology. Wrong  
> topology means that fault domains are not honored. 
> For the end user, it means that two data replicas can end up in the same 
> fault domain and a single failure can cause loss of two, or more, replicas. 
> Cluster would be in the inconsistent state but it would not be aware of that 
> and the whole thing would work as if everything was fine. We can notice that 
> something wrong happened almost only by looking in the log for the error:
> {code}
> LOG.error("The resolve call returned null! Using " + 
> NetworkTopology.DEFAULT_RACK + " for host " + names);
> {code}
>  



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6148) LeaseManager crashes while initiating block recovery

2014-03-24 Thread Kihwal Lee (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6148?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13945473#comment-13945473
 ] 

Kihwal Lee commented on HDFS-6148:
--

It may have something to do with loading fsimage + edits and processing 
under-construction files.  LeaseManager crashes one hour after NN start-up.

> LeaseManager crashes while initiating block recovery
> 
>
> Key: HDFS-6148
> URL: https://issues.apache.org/jira/browse/HDFS-6148
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 2.4.0
>Reporter: Kihwal Lee
>Priority: Blocker
>
> While running branch-2.4, the LeaseManager crashed with an NPE. This does not 
> always happen on block recovery.
> {panel}
> Exception in thread
> "org.apache.hadoop.hdfs.server.namenode.LeaseManager$Monitor@5d66b728"
> java.lang.NullPointerException
> at 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockInfoUnderConstruction$
> 
> ReplicaUnderConstruction.isAlive(BlockInfoUnderConstruction.java:121)
> at 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockInfoUnderConstruction.
> initializeBlockRecovery(BlockInfoUnderConstruction.java:286)
> at 
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.internalReleaseLease(FSNamesystem.java:3746)
> at 
> org.apache.hadoop.hdfs.server.namenode.LeaseManager.checkLeases(LeaseManager.java:474)
> at 
> org.apache.hadoop.hdfs.server.namenode.LeaseManager.access$900(LeaseManager.java:68)
> at 
> org.apache.hadoop.hdfs.server.namenode.LeaseManager$Monitor.run(LeaseManager.java:411)
> at java.lang.Thread.run(Thread.java:722)
> {panel}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6148) LeaseManager crashes while initiating block recovery

2014-03-24 Thread Kihwal Lee (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6148?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13945426#comment-13945426
 ] 

Kihwal Lee commented on HDFS-6148:
--

replicas.size() was non-zero and there was a corresponding 
ReplicaUnderConstruction, but its expectedLocation seemed to be null.  This can 
happen if setExpectedStorageLocations() were called with array of nulls.  This 
might happen if a last block with null locations is turned into a 
BlockInfoUnderConstruction.   There might be other ways though.

> LeaseManager crashes while initiating block recovery
> 
>
> Key: HDFS-6148
> URL: https://issues.apache.org/jira/browse/HDFS-6148
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 2.4.0
>Reporter: Kihwal Lee
>Priority: Blocker
>
> While running branch-2.4, the LeaseManager crashed with an NPE. This does not 
> always happen on block recovery.
> {panel}
> Exception in thread
> "org.apache.hadoop.hdfs.server.namenode.LeaseManager$Monitor@5d66b728"
> java.lang.NullPointerException
> at 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockInfoUnderConstruction$
> 
> ReplicaUnderConstruction.isAlive(BlockInfoUnderConstruction.java:121)
> at 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockInfoUnderConstruction.
> initializeBlockRecovery(BlockInfoUnderConstruction.java:286)
> at 
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.internalReleaseLease(FSNamesystem.java:3746)
> at 
> org.apache.hadoop.hdfs.server.namenode.LeaseManager.checkLeases(LeaseManager.java:474)
> at 
> org.apache.hadoop.hdfs.server.namenode.LeaseManager.access$900(LeaseManager.java:68)
> at 
> org.apache.hadoop.hdfs.server.namenode.LeaseManager$Monitor.run(LeaseManager.java:411)
> at java.lang.Thread.run(Thread.java:722)
> {panel}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6149) Running UTs with testKerberos profile has failures.

2014-03-24 Thread Jinghui Wang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6149?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jinghui Wang updated HDFS-6149:
---

Attachment: HDFS-6149.patch

> Running UTs with testKerberos profile has failures.
> ---
>
> Key: HDFS-6149
> URL: https://issues.apache.org/jira/browse/HDFS-6149
> Project: Hadoop HDFS
>  Issue Type: Test
>  Components: test
>Affects Versions: 2.2.0
>Reporter: Jinghui Wang
>Assignee: Jinghui Wang
> Fix For: 2.3.0
>
> Attachments: HDFS-6149.patch
>
>
> UT failures in TestHttpFSWithKerberos.
> Tests using testDelegationTokenWithinDoAs fail because of the statically set 
> keytab file.
> Test testDelegationTokenHttpFSAccess also fails due the incorrect assumption 
> that CANCELDELEGATIONTOKEN does not require credentials.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


  1   2   >