[jira] [Commented] (HDFS-5804) HDFS NFS Gateway fails to mount and proxy when using Kerberos

2014-01-24 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5804?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13881707#comment-13881707
 ] 

Hadoop QA commented on HDFS-5804:
-

{color:green}+1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12625100/HDFS-5804.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 1 new 
or modified test files.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  The javadoc tool did not generate any 
warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:green}+1 core tests{color}.  The patch passed unit tests in 
hadoop-hdfs-project/hadoop-hdfs-nfs.

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-HDFS-Build/5940//testReport/
Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/5940//console

This message is automatically generated.

> HDFS NFS Gateway fails to mount and proxy when using Kerberos
> -
>
> Key: HDFS-5804
> URL: https://issues.apache.org/jira/browse/HDFS-5804
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: nfs
>Affects Versions: 3.0.0, 2.2.0
>Reporter: Abin Shahab
> Attachments: HDFS-5804.patch, HDFS-5804.patch, HDFS-5804.patch, 
> exception-as-root.log, javadoc-after-patch.log, javadoc-before-patch.log
>
>
> When using HDFS nfs gateway with secure hadoop 
> (hadoop.security.authentication: kerberos), mounting hdfs fails. 
> Additionally, there is no mechanism to support proxy user(nfs needs to proxy 
> as the user invoking commands on the hdfs mount).
> Steps to reproduce:
> 1) start a hadoop cluster with kerberos enabled.
> 2) sudo su -l nfsserver and start an nfs server. This 'nfsserver' account has 
> a an account in kerberos.
> 3) Get the keytab for nfsserver, and issue the following mount command: mount 
> -t nfs -o vers=3,proto=tcp,nolock $server:/  $mount_point
> 4) You'll see in the nfsserver logs that Kerberos is complaining about not 
> having a TGT for root.
> This is the stacktrace: 
> java.io.IOException: Failed on local exception: java.io.IOException: 
> org.apache.hadoop.security.AccessControlException: Client cannot authenticate 
> via:[TOKEN, KERBEROS]; Host Details : local host is: 
> "my-nfs-server-host.com/10.252.4.197"; destination host is: 
> "my-namenode-host.com":8020; 
>   at org.apache.hadoop.net.NetUtils.wrapException(NetUtils.java:764)
>   at org.apache.hadoop.ipc.Client.call(Client.java:1351)
>   at org.apache.hadoop.ipc.Client.call(Client.java:1300)
>   at 
> org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:206)
>   at com.sun.proxy.$Proxy9.getFileLinkInfo(Unknown Source)
>   at sun.reflect.GeneratedMethodAccessor2.invoke(Unknown Source)
>   at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>   at java.lang.reflect.Method.invoke(Method.java:606)
>   at 
> org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:186)
>   at 
> org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:102)
>   at com.sun.proxy.$Proxy9.getFileLinkInfo(Unknown Source)
>   at 
> org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.getFileLinkInfo(ClientNamenodeProtocolTranslatorPB.java:664)
>   at org.apache.hadoop.hdfs.DFSClient.getFileLinkInfo(DFSClient.java:1713)
>   at 
> org.apache.hadoop.hdfs.nfs.nfs3.Nfs3Utils.getFileStatus(Nfs3Utils.java:58)
>   at 
> org.apache.hadoop.hdfs.nfs.nfs3.Nfs3Utils.getFileAttr(Nfs3Utils.java:79)
>   at 
> org.apache.hadoop.hdfs.nfs.nfs3.RpcProgramNfs3.fsinfo(RpcProgramNfs3.java:1643)
>   at 
> org.apache.hadoop.hdfs.nfs.nfs3.RpcProgramNfs3.handleInternal(RpcProgramNfs3.java:1891)
>   at 
> org.apache.hadoop.oncrpc.RpcProgram.messageReceived(RpcProgram.java:143)
>   at 
> org.jboss.netty.channel.SimpleChannelUpstreamHandler.handleUpstream(SimpleChannelUpstreamHandler.java:70)
>   at 
> org.jboss.netty.channel.DefaultChannelPipeline.sendUpstream(DefaultChannelPipeline.java:560)
>   at 
> org.jboss.netty.channel.DefaultChannelPipeline$DefaultChannelHandlerContext.sendUps

[jira] [Commented] (HDFS-5728) [Diskfull] Block recovery will fail if the metafile does not have crc for all chunks of the block

2014-01-24 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5728?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13881686#comment-13881686
 ] 

Hudson commented on HDFS-5728:
--

SUCCESS: Integrated in Hadoop-trunk-Commit #5036 (See 
[https://builds.apache.org/job/Hadoop-trunk-Commit/5036/])
HDFS-5728. Block recovery will fail if the metafile does not have crc for all 
chunks of the block. Contributed by Vinay. (kihwal: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1561223)
* /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/BlockPoolSlice.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestLeaseRecovery.java


> [Diskfull] Block recovery will fail if the metafile does not have crc for all 
> chunks of the block
> -
>
> Key: HDFS-5728
> URL: https://issues.apache.org/jira/browse/HDFS-5728
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: datanode
>Affects Versions: 0.23.10, 2.2.0
>Reporter: Vinay
>Assignee: Vinay
>Priority: Critical
> Fix For: 3.0.0, 2.4.0
>
> Attachments: HDFS-5728.patch, HDFS-5728.patch, HDFS-5728.patch
>
>
> 1. Client (regionsever) has opened stream to write its WAL to HDFS. This is 
> not one time upload, data will be written slowly.
> 2. One of the DataNode got diskfull ( due to some other data filled up disks)
> 3. Unfortunately block was being written to only this datanode in cluster, so 
> client write has also failed.
> 4. After some time disk is made free and all processes are restarted.
> 5. Now HMaster try to recover the file by calling recoverLease. 
> At this time recovery was failing saying file length mismatch.
> When checked,
>  actual block file length: 62484480
>  Calculated block length: 62455808
> This was because, metafile was having crc for only 62455808 bytes, and it 
> considered 62455808 as the block size.
> No matter how many times, recovery was continously failing.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (HDFS-5241) Provide alternate queuing audit logger to reduce logging contention

2014-01-24 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5241?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13881684#comment-13881684
 ] 

Hudson commented on HDFS-5241:
--

SUCCESS: Integrated in Hadoop-trunk-Commit #5036 (See 
[https://builds.apache.org/job/Hadoop-trunk-Commit/5036/])
HDFS-5241.  Provide alternate queuing audit logger to reduce logging contention 
(daryn) (daryn: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1560761)
* /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/DFSConfigKeys.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSNamesystem.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestAuditLogs.java


> Provide alternate queuing audit logger to reduce logging contention
> ---
>
> Key: HDFS-5241
> URL: https://issues.apache.org/jira/browse/HDFS-5241
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: namenode
>Affects Versions: 2.0.0-alpha, 3.0.0
>Reporter: Daryn Sharp
>Assignee: Daryn Sharp
> Attachments: HDFS-5241.patch, HDFS-5241.patch
>
>
> The default audit logger has extremely poor performance.  The internal 
> synchronization of log4j causes massive contention between the call handlers 
> (100 by default) which drastically limits the throughput of the NN.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (HDFS-5788) listLocatedStatus response can be very large

2014-01-24 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5788?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13881679#comment-13881679
 ] 

Hudson commented on HDFS-5788:
--

SUCCESS: Integrated in Hadoop-trunk-Commit #5036 (See 
[https://builds.apache.org/job/Hadoop-trunk-Commit/5036/])
HDFS-5788. listLocatedStatus response can be very large. Contributed by Nathan 
Roberts. (kihwal: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1560750)
* /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSDirectory.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestINodeFile.java


> listLocatedStatus response can be very large
> 
>
> Key: HDFS-5788
> URL: https://issues.apache.org/jira/browse/HDFS-5788
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: namenode
>Affects Versions: 3.0.0, 0.23.10, 2.2.0
>Reporter: Nathan Roberts
>Assignee: Nathan Roberts
> Fix For: 3.0.0, 2.4.0
>
> Attachments: HDFS-5788.patch
>
>
> Currently we limit the size of listStatus requests to a default of 1000 
> entries. This works fine except in the case of listLocatedStatus where the 
> location information can be quite large. As an example, a directory with 7000 
> entries, 4 blocks each, 3 way replication - a listLocatedStatus response is 
> over 1MB. This can chew up very large amounts of memory in the NN if lots of 
> clients try to do this simultaneously.
> Seems like it would be better if we also considered the amount of location 
> information being returned when deciding how many files to return.
> Patch will follow shortly.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (HDFS-5789) Some of snapshot APIs missing checkOperation double check in fsn

2014-01-24 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5789?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13881678#comment-13881678
 ] 

Hudson commented on HDFS-5789:
--

SUCCESS: Integrated in Hadoop-trunk-Commit #5036 (See 
[https://builds.apache.org/job/Hadoop-trunk-Commit/5036/])
HDFS-5789. Some of snapshot APIs missing checkOperation double check in fsn. 
Contributed by Uma Maheswara Rao G. (umamahesh: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1560731)
* /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSNamesystem.java


> Some of snapshot APIs missing checkOperation double check in fsn
> 
>
> Key: HDFS-5789
> URL: https://issues.apache.org/jira/browse/HDFS-5789
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: namenode
>Affects Versions: 3.0.0
>Reporter: Uma Maheswara Rao G
>Assignee: Uma Maheswara Rao G
> Fix For: 3.0.0, 2.3.0
>
> Attachments: HDFS-5789.patch
>
>
> HDFS-4591 introduced double checked for HA state while taking fsn lock.
> checkoperation made before actually taking lock and after the lock again.
> This pattern missed in some of the snapshot APIs and cache management related 
> APIs.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (HDFS-5776) Support 'hedged' reads in DFSClient

2014-01-24 Thread Colin Patrick McCabe (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5776?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13881661#comment-13881661
 ] 

Colin Patrick McCabe commented on HDFS-5776:


I might be misunderstanding, but it seems like this should be a client setting, 
not a datanode setting.  Right?

> Support 'hedged' reads in DFSClient
> ---
>
> Key: HDFS-5776
> URL: https://issues.apache.org/jira/browse/HDFS-5776
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: hdfs-client
>Affects Versions: 3.0.0
>Reporter: Liang Xie
>Assignee: Liang Xie
> Attachments: HDFS-5776-v2.txt, HDFS-5776-v3.txt, HDFS-5776-v4.txt, 
> HDFS-5776-v5.txt, HDFS-5776-v6.txt, HDFS-5776-v7.txt, HDFS-5776-v8.txt, 
> HDFS-5776.txt
>
>
> This is a placeholder of hdfs related stuff backport from 
> https://issues.apache.org/jira/browse/HBASE-7509
> The quorum read ability should be helpful especially to optimize read outliers
> we can utilize "dfs.dfsclient.quorum.read.threshold.millis" & 
> "dfs.dfsclient.quorum.read.threadpool.size" to enable/disable the hedged read 
> ability from client side(e.g. HBase), and by using DFSQuorumReadMetrics, we 
> could export the interested metric valus into client system(e.g. HBase's 
> regionserver metric).
> The core logic is in pread code path, we decide to goto the original 
> fetchBlockByteRange or the new introduced fetchBlockByteRangeSpeculative per 
> the above config items.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (HDFS-5702) FsShell Cli: Add XML based End-to-End test for getfacl and setfacl commands

2014-01-24 Thread Chris Nauroth (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5702?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13881572#comment-13881572
 ] 

Chris Nauroth commented on HDFS-5702:
-

Vinay, in addition to the list above, could you please also add a fourth test?

4. setfacl -x mask::

This test will fail until HADOOP-10277 gets fixed.  I have a patch in progress 
now.

> FsShell Cli: Add XML based End-to-End test for getfacl and setfacl commands
> ---
>
> Key: HDFS-5702
> URL: https://issues.apache.org/jira/browse/HDFS-5702
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: hdfs-client, namenode, security
>Reporter: Vinay
>Assignee: Vinay
> Attachments: HDFS-5702.patch
>
>
> FsShell Cli: Add XML based End-to-End test for getfacl and setfacl commands



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (HDFS-5631) Expose interfaces required by FsDatasetSpi implementations

2014-01-24 Thread David Powell (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5631?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

David Powell updated HDFS-5631:
---

Attachment: HDFS-5631.patch

> Expose interfaces required by FsDatasetSpi implementations
> --
>
> Key: HDFS-5631
> URL: https://issues.apache.org/jira/browse/HDFS-5631
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: datanode
>Affects Versions: 3.0.0
>Reporter: David Powell
>Priority: Minor
> Attachments: HDFS-5631.patch, HDFS-5631.patch
>
>
> This sub-task addresses section 4.1 of the document attached to HDFS-5194,
> the exposure of interfaces needed by a FsDatasetSpi implementation.
> Specifically it makes ChunkChecksum public and BlockMetadataHeader's
> readHeader() and writeHeader() methods public.
> The changes to BlockReaderUtil (and related classes) discussed by section
> 4.1 are only needed if supporting short-circuit, and should be addressed
> as part of an effort to provide such support rather than this JIRA.
> To help ensure these changes are complete and are not regressed in the
> future, tests that gauge the accessibility (though *not* behavior)
> of interfaces needed by a FsDatasetSpi subclass are also included.
> These take the form of a dummy FsDatasetSpi subclass -- a successful
> compilation is effectively a pass.  Trivial unit tests are included so
> that there is something tangible to track.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (HDFS-5631) Expose interfaces required by FsDatasetSpi implementations

2014-01-24 Thread David Powell (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5631?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

David Powell updated HDFS-5631:
---

Target Version/s: 3.0.0
  Status: Patch Available  (was: Open)

> Expose interfaces required by FsDatasetSpi implementations
> --
>
> Key: HDFS-5631
> URL: https://issues.apache.org/jira/browse/HDFS-5631
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: datanode
>Affects Versions: 3.0.0
>Reporter: David Powell
>Priority: Minor
> Attachments: HDFS-5631.patch, HDFS-5631.patch
>
>
> This sub-task addresses section 4.1 of the document attached to HDFS-5194,
> the exposure of interfaces needed by a FsDatasetSpi implementation.
> Specifically it makes ChunkChecksum public and BlockMetadataHeader's
> readHeader() and writeHeader() methods public.
> The changes to BlockReaderUtil (and related classes) discussed by section
> 4.1 are only needed if supporting short-circuit, and should be addressed
> as part of an effort to provide such support rather than this JIRA.
> To help ensure these changes are complete and are not regressed in the
> future, tests that gauge the accessibility (though *not* behavior)
> of interfaces needed by a FsDatasetSpi subclass are also included.
> These take the form of a dummy FsDatasetSpi subclass -- a successful
> compilation is effectively a pass.  Trivial unit tests are included so
> that there is something tangible to track.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (HDFS-5698) Use protobuf to serialize / deserialize FSImage

2014-01-24 Thread Kihwal Lee (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5698?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13881543#comment-13881543
 ] 

Kihwal Lee commented on HDFS-5698:
--

What will be the performance/size implications of this?

> Use protobuf to serialize / deserialize FSImage
> ---
>
> Key: HDFS-5698
> URL: https://issues.apache.org/jira/browse/HDFS-5698
> Project: Hadoop HDFS
>  Issue Type: Improvement
>Reporter: Haohui Mai
>Assignee: Haohui Mai
> Attachments: HDFS-5698.000.patch
>
>
> Currently, the code serializes FSImage using in-house serialization 
> mechanisms. There are a couple disadvantages of the current approach:
> # Mixing the responsibility of reconstruction and serialization / 
> deserialization. The current code paths of serialization / deserialization 
> have spent a lot of effort on maintaining compatibility. What is worse is 
> that they are mixed with the complex logic of reconstructing the namespace, 
> making the code difficult to follow.
> # Poor documentation of the current FSImage format. The format of the FSImage 
> is practically defined by the implementation. An bug in implementation means 
> a bug in the specification. Furthermore, it also makes writing third-party 
> tools quite difficult.
> # Changing schemas is non-trivial. Adding a field in FSImage requires bumping 
> the layout version every time. Bumping out layout version requires (1) the 
> users to explicitly upgrade the clusters, and (2) putting new code to 
> maintain backward compatibility.
> This jira proposes to use protobuf to serialize the FSImage. Protobuf has 
> been used to serialize / deserialize the RPC message in Hadoop.
> Protobuf addresses all the above problems. It clearly separates the 
> responsibility of serialization and reconstructing the namespace. The 
> protobuf files document the current format of the FSImage. The developers now 
> can add optional fields with ease, since the old code can always read the new 
> FSImage.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (HDFS-5830) WebHdfsFileSystem.getFileBlockLocations throws IllegalArgumentException when accessing another cluster.

2014-01-24 Thread Kihwal Lee (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5830?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13881535#comment-13881535
 ] 

Kihwal Lee commented on HDFS-5830:
--

Marking it as a blocker since WebHdfsFileSystem should remain compatible.  
{{toLocatedBlock()}} should use a different version of constructor if cached 
location is null.

> WebHdfsFileSystem.getFileBlockLocations throws IllegalArgumentException when 
> accessing another cluster. 
> 
>
> Key: HDFS-5830
> URL: https://issues.apache.org/jira/browse/HDFS-5830
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: caching, hdfs-client
>Affects Versions: 2.3.0
>Reporter: Yongjun Zhang
>Assignee: Yongjun Zhang
>Priority: Blocker
>
> WebHdfsFileSystem.getFileBlockLocations throws IllegalArgumentException when 
> accessing a another cluster (that doesn't have caching support). 
> java.lang.IllegalArgumentException: cachedLocs should not be null, use a 
> different constructor
> at com.google.common.base.Preconditions.checkArgument(Preconditions.java:88)
> at org.apache.hadoop.hdfs.protocol.LocatedBlock.(LocatedBlock.java:79)
> at org.apache.hadoop.hdfs.web.JsonUtil.toLocatedBlock(JsonUtil.java:414)
> at org.apache.hadoop.hdfs.web.JsonUtil.toLocatedBlockList(JsonUtil.java:446)
> at org.apache.hadoop.hdfs.web.JsonUtil.toLocatedBlocks(JsonUtil.java:479)
> at 
> org.apache.hadoop.hdfs.web.WebHdfsFileSystem.getFileBlockLocations(WebHdfsFileSystem.java:1067)
> at org.apache.hadoop.fs.FileSystem$4.next(FileSystem.java:1812)
> at org.apache.hadoop.fs.FileSystem$4.next(FileSystem.java:1797)



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (HDFS-5830) WebHdfsFileSystem.getFileBlockLocations throws IllegalArgumentException when accessing another cluster.

2014-01-24 Thread Kihwal Lee (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5830?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kihwal Lee updated HDFS-5830:
-

Priority: Blocker  (was: Major)
Target Version/s: 2.3.0

> WebHdfsFileSystem.getFileBlockLocations throws IllegalArgumentException when 
> accessing another cluster. 
> 
>
> Key: HDFS-5830
> URL: https://issues.apache.org/jira/browse/HDFS-5830
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: caching, hdfs-client
>Affects Versions: 2.3.0
>Reporter: Yongjun Zhang
>Assignee: Yongjun Zhang
>Priority: Blocker
>
> WebHdfsFileSystem.getFileBlockLocations throws IllegalArgumentException when 
> accessing a another cluster (that doesn't have caching support). 
> java.lang.IllegalArgumentException: cachedLocs should not be null, use a 
> different constructor
> at com.google.common.base.Preconditions.checkArgument(Preconditions.java:88)
> at org.apache.hadoop.hdfs.protocol.LocatedBlock.(LocatedBlock.java:79)
> at org.apache.hadoop.hdfs.web.JsonUtil.toLocatedBlock(JsonUtil.java:414)
> at org.apache.hadoop.hdfs.web.JsonUtil.toLocatedBlockList(JsonUtil.java:446)
> at org.apache.hadoop.hdfs.web.JsonUtil.toLocatedBlocks(JsonUtil.java:479)
> at 
> org.apache.hadoop.hdfs.web.WebHdfsFileSystem.getFileBlockLocations(WebHdfsFileSystem.java:1067)
> at org.apache.hadoop.fs.FileSystem$4.next(FileSystem.java:1812)
> at org.apache.hadoop.fs.FileSystem$4.next(FileSystem.java:1797)



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (HDFS-5728) [Diskfull] Block recovery will fail if the metafile does not have crc for all chunks of the block

2014-01-24 Thread Kihwal Lee (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5728?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13881518#comment-13881518
 ] 

Kihwal Lee commented on HDFS-5728:
--

I've committed this to trunk and branch-2. Thank you for reporting and working 
on the patch, Vinay. Thanks for the reveiw, Uma.

> [Diskfull] Block recovery will fail if the metafile does not have crc for all 
> chunks of the block
> -
>
> Key: HDFS-5728
> URL: https://issues.apache.org/jira/browse/HDFS-5728
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: datanode
>Affects Versions: 0.23.10, 2.2.0
>Reporter: Vinay
>Assignee: Vinay
>Priority: Critical
> Fix For: 3.0.0, 2.4.0
>
> Attachments: HDFS-5728.patch, HDFS-5728.patch, HDFS-5728.patch
>
>
> 1. Client (regionsever) has opened stream to write its WAL to HDFS. This is 
> not one time upload, data will be written slowly.
> 2. One of the DataNode got diskfull ( due to some other data filled up disks)
> 3. Unfortunately block was being written to only this datanode in cluster, so 
> client write has also failed.
> 4. After some time disk is made free and all processes are restarted.
> 5. Now HMaster try to recover the file by calling recoverLease. 
> At this time recovery was failing saying file length mismatch.
> When checked,
>  actual block file length: 62484480
>  Calculated block length: 62455808
> This was because, metafile was having crc for only 62455808 bytes, and it 
> considered 62455808 as the block size.
> No matter how many times, recovery was continously failing.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (HDFS-5728) [Diskfull] Block recovery will fail if the metafile does not have crc for all chunks of the block

2014-01-24 Thread Kihwal Lee (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5728?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kihwal Lee updated HDFS-5728:
-

   Resolution: Fixed
Fix Version/s: 2.4.0
   3.0.0
 Hadoop Flags: Reviewed
   Status: Resolved  (was: Patch Available)

> [Diskfull] Block recovery will fail if the metafile does not have crc for all 
> chunks of the block
> -
>
> Key: HDFS-5728
> URL: https://issues.apache.org/jira/browse/HDFS-5728
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: datanode
>Affects Versions: 0.23.10, 2.2.0
>Reporter: Vinay
>Assignee: Vinay
>Priority: Critical
> Fix For: 3.0.0, 2.4.0
>
> Attachments: HDFS-5728.patch, HDFS-5728.patch, HDFS-5728.patch
>
>
> 1. Client (regionsever) has opened stream to write its WAL to HDFS. This is 
> not one time upload, data will be written slowly.
> 2. One of the DataNode got diskfull ( due to some other data filled up disks)
> 3. Unfortunately block was being written to only this datanode in cluster, so 
> client write has also failed.
> 4. After some time disk is made free and all processes are restarted.
> 5. Now HMaster try to recover the file by calling recoverLease. 
> At this time recovery was failing saying file length mismatch.
> When checked,
>  actual block file length: 62484480
>  Calculated block length: 62455808
> This was because, metafile was having crc for only 62455808 bytes, and it 
> considered 62455808 as the block size.
> No matter how many times, recovery was continously failing.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (HDFS-5728) [Diskfull] Block recovery will fail if the metafile does not have crc for all chunks of the block

2014-01-24 Thread Kihwal Lee (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5728?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kihwal Lee updated HDFS-5728:
-

Summary: [Diskfull] Block recovery will fail if the metafile does not have 
crc for all chunks of the block  (was: [Diskfull] Block recovery will fail if 
the metafile not having crc for all chunks of the block)

> [Diskfull] Block recovery will fail if the metafile does not have crc for all 
> chunks of the block
> -
>
> Key: HDFS-5728
> URL: https://issues.apache.org/jira/browse/HDFS-5728
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: datanode
>Affects Versions: 0.23.10, 2.2.0
>Reporter: Vinay
>Assignee: Vinay
>Priority: Critical
> Fix For: 3.0.0, 2.4.0
>
> Attachments: HDFS-5728.patch, HDFS-5728.patch, HDFS-5728.patch
>
>
> 1. Client (regionsever) has opened stream to write its WAL to HDFS. This is 
> not one time upload, data will be written slowly.
> 2. One of the DataNode got diskfull ( due to some other data filled up disks)
> 3. Unfortunately block was being written to only this datanode in cluster, so 
> client write has also failed.
> 4. After some time disk is made free and all processes are restarted.
> 5. Now HMaster try to recover the file by calling recoverLease. 
> At this time recovery was failing saying file length mismatch.
> When checked,
>  actual block file length: 62484480
>  Calculated block length: 62455808
> This was because, metafile was having crc for only 62455808 bytes, and it 
> considered 62455808 as the block size.
> No matter how many times, recovery was continously failing.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (HDFS-5728) [Diskfull] Block recovery will fail if the metafile not having crc for all chunks of the block

2014-01-24 Thread Kihwal Lee (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5728?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kihwal Lee updated HDFS-5728:
-

Target Version/s: 2.4.0  (was: 2.4.0, 0.23.11)

> [Diskfull] Block recovery will fail if the metafile not having crc for all 
> chunks of the block
> --
>
> Key: HDFS-5728
> URL: https://issues.apache.org/jira/browse/HDFS-5728
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: datanode
>Affects Versions: 0.23.10, 2.2.0
>Reporter: Vinay
>Assignee: Vinay
>Priority: Critical
> Attachments: HDFS-5728.patch, HDFS-5728.patch, HDFS-5728.patch
>
>
> 1. Client (regionsever) has opened stream to write its WAL to HDFS. This is 
> not one time upload, data will be written slowly.
> 2. One of the DataNode got diskfull ( due to some other data filled up disks)
> 3. Unfortunately block was being written to only this datanode in cluster, so 
> client write has also failed.
> 4. After some time disk is made free and all processes are restarted.
> 5. Now HMaster try to recover the file by calling recoverLease. 
> At this time recovery was failing saying file length mismatch.
> When checked,
>  actual block file length: 62484480
>  Calculated block length: 62455808
> This was because, metafile was having crc for only 62455808 bytes, and it 
> considered 62455808 as the block size.
> No matter how many times, recovery was continously failing.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (HDFS-5728) [Diskfull] Block recovery will fail if the metafile not having crc for all chunks of the block

2014-01-24 Thread Kihwal Lee (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5728?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13881506#comment-13881506
 ] 

Kihwal Lee commented on HDFS-5728:
--

The build server has been down for more than a day, so precommit won't run any 
time soon.
+1 I won't wait for the build server to return. The previous version of patch 
was fine except the lines removed in the latest version.


> [Diskfull] Block recovery will fail if the metafile not having crc for all 
> chunks of the block
> --
>
> Key: HDFS-5728
> URL: https://issues.apache.org/jira/browse/HDFS-5728
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: datanode
>Affects Versions: 0.23.10, 2.2.0
>Reporter: Vinay
>Assignee: Vinay
>Priority: Critical
> Attachments: HDFS-5728.patch, HDFS-5728.patch, HDFS-5728.patch
>
>
> 1. Client (regionsever) has opened stream to write its WAL to HDFS. This is 
> not one time upload, data will be written slowly.
> 2. One of the DataNode got diskfull ( due to some other data filled up disks)
> 3. Unfortunately block was being written to only this datanode in cluster, so 
> client write has also failed.
> 4. After some time disk is made free and all processes are restarted.
> 5. Now HMaster try to recover the file by calling recoverLease. 
> At this time recovery was failing saying file length mismatch.
> When checked,
>  actual block file length: 62484480
>  Calculated block length: 62455808
> This was because, metafile was having crc for only 62455808 bytes, and it 
> considered 62455808 as the block size.
> No matter how many times, recovery was continously failing.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (HDFS-5620) NameNode: implement Global ACL Set as a memory optimization.

2014-01-24 Thread Chris Nauroth (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5620?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chris Nauroth updated HDFS-5620:


Attachment: aclRamSizingEstimates.png
aclRamSizingEstimates

> NameNode: implement Global ACL Set as a memory optimization.
> 
>
> Key: HDFS-5620
> URL: https://issues.apache.org/jira/browse/HDFS-5620
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: namenode
>Affects Versions: HDFS ACLs (HDFS-4685)
>Reporter: Chris Nauroth
>Assignee: Chris Nauroth
> Attachments: HDFS-5620.1.patch, aclRamSizingEstimates, 
> aclRamSizingEstimates.png
>
>
> The {{AclManager}} can maintain a Global ACL Set to store all distinct ACLs 
> in use by the file system.  All inodes that have the same ACL entries can 
> share the same ACL instance.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Resolved] (HDFS-5620) NameNode: implement Global ACL Set as a memory optimization.

2014-01-24 Thread Chris Nauroth (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5620?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chris Nauroth resolved HDFS-5620.
-

Resolution: Won't Fix

I'm resolving this issue as won't fix.  [~wheat9] was right to question whether 
or not we really need this optimization.

Now that ACLs implementation is almost done, I can get a more accurate 
measurement of the memory impact using jmap.  I'm including a plot that 
estimates memory consumption by ACLs in a presumed 50 GB NameNode, trying to 
store up to 100 million inodes.  I plotted 3 different usage scenarios:
# max usage: All inodes have an ACL, and each one has the maximum ACL entries.
# mid usage: Half the inodes have an ACL, and each ACL has 10 entries.
# low usage: 10% of inodes have an ACL, and each ACL has 2 entries.

For each scenario, I plotted the unoptimized memory consumption and also the 
optimized memory consumption, assuming that only 20% of the ACLs are unique.  
Each scenario uses a different line color.  The unoptimized version uses a 
solid line, and the optimized version uses a dashed line.  I'm also attaching 
the GnuPlot script I wrote to generate this.

The plot demonstrates that ACL usage really needs to get quite high (very large 
number of inodes with a very high proportion of them having ACLs) before this 
memory optimization starts to provide a benefit.

I am +1 for skipping this for now.  We can always resurrect this patch if we 
observe a need for it based on real world usage of ACLs.  I'm also going to put 
this information into a new revision of the design doc on HDFS-4685.

BTW, if we do resurrect this patch, then I probably wouldn't commit the exact 
HDFS-5620.1.patch that I put together quickly as a prototype.  The Guava 
interner has a spinlock inside of it, which we don't really need, because this 
would only be accessed under the namesystem write lock anyway.

> NameNode: implement Global ACL Set as a memory optimization.
> 
>
> Key: HDFS-5620
> URL: https://issues.apache.org/jira/browse/HDFS-5620
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: namenode
>Affects Versions: HDFS ACLs (HDFS-4685)
>Reporter: Chris Nauroth
>Assignee: Chris Nauroth
> Attachments: HDFS-5620.1.patch, aclRamSizingEstimates, 
> aclRamSizingEstimates.png
>
>
> The {{AclManager}} can maintain a Global ACL Set to store all distinct ACLs 
> in use by the file system.  All inodes that have the same ACL entries can 
> share the same ACL instance.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (HDFS-5356) MiniDFSCluster shoud close all open FileSystems when shutdown()

2014-01-24 Thread Kihwal Lee (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5356?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kihwal Lee updated HDFS-5356:
-

Target Version/s: 2.4.0  (was: 2.3.0)

> MiniDFSCluster shoud close all open FileSystems when shutdown()
> ---
>
> Key: HDFS-5356
> URL: https://issues.apache.org/jira/browse/HDFS-5356
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: test
>Affects Versions: 3.0.0, 2.2.0
>Reporter: haosdent
>Priority: Critical
> Attachments: HDFS-5356.patch
>
>
> After add some metrics functions to DFSClient, I found that some unit tests 
> relates to metrics are failed. Because MiniDFSCluster are never close open 
> FileSystems, DFSClients are alive after MiniDFSCluster shutdown(). The 
> metrics of DFSClients in DefaultMetricsSystem are still exist and this make 
> other unit tests failed.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (HDFS-5804) HDFS NFS Gateway fails to mount and proxy when using Kerberos

2014-01-24 Thread Abin Shahab (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5804?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13881375#comment-13881375
 ] 

Abin Shahab commented on HDFS-5804:
---

Jing, let me know if you have any feedback on my patch.

> HDFS NFS Gateway fails to mount and proxy when using Kerberos
> -
>
> Key: HDFS-5804
> URL: https://issues.apache.org/jira/browse/HDFS-5804
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: nfs
>Affects Versions: 3.0.0, 2.2.0
>Reporter: Abin Shahab
> Attachments: HDFS-5804.patch, HDFS-5804.patch, HDFS-5804.patch, 
> exception-as-root.log, javadoc-after-patch.log, javadoc-before-patch.log
>
>
> When using HDFS nfs gateway with secure hadoop 
> (hadoop.security.authentication: kerberos), mounting hdfs fails. 
> Additionally, there is no mechanism to support proxy user(nfs needs to proxy 
> as the user invoking commands on the hdfs mount).
> Steps to reproduce:
> 1) start a hadoop cluster with kerberos enabled.
> 2) sudo su -l nfsserver and start an nfs server. This 'nfsserver' account has 
> a an account in kerberos.
> 3) Get the keytab for nfsserver, and issue the following mount command: mount 
> -t nfs -o vers=3,proto=tcp,nolock $server:/  $mount_point
> 4) You'll see in the nfsserver logs that Kerberos is complaining about not 
> having a TGT for root.
> This is the stacktrace: 
> java.io.IOException: Failed on local exception: java.io.IOException: 
> org.apache.hadoop.security.AccessControlException: Client cannot authenticate 
> via:[TOKEN, KERBEROS]; Host Details : local host is: 
> "my-nfs-server-host.com/10.252.4.197"; destination host is: 
> "my-namenode-host.com":8020; 
>   at org.apache.hadoop.net.NetUtils.wrapException(NetUtils.java:764)
>   at org.apache.hadoop.ipc.Client.call(Client.java:1351)
>   at org.apache.hadoop.ipc.Client.call(Client.java:1300)
>   at 
> org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:206)
>   at com.sun.proxy.$Proxy9.getFileLinkInfo(Unknown Source)
>   at sun.reflect.GeneratedMethodAccessor2.invoke(Unknown Source)
>   at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>   at java.lang.reflect.Method.invoke(Method.java:606)
>   at 
> org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:186)
>   at 
> org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:102)
>   at com.sun.proxy.$Proxy9.getFileLinkInfo(Unknown Source)
>   at 
> org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.getFileLinkInfo(ClientNamenodeProtocolTranslatorPB.java:664)
>   at org.apache.hadoop.hdfs.DFSClient.getFileLinkInfo(DFSClient.java:1713)
>   at 
> org.apache.hadoop.hdfs.nfs.nfs3.Nfs3Utils.getFileStatus(Nfs3Utils.java:58)
>   at 
> org.apache.hadoop.hdfs.nfs.nfs3.Nfs3Utils.getFileAttr(Nfs3Utils.java:79)
>   at 
> org.apache.hadoop.hdfs.nfs.nfs3.RpcProgramNfs3.fsinfo(RpcProgramNfs3.java:1643)
>   at 
> org.apache.hadoop.hdfs.nfs.nfs3.RpcProgramNfs3.handleInternal(RpcProgramNfs3.java:1891)
>   at 
> org.apache.hadoop.oncrpc.RpcProgram.messageReceived(RpcProgram.java:143)
>   at 
> org.jboss.netty.channel.SimpleChannelUpstreamHandler.handleUpstream(SimpleChannelUpstreamHandler.java:70)
>   at 
> org.jboss.netty.channel.DefaultChannelPipeline.sendUpstream(DefaultChannelPipeline.java:560)
>   at 
> org.jboss.netty.channel.DefaultChannelPipeline$DefaultChannelHandlerContext.sendUpstream(DefaultChannelPipeline.java:787)
>   at 
> org.jboss.netty.channel.Channels.fireMessageReceived(Channels.java:281)
>   at 
> org.apache.hadoop.oncrpc.RpcUtil$RpcMessageParserStage.messageReceived(RpcUtil.java:132)
>   at 
> org.jboss.netty.channel.SimpleChannelUpstreamHandler.handleUpstream(SimpleChannelUpstreamHandler.java:70)
>   at 
> org.jboss.netty.channel.DefaultChannelPipeline.sendUpstream(DefaultChannelPipeline.java:560)
>   at 
> org.jboss.netty.channel.DefaultChannelPipeline$DefaultChannelHandlerContext.sendUpstream(DefaultChannelPipeline.java:787)
>   at 
> org.jboss.netty.channel.Channels.fireMessageReceived(Channels.java:296)
>   at 
> org.jboss.netty.handler.codec.frame.FrameDecoder.unfoldAndFireMessageReceived(FrameDecoder.java:462)
>   at 
> org.jboss.netty.handler.codec.frame.FrameDecoder.callDecode(FrameDecoder.java:443)
>   at 
> org.jboss.netty.handler.codec.frame.FrameDecoder.messageReceived(FrameDecoder.java:303)
>   at 
> org.jboss.netty.channel.SimpleChannelUpstreamHandler.handleUpstream(SimpleChannelUpstreamHandler.java:70)
>   at 
> org.jboss.netty.channel.DefaultChannelPipeline.sendUpstream(DefaultChannelPipeline.java:560)
>   at 
> org.jboss.netty.channe

[jira] [Updated] (HDFS-5804) HDFS NFS Gateway fails to mount and proxy when using Kerberos

2014-01-24 Thread Abin Shahab (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5804?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Abin Shahab updated HDFS-5804:
--

Attachment: HDFS-5804.patch

Updated documentation and param names

> HDFS NFS Gateway fails to mount and proxy when using Kerberos
> -
>
> Key: HDFS-5804
> URL: https://issues.apache.org/jira/browse/HDFS-5804
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: nfs
>Affects Versions: 3.0.0, 2.2.0
>Reporter: Abin Shahab
> Attachments: HDFS-5804.patch, HDFS-5804.patch, HDFS-5804.patch, 
> exception-as-root.log, javadoc-after-patch.log, javadoc-before-patch.log
>
>
> When using HDFS nfs gateway with secure hadoop 
> (hadoop.security.authentication: kerberos), mounting hdfs fails. 
> Additionally, there is no mechanism to support proxy user(nfs needs to proxy 
> as the user invoking commands on the hdfs mount).
> Steps to reproduce:
> 1) start a hadoop cluster with kerberos enabled.
> 2) sudo su -l nfsserver and start an nfs server. This 'nfsserver' account has 
> a an account in kerberos.
> 3) Get the keytab for nfsserver, and issue the following mount command: mount 
> -t nfs -o vers=3,proto=tcp,nolock $server:/  $mount_point
> 4) You'll see in the nfsserver logs that Kerberos is complaining about not 
> having a TGT for root.
> This is the stacktrace: 
> java.io.IOException: Failed on local exception: java.io.IOException: 
> org.apache.hadoop.security.AccessControlException: Client cannot authenticate 
> via:[TOKEN, KERBEROS]; Host Details : local host is: 
> "my-nfs-server-host.com/10.252.4.197"; destination host is: 
> "my-namenode-host.com":8020; 
>   at org.apache.hadoop.net.NetUtils.wrapException(NetUtils.java:764)
>   at org.apache.hadoop.ipc.Client.call(Client.java:1351)
>   at org.apache.hadoop.ipc.Client.call(Client.java:1300)
>   at 
> org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:206)
>   at com.sun.proxy.$Proxy9.getFileLinkInfo(Unknown Source)
>   at sun.reflect.GeneratedMethodAccessor2.invoke(Unknown Source)
>   at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>   at java.lang.reflect.Method.invoke(Method.java:606)
>   at 
> org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:186)
>   at 
> org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:102)
>   at com.sun.proxy.$Proxy9.getFileLinkInfo(Unknown Source)
>   at 
> org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.getFileLinkInfo(ClientNamenodeProtocolTranslatorPB.java:664)
>   at org.apache.hadoop.hdfs.DFSClient.getFileLinkInfo(DFSClient.java:1713)
>   at 
> org.apache.hadoop.hdfs.nfs.nfs3.Nfs3Utils.getFileStatus(Nfs3Utils.java:58)
>   at 
> org.apache.hadoop.hdfs.nfs.nfs3.Nfs3Utils.getFileAttr(Nfs3Utils.java:79)
>   at 
> org.apache.hadoop.hdfs.nfs.nfs3.RpcProgramNfs3.fsinfo(RpcProgramNfs3.java:1643)
>   at 
> org.apache.hadoop.hdfs.nfs.nfs3.RpcProgramNfs3.handleInternal(RpcProgramNfs3.java:1891)
>   at 
> org.apache.hadoop.oncrpc.RpcProgram.messageReceived(RpcProgram.java:143)
>   at 
> org.jboss.netty.channel.SimpleChannelUpstreamHandler.handleUpstream(SimpleChannelUpstreamHandler.java:70)
>   at 
> org.jboss.netty.channel.DefaultChannelPipeline.sendUpstream(DefaultChannelPipeline.java:560)
>   at 
> org.jboss.netty.channel.DefaultChannelPipeline$DefaultChannelHandlerContext.sendUpstream(DefaultChannelPipeline.java:787)
>   at 
> org.jboss.netty.channel.Channels.fireMessageReceived(Channels.java:281)
>   at 
> org.apache.hadoop.oncrpc.RpcUtil$RpcMessageParserStage.messageReceived(RpcUtil.java:132)
>   at 
> org.jboss.netty.channel.SimpleChannelUpstreamHandler.handleUpstream(SimpleChannelUpstreamHandler.java:70)
>   at 
> org.jboss.netty.channel.DefaultChannelPipeline.sendUpstream(DefaultChannelPipeline.java:560)
>   at 
> org.jboss.netty.channel.DefaultChannelPipeline$DefaultChannelHandlerContext.sendUpstream(DefaultChannelPipeline.java:787)
>   at 
> org.jboss.netty.channel.Channels.fireMessageReceived(Channels.java:296)
>   at 
> org.jboss.netty.handler.codec.frame.FrameDecoder.unfoldAndFireMessageReceived(FrameDecoder.java:462)
>   at 
> org.jboss.netty.handler.codec.frame.FrameDecoder.callDecode(FrameDecoder.java:443)
>   at 
> org.jboss.netty.handler.codec.frame.FrameDecoder.messageReceived(FrameDecoder.java:303)
>   at 
> org.jboss.netty.channel.SimpleChannelUpstreamHandler.handleUpstream(SimpleChannelUpstreamHandler.java:70)
>   at 
> org.jboss.netty.channel.DefaultChannelPipeline.sendUpstream(DefaultChannelPipeline.java:560)
>   at 
> org.jboss.netty.channel.DefaultChannelPipeline.sendUpstream(Defa

[jira] [Updated] (HDFS-5796) The file system browser in the namenode UI requires SPNEGO.

2014-01-24 Thread Kihwal Lee (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5796?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kihwal Lee updated HDFS-5796:
-

Target Version/s: 2.4.0

> The file system browser in the namenode UI requires SPNEGO.
> ---
>
> Key: HDFS-5796
> URL: https://issues.apache.org/jira/browse/HDFS-5796
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 2.4.0
>Reporter: Kihwal Lee
>Priority: Critical
>
> After HDFS-5382, the browser makes webhdfs REST calls directly, requiring 
> SPNEGO to work between user's browser and namenode.  This won't work if the 
> cluster's security infrastructure is isolated from the regular network.  
> Moreover, SPNEGO is not supposed to be required for user-facing web pages.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (HDFS-5776) Support 'hedged' reads in DFSClient

2014-01-24 Thread Arpit Agarwal (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5776?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13881341#comment-13881341
 ] 

Arpit Agarwal commented on HDFS-5776:
-

I reviewed the v8 patch. The implementation of {{hedgedFetchBlockByteRange}} 
looks great. Nice use of synchronization tools to make the code easy to 
understand.

{quote}
how much? 5000? 50? or sth else? i think if enabling this feature 
explicitly, the end user/application should know a little backgroud at least, 
right?
since we don't have any knowledge about end-user's storage configuration, just 
image if they have a fast flash(with HDFS-2832 enabled), say fusionio, probably 
one real disk read only cost tens of microseconds,
{quote}
[~xieliang007]  #2 and #4 from my previous comment remain unaddressed.

Threads are not free. If you really want to provide a user configurable setting 
for the thread count there should be a limit on the order of 64/128. I leave 
the exact number to you. The best approach is to use a small multiple of the 
processor count.

If an app is not well behaved then the absence of limits can create a positive 
feedback loop. The slower the storage layer the more threads will get created 
when the correct behavior under load should be back off. Please add a thread 
count limit or ideally let’s not expose this setting at all.

The same goes for the delay. Please add a lower bound. The exact value is up to 
you. We can always revisit the value if it turns out to be a bottleneck.

{quote}
let's keep it there, it's more easier to understand for developer or end user.
{quote}
I don't think it helps to have these functions and as Jing pointed out there is 
no purpose for it. I think it would be best to leave a single config setting 
i.e. either a boolean or a thread count, and a single method 
{{#isHedgedReadsEnabled}} to query the status of the feature.

{quote}
yes, but the loop is very very light, only when some exceptions like 
AccessControlException/InvalidEncryptionKeyException/InvalidBlockTokenException 
happened, will do extra loop, and all those have a fast quit mechanism, like 
refetchToken/refetchEncryptionKey or disableLegacyBlockReaderLocal, so this 
loop will only be executed just a very few times
...
yes, you had a misunderstanding here. that's why i catch IOException fbae 
around fetchBlockAt. If we don't catch here, there will be always new refetch 
from outside loop and will have a spin loop
{quote}
I still do not understand how you are guarding against multiple refetch. 
Previously these counters were initialized outside any loop, now they are being 
reinitialized inside a loop.

{{chooseDataNode(LocatedBlock block)}} function looks redundant and should be 
removed.

> Support 'hedged' reads in DFSClient
> ---
>
> Key: HDFS-5776
> URL: https://issues.apache.org/jira/browse/HDFS-5776
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: hdfs-client
>Affects Versions: 3.0.0
>Reporter: Liang Xie
>Assignee: Liang Xie
> Attachments: HDFS-5776-v2.txt, HDFS-5776-v3.txt, HDFS-5776-v4.txt, 
> HDFS-5776-v5.txt, HDFS-5776-v6.txt, HDFS-5776-v7.txt, HDFS-5776-v8.txt, 
> HDFS-5776.txt
>
>
> This is a placeholder of hdfs related stuff backport from 
> https://issues.apache.org/jira/browse/HBASE-7509
> The quorum read ability should be helpful especially to optimize read outliers
> we can utilize "dfs.dfsclient.quorum.read.threshold.millis" & 
> "dfs.dfsclient.quorum.read.threadpool.size" to enable/disable the hedged read 
> ability from client side(e.g. HBase), and by using DFSQuorumReadMetrics, we 
> could export the interested metric valus into client system(e.g. HBase's 
> regionserver metric).
> The core logic is in pread code path, we decide to goto the original 
> fetchBlockByteRange or the new introduced fetchBlockByteRangeSpeculative per 
> the above config items.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (HDFS-5709) Improve upgrade with existing files and directories named ".snapshot"

2014-01-24 Thread Andrew Wang (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5709?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13881336#comment-13881336
 ] 

Andrew Wang commented on HDFS-5709:
---

I'm not sure if this adds anything to the discussion, but the current workflow 
in the patch is:

* By default, the config option is not set
* On upgrade, if we hit a .snapshot path, we throw an exception and abort 
telling the user to set the config option
* When the config option is set, there are log prints whenever something is 
renamed

This feels procedurally pretty similar to the alternative of a special startup 
option, and it's only required to be set if we hit a .snapshot path. Most users 
will never need to worry about this config.

I agree that doing renames with the context of the full parent path (e.g. 
/X/.snapshot renames to /X/.myx-snapshot, /Y/.snapshot renames to 
/Y/.myy.snapshot) would be more powerful, but IIUC that implies scanning the 
fsimage and logs initially to enumerate all the .snapshot paths, have the user 
configure their rename rules, then load the fsimage/logs again to do the rename.

Suresh, does any of this influence your view? I'd really like to move forward 
on this issue.

> Improve upgrade with existing files and directories named ".snapshot"
> -
>
> Key: HDFS-5709
> URL: https://issues.apache.org/jira/browse/HDFS-5709
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: namenode
>Affects Versions: 3.0.0, 2.2.0
>Reporter: Andrew Wang
>Assignee: Andrew Wang
>  Labels: snapshots, upgrade
> Attachments: hdfs-5709-1.patch, hdfs-5709-2.patch, hdfs-5709-3.patch, 
> hdfs-5709-4.patch, hdfs-5709-5.patch
>
>
> Right now in trunk, upgrade fails messily if the old fsimage or edits refer 
> to a directory named ".snapshot". We should at least print a better error 
> message (which I believe was the original intention in HDFS-4666), and [~atm] 
> proposed automatically renaming these files and directories.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Moved] (HDFS-5831) TestAuditLogs#testAuditAllowedStat sometimes fails in trunk

2014-01-24 Thread Ted Yu (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5831?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ted Yu moved HBASE-10415 to HDFS-5831:
--

Key: HDFS-5831  (was: HBASE-10415)
Project: Hadoop HDFS  (was: HBase)

> TestAuditLogs#testAuditAllowedStat sometimes fails in trunk
> ---
>
> Key: HDFS-5831
> URL: https://issues.apache.org/jira/browse/HDFS-5831
> Project: Hadoop HDFS
>  Issue Type: Test
>Reporter: Ted Yu
>
> Running TestAuditLogs on Linux, I got:
> {code}
> testAuditAllowedStat[1](org.apache.hadoop.hdfs.server.namenode.TestAuditLogs) 
>  Time elapsed: 6.677 sec  <<< FAILURE!
> java.lang.AssertionError: null
> at org.junit.Assert.fail(Assert.java:92)
> at org.junit.Assert.assertTrue(Assert.java:43)
> at org.junit.Assert.assertNotNull(Assert.java:526)
> at org.junit.Assert.assertNotNull(Assert.java:537)
> at 
> org.apache.hadoop.hdfs.server.namenode.TestAuditLogs.verifyAuditLogsRepeat(TestAuditLogs.java:312)
> at 
> org.apache.hadoop.hdfs.server.namenode.TestAuditLogs.verifyAuditLogs(TestAuditLogs.java:295)
> at 
> org.apache.hadoop.hdfs.server.namenode.TestAuditLogs.testAuditAllowedStat(TestAuditLogs.java:163)
> {code}



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (HDFS-5831) TestAuditLogs#testAuditAllowedStat sometimes fails in trunk

2014-01-24 Thread Ted Yu (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5831?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ted Yu updated HDFS-5831:
-

Attachment: 
5831-org.apache.hadoop.hdfs.server.namenode.TestAuditLogs-output.txt

Test output.

> TestAuditLogs#testAuditAllowedStat sometimes fails in trunk
> ---
>
> Key: HDFS-5831
> URL: https://issues.apache.org/jira/browse/HDFS-5831
> Project: Hadoop HDFS
>  Issue Type: Test
>Reporter: Ted Yu
> Attachments: 
> 5831-org.apache.hadoop.hdfs.server.namenode.TestAuditLogs-output.txt
>
>
> Running TestAuditLogs on Linux, I got:
> {code}
> testAuditAllowedStat[1](org.apache.hadoop.hdfs.server.namenode.TestAuditLogs) 
>  Time elapsed: 6.677 sec  <<< FAILURE!
> java.lang.AssertionError: null
> at org.junit.Assert.fail(Assert.java:92)
> at org.junit.Assert.assertTrue(Assert.java:43)
> at org.junit.Assert.assertNotNull(Assert.java:526)
> at org.junit.Assert.assertNotNull(Assert.java:537)
> at 
> org.apache.hadoop.hdfs.server.namenode.TestAuditLogs.verifyAuditLogsRepeat(TestAuditLogs.java:312)
> at 
> org.apache.hadoop.hdfs.server.namenode.TestAuditLogs.verifyAuditLogs(TestAuditLogs.java:295)
> at 
> org.apache.hadoop.hdfs.server.namenode.TestAuditLogs.testAuditAllowedStat(TestAuditLogs.java:163)
> {code}



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Resolved] (HDFS-5829) TestDNFencingWithReplication fails on branch2

2014-01-24 Thread Mit Desai (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5829?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mit Desai resolved HDFS-5829.
-

Resolution: Cannot Reproduce

> TestDNFencingWithReplication fails on branch2
> -
>
> Key: HDFS-5829
> URL: https://issues.apache.org/jira/browse/HDFS-5829
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 2.2.0
>Reporter: Mit Desai
>
> {noformat}
> Running org.apache.hadoop.hdfs.server.namenode.ha.TestDNFencingWithReplication
> Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.097 sec <<< 
> FAILURE!
> testFencingStress(org.apache.hadoop.hdfs.server.namenode.ha.TestDNFencingWithReplication)
>   Time elapsed: 6 sec  <<< ERROR!
> java.lang.ExceptionInInitializerError
>   at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
>   at 
> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57)
>   at 
> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
>   at java.lang.reflect.Constructor.newInstance(Constructor.java:526)
>   at 
> org.junit.runners.BlockJUnit4ClassRunner.createTest(BlockJUnit4ClassRunner.java:187)
>   at 
> org.junit.runners.BlockJUnit4ClassRunner$1.runReflectiveCall(BlockJUnit4ClassRunner.java:236)
>   at 
> org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:15)
>   at 
> org.junit.runners.BlockJUnit4ClassRunner.methodBlock(BlockJUnit4ClassRunner.java:233)
>   at 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:68)
>   at 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:47)
>   at org.junit.runners.ParentRunner$3.run(ParentRunner.java:231)
>   at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:60)
>   at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:229)
>   at org.junit.runners.ParentRunner.access$000(ParentRunner.java:50)
>   at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:222)
>   at org.junit.runners.ParentRunner.run(ParentRunner.java:300)
>   at 
> org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:252)
>   at 
> org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:141)
>   at 
> org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:112)
>   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>   at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>   at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>   at java.lang.reflect.Method.invoke(Method.java:606)
>   at 
> org.apache.maven.surefire.util.ReflectionUtils.invokeMethodWithArray(ReflectionUtils.java:189)
>   at 
> org.apache.maven.surefire.booter.ProviderFactory$ProviderProxy.invoke(ProviderFactory.java:165)
>   at 
> org.apache.maven.surefire.booter.ProviderFactory.invokeProvider(ProviderFactory.java:85)
>   at 
> org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:115)
>   at 
> org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:75)
> Caused by: java.lang.NullPointerException
>   at 
> org.apache.hadoop.hdfs.server.namenode.ha.TestDNFencingWithReplication.(TestDNFencingWithReplication.java:49)
>   ... 28 more
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (HDFS-5776) Support 'hedged' reads in DFSClient

2014-01-24 Thread Arpit Agarwal (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5776?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13881305#comment-13881305
 ] 

Arpit Agarwal commented on HDFS-5776:
-

{quote}
I don't see why we wouldn't expose this setting. It doesn't give the client the 
ability to do anything bad it couldn't already do. You can already try to open 
a zillion files at once in order to attack the NameNode / DataNodes. Preventing 
denial-of-service attacks is not currently something we try to do. And in the 
future, if we ever do try to prevent denial-of-service attacks, I don't think 
having hedged reads makes that any more or less difficult than it would 
otherwise be.
{quote}
[~cmccabe] I am thinking of carelessly configured settings, not a deliberate 
dos.

> Support 'hedged' reads in DFSClient
> ---
>
> Key: HDFS-5776
> URL: https://issues.apache.org/jira/browse/HDFS-5776
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: hdfs-client
>Affects Versions: 3.0.0
>Reporter: Liang Xie
>Assignee: Liang Xie
> Attachments: HDFS-5776-v2.txt, HDFS-5776-v3.txt, HDFS-5776-v4.txt, 
> HDFS-5776-v5.txt, HDFS-5776-v6.txt, HDFS-5776-v7.txt, HDFS-5776-v8.txt, 
> HDFS-5776.txt
>
>
> This is a placeholder of hdfs related stuff backport from 
> https://issues.apache.org/jira/browse/HBASE-7509
> The quorum read ability should be helpful especially to optimize read outliers
> we can utilize "dfs.dfsclient.quorum.read.threshold.millis" & 
> "dfs.dfsclient.quorum.read.threadpool.size" to enable/disable the hedged read 
> ability from client side(e.g. HBase), and by using DFSQuorumReadMetrics, we 
> could export the interested metric valus into client system(e.g. HBase's 
> regionserver metric).
> The core logic is in pread code path, we decide to goto the original 
> fetchBlockByteRange or the new introduced fetchBlockByteRangeSpeculative per 
> the above config items.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Created] (HDFS-5830) WebHdfsFileSystem.getFileBlockLocations throws IllegalArgumentException when accessing another cluster.

2014-01-24 Thread Yongjun Zhang (JIRA)
Yongjun Zhang created HDFS-5830:
---

 Summary: WebHdfsFileSystem.getFileBlockLocations throws 
IllegalArgumentException when accessing another cluster. 
 Key: HDFS-5830
 URL: https://issues.apache.org/jira/browse/HDFS-5830
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: caching, hdfs-client
Affects Versions: 2.3.0
Reporter: Yongjun Zhang
Assignee: Yongjun Zhang


WebHdfsFileSystem.getFileBlockLocations throws IllegalArgumentException when 
accessing a another cluster (that doesn't have caching support). 

java.lang.IllegalArgumentException: cachedLocs should not be null, use a 
different constructor
at com.google.common.base.Preconditions.checkArgument(Preconditions.java:88)
at org.apache.hadoop.hdfs.protocol.LocatedBlock.(LocatedBlock.java:79)
at org.apache.hadoop.hdfs.web.JsonUtil.toLocatedBlock(JsonUtil.java:414)
at org.apache.hadoop.hdfs.web.JsonUtil.toLocatedBlockList(JsonUtil.java:446)
at org.apache.hadoop.hdfs.web.JsonUtil.toLocatedBlocks(JsonUtil.java:479)
at 
org.apache.hadoop.hdfs.web.WebHdfsFileSystem.getFileBlockLocations(WebHdfsFileSystem.java:1067)
at org.apache.hadoop.fs.FileSystem$4.next(FileSystem.java:1812)
at org.apache.hadoop.fs.FileSystem$4.next(FileSystem.java:1797)




--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (HDFS-5771) Track progress when loading fsimage

2014-01-24 Thread Haohui Mai (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5771?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Haohui Mai updated HDFS-5771:
-

Attachment: HDFS-5771.001.patch

Rebased

> Track progress when loading fsimage
> ---
>
> Key: HDFS-5771
> URL: https://issues.apache.org/jira/browse/HDFS-5771
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>Affects Versions: HDFS-5698 (FSImage in protobuf)
>Reporter: Haohui Mai
>Assignee: Haohui Mai
> Attachments: HDFS-5771.000.patch, HDFS-5771.001.patch
>
>
> The old code that loads the fsimage tracks the progress during loading. This 
> jira proposes to implement the same functionality in the new code which 
> serializes the fsimage using protobuf..



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (HDFS-5776) Support 'hedged' reads in DFSClient

2014-01-24 Thread stack (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5776?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13881281#comment-13881281
 ] 

stack commented on HDFS-5776:
-

So, to enable, we set DFS_DFSCLIENT_HEDGED_READ_THREADPOOL_SIZE in DN config.  
Should the number of threads in the hbase case be greater than 
NUMBER_OF_HBASE_OPEN_FILES (though this is most often an unknown number, one 
that is changing over the life over the hbase process, and up in the thousands 
frequently)?  Otherwise we could set it some 'sensible' number like 16 and then 
just watch the metrics this patch also adds.  If we are too often running the 
requests in the current thread because the executor has none to spare then we 
can up the number of pool threads (though it requires a DN restart, a PITA)?  
That should work for the first cut at this feature.

nit: You could declare and assign in the one go rather than postpone the assign 
to the constructor: HEDGED_READ_METRIC = new DFSHedgedReadMetrics();

What is your thinking regards the boolean enabling/disabling hedge reads in 
DFSClient [~xieliang007]?  On the one hand, there is a problem where the 
setting of pool size is done in DN config yet we have enable/disable hedge 
reads in the API; if the DN config has a pool size set to 0 then hedged reads 
are off (as was noted above), and though we may  'enable' hedge reads in the 
API, we won't be getting the behaviour we think we should be getting.  On the 
other hand, it looks like this boolean could be used 'conserving' resources 
disabling hedged reads on a per request basis though hedged reads have been 
marked globally 'on' in the DN?  Is that your thinking?  I'm inclined to agree 
with the previous reviewers that this may verge on the 'exotic'.  For the first 
cut at this feature, lets have a global on/off switch with number of threads 
being the means of constraining how much hedged reading we do?

Otherwise patch looks great to me.



> Support 'hedged' reads in DFSClient
> ---
>
> Key: HDFS-5776
> URL: https://issues.apache.org/jira/browse/HDFS-5776
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: hdfs-client
>Affects Versions: 3.0.0
>Reporter: Liang Xie
>Assignee: Liang Xie
> Attachments: HDFS-5776-v2.txt, HDFS-5776-v3.txt, HDFS-5776-v4.txt, 
> HDFS-5776-v5.txt, HDFS-5776-v6.txt, HDFS-5776-v7.txt, HDFS-5776-v8.txt, 
> HDFS-5776.txt
>
>
> This is a placeholder of hdfs related stuff backport from 
> https://issues.apache.org/jira/browse/HBASE-7509
> The quorum read ability should be helpful especially to optimize read outliers
> we can utilize "dfs.dfsclient.quorum.read.threshold.millis" & 
> "dfs.dfsclient.quorum.read.threadpool.size" to enable/disable the hedged read 
> ability from client side(e.g. HBase), and by using DFSQuorumReadMetrics, we 
> could export the interested metric valus into client system(e.g. HBase's 
> regionserver metric).
> The core logic is in pread code path, we decide to goto the original 
> fetchBlockByteRange or the new introduced fetchBlockByteRangeSpeculative per 
> the above config items.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Created] (HDFS-5829) TestDNFencingWithReplication fails on branch2

2014-01-24 Thread Mit Desai (JIRA)
Mit Desai created HDFS-5829:
---

 Summary: TestDNFencingWithReplication fails on branch2
 Key: HDFS-5829
 URL: https://issues.apache.org/jira/browse/HDFS-5829
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 2.2.0
Reporter: Mit Desai


{noformat}
Running org.apache.hadoop.hdfs.server.namenode.ha.TestDNFencingWithReplication
Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.097 sec <<< 
FAILURE!
testFencingStress(org.apache.hadoop.hdfs.server.namenode.ha.TestDNFencingWithReplication)
  Time elapsed: 6 sec  <<< ERROR!
java.lang.ExceptionInInitializerError
at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
at 
sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57)
at 
sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
at java.lang.reflect.Constructor.newInstance(Constructor.java:526)
at 
org.junit.runners.BlockJUnit4ClassRunner.createTest(BlockJUnit4ClassRunner.java:187)
at 
org.junit.runners.BlockJUnit4ClassRunner$1.runReflectiveCall(BlockJUnit4ClassRunner.java:236)
at 
org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:15)
at 
org.junit.runners.BlockJUnit4ClassRunner.methodBlock(BlockJUnit4ClassRunner.java:233)
at 
org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:68)
at 
org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:47)
at org.junit.runners.ParentRunner$3.run(ParentRunner.java:231)
at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:60)
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:229)
at org.junit.runners.ParentRunner.access$000(ParentRunner.java:50)
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:222)
at org.junit.runners.ParentRunner.run(ParentRunner.java:300)
at 
org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:252)
at 
org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:141)
at 
org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:112)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:606)
at 
org.apache.maven.surefire.util.ReflectionUtils.invokeMethodWithArray(ReflectionUtils.java:189)
at 
org.apache.maven.surefire.booter.ProviderFactory$ProviderProxy.invoke(ProviderFactory.java:165)
at 
org.apache.maven.surefire.booter.ProviderFactory.invokeProvider(ProviderFactory.java:85)
at 
org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:115)
at 
org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:75)
Caused by: java.lang.NullPointerException
at 
org.apache.hadoop.hdfs.server.namenode.ha.TestDNFencingWithReplication.(TestDNFencingWithReplication.java:49)
... 28 more
{noformat}



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Assigned] (HDFS-5827) Children are not inheriting parent's default ACLs

2014-01-24 Thread Chris Nauroth (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5827?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chris Nauroth reassigned HDFS-5827:
---

Assignee: Chris Nauroth

Assigning to myself for verification as part of working on HDFS-5616.

> Children are not inheriting parent's default ACLs
> -
>
> Key: HDFS-5827
> URL: https://issues.apache.org/jira/browse/HDFS-5827
> Project: Hadoop HDFS
>  Issue Type: Bug
>Reporter: Vinay
>Assignee: Chris Nauroth
>
> Children are not inheriting the parent's default ACLs on creation.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (HDFS-5702) FsShell Cli: Add XML based End-to-End test for getfacl and setfacl commands

2014-01-24 Thread Chris Nauroth (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5702?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13881220#comment-13881220
 ] 

Chris Nauroth commented on HDFS-5702:
-

Excellent work.  Thanks for adding these tests!  As mentioned on HDFS-5827, the 
2 tests failures are expected, because default ACL handling is not yet 
implemented in the NameNode (tracked in HDFS-5616).  I'm planning on committing 
the failing tests anyway, which is acceptable as a temporary state while we're 
isolated on a feature branch.

Just a few minor comments:
# It looks like {{TestAclCLI#fs}} isn't needed for anything, so let's remove 
that and the import of {{FileSystem}}.
# Let's add 3 more tests:
## getfacl -R
## setfacl -R
## setfacl --set (replacing the whole ACL)

After that, I think this will be ready to commit.

> FsShell Cli: Add XML based End-to-End test for getfacl and setfacl commands
> ---
>
> Key: HDFS-5702
> URL: https://issues.apache.org/jira/browse/HDFS-5702
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: hdfs-client, namenode, security
>Reporter: Vinay
>Assignee: Vinay
> Attachments: HDFS-5702.patch
>
>
> FsShell Cli: Add XML based End-to-End test for getfacl and setfacl commands



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (HDFS-5827) Children are not inheriting parent's default ACLs

2014-01-24 Thread Chris Nauroth (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5827?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13881194#comment-13881194
 ] 

Chris Nauroth commented on HDFS-5827:
-

Thanks, Vinay.  You're right that this is incorrect behavior.  This is 
happening because default ACL handling hasn't been implemented on the NameNode 
side yet.  This is tracked in HDFS-5616, and I'm planning on getting it done in 
the next few days.  Until then, this particular test won't pass.

Do you mind if we resolve this as duplicate?  For HDFS-5702, I think we can 
just commit the failing tests, and then I'll verify later that my HDFS-5616 
patch fixes them.  We're isolated on a feature branch, so having a couple of 
failing tests for a few days won't harm anyone else.  Sound good?

> Children are not inheriting parent's default ACLs
> -
>
> Key: HDFS-5827
> URL: https://issues.apache.org/jira/browse/HDFS-5827
> Project: Hadoop HDFS
>  Issue Type: Bug
>Reporter: Vinay
>
> Children are not inheriting the parent's default ACLs on creation.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Assigned] (HDFS-5616) NameNode: implement default ACL handling.

2014-01-24 Thread Chris Nauroth (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5616?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chris Nauroth reassigned HDFS-5616:
---

Assignee: Chris Nauroth

> NameNode: implement default ACL handling.
> -
>
> Key: HDFS-5616
> URL: https://issues.apache.org/jira/browse/HDFS-5616
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: namenode
>Affects Versions: HDFS ACLs (HDFS-4685)
>Reporter: Chris Nauroth
>Assignee: Chris Nauroth
>
> Implement and test handling of default ACLs within NameNode.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Created] (HDFS-5828) BlockPlacementPolicyWithNodeGroup can place multiple replicas on the same node group when dfs.namenode.avoid.write.stale.datanode is true

2014-01-24 Thread Buddy (JIRA)
Buddy created HDFS-5828:
---

 Summary: BlockPlacementPolicyWithNodeGroup can place multiple 
replicas on the same node group when dfs.namenode.avoid.write.stale.datanode is 
true
 Key: HDFS-5828
 URL: https://issues.apache.org/jira/browse/HDFS-5828
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: namenode
Affects Versions: 2.4.0
Reporter: Buddy


When placing replicas using the replica placement policy 
BlockPlacementPolicyWithNodeGroup, the number of targets returned should be 
less than or equal to the number of node groups and no node group should get 
two replicas of the same block. The Junit test 
TestReplicationPolicyWithNodeGroup.testChooseMoreTargetsThanNodeGroups verifies 
this.

However, if the conf property "dfs.namenode.avoid.write.stale.datanode" is set 
to true, then block placement policy will return more targets than node groups 
when the number of replicas requested exceeds the number of node groups.

This can be seen by putting:
   CONF.setBoolean(DFS_NAMENODE_AVOID_STALE_DATANODE_FOR_WRITE_KEY, true);

in the setup method for TestReplicationPolicyWithNodeGroup. This will cause 
testChooseMoreTargetsThanNodeGroups to fail.




--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (HDFS-5343) When cat command is issued on snapshot files getting unexpected result

2014-01-24 Thread Uma Maheswara Rao G (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5343?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13880975#comment-13880975
 ] 

Uma Maheswara Rao G commented on HDFS-5343:
---

Seems like jenkins url not accessible. Any one knows, whether jenkins down?

> When cat command is issued on snapshot files getting unexpected result
> --
>
> Key: HDFS-5343
> URL: https://issues.apache.org/jira/browse/HDFS-5343
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: hdfs-client
>Reporter: sathish
>Assignee: sathish
> Attachments: HDFS-5343-0003.patch, HDFS-5343-0004.patch, 
> HDFS-5343-0005.patch, HDFS-5343-0006.patch, HDFS-5343-002.patch
>
>
> first if we create one file with some file length and take the snapshot of 
> that file,and again append some data through append method to that file,then 
> if we do cat command operation on snapshot of that file,in general it should 
> dispaly the data what we added with create operation,but it is displaying the 
> total data i.e. create +_ appended data.
> but if we do the same operation and if we read the contents of snapshot file 
> through input stream it is just displaying the data created in snapshoted 
> files.
> in this the behaviour of cat command and reading through inputstream is 
> getting different



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (HDFS-5702) FsShell Cli: Add XML based End-to-End test for getfacl and setfacl commands

2014-01-24 Thread Vinay (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5702?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vinay updated HDFS-5702:


Attachment: HDFS-5702.patch

Attaching the patch with some of the initial tests. Please review
Among added tests last 2 tests will fail due to  HDFS-5827.

Please suggest if some more tests needs to be added.

> FsShell Cli: Add XML based End-to-End test for getfacl and setfacl commands
> ---
>
> Key: HDFS-5702
> URL: https://issues.apache.org/jira/browse/HDFS-5702
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: hdfs-client, namenode, security
>Reporter: Vinay
>Assignee: Vinay
> Attachments: HDFS-5702.patch
>
>
> FsShell Cli: Add XML based End-to-End test for getfacl and setfacl commands



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (HDFS-5827) Children are not inheriting parent's default ACLs

2014-01-24 Thread Vinay (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5827?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13880942#comment-13880942
 ] 

Vinay commented on HDFS-5827:
-

Hi [~cnauroth], Could you take a look at this.
This I found while writing xml tests. 
Please correct me if I am wrong.

> Children are not inheriting parent's default ACLs
> -
>
> Key: HDFS-5827
> URL: https://issues.apache.org/jira/browse/HDFS-5827
> Project: Hadoop HDFS
>  Issue Type: Bug
>Reporter: Vinay
>
> Children are not inheriting the parent's default ACLs on creation.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (HDFS-5827) Children are not inheriting parent's default ACLs

2014-01-24 Thread Vinay (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5827?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vinay updated HDFS-5827:


Description: 
Children are not inheriting the parent's default ACLs on creation.


  was:
Children are not inheriting the parent's default ACLs on creation.

Following is the ACLs on parent and child in HDFS
{noformat}# file: /dir1
# owner: vinay
# group: supergroup
user::rwx
mask::r-x
other::r-x
default:user::rwx
default:user:charlie:r-x
default:group::r-x
default:group:admin:rwx
default:mask::rwx
default:other::r-x

# file: /dir1/dir2
# owner: vinay
# group: supergroup
user::rwx
group::r-x
other::r-x

# file: /dir1/file
# owner: vinay
# group: supergroup
user::rw-
group::r--
other::r--{noformat}


Following is the output in linux ACL
# file: testAcl
# owner: vinay
# group: users
user::rwx
user:vinay:r--
group::r-x
group:users:r-x
mask::r-x
other::r-x
default:user::rwx
default:user:vinay:r-x
default:group::r-x
default:group:users:rwx
default:mask::rwx
default:other::r-x

# file: testAcl/hello
# owner: vinay
# group: users
user::rw-
user:vinay:r-x  #effective:r--
group::r-x  #effective:r--
group:users:rwx #effective:rw-
mask::rw-
other::r--
{noformat}


> Children are not inheriting parent's default ACLs
> -
>
> Key: HDFS-5827
> URL: https://issues.apache.org/jira/browse/HDFS-5827
> Project: Hadoop HDFS
>  Issue Type: Bug
>Reporter: Vinay
>
> Children are not inheriting the parent's default ACLs on creation.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (HDFS-5827) Children are not inheriting parent's default ACLs

2014-01-24 Thread Vinay (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5827?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13880940#comment-13880940
 ] 

Vinay commented on HDFS-5827:
-

Following is the ACLs on parent and child in HDFS
{noformat}# file: /dir1
# owner: vinay
# group: supergroup
user::rwx
mask::r-x
other::r-x
default:user::rwx
default:user:charlie:r-x
default:group::r-x
default:group:admin:rwx
default:mask::rwx
default:other::r-x

# file: /dir1/dir2
# owner: vinay
# group: supergroup
user::rwx
group::r-x
other::r-x

# file: /dir1/file
# owner: vinay
# group: supergroup
user::rw-
group::r--
other::r--{noformat}


Following is the output in linux ACL
{noformat}
# file: testAcl
# owner: vinay
# group: users
user::rwx
user:vinay:r--
group::r-x
group:users:r-x
mask::r-x
other::r-x
default:user::rwx
default:user:vinay:r-x
default:group::r-x
default:group:users:rwx
default:mask::rwx
default:other::r-x

# file: testAcl/hello
# owner: vinay
# group: users
user::rw-
user:vinay:r-x  #effective:r--
group::r-x  #effective:r--
group:users:rwx #effective:rw-
mask::rw-
other::r--
{noformat}

> Children are not inheriting parent's default ACLs
> -
>
> Key: HDFS-5827
> URL: https://issues.apache.org/jira/browse/HDFS-5827
> Project: Hadoop HDFS
>  Issue Type: Bug
>Reporter: Vinay
>
> Children are not inheriting the parent's default ACLs on creation.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Created] (HDFS-5827) Children are not inheriting parent's default ACLs

2014-01-24 Thread Vinay (JIRA)
Vinay created HDFS-5827:
---

 Summary: Children are not inheriting parent's default ACLs
 Key: HDFS-5827
 URL: https://issues.apache.org/jira/browse/HDFS-5827
 Project: Hadoop HDFS
  Issue Type: Bug
Reporter: Vinay


Children are not inheriting the parent's default ACLs on creation.

Following is the ACLs on parent and child in HDFS
{noformat}# file: /dir1
# owner: vinay
# group: supergroup
user::rwx
mask::r-x
other::r-x
default:user::rwx
default:user:charlie:r-x
default:group::r-x
default:group:admin:rwx
default:mask::rwx
default:other::r-x

# file: /dir1/dir2
# owner: vinay
# group: supergroup
user::rwx
group::r-x
other::r-x

# file: /dir1/file
# owner: vinay
# group: supergroup
user::rw-
group::r--
other::r--{noformat}


Following is the output in linux ACL
# file: testAcl
# owner: vinay
# group: users
user::rwx
user:vinay:r--
group::r-x
group:users:r-x
mask::r-x
other::r-x
default:user::rwx
default:user:vinay:r-x
default:group::r-x
default:group:users:rwx
default:mask::rwx
default:other::r-x

# file: testAcl/hello
# owner: vinay
# group: users
user::rw-
user:vinay:r-x  #effective:r--
group::r-x  #effective:r--
group:users:rwx #effective:rw-
mask::rw-
other::r--
{noformat}



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (HDFS-5343) When cat command is issued on snapshot files getting unexpected result

2014-01-24 Thread Uma Maheswara Rao G (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5343?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13880893#comment-13880893
 ] 

Uma Maheswara Rao G commented on HDFS-5343:
---

+1, Patch looks good.   Pending jenkins report.

> When cat command is issued on snapshot files getting unexpected result
> --
>
> Key: HDFS-5343
> URL: https://issues.apache.org/jira/browse/HDFS-5343
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: hdfs-client
>Reporter: sathish
>Assignee: sathish
> Attachments: HDFS-5343-0003.patch, HDFS-5343-0004.patch, 
> HDFS-5343-0005.patch, HDFS-5343-0006.patch, HDFS-5343-002.patch
>
>
> first if we create one file with some file length and take the snapshot of 
> that file,and again append some data through append method to that file,then 
> if we do cat command operation on snapshot of that file,in general it should 
> dispaly the data what we added with create operation,but it is displaying the 
> total data i.e. create +_ appended data.
> but if we do the same operation and if we read the contents of snapshot file 
> through input stream it is just displaying the data created in snapshoted 
> files.
> in this the behaviour of cat command and reading through inputstream is 
> getting different



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (HDFS-5343) When cat command is issued on snapshot files getting unexpected result

2014-01-24 Thread Uma Maheswara Rao G (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5343?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Uma Maheswara Rao G updated HDFS-5343:
--

Status: Patch Available  (was: Open)

> When cat command is issued on snapshot files getting unexpected result
> --
>
> Key: HDFS-5343
> URL: https://issues.apache.org/jira/browse/HDFS-5343
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: hdfs-client
>Reporter: sathish
>Assignee: sathish
> Attachments: HDFS-5343-0003.patch, HDFS-5343-0004.patch, 
> HDFS-5343-0005.patch, HDFS-5343-0006.patch, HDFS-5343-002.patch
>
>
> first if we create one file with some file length and take the snapshot of 
> that file,and again append some data through append method to that file,then 
> if we do cat command operation on snapshot of that file,in general it should 
> dispaly the data what we added with create operation,but it is displaying the 
> total data i.e. create +_ appended data.
> but if we do the same operation and if we read the contents of snapshot file 
> through input stream it is just displaying the data created in snapshoted 
> files.
> in this the behaviour of cat command and reading through inputstream is 
> getting different



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (HDFS-5343) When cat command is issued on snapshot files getting unexpected result

2014-01-24 Thread sathish (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5343?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

sathish updated HDFS-5343:
--

Attachment: HDFS-5343-0006.patch

I updated the patch as per your comments,please review the patch

> When cat command is issued on snapshot files getting unexpected result
> --
>
> Key: HDFS-5343
> URL: https://issues.apache.org/jira/browse/HDFS-5343
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: hdfs-client
>Reporter: sathish
>Assignee: sathish
> Attachments: HDFS-5343-0003.patch, HDFS-5343-0004.patch, 
> HDFS-5343-0005.patch, HDFS-5343-0006.patch, HDFS-5343-002.patch
>
>
> first if we create one file with some file length and take the snapshot of 
> that file,and again append some data through append method to that file,then 
> if we do cat command operation on snapshot of that file,in general it should 
> dispaly the data what we added with create operation,but it is displaying the 
> total data i.e. create +_ appended data.
> but if we do the same operation and if we read the contents of snapshot file 
> through input stream it is just displaying the data created in snapshoted 
> files.
> in this the behaviour of cat command and reading through inputstream is 
> getting different



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (HDFS-5343) When cat command is issued on snapshot files getting unexpected result

2014-01-24 Thread Uma Maheswara Rao G (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5343?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13880862#comment-13880862
 ] 

Uma Maheswara Rao G commented on HDFS-5343:
---

{code}
ToolRunner.run(conf, shell, new String[] { "-cat",
+"/TestSnapshotFileLength/sub1/.snapshot/snapshot1/file1" });
+assertEquals(bao.size(), BLOCKSIZE);
+System.setOut(psBackup);
{code}

here System.setOut must be in try finally. Keep ToolRunner execution in try and 
keep System.setOut in finally. Previous comment said about unnecessary catching 
of exception.

> When cat command is issued on snapshot files getting unexpected result
> --
>
> Key: HDFS-5343
> URL: https://issues.apache.org/jira/browse/HDFS-5343
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: hdfs-client
>Reporter: sathish
>Assignee: sathish
> Attachments: HDFS-5343-0003.patch, HDFS-5343-0004.patch, 
> HDFS-5343-0005.patch, HDFS-5343-002.patch
>
>
> first if we create one file with some file length and take the snapshot of 
> that file,and again append some data through append method to that file,then 
> if we do cat command operation on snapshot of that file,in general it should 
> dispaly the data what we added with create operation,but it is displaying the 
> total data i.e. create +_ appended data.
> but if we do the same operation and if we read the contents of snapshot file 
> through input stream it is just displaying the data created in snapshoted 
> files.
> in this the behaviour of cat command and reading through inputstream is 
> getting different



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (HDFS-5343) When cat command is issued on snapshot files getting unexpected result

2014-01-24 Thread sathish (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5343?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

sathish updated HDFS-5343:
--

Attachment: HDFS-5343-0005.patch

please review the patch

> When cat command is issued on snapshot files getting unexpected result
> --
>
> Key: HDFS-5343
> URL: https://issues.apache.org/jira/browse/HDFS-5343
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: hdfs-client
>Reporter: sathish
>Assignee: sathish
> Attachments: HDFS-5343-0003.patch, HDFS-5343-0004.patch, 
> HDFS-5343-0005.patch, HDFS-5343-002.patch
>
>
> first if we create one file with some file length and take the snapshot of 
> that file,and again append some data through append method to that file,then 
> if we do cat command operation on snapshot of that file,in general it should 
> dispaly the data what we added with create operation,but it is displaying the 
> total data i.e. create +_ appended data.
> but if we do the same operation and if we read the contents of snapshot file 
> through input stream it is just displaying the data created in snapshoted 
> files.
> in this the behaviour of cat command and reading through inputstream is 
> getting different



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (HDFS-5343) When cat command is issued on snapshot files getting unexpected result

2014-01-24 Thread sathish (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5343?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13880859#comment-13880859
 ] 

sathish commented on HDFS-5343:
---

yes correct ,no need to catch the exception here,in the previous patch itself i 
updated that comment
but unfortunately it came in latest patch.
I will update my patch.

> When cat command is issued on snapshot files getting unexpected result
> --
>
> Key: HDFS-5343
> URL: https://issues.apache.org/jira/browse/HDFS-5343
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: hdfs-client
>Reporter: sathish
>Assignee: sathish
> Attachments: HDFS-5343-0003.patch, HDFS-5343-0004.patch, 
> HDFS-5343-002.patch
>
>
> first if we create one file with some file length and take the snapshot of 
> that file,and again append some data through append method to that file,then 
> if we do cat command operation on snapshot of that file,in general it should 
> dispaly the data what we added with create operation,but it is displaying the 
> total data i.e. create +_ appended data.
> but if we do the same operation and if we read the contents of snapshot file 
> through input stream it is just displaying the data created in snapshoted 
> files.
> in this the behaviour of cat command and reading through inputstream is 
> getting different



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (HDFS-5343) When cat command is issued on snapshot files getting unexpected result

2014-01-24 Thread Uma Maheswara Rao G (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5343?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13880849#comment-13880849
 ] 

Uma Maheswara Rao G commented on HDFS-5343:
---

Thanks for the update.
Taking from my previous comment:
{code}
try {
+ToolRunner.run(conf, shell, new String[] { "-cat",
+"/TestSnapshotFileLength/sub1/.snapshot/snapshot1/file1" });
+} catch (Exception e) {
+e.printStackTrace();
+}
{code}
I think you need not catch exception here right?

Could you please answer above question?

> When cat command is issued on snapshot files getting unexpected result
> --
>
> Key: HDFS-5343
> URL: https://issues.apache.org/jira/browse/HDFS-5343
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: hdfs-client
>Reporter: sathish
>Assignee: sathish
> Attachments: HDFS-5343-0003.patch, HDFS-5343-0004.patch, 
> HDFS-5343-002.patch
>
>
> first if we create one file with some file length and take the snapshot of 
> that file,and again append some data through append method to that file,then 
> if we do cat command operation on snapshot of that file,in general it should 
> dispaly the data what we added with create operation,but it is displaying the 
> total data i.e. create +_ appended data.
> but if we do the same operation and if we read the contents of snapshot file 
> through input stream it is just displaying the data created in snapshoted 
> files.
> in this the behaviour of cat command and reading through inputstream is 
> getting different



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (HDFS-5343) When cat command is issued on snapshot files getting unexpected result

2014-01-24 Thread sathish (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5343?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13880830#comment-13880830
 ] 

sathish commented on HDFS-5343:
---

Thanks UMA for reviewing the patch.
I updated the patch as per your comments.please review the patch

> When cat command is issued on snapshot files getting unexpected result
> --
>
> Key: HDFS-5343
> URL: https://issues.apache.org/jira/browse/HDFS-5343
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: hdfs-client
>Reporter: sathish
>Assignee: sathish
> Attachments: HDFS-5343-0003.patch, HDFS-5343-0004.patch, 
> HDFS-5343-002.patch
>
>
> first if we create one file with some file length and take the snapshot of 
> that file,and again append some data through append method to that file,then 
> if we do cat command operation on snapshot of that file,in general it should 
> dispaly the data what we added with create operation,but it is displaying the 
> total data i.e. create +_ appended data.
> but if we do the same operation and if we read the contents of snapshot file 
> through input stream it is just displaying the data created in snapshoted 
> files.
> in this the behaviour of cat command and reading through inputstream is 
> getting different



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (HDFS-5343) When cat command is issued on snapshot files getting unexpected result

2014-01-24 Thread sathish (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5343?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

sathish updated HDFS-5343:
--

Attachment: HDFS-5343-0004.patch

> When cat command is issued on snapshot files getting unexpected result
> --
>
> Key: HDFS-5343
> URL: https://issues.apache.org/jira/browse/HDFS-5343
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: hdfs-client
>Reporter: sathish
>Assignee: sathish
> Attachments: HDFS-5343-0003.patch, HDFS-5343-0004.patch, 
> HDFS-5343-002.patch
>
>
> first if we create one file with some file length and take the snapshot of 
> that file,and again append some data through append method to that file,then 
> if we do cat command operation on snapshot of that file,in general it should 
> dispaly the data what we added with create operation,but it is displaying the 
> total data i.e. create +_ appended data.
> but if we do the same operation and if we read the contents of snapshot file 
> through input stream it is just displaying the data created in snapshoted 
> files.
> in this the behaviour of cat command and reading through inputstream is 
> getting different



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (HDFS-5588) Incorrect values of trash configuration and trash emptier state in namenode log

2014-01-24 Thread Suresh Srinivas (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5588?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Suresh Srinivas updated HDFS-5588:
--

  Resolution: Duplicate
Release Note: This has already been fixed by HDFS-5560
  Status: Resolved  (was: Patch Available)

> Incorrect values of trash configuration and trash emptier state in namenode 
> log
> ---
>
> Key: HDFS-5588
> URL: https://issues.apache.org/jira/browse/HDFS-5588
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: hdfs-client, namenode
>Affects Versions: 2.2.0
>Reporter: Adam Kawa
> Attachments: HDFS-5588-1.patch
>
>
> Values of trash configuration and trash emptier state in namenode log are 
> displayed in milliseconds (but it should be seconds). This is very confusing 
> because it makes you feel that Trash does not remove data or the meaning of 
> configuration settings changed.
> {code}
> // org.apache.hadoop.fs.TrashPolicyDefault
>   @Override
>   public void initialize(Configuration conf, FileSystem fs, Path home) {
> ...
> this.deletionInterval = (long)(conf.getFloat(
> FS_TRASH_INTERVAL_KEY, FS_TRASH_INTERVAL_DEFAULT)
> * MSECS_PER_MINUTE);
> this.emptierInterval = (long)(conf.getFloat(
> FS_TRASH_CHECKPOINT_INTERVAL_KEY, 
> FS_TRASH_CHECKPOINT_INTERVAL_DEFAULT)
> * MSECS_PER_MINUTE);
> LOG.info("Namenode trash configuration: Deletion interval = " +
>  this.deletionInterval + " minutes, Emptier interval = " +
>  this.emptierInterval + " minutes.");
>}
> {code}
> this.deletionInterval and this.emptierInterval are in miliseconds, but the 
> LOG.info tells that they are in minutes.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)