[jira] [Updated] (HDFS-4066) Fix compile warnings for libwebhdfs

2014-11-30 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4066?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HDFS-4066:

Fix Version/s: (was: 2.6.0)
   2.7.0

> Fix compile warnings for libwebhdfs
> ---
>
> Key: HDFS-4066
> URL: https://issues.apache.org/jira/browse/HDFS-4066
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: webhdfs
>Affects Versions: 3.0.0
>Reporter: Jing Zhao
>Assignee: Jing Zhao
> Fix For: 2.7.0
>
>
> The compile of libwebhdfs still generates a bunch of warnings, which need to 
> be fixed.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-7067) ClassCastException while using a key created by keytool to create encryption zone.

2014-11-30 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-7067?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HDFS-7067:

Fix Version/s: (was: 2.6.0)
   2.7.0

> ClassCastException while using a key created by keytool to create encryption 
> zone. 
> ---
>
> Key: HDFS-7067
> URL: https://issues.apache.org/jira/browse/HDFS-7067
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: encryption
>Affects Versions: 2.6.0
>Reporter: Yi Yao
>Priority: Minor
> Fix For: 2.7.0
>
>
> I'm using transparent encryption. If I create a key for KMS keystore via 
> keytool and use the key to create an encryption zone. I get a 
> ClassCastException rather than an exception with decent error message. I know 
> we should use 'hadoop key create' to create a key. It's better to provide an 
> decent error message to remind user to use the right way to create a KMS key.
> [LOG]
> ERROR[user=hdfs] Method:'GET' Exception:'java.lang.ClassCastException: 
> javax.crypto.spec.SecretKeySpec cannot be cast to 
> org.apache.hadoop.crypto.key.JavaKeyStoreProvider$KeyMetadata'



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-5687) Problem in accessing NN JSP page

2014-11-30 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5687?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HDFS-5687:

Fix Version/s: (was: 2.6.0)
   2.7.0

> Problem in accessing NN JSP page
> 
>
> Key: HDFS-5687
> URL: https://issues.apache.org/jira/browse/HDFS-5687
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: hdfs-client
>Affects Versions: 2.3.0
>Reporter: sathish
>Assignee: sathish
>Priority: Minor
> Fix For: 2.7.0
>
> Attachments: HDFS-5687-0001.patch
>
>
> In NN UI page After clicking the browse File System page,from that page,if 
> you click GO Back TO DFS HOME ICon it is not accessing the dfshealth.jsp page
> NN http URL is http://nnaddr///nninfoaddr/dfshealth.jsp,it is coming like 
> this,due to this i think it is not browsing that page
> It should be http://nninfoaddr/dfshealth.jsp/ like this



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-6694) TestPipelinesFailover.testPipelineRecoveryStress tests fail intermittently with various symptoms

2014-11-30 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6694?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HDFS-6694:

Fix Version/s: (was: 2.6.0)
   2.7.0

> TestPipelinesFailover.testPipelineRecoveryStress tests fail intermittently 
> with various symptoms
> 
>
> Key: HDFS-6694
> URL: https://issues.apache.org/jira/browse/HDFS-6694
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 3.0.0
>Reporter: Yongjun Zhang
>Assignee: Yongjun Zhang
>Priority: Critical
> Fix For: 2.7.0
>
> Attachments: HDFS-6694.001.dbg.patch, HDFS-6694.001.dbg.patch, 
> HDFS-6694.001.dbg.patch, HDFS-6694.002.dbg.patch, 
> org.apache.hadoop.hdfs.server.namenode.ha.TestPipelinesFailover-output.txt, 
> org.apache.hadoop.hdfs.server.namenode.ha.TestPipelinesFailover-output.txt, 
> org.apache.hadoop.hdfs.server.namenode.ha.TestPipelinesFailover.txt
>
>
> TestPipelinesFailover.testPipelineRecoveryStress tests fail intermittently 
> with various symptoms. Typical failures are described in first comment.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-6940) Initial refactoring to allow ConsensusNode implementation

2014-11-30 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6940?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HDFS-6940:

Fix Version/s: (was: 2.6.0)
   2.7.0

> Initial refactoring to allow ConsensusNode implementation
> -
>
> Key: HDFS-6940
> URL: https://issues.apache.org/jira/browse/HDFS-6940
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: namenode
>Affects Versions: 2.0.6-alpha, 2.5.0
>Reporter: Konstantin Shvachko
>Assignee: Konstantin Shvachko
> Fix For: 2.7.0
>
> Attachments: HDFS-6940.patch
>
>
> Minor refactoring of FSNamesystem to open private methods that are needed for 
> CNode implementation.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-7273) Add start and stop wrapper scripts for mover

2014-11-30 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-7273?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HDFS-7273:

Fix Version/s: (was: 2.6.0)
   2.7.0

> Add start and stop wrapper scripts for mover
> 
>
> Key: HDFS-7273
> URL: https://issues.apache.org/jira/browse/HDFS-7273
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: balancer & mover
>Affects Versions: 2.6.0
>Reporter: Benoy Antony
>Assignee: Benoy Antony
>Priority: Minor
> Fix For: 2.7.0
>
>
> Similar to balancer , we need start/stop mover scripts to run data migration 
> tool as a daemon.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-6300) Allows to run multiple balancer simultaneously

2014-11-30 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6300?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HDFS-6300:

Fix Version/s: (was: 2.6.0)
   2.7.0

> Allows to run multiple balancer simultaneously
> --
>
> Key: HDFS-6300
> URL: https://issues.apache.org/jira/browse/HDFS-6300
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: balancer & mover
>Reporter: Rakesh R
>Assignee: Rakesh R
> Fix For: 2.7.0
>
> Attachments: HDFS-6300.patch
>
>
> Javadoc of "Balancer.java" says, it will not allow to run second balancer if 
> the first one is in progress. But I've noticed multiple can run together and 
> balancer.id implementation is not safe guarding.
> {code}
>  * Another balancer is running. Exiting...
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-3907) Allow multiple users for local block readers

2014-11-30 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3907?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HDFS-3907:

Fix Version/s: (was: 2.6.0)
   2.7.0

> Allow multiple users for local block readers
> 
>
> Key: HDFS-3907
> URL: https://issues.apache.org/jira/browse/HDFS-3907
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: datanode
>Affects Versions: 1.0.0, 2.0.0-alpha
>Reporter: Eli Collins
>Assignee: Eli Collins
> Fix For: 2.7.0
>
> Attachments: hdfs-3907.txt
>
>
> The {{dfs.block.local-path-access.user}} config added in HDFS-2246 only 
> supports a single user, however as long as blocks are group readable by more 
> than one user the feature could be used by multiple users, to support this we 
> just need to allow both to be configured. In practice this allows us to also 
> support HBase where the client (RS) runs as the hbase system user and the DN 
> runs as hdfs system user. I think this should work secure as well since we're 
> not using impersonation in the HBase case.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-5160) MiniDFSCluster webui does not work

2014-11-30 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5160?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HDFS-5160:

Fix Version/s: (was: 2.6.0)
   2.7.0

> MiniDFSCluster webui does not work
> --
>
> Key: HDFS-5160
> URL: https://issues.apache.org/jira/browse/HDFS-5160
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: test
>Affects Versions: 2.1.0-beta
>Reporter: Alejandro Abdelnur
> Fix For: 2.7.0
>
>
> The webui does not work, when going to http://localhost:50070 you get:
> {code}
> Directory: /
> webapps/  102 bytes   Sep 4, 2013 9:32:55 AM
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-3761) Uses of AuthenticatedURL should set the configured timeouts

2014-11-30 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3761?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HDFS-3761:

Fix Version/s: (was: 2.6.0)
   2.7.0

> Uses of AuthenticatedURL should set the configured timeouts
> ---
>
> Key: HDFS-3761
> URL: https://issues.apache.org/jira/browse/HDFS-3761
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 2.0.0-alpha, 2.0.2-alpha
>Reporter: Alejandro Abdelnur
>Priority: Critical
> Fix For: 2.7.0
>
>
> Similar to {{URLUtils}}, the {{AuthenticatedURL}} should set the configured 
> timeouts in URL connection instances. HADOOP-8644 is introducing a 
> {{ConnectionConfigurator}} interface to the {{AuthenticatedURL}} class, it 
> should be done via that interface.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-6833) DirectoryScanner should not register a deleting block with memory of DataNode

2014-11-18 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6833?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HDFS-6833:

Target Version/s: 2.6.1

> DirectoryScanner should not register a deleting block with memory of DataNode
> -
>
> Key: HDFS-6833
> URL: https://issues.apache.org/jira/browse/HDFS-6833
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: datanode
>Affects Versions: 3.0.0, 2.5.0, 2.5.1
>Reporter: Shinichi Yamashita
>Assignee: Shinichi Yamashita
>Priority: Critical
> Attachments: HDFS-6833-6-2.patch, HDFS-6833-6-3.patch, 
> HDFS-6833-6.patch, HDFS-6833-7-2.patch, HDFS-6833-7.patch, HDFS-6833.8.patch, 
> HDFS-6833.9.patch, HDFS-6833.patch, HDFS-6833.patch, HDFS-6833.patch, 
> HDFS-6833.patch, HDFS-6833.patch
>
>
> When a block is deleted in DataNode, the following messages are usually 
> output.
> {code}
> 2014-08-07 17:53:11,606 INFO 
> org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetAsyncDiskService:
>  Scheduling blk_1073741825_1001 file 
> /hadoop/data1/dfs/data/current/BP-1887080305-172.28.0.101-1407398838872/current/finalized/subdir0/subdir0/blk_1073741825
>  for deletion
> 2014-08-07 17:53:11,617 INFO 
> org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetAsyncDiskService:
>  Deleted BP-1887080305-172.28.0.101-1407398838872 blk_1073741825_1001 file 
> /hadoop/data1/dfs/data/current/BP-1887080305-172.28.0.101-1407398838872/current/finalized/subdir0/subdir0/blk_1073741825
> {code}
> However, DirectoryScanner may be executed when DataNode deletes the block in 
> the current implementation. And the following messsages are output.
> {code}
> 2014-08-07 17:53:30,519 INFO 
> org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetAsyncDiskService:
>  Scheduling blk_1073741825_1001 file 
> /hadoop/data1/dfs/data/current/BP-1887080305-172.28.0.101-1407398838872/current/finalized/subdir0/subdir0/blk_1073741825
>  for deletion
> 2014-08-07 17:53:31,426 INFO 
> org.apache.hadoop.hdfs.server.datanode.DirectoryScanner: BlockPool 
> BP-1887080305-172.28.0.101-1407398838872 Total blocks: 1, missing metadata 
> files:0, missing block files:0, missing blocks in memory:1, mismatched 
> blocks:0
> 2014-08-07 17:53:31,426 WARN 
> org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl: Added 
> missing block to memory FinalizedReplica, blk_1073741825_1001, FINALIZED
>   getNumBytes() = 21230663
>   getBytesOnDisk()  = 21230663
>   getVisibleLength()= 21230663
>   getVolume()   = /hadoop/data1/dfs/data/current
>   getBlockFile()= 
> /hadoop/data1/dfs/data/current/BP-1887080305-172.28.0.101-1407398838872/current/finalized/subdir0/subdir0/blk_1073741825
>   unlinked  =false
> 2014-08-07 17:53:31,531 INFO 
> org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetAsyncDiskService:
>  Deleted BP-1887080305-172.28.0.101-1407398838872 blk_1073741825_1001 file 
> /hadoop/data1/dfs/data/current/BP-1887080305-172.28.0.101-1407398838872/current/finalized/subdir0/subdir0/blk_1073741825
> {code}
> Deleting block information is registered in DataNode's memory.
> And when DataNode sends a block report, NameNode receives wrong block 
> information.
> For example, when we execute recommission or change the number of 
> replication, NameNode may delete the right block as "ExcessReplicate" by this 
> problem.
> And "Under-Replicated Blocks" and "Missing Blocks" occur.
> When DataNode run DirectoryScanner, DataNode should not register a deleting 
> block.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-7274) Disable SSLv3 in HttpFS

2014-11-13 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-7274?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HDFS-7274:

Fix Version/s: 2.6.0

> Disable SSLv3 in HttpFS
> ---
>
> Key: HDFS-7274
> URL: https://issues.apache.org/jira/browse/HDFS-7274
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: webhdfs
>Affects Versions: 2.6.0
>Reporter: Robert Kanter
>Assignee: Robert Kanter
>Priority: Blocker
> Fix For: 2.6.0, 2.5.2
>
> Attachments: HDFS-7274.patch, HDFS-7274.patch
>
>
> We should disable SSLv3 in HttpFS to protect against the POODLEbleed 
> vulnerability.
> See 
> [CVE-2014-3566|http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2014-3566]
> We have {{sslProtocol="TLS"}} set to only allow TLS in ssl-server.xml, but 
> when I checked, I could still connect with SSLv3.  There documentation is 
> somewhat unclear in the tomcat configs between {{sslProtocol}}, 
> {{sslProtocols}}, and {{sslEnabledProtocols}} and what each value they take 
> does exactly.  From what I can gather, {{sslProtocol="TLS"}} actually 
> includes SSLv3 and the only way to fix this is to explicitly list which TLS 
> versions we support.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-7391) Renable SSLv2Hello in HttpFS

2014-11-12 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-7391?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HDFS-7391:

   Resolution: Fixed
Fix Version/s: 2.6.0
   Status: Resolved  (was: Patch Available)

I just committed this to branch-2.6. Thanks [~rkanter]!



[~kasha] - I wasn't sure if you wanted this in branch-2.5, so I haven't 
cherry-picked it. Please do so if you want to. Thanks.

> Renable SSLv2Hello in HttpFS
> 
>
> Key: HDFS-7391
> URL: https://issues.apache.org/jira/browse/HDFS-7391
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: webhdfs
>Affects Versions: 2.6.0, 2.5.2
>Reporter: Robert Kanter
>Assignee: Robert Kanter
>Priority: Blocker
> Fix For: 2.6.0
>
> Attachments: HDFS-7391-branch-2.5.patch, HDFS-7391.patch
>
>
> We should re-enable "SSLv2Hello", which is required for older clients (e.g. 
> Java 6 with openssl 0.9.8x) so they can't connect without it. Just to be 
> clear, it does not mean SSLv2, which is insecure.
> I couldn't simply do an addendum patch on HDFS-7274 because it's already been 
> closed.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-7391) Renable SSLv2Hello in HttpFS

2014-11-12 Thread Arun C Murthy (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-7391?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14209148#comment-14209148
 ] 

Arun C Murthy commented on HDFS-7391:
-

[~kasha] you just saved me some typing chores, I'll take care of this. Thanks!

> Renable SSLv2Hello in HttpFS
> 
>
> Key: HDFS-7391
> URL: https://issues.apache.org/jira/browse/HDFS-7391
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: webhdfs
>Affects Versions: 2.6.0, 2.5.2
>Reporter: Robert Kanter
>Assignee: Robert Kanter
>Priority: Blocker
> Attachments: HDFS-7391-branch-2.5.patch, HDFS-7391.patch
>
>
> We should re-enable "SSLv2Hello", which is required for older clients (e.g. 
> Java 6 with openssl 0.9.8x) so they can't connect without it. Just to be 
> clear, it does not mean SSLv2, which is insecure.
> I couldn't simply do an addendum patch on HDFS-7274 because it's already been 
> closed.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-7391) Renable SSLv2Hello in HttpFS

2014-11-12 Thread Arun C Murthy (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-7391?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14209030#comment-14209030
 ] 

Arun C Murthy commented on HDFS-7391:
-

Sounds good, thanks [~rkanter] and [~ywskycn]!

I'll commit both shortly for RC1.

> Renable SSLv2Hello in HttpFS
> 
>
> Key: HDFS-7391
> URL: https://issues.apache.org/jira/browse/HDFS-7391
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: webhdfs
>Affects Versions: 2.6.0, 2.5.2
>Reporter: Robert Kanter
>Assignee: Robert Kanter
>Priority: Blocker
> Attachments: HDFS-7391-branch-2.5.patch, HDFS-7391.patch
>
>
> We should re-enable "SSLv2Hello", which is required for older clients (e.g. 
> Java 6 with openssl 0.9.8x) so they can't connect without it. Just to be 
> clear, it does not mean SSLv2, which is insecure.
> I couldn't simply do an addendum patch on HDFS-7274 because it's already been 
> closed.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-7391) Renable SSLv2Hello in HttpFS

2014-11-12 Thread Arun C Murthy (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-7391?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14208818#comment-14208818
 ] 

Arun C Murthy commented on HDFS-7391:
-

[~rkanter] - Is this and the addendum patch for HADOOP-11217 the only ones for 
SSLv2Hello? I thought there were 3?

I'll commit both later today if there are no objections; others please commit 
to branch-2/branch-2.6/branch-2.6.0 if you get to it before me. Thanks.

> Renable SSLv2Hello in HttpFS
> 
>
> Key: HDFS-7391
> URL: https://issues.apache.org/jira/browse/HDFS-7391
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: webhdfs
>Affects Versions: 2.6.0, 2.5.2
>Reporter: Robert Kanter
>Assignee: Robert Kanter
>Priority: Blocker
> Attachments: HDFS-7391-branch-2.5.patch, HDFS-7391.patch
>
>
> We should re-enable "SSLv2Hello", which is required for older clients (e.g. 
> Java 6 with openssl 0.9.8x) so they can't connect without it. Just to be 
> clear, it does not mean SSLv2, which is insecure.
> I couldn't simply do an addendum patch on HDFS-7274 because it's already been 
> closed.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-7383) DataNode.requestShortCircuitFdsForRead may throw NullPointerException

2014-11-09 Thread Arun C Murthy (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-7383?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14204245#comment-14204245
 ] 

Arun C Murthy commented on HDFS-7383:
-

[~sureshms] this might be bad enough that we want it in 2.6? Thanks.

> DataNode.requestShortCircuitFdsForRead may throw NullPointerException
> -
>
> Key: HDFS-7383
> URL: https://issues.apache.org/jira/browse/HDFS-7383
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: datanode
>Reporter: Tsz Wo Nicholas Sze
>Assignee: Tsz Wo Nicholas Sze
> Fix For: 2.7.0
>
> Attachments: h7383_20141108.patch
>
>
> {noformat}
> java.lang.NullPointerException
>   at 
> org.apache.hadoop.hdfs.server.datanode.DataNode.requestShortCircuitFdsForRead(DataNode.java:1525)
>   at 
> org.apache.hadoop.hdfs.server.datanode.DataXceiver.requestShortCircuitFds(DataXceiver.java:286)
>   at 
> org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opRequestShortCircuitFds(Receiver.java:185)
>   at 
> org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:89)
>   at 
> org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:234)
>   at java.lang.Thread.run(Thread.java:745)
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-6573) TestDeadDatanode#testDeadDatanode fails if debug flag is turned on for the unit test

2014-06-30 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6573?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HDFS-6573:


Fix Version/s: (was: 2.4.1)
   2.5.0

> TestDeadDatanode#testDeadDatanode fails if debug flag is turned on for the 
> unit test
> 
>
> Key: HDFS-6573
> URL: https://issues.apache.org/jira/browse/HDFS-6573
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: test
>Affects Versions: 2.2.0, 2.3.0, 2.4.0
>Reporter: Jinghui Wang
>Assignee: Jinghui Wang
>Priority: Minor
> Fix For: 2.5.0
>
> Attachments: HDFS-6573.patch
>
>
> o.a.h.hdfs.server.namenode.NameNodeRpcServer#blockReport calls a function in 
> for debug mode, which uses the StorageBlockReport data passed in by the test 
> TestDeadDatanode#testDeadDatanode. The report block list is an long array of 
> size 3 in the test, but the blockReport expects the format of long array of 
> size 5.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6527) Edit log corruption due to defered INode removal

2014-06-20 Thread Arun C Murthy (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6527?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14039716#comment-14039716
 ] 

Arun C Murthy commented on HDFS-6527:
-

Changed fix version to 2.4.1 since it got merged in for rc1.

> Edit log corruption due to defered INode removal
> 
>
> Key: HDFS-6527
> URL: https://issues.apache.org/jira/browse/HDFS-6527
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 2.4.0
>Reporter: Kihwal Lee
>Assignee: Kihwal Lee
>Priority: Blocker
> Fix For: 2.4.1
>
> Attachments: HDFS-6527.branch-2.4.patch, HDFS-6527.trunk.patch, 
> HDFS-6527.v2.patch, HDFS-6527.v3.patch, HDFS-6527.v4.patch, HDFS-6527.v5.patch
>
>
> We have seen a SBN crashing with the following error:
> {panel}
> \[Edit log tailer\] ERROR namenode.FSEditLogLoader:
> Encountered exception on operation AddBlockOp
> [path=/xxx,
> penultimateBlock=NULL, lastBlock=blk_111_111, RpcClientId=,
> RpcCallId=-2]
> java.io.FileNotFoundException: File does not exist: /xxx
> {panel}
> This was caused by the deferred removal of deleted inodes from the inode map. 
> Since getAdditionalBlock() acquires FSN read lock and then write lock, a 
> deletion can happen in between. Because of deferred inode removal outside FSN 
> write lock, getAdditionalBlock() can get the deleted inode from the inode map 
> with FSN write lock held. This allow addition of a block to a deleted file.
> As a result, the edit log will contain OP_ADD, OP_DELETE, followed by
>  OP_ADD_BLOCK.  This cannot be replayed by NN, so NN doesn't start up or SBN 
> crashes.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6527) Edit log corruption due to defered INode removal

2014-06-20 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6527?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HDFS-6527:


Fix Version/s: (was: 2.5.0)
   2.4.1

> Edit log corruption due to defered INode removal
> 
>
> Key: HDFS-6527
> URL: https://issues.apache.org/jira/browse/HDFS-6527
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 2.4.0
>Reporter: Kihwal Lee
>Assignee: Kihwal Lee
>Priority: Blocker
> Fix For: 2.4.1
>
> Attachments: HDFS-6527.branch-2.4.patch, HDFS-6527.trunk.patch, 
> HDFS-6527.v2.patch, HDFS-6527.v3.patch, HDFS-6527.v4.patch, HDFS-6527.v5.patch
>
>
> We have seen a SBN crashing with the following error:
> {panel}
> \[Edit log tailer\] ERROR namenode.FSEditLogLoader:
> Encountered exception on operation AddBlockOp
> [path=/xxx,
> penultimateBlock=NULL, lastBlock=blk_111_111, RpcClientId=,
> RpcCallId=-2]
> java.io.FileNotFoundException: File does not exist: /xxx
> {panel}
> This was caused by the deferred removal of deleted inodes from the inode map. 
> Since getAdditionalBlock() acquires FSN read lock and then write lock, a 
> deletion can happen in between. Because of deferred inode removal outside FSN 
> write lock, getAdditionalBlock() can get the deleted inode from the inode map 
> with FSN write lock held. This allow addition of a block to a deleted file.
> As a result, the edit log will contain OP_ADD, OP_DELETE, followed by
>  OP_ADD_BLOCK.  This cannot be replayed by NN, so NN doesn't start up or SBN 
> crashes.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-3761) Uses of AuthenticatedURL should set the configured timeouts

2014-04-10 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3761?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HDFS-3761:


Fix Version/s: (was: 2.4.0)
   2.5.0

> Uses of AuthenticatedURL should set the configured timeouts
> ---
>
> Key: HDFS-3761
> URL: https://issues.apache.org/jira/browse/HDFS-3761
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 2.0.0-alpha, 2.0.2-alpha
>Reporter: Alejandro Abdelnur
>Priority: Critical
> Fix For: 2.5.0
>
>
> Similar to {{URLUtils}}, the {{AuthenticatedURL}} should set the configured 
> timeouts in URL connection instances. HADOOP-8644 is introducing a 
> {{ConnectionConfigurator}} interface to the {{AuthenticatedURL}} class, it 
> should be done via that interface.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-3828) Block Scanner rescans blocks too frequently

2014-04-10 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3828?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HDFS-3828:


Fix Version/s: (was: 2.4.0)
   2.5.0

> Block Scanner rescans blocks too frequently
> ---
>
> Key: HDFS-3828
> URL: https://issues.apache.org/jira/browse/HDFS-3828
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 0.23.0
>Reporter: Andy Isaacson
>Assignee: Andy Isaacson
> Fix For: 2.5.0
>
> Attachments: hdfs-3828-1.txt, hdfs-3828-2.txt, hdfs-3828-3.txt, 
> hdfs3828.txt
>
>
> {{BlockPoolSliceScanner#scan}} calls cleanUp every time it's invoked from 
> {{DataBlockScanner#run}} via {{scanBlockPoolSlice}}.  But cleanUp 
> unconditionally roll()s the verificationLogs, so after two iterations we have 
> lost the first iteration of block verification times.  As a result a cluster 
> with just one block repeatedly rescans it every 10 seconds:
> {noformat}
> 2012-08-16 15:59:57,884 INFO  datanode.BlockPoolSliceScanner 
> (BlockPoolSliceScanner.java:verifyBlock(391)) - Verification succeeded for 
> BP-2101131164-172.29.122.91-1337906886255:blk_7919273167187535506_4915
> 2012-08-16 16:00:07,904 INFO  datanode.BlockPoolSliceScanner 
> (BlockPoolSliceScanner.java:verifyBlock(391)) - Verification succeeded for 
> BP-2101131164-172.29.122.91-1337906886255:blk_7919273167187535506_4915
> 2012-08-16 16:00:17,925 INFO  datanode.BlockPoolSliceScanner 
> (BlockPoolSliceScanner.java:verifyBlock(391)) - Verification succeeded for 
> BP-2101131164-172.29.122.91-1337906886255:blk_7919273167187535506_4915
> {noformat}
> {quote}
> To fix this, we need to avoid roll()ing the logs multiple times per period.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-3592) libhdfs should expose ClientProtocol::mkdirs2

2014-04-10 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3592?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HDFS-3592:


Fix Version/s: (was: 2.4.0)
   2.5.0

> libhdfs should expose ClientProtocol::mkdirs2
> -
>
> Key: HDFS-3592
> URL: https://issues.apache.org/jira/browse/HDFS-3592
> Project: Hadoop HDFS
>  Issue Type: Improvement
>Affects Versions: 2.0.0-alpha, 3.0.0
>Reporter: Colin Patrick McCabe
>Assignee: Colin Patrick McCabe
>Priority: Minor
> Fix For: 3.0.0, 2.5.0
>
>
> It would be nice if libhdfs exposed mkdirs2.  This version of mkdirs is much 
> more verbose about any errors that occur-- it throws AccessControlException, 
> FileAlreadyExists, FileNotFoundException, ParentNotDirectoryException, etc.
> The original mkdirs just throws IOException if anything goes wrong.
> For something like fuse_dfs, it is very important to return the correct errno 
> code when an error has occurred.  mkdirs2 would allow us to do that.
> I'm not sure if we should just change hdfsMkdirs to use mkdirs2, or add an 
> hdfsMkdirs2.  Probably the latter, but the former course would maintain "bug 
> compatibility" with ancient releases-- if that is important.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-3907) Allow multiple users for local block readers

2014-04-10 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3907?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HDFS-3907:


Fix Version/s: (was: 2.4.0)
   2.5.0

> Allow multiple users for local block readers
> 
>
> Key: HDFS-3907
> URL: https://issues.apache.org/jira/browse/HDFS-3907
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: datanode
>Affects Versions: 1.0.0, 2.0.0-alpha
>Reporter: Eli Collins
>Assignee: Eli Collins
> Fix For: 2.5.0
>
> Attachments: hdfs-3907.txt
>
>
> The {{dfs.block.local-path-access.user}} config added in HDFS-2246 only 
> supports a single user, however as long as blocks are group readable by more 
> than one user the feature could be used by multiple users, to support this we 
> just need to allow both to be configured. In practice this allows us to also 
> support HBase where the client (RS) runs as the hbase system user and the DN 
> runs as hdfs system user. I think this should work secure as well since we're 
> not using impersonation in the HBase case.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-3988) HttpFS can't do GETDELEGATIONTOKEN without a prior authenticated request

2014-04-10 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3988?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HDFS-3988:


Fix Version/s: (was: 2.4.0)
   2.5.0

> HttpFS can't do GETDELEGATIONTOKEN without a prior authenticated request
> 
>
> Key: HDFS-3988
> URL: https://issues.apache.org/jira/browse/HDFS-3988
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 2.0.2-alpha
>Reporter: Alejandro Abdelnur
>Assignee: Alejandro Abdelnur
> Fix For: 2.5.0
>
>
> A request to obtain a delegation token cannot it initiate an authentication 
> sequence, it must be accompanied by an auth cookie obtained in a prev request 
> using a different operation.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-4754) Add an API in the namenode to mark a datanode as stale

2014-04-10 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4754?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HDFS-4754:


Fix Version/s: (was: 2.4.0)
   2.5.0

> Add an API in the namenode to mark a datanode as stale
> --
>
> Key: HDFS-4754
> URL: https://issues.apache.org/jira/browse/HDFS-4754
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: hdfs-client, namenode
>Reporter: Nicolas Liochon
>Assignee: Nicolas Liochon
>Priority: Critical
> Fix For: 3.0.0, 2.5.0
>
> Attachments: 4754.v1.patch, 4754.v2.patch, 4754.v4.patch, 
> 4754.v4.patch
>
>
> There is a detection of the stale datanodes in HDFS since HDFS-3703, with a 
> timeout, defaulted to 30s.
> There are two reasons to add an API to mark a node as stale even if the 
> timeout is not yet reached:
>  1) ZooKeeper can detect that a client is dead at any moment. So, for HBase, 
> we sometimes start the recovery before a node is marked staled. (even with 
> reasonable settings as: stale: 20s; HBase ZK timeout: 30s
>  2) Some third parties could detect that a node is dead before the timeout, 
> hence saving us the cost of retrying. An example or such hw is Arista, 
> presented here by [~tsuna] 
> http://tsunanet.net/~tsuna/fsf-hbase-meetup-april13.pdf, and confirmed in 
> HBASE-6290.
> As usual, even if the node is dead it can comeback before the 10 minutes 
> limit. So I would propose to set a timebound. The API would be
> namenode.markStale(String ipAddress, int port, long durationInMs);
> After durationInMs, the namenode would again rely only on its heartbeat to 
> decide.
> Thoughts?
> If there is no objections, and if nobody in the hdfs dev team has the time to 
> spend some time on it, I will give it a try for branch 2 & 3.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-4489) Use InodeID as as an identifier of a file in HDFS protocols and APIs

2014-04-10 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4489?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HDFS-4489:


Fix Version/s: (was: 2.4.0)
   2.5.0

> Use InodeID as as an identifier of a file in HDFS protocols and APIs
> 
>
> Key: HDFS-4489
> URL: https://issues.apache.org/jira/browse/HDFS-4489
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: namenode
>Reporter: Brandon Li
>Assignee: Brandon Li
> Fix For: 2.5.0
>
> Attachments: 4434.optimized.patch
>
>
> The benefit of using InodeID to uniquely identify a file can be multiple 
> folds. Here are a few of them:
> 1. uniquely identify a file cross rename, related JIRAs include HDFS-4258, 
> HDFS-4437.
> 2. modification checks in tools like distcp. Since a file could have been 
> replaced or renamed to, the file name and size combination is no t reliable, 
> but the combination of file id and size is unique.
> 3. id based protocol support (e.g., NFS)
> 4. to make the pluggable block placement policy use fileid instead of 
> filename (HDFS-385).



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-5160) MiniDFSCluster webui does not work

2014-04-10 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5160?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HDFS-5160:


Fix Version/s: (was: 2.4.0)
   2.5.0

> MiniDFSCluster webui does not work
> --
>
> Key: HDFS-5160
> URL: https://issues.apache.org/jira/browse/HDFS-5160
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: test
>Affects Versions: 2.1.0-beta
>Reporter: Alejandro Abdelnur
> Fix For: 2.5.0
>
>
> The webui does not work, when going to http://localhost:50070 you get:
> {code}
> Directory: /
> webapps/  102 bytes   Sep 4, 2013 9:32:55 AM
> {code}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-4066) Fix compile warnings for libwebhdfs

2014-04-10 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4066?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HDFS-4066:


Fix Version/s: (was: 2.4.0)
   2.5.0

> Fix compile warnings for libwebhdfs
> ---
>
> Key: HDFS-4066
> URL: https://issues.apache.org/jira/browse/HDFS-4066
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: webhdfs
>Affects Versions: 3.0.0
>Reporter: Jing Zhao
>Assignee: Jing Zhao
> Fix For: 2.5.0
>
>
> The compile of libwebhdfs still generates a bunch of warnings, which need to 
> be fixed.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-4198) the shell script error for Cygwin on windows7

2014-04-10 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4198?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HDFS-4198:


Fix Version/s: (was: 2.4.0)
   2.5.0

> the shell script error for Cygwin on windows7
> -
>
> Key: HDFS-4198
> URL: https://issues.apache.org/jira/browse/HDFS-4198
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: scripts
>Affects Versions: 2.0.2-alpha
> Environment: windows7 ,cygwin.
>Reporter: Han Hui Wen 
> Fix For: 2.5.0
>
>
> See the following 
> [comment|https://issues.apache.org/jira/browse/HDFS-4198?focusedCommentId=13498818&page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-13498818]
>  for detailed description.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-5098) Enhance FileSystem.Statistics to have locality information

2014-04-10 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5098?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HDFS-5098:


Fix Version/s: (was: 2.4.0)
   2.5.0

> Enhance FileSystem.Statistics to have locality information
> --
>
> Key: HDFS-5098
> URL: https://issues.apache.org/jira/browse/HDFS-5098
> Project: Hadoop HDFS
>  Issue Type: Improvement
>Reporter: Bikas Saha
>Assignee: Suresh Srinivas
> Fix For: 2.5.0
>
>
> Currently in MR/Tez we dont have a good and accurate means to detect how much 
> the the IO was actually done locally. Getting this information from the 
> source of truth would be much better.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-4286) Changes from BOOKKEEPER-203 broken capability of including bookkeeper-server jar in hidden package of BKJM

2014-04-10 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4286?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HDFS-4286:


Fix Version/s: (was: 2.4.0)
   2.5.0

> Changes from BOOKKEEPER-203 broken capability of including bookkeeper-server 
> jar in hidden package of BKJM
> --
>
> Key: HDFS-4286
> URL: https://issues.apache.org/jira/browse/HDFS-4286
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>Reporter: Vinayakumar B
> Fix For: 3.0.0, 2.5.0
>
>
> BOOKKEEPER-203 introduced changes to LedgerLayout to include 
> ManagerFactoryClass instead of ManagerFactoryName.
> So because of this, BKJM cannot shade the bookkeeper-server jar inside BKJM 
> jar
> LAYOUT znode created by BookieServer is not readable by the BKJM as it have 
> classes in hidden packages. (same problem vice versa)



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-5687) Problem in accessing NN JSP page

2014-04-10 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5687?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HDFS-5687:


Fix Version/s: (was: 2.4.0)
   2.5.0

> Problem in accessing NN JSP page
> 
>
> Key: HDFS-5687
> URL: https://issues.apache.org/jira/browse/HDFS-5687
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: hdfs-client
>Affects Versions: 2.3.0
>Reporter: sathish
>Assignee: sathish
>Priority: Minor
> Fix For: 2.5.0
>
> Attachments: HDFS-5687-0001.patch
>
>
> In NN UI page After clicking the browse File System page,from that page,if 
> you click GO Back TO DFS HOME ICon it is not accessing the dfshealth.jsp page
> NN http URL is http://nnaddr///nninfoaddr/dfshealth.jsp,it is coming like 
> this,due to this i think it is not browsing that page
> It should be http://nninfoaddr/dfshealth.jsp/ like this



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6143) WebHdfsFileSystem open should throw FileNotFoundException for non-existing paths

2014-04-07 Thread Arun C Murthy (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6143?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13961906#comment-13961906
 ] 

Arun C Murthy commented on HDFS-6143:
-

One concern: what is the impact on compatibility with 2.2 etc?

> WebHdfsFileSystem open should throw FileNotFoundException for non-existing 
> paths
> 
>
> Key: HDFS-6143
> URL: https://issues.apache.org/jira/browse/HDFS-6143
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 2.3.0
>Reporter: Gera Shegalov
>Assignee: Gera Shegalov
>Priority: Blocker
> Attachments: HDFS-6143-branch-2.4.0.v01.patch, 
> HDFS-6143-trunk-after-HDFS-5570.v01.patch, HDFS-6143.v01.patch, 
> HDFS-6143.v02.patch, HDFS-6143.v03.patch, HDFS-6143.v04.patch, 
> HDFS-6143.v04.patch, HDFS-6143.v05.patch, HDFS-6143.v06.patch
>
>
> WebHdfsFileSystem.open and HftpFileSystem.open incorrectly handles 
> non-existing paths. 
> - 'open', does not really open anything, i.e., it does not contact the 
> server, and therefore cannot discover FileNotFound, it's deferred until next 
> read. It's counterintuitive and not how local FS or HDFS work. In POSIX you 
> get ENOENT on open. 
> [LzoInputFormat.getSplits|https://github.com/kevinweil/elephant-bird/blob/master/core/src/main/java/com/twitter/elephantbird/mapreduce/input/LzoInputFormat.java]
>  is an example of the code that's broken because of this.
> - On the server side, FileDataServlet incorrectly sends SC_BAD_REQUEST 
> instead of SC_NOT_FOUND for non-exitsing paths



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6143) (WebHdfs|Hftp)FileSystem open should throw FileNotFoundException for non-existing paths

2014-03-31 Thread Arun C Murthy (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6143?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13955014#comment-13955014
 ] 

Arun C Murthy commented on HDFS-6143:
-

[~jira.shegalov] Saw this too late - I'm moving this to 2.4.1, let's try get 
this in asap. I have rc0 going out, we can re-target 2.4.0 if I need to spin 
another rc. Thanks.

> (WebHdfs|Hftp)FileSystem open should throw FileNotFoundException for 
> non-existing paths
> ---
>
> Key: HDFS-6143
> URL: https://issues.apache.org/jira/browse/HDFS-6143
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 2.3.0
>Reporter: Gera Shegalov
>Assignee: Gera Shegalov
>Priority: Blocker
> Attachments: HDFS-6143.v01.patch, HDFS-6143.v02.patch, 
> HDFS-6143.v03.patch, HDFS-6143.v04.patch
>
>
> HftpFileSystem.open incorrectly handles non-existing paths. 
> - 'open', does not really open anything, i.e., it does not contact the 
> server, and therefore cannot discover FileNotFound, it's deferred until next 
> read. It's counterintuitive and not how local FS or HDFS work. In POSIX you 
> get ENOENT on open. 
> [LzoInputFormat.getSplits|https://github.com/kevinweil/elephant-bird/blob/master/core/src/main/java/com/twitter/elephantbird/mapreduce/input/LzoInputFormat.java]
>  is an example of the code that's broken because of this.
> - On the server side, FileDataServlet incorrectly sends SC_BAD_REQUEST 
> instead of SC_NOT_FOUND for non-exitsing paths



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6143) (WebHdfs|Hftp)FileSystem open should throw FileNotFoundException for non-existing paths

2014-03-31 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6143?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HDFS-6143:


Target Version/s: 2.4.1  (was: 2.4.0)

> (WebHdfs|Hftp)FileSystem open should throw FileNotFoundException for 
> non-existing paths
> ---
>
> Key: HDFS-6143
> URL: https://issues.apache.org/jira/browse/HDFS-6143
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 2.3.0
>Reporter: Gera Shegalov
>Assignee: Gera Shegalov
>Priority: Blocker
> Attachments: HDFS-6143.v01.patch, HDFS-6143.v02.patch, 
> HDFS-6143.v03.patch, HDFS-6143.v04.patch
>
>
> HftpFileSystem.open incorrectly handles non-existing paths. 
> - 'open', does not really open anything, i.e., it does not contact the 
> server, and therefore cannot discover FileNotFound, it's deferred until next 
> read. It's counterintuitive and not how local FS or HDFS work. In POSIX you 
> get ENOENT on open. 
> [LzoInputFormat.getSplits|https://github.com/kevinweil/elephant-bird/blob/master/core/src/main/java/com/twitter/elephantbird/mapreduce/input/LzoInputFormat.java]
>  is an example of the code that's broken because of this.
> - On the server side, FileDataServlet incorrectly sends SC_BAD_REQUEST 
> instead of SC_NOT_FOUND for non-exitsing paths



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-4564) Webhdfs returns incorrect http response codes for denied operations

2014-03-31 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4564?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HDFS-4564:


   Resolution: Fixed
Fix Version/s: 2.4.0
   Status: Resolved  (was: Patch Available)

I just committed this. Thanks [~daryn] (and [~jingzhao] for the review)!

> Webhdfs returns incorrect http response codes for denied operations
> ---
>
> Key: HDFS-4564
> URL: https://issues.apache.org/jira/browse/HDFS-4564
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: webhdfs
>Affects Versions: 0.23.0, 2.0.0-alpha, 3.0.0
>Reporter: Daryn Sharp
>Assignee: Daryn Sharp
>Priority: Blocker
> Fix For: 2.4.0
>
> Attachments: HDFS-4564.branch-23.patch, HDFS-4564.branch-23.patch, 
> HDFS-4564.branch-23.patch, HDFS-4564.patch, HDFS-4564.patch
>
>
> Webhdfs is returning 401 (Unauthorized) instead of 403 (Forbidden) when it's 
> denying operations.  Examples including rejecting invalid proxy user attempts 
> and renew/cancel with an invalid user.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-4564) Webhdfs returns incorrect http response codes for denied operations

2014-03-31 Thread Arun C Murthy (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-4564?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13954997#comment-13954997
 ] 

Arun C Murthy commented on HDFS-4564:
-

I'm going to commit this, based on [~jingzhao]'s +1 - I'd like to roll 2.4 rc0 
asap since my availability next few days is limited with the Hadoop Summit.

> Webhdfs returns incorrect http response codes for denied operations
> ---
>
> Key: HDFS-4564
> URL: https://issues.apache.org/jira/browse/HDFS-4564
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: webhdfs
>Affects Versions: 0.23.0, 2.0.0-alpha, 3.0.0
>Reporter: Daryn Sharp
>Assignee: Daryn Sharp
>Priority: Blocker
> Attachments: HDFS-4564.branch-23.patch, HDFS-4564.branch-23.patch, 
> HDFS-4564.branch-23.patch, HDFS-4564.patch, HDFS-4564.patch
>
>
> Webhdfs is returning 401 (Unauthorized) instead of 403 (Forbidden) when it's 
> denying operations.  Examples including rejecting invalid proxy user attempts 
> and renew/cancel with an invalid user.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-4564) Webhdfs returns incorrect http response codes for denied operations

2014-03-29 Thread Arun C Murthy (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-4564?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13954374#comment-13954374
 ] 

Arun C Murthy commented on HDFS-4564:
-

[~daryn] - Can you pls clarify some of the above review comments? I'd like to 
move fwd with 2.4 ASAP. Tx!

> Webhdfs returns incorrect http response codes for denied operations
> ---
>
> Key: HDFS-4564
> URL: https://issues.apache.org/jira/browse/HDFS-4564
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: webhdfs
>Affects Versions: 0.23.0, 2.0.0-alpha, 3.0.0
>Reporter: Daryn Sharp
>Assignee: Daryn Sharp
>Priority: Blocker
> Attachments: HDFS-4564.branch-23.patch, HDFS-4564.branch-23.patch, 
> HDFS-4564.branch-23.patch, HDFS-4564.patch, HDFS-4564.patch
>
>
> Webhdfs is returning 401 (Unauthorized) instead of 403 (Forbidden) when it's 
> denying operations.  Examples including rejecting invalid proxy user attempts 
> and renew/cancel with an invalid user.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-5840) Follow-up to HDFS-5138 to improve error handling during partial upgrade failures

2014-03-22 Thread Arun C Murthy (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5840?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13944061#comment-13944061
 ] 

Arun C Murthy commented on HDFS-5840:
-

[~sureshms] / [~atm] - Should this be a blocker for 2.4? Thanks.

> Follow-up to HDFS-5138 to improve error handling during partial upgrade 
> failures
> 
>
> Key: HDFS-5840
> URL: https://issues.apache.org/jira/browse/HDFS-5840
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: namenode
>Affects Versions: 3.0.0
>Reporter: Aaron T. Myers
>Assignee: Aaron T. Myers
> Fix For: 3.0.0
>
> Attachments: HDFS-5840.patch
>
>
> Suresh posted some good comment in HDFS-5138 after that patch had already 
> been committed to trunk. This JIRA is to address those. See the first comment 
> of this JIRA for the full content of the review.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-5138) Support HDFS upgrade in HA

2014-03-22 Thread Arun C Murthy (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5138?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13944060#comment-13944060
 ] 

Arun C Murthy commented on HDFS-5138:
-

[~jingzhao] / [~sureshms] - Should we close this one and mark HDFS-6135 as the 
blocker? Thanks.

> Support HDFS upgrade in HA
> --
>
> Key: HDFS-5138
> URL: https://issues.apache.org/jira/browse/HDFS-5138
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 2.1.1-beta
>Reporter: Kihwal Lee
>Assignee: Aaron T. Myers
>Priority: Blocker
> Fix For: 3.0.0
>
> Attachments: HDFS-5138.patch, HDFS-5138.patch, HDFS-5138.patch, 
> HDFS-5138.patch, HDFS-5138.patch, HDFS-5138.patch, HDFS-5138.patch, 
> HDFS-5138.patch, HDFS-5138.patch, HDFS-5138.patch, HDFS-5138.patch, 
> hdfs-5138-branch-2.txt
>
>
> With HA enabled, NN wo't start with "-upgrade". Since there has been a layout 
> version change between 2.0.x and 2.1.x, starting NN in upgrade mode was 
> necessary when deploying 2.1.x to an existing 2.0.x cluster. But the only way 
> to get around this was to disable HA and upgrade. 
> The NN and the cluster cannot be flipped back to HA until the upgrade is 
> finalized. If HA is disabled only on NN for layout upgrade and HA is turned 
> back on without involving DNs, things will work, but finaliizeUpgrade won't 
> work (the NN is in HA and it cannot be in upgrade mode) and DN's upgrade 
> snapshots won't get removed.
> We will need a different ways of doing layout upgrade and upgrade snapshot.  
> I am marking this as a 2.1.1-beta blocker based on feedback from others.  If 
> there is a reasonable workaround that does not increase maintenance window 
> greatly, we can lower its priority from blocker to critical.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-4564) Webhdfs returns incorrect http response codes for denied operations

2014-03-12 Thread Arun C Murthy (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-4564?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13932008#comment-13932008
 ] 

Arun C Murthy commented on HDFS-4564:
-

How is this looking [~daryn]? Thanks.

(Doing a pass over 2.4 blockers)

> Webhdfs returns incorrect http response codes for denied operations
> ---
>
> Key: HDFS-4564
> URL: https://issues.apache.org/jira/browse/HDFS-4564
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: webhdfs
>Affects Versions: 0.23.0, 2.0.0-alpha, 3.0.0
>Reporter: Daryn Sharp
>Assignee: Daryn Sharp
>Priority: Blocker
> Attachments: HDFS-4564.branch-23.patch, HDFS-4564.branch-23.patch, 
> HDFS-4564.branch-23.patch, HDFS-4564.patch
>
>
> Webhdfs is returning 401 (Unauthorized) instead of 403 (Forbidden) when it's 
> denying operations.  Examples including rejecting invalid proxy user attempts 
> and renew/cancel with an invalid user.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-3907) Allow multiple users for local block readers

2014-02-24 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3907?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HDFS-3907:


Fix Version/s: (was: 2.3.0)
   2.4.0

> Allow multiple users for local block readers
> 
>
> Key: HDFS-3907
> URL: https://issues.apache.org/jira/browse/HDFS-3907
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: datanode
>Affects Versions: 1.0.0, 2.0.0-alpha
>Reporter: Eli Collins
>Assignee: Eli Collins
> Fix For: 2.4.0
>
> Attachments: hdfs-3907.txt
>
>
> The {{dfs.block.local-path-access.user}} config added in HDFS-2246 only 
> supports a single user, however as long as blocks are group readable by more 
> than one user the feature could be used by multiple users, to support this we 
> just need to allow both to be configured. In practice this allows us to also 
> support HBase where the client (RS) runs as the hbase system user and the DN 
> runs as hdfs system user. I think this should work secure as well since we're 
> not using impersonation in the HBase case.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (HDFS-4198) the shell script error for Cygwin on windows7

2014-02-24 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4198?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HDFS-4198:


Fix Version/s: (was: 2.3.0)
   2.4.0

> the shell script error for Cygwin on windows7
> -
>
> Key: HDFS-4198
> URL: https://issues.apache.org/jira/browse/HDFS-4198
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: scripts
>Affects Versions: 2.0.2-alpha
> Environment: windows7 ,cygwin.
>Reporter: Han Hui Wen 
> Fix For: 2.4.0
>
>
> See the following 
> [comment|https://issues.apache.org/jira/browse/HDFS-4198?focusedCommentId=13498818&page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-13498818]
>  for detailed description.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (HDFS-5522) Datanode disk error check may be incorrectly skipped

2014-02-24 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5522?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HDFS-5522:


Fix Version/s: (was: 2.3.0)
   2.4.0

> Datanode disk error check may be incorrectly skipped
> 
>
> Key: HDFS-5522
> URL: https://issues.apache.org/jira/browse/HDFS-5522
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 0.23.9, 2.2.0
>Reporter: Kihwal Lee
> Fix For: 0.23.11, 2.4.0
>
>
> After HDFS-4581 and HDFS-4699, {{checkDiskError()}} is not called when 
> network errors occur during processing data node requests.  This appears to 
> create problems when a disk is having problems, but not failing I/O soon. 
> If I/O hangs for a long time, network read/write may timeout first and the 
> peer may close the connection. Although the error was caused by a faulty 
> local disk, disk check is not being carried out in this case. 



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (HDFS-4066) Fix compile warnings for libwebhdfs

2014-02-24 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4066?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HDFS-4066:


Fix Version/s: (was: 2.3.0)
   2.4.0

> Fix compile warnings for libwebhdfs
> ---
>
> Key: HDFS-4066
> URL: https://issues.apache.org/jira/browse/HDFS-4066
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: webhdfs
>Affects Versions: 3.0.0
>Reporter: Jing Zhao
>Assignee: Jing Zhao
> Fix For: 2.4.0
>
>
> The compile of libwebhdfs still generates a bunch of warnings, which need to 
> be fixed.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (HDFS-5160) MiniDFSCluster webui does not work

2014-02-24 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5160?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HDFS-5160:


Fix Version/s: (was: 2.3.0)
   2.4.0

> MiniDFSCluster webui does not work
> --
>
> Key: HDFS-5160
> URL: https://issues.apache.org/jira/browse/HDFS-5160
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: test
>Affects Versions: 2.1.0-beta
>Reporter: Alejandro Abdelnur
> Fix For: 2.4.0
>
>
> The webui does not work, when going to http://localhost:50070 you get:
> {code}
> Directory: /
> webapps/  102 bytes   Sep 4, 2013 9:32:55 AM
> {code}



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (HDFS-5244) TestNNStorageRetentionManager#testPurgeMultipleDirs fails because incorrectly expects Hashmap values to have order.

2014-02-24 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5244?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HDFS-5244:


Fix Version/s: (was: 2.3.0)
   2.4.0

> TestNNStorageRetentionManager#testPurgeMultipleDirs fails because incorrectly 
> expects Hashmap values to have order. 
> 
>
> Key: HDFS-5244
> URL: https://issues.apache.org/jira/browse/HDFS-5244
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: test
>Affects Versions: 2.1.0-beta
> Environment: Red Hat Enterprise 6 with Sun Java 1.7 and IBM java 1.6
>Reporter: Jinghui Wang
>Assignee: Jinghui Wang
> Fix For: 3.0.0, 2.1.0-beta, 2.4.0
>
> Attachments: HDFS-5244.patch
>
>
> The test o.a.h.hdfs.server.namenode.TestNNStorageRetentionManager uses a 
> HashMap(dirRoots) to store the root storages to be mocked for the purging 
> test, which does not have any predictable order. The directories needs be 
> purged are stored in a LinkedHashSet, which has a predictable order. So, when 
> the directories get mocked for the test, they could be already out of
> the order that they were added. Thus, the order that the directories were
> actually purged and the order of them being added to the LinkedHashList could
> be different and cause the test to fail.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (HDFS-3761) Uses of AuthenticatedURL should set the configured timeouts

2014-02-24 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3761?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HDFS-3761:


Fix Version/s: (was: 2.3.0)
   2.4.0

> Uses of AuthenticatedURL should set the configured timeouts
> ---
>
> Key: HDFS-3761
> URL: https://issues.apache.org/jira/browse/HDFS-3761
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 2.0.0-alpha, 2.0.2-alpha
>Reporter: Alejandro Abdelnur
>Priority: Critical
> Fix For: 2.4.0
>
>
> Similar to {{URLUtils}}, the {{AuthenticatedURL}} should set the configured 
> timeouts in URL connection instances. HADOOP-8644 is introducing a 
> {{ConnectionConfigurator}} interface to the {{AuthenticatedURL}} class, it 
> should be done via that interface.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (HDFS-3592) libhdfs should expose ClientProtocol::mkdirs2

2014-02-24 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3592?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HDFS-3592:


Fix Version/s: (was: 2.3.0)
   2.4.0

> libhdfs should expose ClientProtocol::mkdirs2
> -
>
> Key: HDFS-3592
> URL: https://issues.apache.org/jira/browse/HDFS-3592
> Project: Hadoop HDFS
>  Issue Type: Improvement
>Affects Versions: 2.0.0-alpha, 3.0.0
>Reporter: Colin Patrick McCabe
>Assignee: Colin Patrick McCabe
>Priority: Minor
> Fix For: 3.0.0, 2.4.0
>
>
> It would be nice if libhdfs exposed mkdirs2.  This version of mkdirs is much 
> more verbose about any errors that occur-- it throws AccessControlException, 
> FileAlreadyExists, FileNotFoundException, ParentNotDirectoryException, etc.
> The original mkdirs just throws IOException if anything goes wrong.
> For something like fuse_dfs, it is very important to return the correct errno 
> code when an error has occurred.  mkdirs2 would allow us to do that.
> I'm not sure if we should just change hdfsMkdirs to use mkdirs2, or add an 
> hdfsMkdirs2.  Probably the latter, but the former course would maintain "bug 
> compatibility" with ancient releases-- if that is important.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (HDFS-4489) Use InodeID as as an identifier of a file in HDFS protocols and APIs

2014-02-24 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4489?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HDFS-4489:


Fix Version/s: (was: 2.3.0)
   2.4.0

> Use InodeID as as an identifier of a file in HDFS protocols and APIs
> 
>
> Key: HDFS-4489
> URL: https://issues.apache.org/jira/browse/HDFS-4489
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: namenode
>Reporter: Brandon Li
>Assignee: Brandon Li
> Fix For: 2.4.0
>
> Attachments: 4434.optimized.patch
>
>
> The benefit of using InodeID to uniquely identify a file can be multiple 
> folds. Here are a few of them:
> 1. uniquely identify a file cross rename, related JIRAs include HDFS-4258, 
> HDFS-4437.
> 2. modification checks in tools like distcp. Since a file could have been 
> replaced or renamed to, the file name and size combination is no t reliable, 
> but the combination of file id and size is unique.
> 3. id based protocol support (e.g., NFS)
> 4. to make the pluggable block placement policy use fileid instead of 
> filename (HDFS-385).



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (HDFS-3988) HttpFS can't do GETDELEGATIONTOKEN without a prior authenticated request

2014-02-24 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3988?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HDFS-3988:


Fix Version/s: (was: 2.3.0)
   2.4.0

> HttpFS can't do GETDELEGATIONTOKEN without a prior authenticated request
> 
>
> Key: HDFS-3988
> URL: https://issues.apache.org/jira/browse/HDFS-3988
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 2.0.2-alpha
>Reporter: Alejandro Abdelnur
>Assignee: Alejandro Abdelnur
> Fix For: 2.4.0
>
>
> A request to obtain a delegation token cannot it initiate an authentication 
> sequence, it must be accompanied by an auth cookie obtained in a prev request 
> using a different operation.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (HDFS-5687) Problem in accessing NN JSP page

2014-02-24 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5687?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HDFS-5687:


Fix Version/s: (was: 2.3.0)
   2.4.0

> Problem in accessing NN JSP page
> 
>
> Key: HDFS-5687
> URL: https://issues.apache.org/jira/browse/HDFS-5687
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: hdfs-client
>Affects Versions: 2.3.0
>Reporter: sathish
>Assignee: sathish
>Priority: Minor
> Fix For: 2.4.0
>
> Attachments: HDFS-5687-0001.patch
>
>
> In NN UI page After clicking the browse File System page,from that page,if 
> you click GO Back TO DFS HOME ICon it is not accessing the dfshealth.jsp page
> NN http URL is http://nnaddr///nninfoaddr/dfshealth.jsp,it is coming like 
> this,due to this i think it is not browsing that page
> It should be http://nninfoaddr/dfshealth.jsp/ like this



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (HDFS-4286) Changes from BOOKKEEPER-203 broken capability of including bookkeeper-server jar in hidden package of BKJM

2014-02-24 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4286?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HDFS-4286:


Fix Version/s: (was: 2.3.0)
   2.4.0

> Changes from BOOKKEEPER-203 broken capability of including bookkeeper-server 
> jar in hidden package of BKJM
> --
>
> Key: HDFS-4286
> URL: https://issues.apache.org/jira/browse/HDFS-4286
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>Reporter: Vinayakumar B
> Fix For: 3.0.0, 2.4.0
>
>
> BOOKKEEPER-203 introduced changes to LedgerLayout to include 
> ManagerFactoryClass instead of ManagerFactoryName.
> So because of this, BKJM cannot shade the bookkeeper-server jar inside BKJM 
> jar
> LAYOUT znode created by BookieServer is not readable by the BKJM as it have 
> classes in hidden packages. (same problem vice versa)



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (HDFS-2802) Support for RW/RO snapshots in HDFS

2014-02-24 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-2802?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HDFS-2802:


Fix Version/s: (was: 2.3.0)
   2.4.0

> Support for RW/RO snapshots in HDFS
> ---
>
> Key: HDFS-2802
> URL: https://issues.apache.org/jira/browse/HDFS-2802
> Project: Hadoop HDFS
>  Issue Type: New Feature
>  Components: namenode
>Reporter: Hari Mankude
>Assignee: Tsz Wo (Nicholas), SZE
> Fix For: 2.4.0
>
> Attachments: 2802.diff, 2802.patch, 2802.patch, 
> HDFS-2802-meeting-minutes-121101.txt, HDFS-2802.20121101.patch, 
> HDFS-2802.branch-2.0510.patch, HDFSSnapshotsDesign.pdf, Snapshots.pdf, 
> Snapshots20120429.pdf, Snapshots20121018.pdf, Snapshots20121030.pdf, 
> h2802_20130417.patch, h2802_20130422.patch, h2802_20130423.patch, 
> h2802_20130425.patch, h2802_20130426.patch, h2802_20130430.patch, 
> h2802_20130430b.patch, h2802_20130508.patch, h2802_20130509.patch, 
> snap.patch, snapshot-design.pdf, snapshot-design.tex, snapshot-one-pager.pdf, 
> snapshot-testplan.pdf
>
>
> Snapshots are point in time images of parts of the filesystem or the entire 
> filesystem. Snapshots can be a read-only or a read-write point in time copy 
> of the filesystem. There are several use cases for snapshots in HDFS. I will 
> post a detailed write-up soon with with more information.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (HDFS-3746) Invalid counter tag in HDFS balancer which lead to infinite loop

2014-02-24 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3746?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HDFS-3746:


Fix Version/s: (was: 2.3.0)
   2.4.0

> Invalid counter tag in HDFS balancer which lead to infinite loop
> 
>
> Key: HDFS-3746
> URL: https://issues.apache.org/jira/browse/HDFS-3746
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: balancer
>Affects Versions: 2.0.0-alpha, 3.0.0
>Reporter: meng gong
>Assignee: meng gong
>  Labels: patch
> Fix For: 3.0.0, 2.4.0
>
>
> Everytime banlancer try to move a block cross the rack. For every 
> NameNodeConnector it will instance a new banlancer. The tag 
> notChangedIterations is reseted to 0. Then it won't reach 5 and exit the 
> thread for not moved in 5 consecutive iterations. This lead to a infinite loop



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (HDFS-2832) Enable support for heterogeneous storages in HDFS

2014-02-24 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-2832?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HDFS-2832:


Fix Version/s: (was: 2.3.0)
   2.4.0

> Enable support for heterogeneous storages in HDFS
> -
>
> Key: HDFS-2832
> URL: https://issues.apache.org/jira/browse/HDFS-2832
> Project: Hadoop HDFS
>  Issue Type: New Feature
>Affects Versions: 0.24.0
>Reporter: Suresh Srinivas
>Assignee: Suresh Srinivas
> Fix For: 3.0.0, 2.4.0
>
> Attachments: 20130813-HeterogeneousStorage.pdf, 
> 20131125-HeterogeneousStorage-TestPlan.pdf, 
> 20131125-HeterogeneousStorage.pdf, 
> 20131202-HeterogeneousStorage-TestPlan.pdf, 
> 20131203-HeterogeneousStorage-TestPlan.pdf, H2832_20131107.patch, 
> editsStored, editsStored, h2832_20131023.patch, h2832_20131023b.patch, 
> h2832_20131025.patch, h2832_20131028.patch, h2832_20131028b.patch, 
> h2832_20131029.patch, h2832_20131103.patch, h2832_20131104.patch, 
> h2832_20131105.patch, h2832_20131107b.patch, h2832_20131108.patch, 
> h2832_20131110.patch, h2832_20131110b.patch, h2832_2013.patch, 
> h2832_20131112.patch, h2832_20131112b.patch, h2832_20131114.patch, 
> h2832_20131118.patch, h2832_20131119.patch, h2832_20131119b.patch, 
> h2832_20131121.patch, h2832_20131122.patch, h2832_20131122b.patch, 
> h2832_20131123.patch, h2832_20131124.patch, h2832_20131202.patch, 
> h2832_20131203.patch, h2832_20131210.patch, h2832_20131211.patch, 
> h2832_20131211b.patch, h2832_branch-2_20131226.patch, 
> h2832_branch-2_20140103.patch
>
>
> HDFS currently supports configuration where storages are a list of 
> directories. Typically each of these directories correspond to a volume with 
> its own file system. All these directories are homogeneous and therefore 
> identified as a single storage at the namenode. I propose, change to the 
> current model where Datanode * is a * storage, to Datanode * is a collection 
> * of strorages. 



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (HDFS-5098) Enhance FileSystem.Statistics to have locality information

2014-02-24 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5098?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HDFS-5098:


Fix Version/s: (was: 2.3.0)
   2.4.0

> Enhance FileSystem.Statistics to have locality information
> --
>
> Key: HDFS-5098
> URL: https://issues.apache.org/jira/browse/HDFS-5098
> Project: Hadoop HDFS
>  Issue Type: Improvement
>Reporter: Bikas Saha
>Assignee: Suresh Srinivas
> Fix For: 2.4.0
>
>
> Currently in MR/Tez we dont have a good and accurate means to detect how much 
> the the IO was actually done locally. Getting this information from the 
> source of truth would be much better.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (HDFS-3747) balancer-Test failed because of time out

2014-02-24 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3747?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HDFS-3747:


Fix Version/s: (was: 2.3.0)
   2.4.0

>  balancer-Test failed because of time out
> -
>
> Key: HDFS-3747
> URL: https://issues.apache.org/jira/browse/HDFS-3747
> Project: Hadoop HDFS
>  Issue Type: Test
>  Components: balancer
>Affects Versions: 2.0.0-alpha, 3.0.0
>Reporter: meng gong
>Assignee: meng gong
>  Labels: test
> Fix For: 3.0.0, 2.4.0
>
> Attachments: TestCaseForBanlancer.java
>
>
> When run a given test case for Banlancer Test. In the test the banlancer 
> thread try to move some block cross the rack but it can't find any available 
> blocks in the source rack. Then the thread won't interrupt until the tag 
> isTimeUp reaches 20min. But maven judges the test failed because the thread 
> have runned for 15min.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (HDFS-4754) Add an API in the namenode to mark a datanode as stale

2014-02-24 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4754?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HDFS-4754:


Fix Version/s: (was: 2.3.0)
   2.4.0

> Add an API in the namenode to mark a datanode as stale
> --
>
> Key: HDFS-4754
> URL: https://issues.apache.org/jira/browse/HDFS-4754
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: hdfs-client, namenode
>Reporter: Nicolas Liochon
>Assignee: Nicolas Liochon
>Priority: Critical
> Fix For: 3.0.0, 2.4.0
>
> Attachments: 4754.v1.patch, 4754.v2.patch, 4754.v4.patch, 
> 4754.v4.patch
>
>
> There is a detection of the stale datanodes in HDFS since HDFS-3703, with a 
> timeout, defaulted to 30s.
> There are two reasons to add an API to mark a node as stale even if the 
> timeout is not yet reached:
>  1) ZooKeeper can detect that a client is dead at any moment. So, for HBase, 
> we sometimes start the recovery before a node is marked staled. (even with 
> reasonable settings as: stale: 20s; HBase ZK timeout: 30s
>  2) Some third parties could detect that a node is dead before the timeout, 
> hence saving us the cost of retrying. An example or such hw is Arista, 
> presented here by [~tsuna] 
> http://tsunanet.net/~tsuna/fsf-hbase-meetup-april13.pdf, and confirmed in 
> HBASE-6290.
> As usual, even if the node is dead it can comeback before the 10 minutes 
> limit. So I would propose to set a timebound. The API would be
> namenode.markStale(String ipAddress, int port, long durationInMs);
> After durationInMs, the namenode would again rely only on its heartbeat to 
> decide.
> Thoughts?
> If there is no objections, and if nobody in the hdfs dev team has the time to 
> spend some time on it, I will give it a try for branch 2 & 3.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (HDFS-3828) Block Scanner rescans blocks too frequently

2014-02-24 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3828?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HDFS-3828:


Fix Version/s: (was: 2.3.0)
   2.4.0

> Block Scanner rescans blocks too frequently
> ---
>
> Key: HDFS-3828
> URL: https://issues.apache.org/jira/browse/HDFS-3828
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 0.23.0
>Reporter: Andy Isaacson
>Assignee: Andy Isaacson
> Fix For: 2.4.0
>
> Attachments: hdfs-3828-1.txt, hdfs-3828-2.txt, hdfs-3828-3.txt, 
> hdfs3828.txt
>
>
> {{BlockPoolSliceScanner#scan}} calls cleanUp every time it's invoked from 
> {{DataBlockScanner#run}} via {{scanBlockPoolSlice}}.  But cleanUp 
> unconditionally roll()s the verificationLogs, so after two iterations we have 
> lost the first iteration of block verification times.  As a result a cluster 
> with just one block repeatedly rescans it every 10 seconds:
> {noformat}
> 2012-08-16 15:59:57,884 INFO  datanode.BlockPoolSliceScanner 
> (BlockPoolSliceScanner.java:verifyBlock(391)) - Verification succeeded for 
> BP-2101131164-172.29.122.91-1337906886255:blk_7919273167187535506_4915
> 2012-08-16 16:00:07,904 INFO  datanode.BlockPoolSliceScanner 
> (BlockPoolSliceScanner.java:verifyBlock(391)) - Verification succeeded for 
> BP-2101131164-172.29.122.91-1337906886255:blk_7919273167187535506_4915
> 2012-08-16 16:00:17,925 INFO  datanode.BlockPoolSliceScanner 
> (BlockPoolSliceScanner.java:verifyBlock(391)) - Verification succeeded for 
> BP-2101131164-172.29.122.91-1337906886255:blk_7919273167187535506_4915
> {noformat}
> {quote}
> To fix this, we need to avoid roll()ing the logs multiple times per period.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (HDFS-4564) Webhdfs returns incorrect http response codes for denied operations

2014-02-04 Thread Arun C Murthy (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-4564?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13891167#comment-13891167
 ] 

Arun C Murthy commented on HDFS-4564:
-

Ok, thanks [~daryn]!

> Webhdfs returns incorrect http response codes for denied operations
> ---
>
> Key: HDFS-4564
> URL: https://issues.apache.org/jira/browse/HDFS-4564
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: webhdfs
>Affects Versions: 0.23.0, 2.0.0-alpha, 3.0.0
>Reporter: Daryn Sharp
>Assignee: Daryn Sharp
>Priority: Blocker
> Attachments: HDFS-4564.branch-23.patch, HDFS-4564.branch-23.patch, 
> HDFS-4564.patch
>
>
> Webhdfs is returning 401 (Unauthorized) instead of 403 (Forbidden) when it's 
> denying operations.  Examples including rejecting invalid proxy user attempts 
> and renew/cancel with an invalid user.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (HDFS-4564) Webhdfs returns incorrect http response codes for denied operations

2014-02-04 Thread Arun C Murthy (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-4564?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13891086#comment-13891086
 ] 

Arun C Murthy commented on HDFS-4564:
-

[~daryn] - Is this close? Thanks.

> Webhdfs returns incorrect http response codes for denied operations
> ---
>
> Key: HDFS-4564
> URL: https://issues.apache.org/jira/browse/HDFS-4564
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: webhdfs
>Affects Versions: 0.23.0, 2.0.0-alpha, 3.0.0
>Reporter: Daryn Sharp
>Assignee: Daryn Sharp
>Priority: Blocker
> Attachments: HDFS-4564.branch-23.patch, HDFS-4564.branch-23.patch, 
> HDFS-4564.patch
>
>
> Webhdfs is returning 401 (Unauthorized) instead of 403 (Forbidden) when it's 
> denying operations.  Examples including rejecting invalid proxy user attempts 
> and renew/cancel with an invalid user.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (HDFS-4564) Webhdfs returns incorrect http response codes for denied operations

2014-02-04 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4564?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HDFS-4564:


Target Version/s: 2.3.0  (was: 3.0.0, 2.3.0)

> Webhdfs returns incorrect http response codes for denied operations
> ---
>
> Key: HDFS-4564
> URL: https://issues.apache.org/jira/browse/HDFS-4564
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: webhdfs
>Affects Versions: 0.23.0, 2.0.0-alpha, 3.0.0
>Reporter: Daryn Sharp
>Assignee: Daryn Sharp
>Priority: Blocker
> Attachments: HDFS-4564.branch-23.patch, HDFS-4564.branch-23.patch, 
> HDFS-4564.patch
>
>
> Webhdfs is returning 401 (Unauthorized) instead of 403 (Forbidden) when it's 
> denying operations.  Examples including rejecting invalid proxy user attempts 
> and renew/cancel with an invalid user.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (HDFS-4949) Centralized cache management in HDFS

2013-12-23 Thread Arun C Murthy (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-4949?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13855882#comment-13855882
 ] 

Arun C Murthy commented on HDFS-4949:
-

We've been discussing through a proposal where-in we can leverage YARN's 
resource-management *and* workload-management capabilities (via delegation of 
resources, in this case RAM to HDFS) to provide a more general cache 
administration in YARN-1488.

> Centralized cache management in HDFS
> 
>
> Key: HDFS-4949
> URL: https://issues.apache.org/jira/browse/HDFS-4949
> Project: Hadoop HDFS
>  Issue Type: New Feature
>  Components: datanode, namenode
>Affects Versions: 3.0.0, 2.4.0
>Reporter: Andrew Wang
>Assignee: Andrew Wang
> Attachments: HDFS-4949-consolidated.patch, 
> caching-design-doc-2013-07-02.pdf, caching-design-doc-2013-08-09.pdf, 
> caching-design-doc-2013-10-24.pdf, caching-testplan.pdf
>
>
> HDFS currently has no support for managing or exposing in-memory caches at 
> datanodes. This makes it harder for higher level application frameworks like 
> Hive, Pig, and Impala to effectively use cluster memory, because they cannot 
> explicitly cache important datasets or place their tasks for memory locality.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (HDFS-5307) Support both HTTP and HTTPS in jsp pages

2013-10-07 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5307?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HDFS-5307:


Fix Version/s: (was: 2.2.0)
   2.2.1

> Support both HTTP and HTTPS in jsp pages
> 
>
> Key: HDFS-5307
> URL: https://issues.apache.org/jira/browse/HDFS-5307
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>Reporter: Haohui Mai
>Assignee: Haohui Mai
> Fix For: 2.2.1
>
> Attachments: HDFS-5307.000.patch
>
>
> Adopt the new APIs provided by HDFS-5306 to support both HTTP and HTTPS in 
> the jsp pages.



--
This message was sent by Atlassian JIRA
(v6.1#6144)


[jira] [Updated] (HDFS-5317) Go back to DFS Home link does not work on datanode webUI

2013-10-07 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5317?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HDFS-5317:


Fix Version/s: (was: 2.2.0)
   2.2.1

> Go back to DFS Home link does not work on datanode webUI
> 
>
> Key: HDFS-5317
> URL: https://issues.apache.org/jira/browse/HDFS-5317
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>Reporter: Suresh Srinivas
>Assignee: Haohui Mai
>Priority: Critical
> Fix For: 2.2.1
>
> Attachments: HDFS-5317.000.patch, HDFS-5317.001.patch
>
>
> Here is how to reproduce this problem:
> # Goto https://namenode:https_port/dfshealth.jsp
> # Click on "Live Nodes" link
> # On the Live Nodes page, click one of the links for the datanodes. This link 
> has always hardcodes the http link instead of http/https based on the context 
> of the webpage.
> On datanode webUI page, the link "Go back to DFS home" does not work.



--
This message was sent by Atlassian JIRA
(v6.1#6144)


[jira] [Updated] (HDFS-5291) Standby namenode after transition to active goes into safemode

2013-10-07 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5291?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HDFS-5291:


Fix Version/s: (was: 2.2.0)
   2.2.1

> Standby namenode after transition to active goes into safemode
> --
>
> Key: HDFS-5291
> URL: https://issues.apache.org/jira/browse/HDFS-5291
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: ha
>Affects Versions: 2.1.1-beta
>Reporter: Arpit Gupta
>Assignee: Jing Zhao
>Priority: Critical
> Fix For: 2.2.1
>
> Attachments: HDFS-5291.000.patch, HDFS-5291.001.patch, 
> HDFS-5291.002.patch, HDFS-5291.003.patch, HDFS-5291.003.patch, nn.log
>
>
> Some log snippets
> standby state to active transition
> {code}
> 2013-10-02 00:13:49,482 INFO  ipc.Server (Server.java:run(2068)) - IPC Server 
> handler 69 on 8020, call 
> org.apache.hadoop.hdfs.protocol.ClientProtocol.renewLease from IP:33911 
> Call#1483 Retry#1: error: org.apache.hadoop.ipc.StandbyException: Operation 
> category WRITE is not supported in state standby
> 2013-10-02 00:13:49,689 INFO  ipc.Server (Server.java:saslProcess(1342)) - 
> Auth successful for nn/hostn...@example.com (auth:SIMPLE)
> 2013-10-02 00:13:49,696 INFO  authorize.ServiceAuthorizationManager 
> (ServiceAuthorizationManager.java:authorize(111)) - Authorization successful 
> for nn/hostn...@example.com (auth:KERBEROS) for protocol=interface 
> org.apache.hadoop.ha.HAServiceProtocol
> 2013-10-02 00:13:49,700 INFO  namenode.FSNamesystem 
> (FSNamesystem.java:stopStandbyServices(1013)) - Stopping services started for 
> standby state
> 2013-10-02 00:13:49,701 WARN  ha.EditLogTailer 
> (EditLogTailer.java:doWork(336)) - Edit log tailer interrupted
> java.lang.InterruptedException: sleep interrupted
> at java.lang.Thread.sleep(Native Method)
> at 
> org.apache.hadoop.hdfs.server.namenode.ha.EditLogTailer$EditLogTailerThread.doWork(EditLogTailer.java:334)
> at 
> org.apache.hadoop.hdfs.server.namenode.ha.EditLogTailer$EditLogTailerThread.access$200(EditLogTailer.java:279)
> at 
> org.apache.hadoop.hdfs.server.namenode.ha.EditLogTailer$EditLogTailerThread$1.run(EditLogTailer.java:296)
> at java.security.AccessController.doPrivileged(Native Method)
> at javax.security.auth.Subject.doAs(Subject.java:356)
> at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1463)
> at 
> org.apache.hadoop.security.SecurityUtil.doAsLoginUserOrFatal(SecurityUtil.java:454)
> at org.apache.hadoop.hdfs.server.namenode.ha.EditLogTailer$EditLogTail
> 2013-10-02 00:13:49,704 INFO  namenode.FSNamesystem 
> (FSNamesystem.java:startActiveServices(885)) - Starting services required for 
> active state
> 2013-10-02 00:13:49,719 INFO  client.QuorumJournalManager 
> (QuorumJournalManager.java:recoverUnfinalizedSegments(419)) - Starting 
> recovery process for unclosed journal segments...
> 2013-10-02 00:13:49,755 INFO  ipc.Server (Server.java:saslProcess(1342)) - 
> Auth successful for hbase/hostn...@example.com (auth:SIMPLE)
> 2013-10-02 00:13:49,761 INFO  authorize.ServiceAuthorizationManager 
> (ServiceAuthorizationManager.java:authorize(111)) - Authorization successful 
> for hbase/hostn...@example.com (auth:KERBEROS) for protocol=interface 
> org.apache.hadoop.hdfs.protocol.ClientProtocol
> 2013-10-02 00:13:49,839 INFO  client.QuorumJournalManager 
> (QuorumJournalManager.java:recoverUnfinalizedSegments(421)) - Successfully 
> started new epoch 85
> 2013-10-02 00:13:49,839 INFO  client.QuorumJournalManager 
> (QuorumJournalManager.java:recoverUnclosedSegment(249)) - Beginning recovery 
> of unclosed segment starting at txid 887112
> 2013-10-02 00:13:49,874 INFO  client.QuorumJournalManager 
> (QuorumJournalManager.java:recoverUnclosedSegment(258)) - Recovery prepare 
> phase complete. Responses:
> IP:8485: segmentState { startTxId: 887112 endTxId: 887531 isInProgress: true 
> } lastWriterEpoch: 84 lastCommittedTxId: 887530
> 172.18.145.97:8485: segmentState { startTxId: 887112 endTxId: 887531 
> isInProgress: true } lastWriterEpoch: 84 lastCommittedTxId: 887530
> 2013-10-02 00:13:49,875 INFO  client.QuorumJournalManager 
> (QuorumJournalManager.java:recover
> {code}
> And then we get into safemode
> {code}
> Construction[IP:1019|RBW]]} size 0
> 2013-10-02 00:13:50,277 INFO  BlockStateChange 
> (BlockManager.java:logAddStoredBlock(2237)) - BLOCK* addStoredBlock: blockMap 
> updated: IP:1019 is added to blk_IP157{blockUCState=UNDER_CONSTRUCTION, 
> primaryNodeIndex=-1, replicas=[ReplicaUnderConstruction[IP:1019|RBW], 
> ReplicaUnderConstruction[172.18.145.96:1019|RBW], ReplicaUnde
> rConstruction[IP:1019|RBW]]} size 0
> 2013-10-02 00:13:50,279 INFO  hdfs.StateChange 
> (FSNamesystem.

[jira] [Commented] (HDFS-4817) make HDFS advisory caching configurable on a per-file basis

2013-10-06 Thread Arun C Murthy (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-4817?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13787917#comment-13787917
 ] 

Arun C Murthy commented on HDFS-4817:
-

[~cmccabe] In future, if you backport patches, please ensure you reflect it 
appropriately in all branches.

For e.g. 

HDFS-4817.  Make HDFS advisory caching configurable on a per-file basis.
(Colin Patrick McCabe)

shows up in Release 2.3.0 as an Improvement in trunk/branch-2 while it shows up 
as a New Feature in branch-2.1-beta.

This will reduce the burden on making releases, makes it consistent for folks 
to figure release etc. Thanks.

> make HDFS advisory caching configurable on a per-file basis
> ---
>
> Key: HDFS-4817
> URL: https://issues.apache.org/jira/browse/HDFS-4817
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: hdfs-client
>Affects Versions: 3.0.0
>Reporter: Colin Patrick McCabe
>Assignee: Colin Patrick McCabe
>Priority: Minor
> Fix For: 2.2.0
>
> Attachments: HDFS-4817.001.patch, HDFS-4817.002.patch, 
> HDFS-4817.004.patch, HDFS-4817.006.patch, HDFS-4817.007.patch, 
> HDFS-4817.008.patch, HDFS-4817.009.patch, HDFS-4817.010.patch, 
> HDFS-4817-b2.1.001.patch
>
>
> HADOOP-7753 and related JIRAs introduced some performance optimizations for 
> the DataNode.  One of them was readahead.  When readahead is enabled, the 
> DataNode starts reading the next bytes it thinks it will need in the block 
> file, before the client requests them.  This helps hide the latency of 
> rotational media and send larger reads down to the device.  Another 
> optimization was "drop-behind."  Using this optimization, we could remove 
> files from the Linux page cache after they were no longer needed.
> Using {{dfs.datanode.drop.cache.behind.writes}} and 
> {{dfs.datanode.drop.cache.behind.reads}} can improve performance  
> substantially on many MapReduce jobs.  In our internal benchmarks, we have 
> seen speedups of 40% on certain workloads.  The reason is because if we know 
> the block data will not be read again any time soon, keeping it out of memory 
> allows more memory to be used by the other processes on the system.  See 
> HADOOP-7714 for more benchmarks.
> We would like to turn on these configurations on a per-file or per-client 
> basis, rather than on the DataNode as a whole.  This will allow more users to 
> actually make use of them.  It would also be good to add unit tests for the 
> drop-cache code path, to ensure that it is functioning as we expect.



--
This message was sent by Atlassian JIRA
(v6.1#6144)


[jira] [Updated] (HDFS-5251) Race between the initialization of NameNode and the http server

2013-09-24 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5251?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HDFS-5251:


Fix Version/s: (was: 2.2.0)
   2.1.2-beta

> Race between the initialization of NameNode and the http server
> ---
>
> Key: HDFS-5251
> URL: https://issues.apache.org/jira/browse/HDFS-5251
> Project: Hadoop HDFS
>  Issue Type: Bug
>Reporter: Haohui Mai
>Assignee: Haohui Mai
> Fix For: 2.1.2-beta
>
> Attachments: HDFS-5251.000.patch
>
>
> The constructor of NameNode starts a HTTP server before the FSNameSystem is 
> initialized. Currently there is a race where the HTTP server can access the 
> uninitialized namesystem variable, throwing a NullPointerException.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-5139) Remove redundant -R option from setrep

2013-09-24 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5139?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HDFS-5139:


Fix Version/s: (was: 2.2.0)
   2.1.2-beta

> Remove redundant -R option from setrep
> --
>
> Key: HDFS-5139
> URL: https://issues.apache.org/jira/browse/HDFS-5139
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: tools
>Affects Versions: 3.0.0, 1.3.0
>Reporter: Arpit Agarwal
>Assignee: Arpit Agarwal
> Fix For: 2.1.2-beta
>
> Attachments: HDFS-5139.01.patch, HDFS-5139.02.patch, 
> HDFS-5139.03.patch, HDFS-5139.04.patch
>
>
> The -R option to setrep is redundant because it is required for directory 
> targets and ignored for file targets.
> We can just remove the option and make -R the default for directories.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-4754) Add an API in the namenode to mark a datanode as stale

2013-09-24 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4754?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HDFS-4754:


Fix Version/s: (was: 2.1.1-beta)
   2.1.2-beta

> Add an API in the namenode to mark a datanode as stale
> --
>
> Key: HDFS-4754
> URL: https://issues.apache.org/jira/browse/HDFS-4754
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: hdfs-client, namenode
>Reporter: Nicolas Liochon
>Assignee: Nicolas Liochon
>Priority: Critical
> Fix For: 3.0.0, 2.1.2-beta
>
> Attachments: 4754.v1.patch, 4754.v2.patch, 4754.v4.patch, 
> 4754.v4.patch
>
>
> There is a detection of the stale datanodes in HDFS since HDFS-3703, with a 
> timeout, defaulted to 30s.
> There are two reasons to add an API to mark a node as stale even if the 
> timeout is not yet reached:
>  1) ZooKeeper can detect that a client is dead at any moment. So, for HBase, 
> we sometimes start the recovery before a node is marked staled. (even with 
> reasonable settings as: stale: 20s; HBase ZK timeout: 30s
>  2) Some third parties could detect that a node is dead before the timeout, 
> hence saving us the cost of retrying. An example or such hw is Arista, 
> presented here by [~tsuna] 
> http://tsunanet.net/~tsuna/fsf-hbase-meetup-april13.pdf, and confirmed in 
> HBASE-6290.
> As usual, even if the node is dead it can comeback before the 10 minutes 
> limit. So I would propose to set a timebound. The API would be
> namenode.markStale(String ipAddress, int port, long durationInMs);
> After durationInMs, the namenode would again rely only on its heartbeat to 
> decide.
> Thoughts?
> If there is no objections, and if nobody in the hdfs dev team has the time to 
> spend some time on it, I will give it a try for branch 2 & 3.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-5098) Enhance FileSystem.Statistics to have locality information

2013-09-24 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5098?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HDFS-5098:


Fix Version/s: (was: 2.1.1-beta)
   2.1.2-beta

> Enhance FileSystem.Statistics to have locality information
> --
>
> Key: HDFS-5098
> URL: https://issues.apache.org/jira/browse/HDFS-5098
> Project: Hadoop HDFS
>  Issue Type: Improvement
>Reporter: Bikas Saha
>Assignee: Suresh Srinivas
> Fix For: 2.1.2-beta
>
>
> Currently in MR/Tez we dont have a good and accurate means to detect how much 
> the the IO was actually done locally. Getting this information from the 
> source of truth would be much better.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-5244) TestNNStorageRetentionManager#testPurgeMultipleDirs fails because incorrectly expects Hashmap values to have order.

2013-09-24 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5244?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HDFS-5244:


Fix Version/s: (was: 2.1.1-beta)
   2.1.2-beta

> TestNNStorageRetentionManager#testPurgeMultipleDirs fails because incorrectly 
> expects Hashmap values to have order. 
> 
>
> Key: HDFS-5244
> URL: https://issues.apache.org/jira/browse/HDFS-5244
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: test
>Affects Versions: 2.1.0-beta
> Environment: Red Hat Enterprise 6 with Sun Java 1.7 and IBM java 1.6
>Reporter: Jinghui Wang
> Fix For: 3.0.0, 2.1.0-beta, 2.1.2-beta
>
> Attachments: HDFS-5244.patch
>
>
> The test o.a.h.hdfs.server.namenode.TestNNStorageRetentionManager uses a 
> HashMap(dirRoots) to store the root storages to be mocked for the purging 
> test, which does not have any predictable order. The directories needs be 
> purged are stored in a LinkedHashSet, which has a predictable order. So, when 
> the directories get mocked for the test, they could be already out of
> the order that they were added. Thus, the order that the directories were
> actually purged and the order of them being added to the LinkedHashList could
> be different and cause the test to fail.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-5160) MiniDFSCluster webui does not work

2013-09-24 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5160?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HDFS-5160:


Fix Version/s: (was: 2.1.1-beta)
   2.1.2-beta

> MiniDFSCluster webui does not work
> --
>
> Key: HDFS-5160
> URL: https://issues.apache.org/jira/browse/HDFS-5160
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: test
>Affects Versions: 2.1.0-beta
>Reporter: Alejandro Abdelnur
> Fix For: 2.1.2-beta
>
>
> The webui does not work, when going to http://localhost:50070 you get:
> {code}
> Directory: /
> webapps/  102 bytes   Sep 4, 2013 9:32:55 AM
> {code}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-5246) Hadoop nfs server binds to port 2049 which is the same as Linux nfs server

2013-09-24 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5246?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HDFS-5246:


Fix Version/s: (was: 2.1.1-beta)
   2.1.2-beta

> Hadoop nfs server binds to port 2049 which is the same as Linux nfs server
> --
>
> Key: HDFS-5246
> URL: https://issues.apache.org/jira/browse/HDFS-5246
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: nfs
>Affects Versions: 2.1.0-beta
> Environment: Red Hat Enterprise 6 with Sun Java 1.7 and IBM Java 1.6
>Reporter: Jinghui Wang
> Fix For: 3.0.0, 2.1.2-beta
>
> Attachments: HDFS-5246-2.patch, HDFS-5246-3.patch, HDFS-5246.patch
>
>
> Hadoop nfs binds the nfs server to port 2049,
> which is also the default port that Linux nfs uses. If Linux nfs is already 
> running on the machine then Hadoop nfs will not be albe to start.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-5228) The RemoteIterator returned by DistributedFileSystem.listFiles(..) may throw NPE

2013-09-24 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5228?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HDFS-5228:


Fix Version/s: (was: 2.1.1-beta)
   2.1.2-beta

> The RemoteIterator returned by DistributedFileSystem.listFiles(..) may throw 
> NPE
> 
>
> Key: HDFS-5228
> URL: https://issues.apache.org/jira/browse/HDFS-5228
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: hdfs-client
>Affects Versions: 2.1.0-beta
>Reporter: Tsz Wo (Nicholas), SZE
>Assignee: Tsz Wo (Nicholas), SZE
>Priority: Blocker
> Fix For: 2.1.2-beta
>
> Attachments: h5228_20130919.patch, h5228_20130919_test.patch, 
> HDFS-5228.2.patch
>
>
> Get a RemoteIterator from DistributedFileSystem.listFiles(..) with a relative 
> path.  Then, it will result a NullPointerException when calling hasNext() 
> from the RemoteIterator.
> This bug was discovered by Arnaud:
> http://hortonworks.com/community/forums/topic/new-bug-in-hdfs-listfiles-method/

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-5225) datanode keeps logging the same 'is no longer in the dataset' message over and over again

2013-09-24 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5225?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HDFS-5225:


Target Version/s: 2.1.2-beta

> datanode keeps logging the same 'is no longer in the dataset' message over 
> and over again
> -
>
> Key: HDFS-5225
> URL: https://issues.apache.org/jira/browse/HDFS-5225
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: datanode
>Affects Versions: 2.1.1-beta
>Reporter: Roman Shaposhnik
>Priority: Blocker
> Attachments: HDFS-5225.1.patch, HDFS-5225-reproduce.1.txt
>
>
> I was running the usual Bigtop testing on 2.1.1-beta RC1 with the following 
> configuration: 4 nodes fully distributed cluster with security on.
> All of a sudden my DN ate up all of the space in /var/log logging the 
> following message repeatedly:
> {noformat}
> 2013-09-18 20:51:12,046 INFO 
> org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner: 
> BP-1884637155-10.224.158.152-1379524544853:blk_1073742189_1369 is no longer 
> in the dataset
> {noformat}
> It wouldn't answer to a jstack and jstack -F ended up being useless.
> Here's what I was able to find in the NameNode logs regarding this block ID:
> {noformat}
> fgrep -rI 'blk_1073742189' hadoop-hdfs-namenode-ip-10-224-158-152.log
> 2013-09-18 18:03:16,972 INFO org.apache.hadoop.hdfs.StateChange: BLOCK* 
> allocateBlock: 
> /user/jenkins/testAppendInputWedSep18180222UTC2013/test4.fileWedSep18180222UTC2013._COPYING_.
>  BP-1884637155-10.224.158.152-1379524544853 
> blk_1073742189_1369{blockUCState=UNDER_CONSTRUCTION, primaryNodeIndex=-1, 
> replicas=[ReplicaUnderConstruction[10.83.107.80:1004|RBW], 
> ReplicaUnderConstruction[10.34.74.206:1004|RBW], 
> ReplicaUnderConstruction[10.224.158.152:1004|RBW]]}
> 2013-09-18 18:03:17,222 INFO BlockStateChange: BLOCK* addStoredBlock: 
> blockMap updated: 10.224.158.152:1004 is added to 
> blk_1073742189_1369{blockUCState=UNDER_CONSTRUCTION, primaryNodeIndex=-1, 
> replicas=[ReplicaUnderConstruction[10.83.107.80:1004|RBW], 
> ReplicaUnderConstruction[10.34.74.206:1004|RBW], 
> ReplicaUnderConstruction[10.224.158.152:1004|RBW]]} size 0
> 2013-09-18 18:03:17,222 INFO BlockStateChange: BLOCK* addStoredBlock: 
> blockMap updated: 10.34.74.206:1004 is added to 
> blk_1073742189_1369{blockUCState=UNDER_CONSTRUCTION, primaryNodeIndex=-1, 
> replicas=[ReplicaUnderConstruction[10.83.107.80:1004|RBW], 
> ReplicaUnderConstruction[10.34.74.206:1004|RBW], 
> ReplicaUnderConstruction[10.224.158.152:1004|RBW]]} size 0
> 2013-09-18 18:03:17,224 INFO BlockStateChange: BLOCK* addStoredBlock: 
> blockMap updated: 10.83.107.80:1004 is added to 
> blk_1073742189_1369{blockUCState=UNDER_CONSTRUCTION, primaryNodeIndex=-1, 
> replicas=[ReplicaUnderConstruction[10.83.107.80:1004|RBW], 
> ReplicaUnderConstruction[10.34.74.206:1004|RBW], 
> ReplicaUnderConstruction[10.224.158.152:1004|RBW]]} size 0
> 2013-09-18 18:03:17,899 INFO 
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: 
> updatePipeline(block=BP-1884637155-10.224.158.152-1379524544853:blk_1073742189_1369,
>  newGenerationStamp=1370, newLength=1048576, newNodes=[10.83.107.80:1004, 
> 10.34.74.206:1004, 10.224.158.152:1004], 
> clientName=DFSClient_NONMAPREDUCE_-450304083_1)
> 2013-09-18 18:03:17,904 INFO 
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: 
> updatePipeline(BP-1884637155-10.224.158.152-1379524544853:blk_1073742189_1369)
>  successfully to 
> BP-1884637155-10.224.158.152-1379524544853:blk_1073742189_1370
> 2013-09-18 18:03:18,540 INFO 
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: 
> updatePipeline(block=BP-1884637155-10.224.158.152-1379524544853:blk_1073742189_1370,
>  newGenerationStamp=1371, newLength=2097152, newNodes=[10.83.107.80:1004, 
> 10.34.74.206:1004, 10.224.158.152:1004], 
> clientName=DFSClient_NONMAPREDUCE_-450304083_1)
> 2013-09-18 18:03:18,548 INFO 
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: 
> updatePipeline(BP-1884637155-10.224.158.152-1379524544853:blk_1073742189_1370)
>  successfully to 
> BP-1884637155-10.224.158.152-1379524544853:blk_1073742189_1371
> 2013-09-18 18:03:26,150 INFO BlockStateChange: BLOCK* addToInvalidates: 
> blk_1073742189_1371 10.83.107.80:1004 10.34.74.206:1004 10.224.158.152:1004 
> 2013-09-18 18:03:26,847 INFO org.apache.hadoop.hdfs.StateChange: BLOCK* 
> InvalidateBlocks: ask 10.34.74.206:1004 to delete [blk_1073742178_1359, 
> blk_1073742183_1362, blk_1073742184_1363, blk_1073742186_1366, 
> blk_1073742188_1368, blk_1073742189_1371]
> 2013-09-18 18:03:29,848 INFO org.apache.hadoop.hdfs.StateChange: BLOCK* 
> InvalidateBlocks: ask 10.224.158.152:1004 to delete [blk_1073742177_1353, 
> blk_1073742178_1359, blk_1073742179_1355, blk_1073742180_1356, 
> blk_1073742181_1358, blk_1073742182_1361, blk_10737

[jira] [Updated] (HDFS-5225) datanode keeps logging the same 'is no longer in the dataset' message over and over again

2013-09-24 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5225?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HDFS-5225:


Priority: Blocker  (was: Major)

> datanode keeps logging the same 'is no longer in the dataset' message over 
> and over again
> -
>
> Key: HDFS-5225
> URL: https://issues.apache.org/jira/browse/HDFS-5225
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: datanode
>Affects Versions: 2.1.1-beta
>Reporter: Roman Shaposhnik
>Priority: Blocker
> Attachments: HDFS-5225.1.patch, HDFS-5225-reproduce.1.txt
>
>
> I was running the usual Bigtop testing on 2.1.1-beta RC1 with the following 
> configuration: 4 nodes fully distributed cluster with security on.
> All of a sudden my DN ate up all of the space in /var/log logging the 
> following message repeatedly:
> {noformat}
> 2013-09-18 20:51:12,046 INFO 
> org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner: 
> BP-1884637155-10.224.158.152-1379524544853:blk_1073742189_1369 is no longer 
> in the dataset
> {noformat}
> It wouldn't answer to a jstack and jstack -F ended up being useless.
> Here's what I was able to find in the NameNode logs regarding this block ID:
> {noformat}
> fgrep -rI 'blk_1073742189' hadoop-hdfs-namenode-ip-10-224-158-152.log
> 2013-09-18 18:03:16,972 INFO org.apache.hadoop.hdfs.StateChange: BLOCK* 
> allocateBlock: 
> /user/jenkins/testAppendInputWedSep18180222UTC2013/test4.fileWedSep18180222UTC2013._COPYING_.
>  BP-1884637155-10.224.158.152-1379524544853 
> blk_1073742189_1369{blockUCState=UNDER_CONSTRUCTION, primaryNodeIndex=-1, 
> replicas=[ReplicaUnderConstruction[10.83.107.80:1004|RBW], 
> ReplicaUnderConstruction[10.34.74.206:1004|RBW], 
> ReplicaUnderConstruction[10.224.158.152:1004|RBW]]}
> 2013-09-18 18:03:17,222 INFO BlockStateChange: BLOCK* addStoredBlock: 
> blockMap updated: 10.224.158.152:1004 is added to 
> blk_1073742189_1369{blockUCState=UNDER_CONSTRUCTION, primaryNodeIndex=-1, 
> replicas=[ReplicaUnderConstruction[10.83.107.80:1004|RBW], 
> ReplicaUnderConstruction[10.34.74.206:1004|RBW], 
> ReplicaUnderConstruction[10.224.158.152:1004|RBW]]} size 0
> 2013-09-18 18:03:17,222 INFO BlockStateChange: BLOCK* addStoredBlock: 
> blockMap updated: 10.34.74.206:1004 is added to 
> blk_1073742189_1369{blockUCState=UNDER_CONSTRUCTION, primaryNodeIndex=-1, 
> replicas=[ReplicaUnderConstruction[10.83.107.80:1004|RBW], 
> ReplicaUnderConstruction[10.34.74.206:1004|RBW], 
> ReplicaUnderConstruction[10.224.158.152:1004|RBW]]} size 0
> 2013-09-18 18:03:17,224 INFO BlockStateChange: BLOCK* addStoredBlock: 
> blockMap updated: 10.83.107.80:1004 is added to 
> blk_1073742189_1369{blockUCState=UNDER_CONSTRUCTION, primaryNodeIndex=-1, 
> replicas=[ReplicaUnderConstruction[10.83.107.80:1004|RBW], 
> ReplicaUnderConstruction[10.34.74.206:1004|RBW], 
> ReplicaUnderConstruction[10.224.158.152:1004|RBW]]} size 0
> 2013-09-18 18:03:17,899 INFO 
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: 
> updatePipeline(block=BP-1884637155-10.224.158.152-1379524544853:blk_1073742189_1369,
>  newGenerationStamp=1370, newLength=1048576, newNodes=[10.83.107.80:1004, 
> 10.34.74.206:1004, 10.224.158.152:1004], 
> clientName=DFSClient_NONMAPREDUCE_-450304083_1)
> 2013-09-18 18:03:17,904 INFO 
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: 
> updatePipeline(BP-1884637155-10.224.158.152-1379524544853:blk_1073742189_1369)
>  successfully to 
> BP-1884637155-10.224.158.152-1379524544853:blk_1073742189_1370
> 2013-09-18 18:03:18,540 INFO 
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: 
> updatePipeline(block=BP-1884637155-10.224.158.152-1379524544853:blk_1073742189_1370,
>  newGenerationStamp=1371, newLength=2097152, newNodes=[10.83.107.80:1004, 
> 10.34.74.206:1004, 10.224.158.152:1004], 
> clientName=DFSClient_NONMAPREDUCE_-450304083_1)
> 2013-09-18 18:03:18,548 INFO 
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: 
> updatePipeline(BP-1884637155-10.224.158.152-1379524544853:blk_1073742189_1370)
>  successfully to 
> BP-1884637155-10.224.158.152-1379524544853:blk_1073742189_1371
> 2013-09-18 18:03:26,150 INFO BlockStateChange: BLOCK* addToInvalidates: 
> blk_1073742189_1371 10.83.107.80:1004 10.34.74.206:1004 10.224.158.152:1004 
> 2013-09-18 18:03:26,847 INFO org.apache.hadoop.hdfs.StateChange: BLOCK* 
> InvalidateBlocks: ask 10.34.74.206:1004 to delete [blk_1073742178_1359, 
> blk_1073742183_1362, blk_1073742184_1363, blk_1073742186_1366, 
> blk_1073742188_1368, blk_1073742189_1371]
> 2013-09-18 18:03:29,848 INFO org.apache.hadoop.hdfs.StateChange: BLOCK* 
> InvalidateBlocks: ask 10.224.158.152:1004 to delete [blk_1073742177_1353, 
> blk_1073742178_1359, blk_1073742179_1355, blk_1073742180_1356, 
> blk_1073742181_1358, blk_1073742182_1361, blk_10

[jira] [Updated] (HDFS-5138) Support HDFS upgrade in HA

2013-09-16 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5138?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HDFS-5138:


Target Version/s:   (was: 2.1.1-beta)

> Support HDFS upgrade in HA
> --
>
> Key: HDFS-5138
> URL: https://issues.apache.org/jira/browse/HDFS-5138
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 2.1.1-beta
>Reporter: Kihwal Lee
>Priority: Blocker
>
> With HA enabled, NN wo't start with "-upgrade". Since there has been a layout 
> version change between 2.0.x and 2.1.x, starting NN in upgrade mode was 
> necessary when deploying 2.1.x to an existing 2.0.x cluster. But the only way 
> to get around this was to disable HA and upgrade. 
> The NN and the cluster cannot be flipped back to HA until the upgrade is 
> finalized. If HA is disabled only on NN for layout upgrade and HA is turned 
> back on without involving DNs, things will work, but finaliizeUpgrade won't 
> work (the NN is in HA and it cannot be in upgrade mode) and DN's upgrade 
> snapshots won't get removed.
> We will need a different ways of doing layout upgrade and upgrade snapshot.  
> I am marking this as a 2.1.1-beta blocker based on feedback from others.  If 
> there is a reasonable workaround that does not increase maintenance window 
> greatly, we can lower its priority from blocker to critical.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-3983) Hftp should support both SPNEGO and KSSL

2013-09-16 Thread Arun C Murthy (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3983?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13768351#comment-13768351
 ] 

Arun C Murthy commented on HDFS-3983:
-

Any update on this? Is this ready to go? Thanks.

> Hftp should support both SPNEGO and KSSL
> 
>
> Key: HDFS-3983
> URL: https://issues.apache.org/jira/browse/HDFS-3983
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: security
>Affects Versions: 2.0.0-alpha
>Reporter: Eli Collins
>Assignee: Eli Collins
>Priority: Blocker
> Attachments: hdfs-3983.txt, hdfs-3983.txt
>
>
> Hftp currently doesn't work against a secure cluster unless you configure 
> {{dfs.https.port}} to be the http port, otherwise the client can't fetch 
> tokens:
> {noformat}
> $ hadoop fs -ls hftp://c1225.hal.cloudera.com:50070/
> 12/09/26 18:02:00 INFO fs.FileSystem: Couldn't get a delegation token from 
> http://c1225.hal.cloudera.com:50470 using http.
> ls: Security enabled but user not authenticated by filter
> {noformat}
> This is due to Hftp still using the https port. Post HDFS-2617 it should use 
> the regular http port. Hsftp should still use the secure port, however now 
> that we have HADOOP-8581 it's worth considering removing Hsftp entirely. I'll 
> start a separate thread about that.  

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Resolved] (HDFS-4952) dfs -ls hftp:// fails on secure hadoop2 cluster

2013-09-13 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4952?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy resolved HDFS-4952.
-

Resolution: Fixed

Duplicate of HDFS-3983

> dfs -ls hftp:// fails on secure hadoop2 cluster
> ---
>
> Key: HDFS-4952
> URL: https://issues.apache.org/jira/browse/HDFS-4952
> Project: Hadoop HDFS
>  Issue Type: Bug
>Reporter: Yesha Vora
>Priority: Blocker
>
> Running :hadoop dfs -ls hftp://:/A
> WARN fs.FileSystem: Couldn't connect to http://:50470, assuming 
> security is disabled
> ls: Security enabled but user not authenticated by filter

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-4952) dfs -ls hftp:// fails on secure hadoop2 cluster

2013-09-13 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4952?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HDFS-4952:


Target Version/s:   (was: 2.1.1-beta)

> dfs -ls hftp:// fails on secure hadoop2 cluster
> ---
>
> Key: HDFS-4952
> URL: https://issues.apache.org/jira/browse/HDFS-4952
> Project: Hadoop HDFS
>  Issue Type: Bug
>Reporter: Yesha Vora
>Priority: Blocker
>
> Running :hadoop dfs -ls hftp://:/A
> WARN fs.FileSystem: Couldn't connect to http://:50470, assuming 
> security is disabled
> ls: Security enabled but user not authenticated by filter

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-5138) Support HDFS upgrade in HA

2013-09-12 Thread Arun C Murthy (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5138?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13765519#comment-13765519
 ] 

Arun C Murthy commented on HDFS-5138:
-

Guys, any luck? I'd like to get 2.1.1 out soon... thanks!

> Support HDFS upgrade in HA
> --
>
> Key: HDFS-5138
> URL: https://issues.apache.org/jira/browse/HDFS-5138
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 2.1.1-beta
>Reporter: Kihwal Lee
>Priority: Blocker
>
> With HA enabled, NN wo't start with "-upgrade". Since there has been a layout 
> version change between 2.0.x and 2.1.x, starting NN in upgrade mode was 
> necessary when deploying 2.1.x to an existing 2.0.x cluster. But the only way 
> to get around this was to disable HA and upgrade. 
> The NN and the cluster cannot be flipped back to HA until the upgrade is 
> finalized. If HA is disabled only on NN for layout upgrade and HA is turned 
> back on without involving DNs, things will work, but finaliizeUpgrade won't 
> work (the NN is in HA and it cannot be in upgrade mode) and DN's upgrade 
> snapshots won't get removed.
> We will need a different ways of doing layout upgrade and upgrade snapshot.  
> I am marking this as a 2.1.1-beta blocker based on feedback from others.  If 
> there is a reasonable workaround that does not increase maintenance window 
> greatly, we can lower its priority from blocker to critical.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-3981) access time is set without holding FSNamesystem write lock

2013-08-27 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3981?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HDFS-3981:


Fix Version/s: (was: 2.1.0-beta)
   2.3.0

> access time is set without holding FSNamesystem write lock
> --
>
> Key: HDFS-3981
> URL: https://issues.apache.org/jira/browse/HDFS-3981
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: namenode
>Affects Versions: 0.23.3, 2.0.3-alpha, 0.23.5
>Reporter: Xiaobo Peng
>Assignee: Xiaobo Peng
> Fix For: 3.0.0, 2.3.0
>
> Attachments: HDFS-3981-branch-0.23.4.patch, 
> HDFS-3981-branch-0.23.patch, HDFS-3981-branch-2.patch, HDFS-3981-trunk.patch, 
> hdfs-3981.txt
>
>
> Incorrect condition in {{FSNamesystem.getBlockLocatoins()}} can lead to 
> updating times without write lock. In most cases this condition will force 
> {{FSNamesystem.getBlockLocatoins()}} to hold write lock, even if times do not 
> need to be updated.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-3761) Uses of AuthenticatedURL should set the configured timeouts

2013-08-27 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3761?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HDFS-3761:


Fix Version/s: (was: 2.1.0-beta)
   2.3.0

> Uses of AuthenticatedURL should set the configured timeouts
> ---
>
> Key: HDFS-3761
> URL: https://issues.apache.org/jira/browse/HDFS-3761
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 2.0.0-alpha, 2.0.2-alpha
>Reporter: Alejandro Abdelnur
>Priority: Critical
> Fix For: 2.3.0
>
>
> Similar to {{URLUtils}}, the {{AuthenticatedURL}} should set the configured 
> timeouts in URL connection instances. HADOOP-8644 is introducing a 
> {{ConnectionConfigurator}} interface to the {{AuthenticatedURL}} class, it 
> should be done via that interface.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-4198) the shell script error for Cygwin on windows7

2013-08-27 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4198?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HDFS-4198:


Fix Version/s: (was: 2.1.0-beta)
   2.3.0

> the shell script error for Cygwin on windows7
> -
>
> Key: HDFS-4198
> URL: https://issues.apache.org/jira/browse/HDFS-4198
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: scripts
>Affects Versions: 2.0.2-alpha
> Environment: windows7 ,cygwin.
>Reporter: Han Hui Wen 
> Fix For: 2.3.0
>
>
> See the following 
> [comment|https://issues.apache.org/jira/browse/HDFS-4198?focusedCommentId=13498818&page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-13498818]
>  for detailed description.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-4066) Fix compile warnings for libwebhdfs

2013-08-27 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4066?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HDFS-4066:


Fix Version/s: (was: 2.1.0-beta)
   2.3.0

> Fix compile warnings for libwebhdfs
> ---
>
> Key: HDFS-4066
> URL: https://issues.apache.org/jira/browse/HDFS-4066
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: webhdfs
>Affects Versions: 3.0.0
>Reporter: Jing Zhao
>Assignee: Jing Zhao
> Fix For: 2.3.0
>
>
> The compile of libwebhdfs still generates a bunch of warnings, which need to 
> be fixed.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-4940) namenode OOMs under Bigtop's TestCLI

2013-08-27 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4940?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HDFS-4940:


Fix Version/s: (was: 2.1.0-beta)
   2.3.0

> namenode OOMs under Bigtop's TestCLI
> 
>
> Key: HDFS-4940
> URL: https://issues.apache.org/jira/browse/HDFS-4940
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 2.1.0-beta
>Reporter: Roman Shaposhnik
>Assignee: Roman Shaposhnik
>Priority: Blocker
> Fix For: 2.3.0
>
>
> Bigtop's TestCLI when executed against Hadoop 2.1.0 seems to make it OOM 
> quite reliably regardless of the heap size settings. I'm attaching a heap 
> dump URL. Alliteratively anybody can just take Bigtop's tests, compiled them 
> against Hadoop 2.1.0 bits and try to reproduce it.  

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-3691) UserGroupInformation#spawnAutoRenewalThreadForUserCreds tries to renew even if the kerberos ticket cache is non-renewable

2013-08-27 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3691?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HDFS-3691:


Fix Version/s: (was: 2.1.0-beta)
   2.3.0

> UserGroupInformation#spawnAutoRenewalThreadForUserCreds tries to renew even 
> if the kerberos ticket cache is non-renewable
> -
>
> Key: HDFS-3691
> URL: https://issues.apache.org/jira/browse/HDFS-3691
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: namenode
>Affects Versions: 2.0.1-alpha
>Reporter: Colin Patrick McCabe
>Priority: Minor
> Fix For: 2.3.0
>
>
> UserGroupInformation#spawnAutoRenewalThreadForUserCreds tries to renew user 
> credentials.  However, it does this even if the kerberos ticket cache in 
> question is non-renewable.
> This leads to an annoying error message being printed out all the time.
> {code}
> cmccabe@keter:/h> klist
> Ticket cache: FILE:/tmp/krb5cc_1014
> Default principal: hdfs/ke...@cloudera.com
> Valid starting ExpiresService principal
> 07/18/12 15:24:15  07/19/12 15:24:13  krbtgt/cloudera@cloudera.com
> {code}
> {code}
> cmccabe@keter:/h> ./bin/hadoop fs -ls /
> 15:21:39,882  WARN UserGroupInformation:739 - Exception encountered while 
> running the renewal command. Aborting renew thread. 
> org.apache.hadoop.util.Shell$ExitCodeException: kinit: KDC can't fulfill 
> requested option while renewing credentials
> Found 3 items
> -rw-r--r--   3 cmccabe users   0 2012-07-09 17:15 /b
> -rw-r--r--   3 hdfssupergroup  0 2012-07-09 17:17 /c
> drwxrwxrwx   - cmccabe audio   0 2012-07-19 11:25 /tmp
> {code}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-3828) Block Scanner rescans blocks too frequently

2013-08-27 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3828?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HDFS-3828:


Fix Version/s: (was: 2.1.0-beta)
   2.3.0

> Block Scanner rescans blocks too frequently
> ---
>
> Key: HDFS-3828
> URL: https://issues.apache.org/jira/browse/HDFS-3828
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 0.23.0
>Reporter: Andy Isaacson
>Assignee: Andy Isaacson
> Fix For: 2.3.0
>
> Attachments: hdfs-3828-1.txt, hdfs-3828-2.txt, hdfs-3828-3.txt, 
> hdfs3828.txt
>
>
> {{BlockPoolSliceScanner#scan}} calls cleanUp every time it's invoked from 
> {{DataBlockScanner#run}} via {{scanBlockPoolSlice}}.  But cleanUp 
> unconditionally roll()s the verificationLogs, so after two iterations we have 
> lost the first iteration of block verification times.  As a result a cluster 
> with just one block repeatedly rescans it every 10 seconds:
> {noformat}
> 2012-08-16 15:59:57,884 INFO  datanode.BlockPoolSliceScanner 
> (BlockPoolSliceScanner.java:verifyBlock(391)) - Verification succeeded for 
> BP-2101131164-172.29.122.91-1337906886255:blk_7919273167187535506_4915
> 2012-08-16 16:00:07,904 INFO  datanode.BlockPoolSliceScanner 
> (BlockPoolSliceScanner.java:verifyBlock(391)) - Verification succeeded for 
> BP-2101131164-172.29.122.91-1337906886255:blk_7919273167187535506_4915
> 2012-08-16 16:00:17,925 INFO  datanode.BlockPoolSliceScanner 
> (BlockPoolSliceScanner.java:verifyBlock(391)) - Verification succeeded for 
> BP-2101131164-172.29.122.91-1337906886255:blk_7919273167187535506_4915
> {noformat}
> {quote}
> To fix this, we need to avoid roll()ing the logs multiple times per period.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-2802) Support for RW/RO snapshots in HDFS

2013-08-27 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-2802?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HDFS-2802:


Fix Version/s: (was: 2.1.0-beta)
   2.3.0

> Support for RW/RO snapshots in HDFS
> ---
>
> Key: HDFS-2802
> URL: https://issues.apache.org/jira/browse/HDFS-2802
> Project: Hadoop HDFS
>  Issue Type: New Feature
>  Components: namenode
>Reporter: Hari Mankude
>Assignee: Tsz Wo (Nicholas), SZE
> Fix For: 2.3.0
>
> Attachments: 2802.diff, 2802.patch, 2802.patch, h2802_20130417.patch, 
> h2802_20130422.patch, h2802_20130423.patch, h2802_20130425.patch, 
> h2802_20130426.patch, h2802_20130430b.patch, h2802_20130430.patch, 
> h2802_20130508.patch, h2802_20130509.patch, HDFS-2802.20121101.patch, 
> HDFS-2802.branch-2.0510.patch, HDFS-2802-meeting-minutes-121101.txt, 
> HDFSSnapshotsDesign.pdf, snap.patch, snapshot-design.pdf, 
> snapshot-design.tex, snapshot-one-pager.pdf, Snapshots20120429.pdf, 
> Snapshots20121018.pdf, Snapshots20121030.pdf, Snapshots.pdf, 
> snapshot-testplan.pdf
>
>
> Snapshots are point in time images of parts of the filesystem or the entire 
> filesystem. Snapshots can be a read-only or a read-write point in time copy 
> of the filesystem. There are several use cases for snapshots in HDFS. I will 
> post a detailed write-up soon with with more information.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-3746) Invalid counter tag in HDFS balancer which lead to infinite loop

2013-08-27 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3746?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HDFS-3746:


Fix Version/s: (was: 2.1.0-beta)
   2.3.0

> Invalid counter tag in HDFS balancer which lead to infinite loop
> 
>
> Key: HDFS-3746
> URL: https://issues.apache.org/jira/browse/HDFS-3746
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: balancer
>Affects Versions: 2.0.0-alpha, 3.0.0
>Reporter: meng gong
>Assignee: meng gong
>  Labels: patch
> Fix For: 3.0.0, 2.3.0
>
>
> Everytime banlancer try to move a block cross the rack. For every 
> NameNodeConnector it will instance a new banlancer. The tag 
> notChangedIterations is reseted to 0. Then it won't reach 5 and exit the 
> thread for not moved in 5 consecutive iterations. This lead to a infinite loop

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-3907) Allow multiple users for local block readers

2013-08-27 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3907?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HDFS-3907:


Fix Version/s: (was: 2.1.0-beta)
   2.3.0

> Allow multiple users for local block readers
> 
>
> Key: HDFS-3907
> URL: https://issues.apache.org/jira/browse/HDFS-3907
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: datanode
>Affects Versions: 1.0.0, 2.0.0-alpha
>Reporter: Eli Collins
>Assignee: Eli Collins
> Fix For: 2.3.0
>
> Attachments: hdfs-3907.txt
>
>
> The {{dfs.block.local-path-access.user}} config added in HDFS-2246 only 
> supports a single user, however as long as blocks are group readable by more 
> than one user the feature could be used by multiple users, to support this we 
> just need to allow both to be configured. In practice this allows us to also 
> support HBase where the client (RS) runs as the hbase system user and the DN 
> runs as hdfs system user. I think this should work secure as well since we're 
> not using impersonation in the HBase case.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-3592) libhdfs should expose ClientProtocol::mkdirs2

2013-08-27 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3592?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HDFS-3592:


Fix Version/s: (was: 2.1.0-beta)
   2.3.0

> libhdfs should expose ClientProtocol::mkdirs2
> -
>
> Key: HDFS-3592
> URL: https://issues.apache.org/jira/browse/HDFS-3592
> Project: Hadoop HDFS
>  Issue Type: Improvement
>Affects Versions: 2.0.0-alpha, 3.0.0
>Reporter: Colin Patrick McCabe
>Assignee: Colin Patrick McCabe
>Priority: Minor
> Fix For: 3.0.0, 2.3.0
>
>
> It would be nice if libhdfs exposed mkdirs2.  This version of mkdirs is much 
> more verbose about any errors that occur-- it throws AccessControlException, 
> FileAlreadyExists, FileNotFoundException, ParentNotDirectoryException, etc.
> The original mkdirs just throws IOException if anything goes wrong.
> For something like fuse_dfs, it is very important to return the correct errno 
> code when an error has occurred.  mkdirs2 would allow us to do that.
> I'm not sure if we should just change hdfsMkdirs to use mkdirs2, or add an 
> hdfsMkdirs2.  Probably the latter, but the former course would maintain "bug 
> compatibility" with ancient releases-- if that is important.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-3987) add support for encrypted webhdfs

2013-08-27 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3987?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HDFS-3987:


Fix Version/s: (was: 2.1.0-beta)
   2.3.0

> add support for encrypted webhdfs
> -
>
> Key: HDFS-3987
> URL: https://issues.apache.org/jira/browse/HDFS-3987
> Project: Hadoop HDFS
>  Issue Type: New Feature
>Affects Versions: 2.0.2-alpha
>Reporter: Alejandro Abdelnur
> Fix For: 2.3.0
>
>
> This is a follow up of HDFS-3983.
> We should have a new filesystem client impl/binding for encrypted WebHDFS, 
> i.e. *webhdfss://*
> On the server side, webhdfs and httpfs we should only need to start the 
> service on a secured (HTTPS) endpoint.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-4286) Changes from BOOKKEEPER-203 broken capability of including bookkeeper-server jar in hidden package of BKJM

2013-08-27 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4286?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HDFS-4286:


Fix Version/s: (was: 2.1.0-beta)
   2.3.0

> Changes from BOOKKEEPER-203 broken capability of including bookkeeper-server 
> jar in hidden package of BKJM
> --
>
> Key: HDFS-4286
> URL: https://issues.apache.org/jira/browse/HDFS-4286
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>Reporter: Vinay
> Fix For: 3.0.0, 2.3.0
>
>
> BOOKKEEPER-203 introduced changes to LedgerLayout to include 
> ManagerFactoryClass instead of ManagerFactoryName.
> So because of this, BKJM cannot shade the bookkeeper-server jar inside BKJM 
> jar
> LAYOUT znode created by BookieServer is not readable by the BKJM as it have 
> classes in hidden packages. (same problem vice versa)

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-3988) HttpFS can't do GETDELEGATIONTOKEN without a prior authenticated request

2013-08-27 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3988?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HDFS-3988:


Fix Version/s: (was: 2.1.0-beta)
   2.3.0

> HttpFS can't do GETDELEGATIONTOKEN without a prior authenticated request
> 
>
> Key: HDFS-3988
> URL: https://issues.apache.org/jira/browse/HDFS-3988
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 2.0.2-alpha
>Reporter: Alejandro Abdelnur
>Assignee: Alejandro Abdelnur
> Fix For: 2.3.0
>
>
> A request to obtain a delegation token cannot it initiate an authentication 
> sequence, it must be accompanied by an auth cookie obtained in a prev request 
> using a different operation.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-4559) WebHDFS does not allow resolution of Symlinks

2013-08-27 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4559?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HDFS-4559:


Fix Version/s: (was: 2.1.0-beta)
   2.3.0

> WebHDFS does not allow resolution of Symlinks
> -
>
> Key: HDFS-4559
> URL: https://issues.apache.org/jira/browse/HDFS-4559
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 2.0.3-alpha
>Reporter: Plamen Jeliazkov
>Assignee: Plamen Jeliazkov
> Fix For: 3.0.0, 2.3.0
>
>
> WebHDFS allows you to create symlinks via the CREATESYMLINK operation, but 
> the GETFILEINFO operation specifically calls the getFileInfo() method of the 
> NameNodeRpcServer which does not resolve symlinks.
> I propose adding a parameter to GETFILEINFO such that if true will call 
> getFileLinkInfo() rather than getFileInfo() which will resolve any symlinks.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-4489) Use InodeID as as an identifier of a file in HDFS protocols and APIs

2013-08-27 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4489?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HDFS-4489:


Fix Version/s: (was: 2.1.0-beta)
   2.3.0

> Use InodeID as as an identifier of a file in HDFS protocols and APIs
> 
>
> Key: HDFS-4489
> URL: https://issues.apache.org/jira/browse/HDFS-4489
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: namenode
>Reporter: Brandon Li
>Assignee: Brandon Li
> Fix For: 2.3.0
>
> Attachments: 4434.optimized.patch
>
>
> The benefit of using InodeID to uniquely identify a file can be multiple 
> folds. Here are a few of them:
> 1. uniquely identify a file cross rename, related JIRAs include HDFS-4258, 
> HDFS-4437.
> 2. modification checks in tools like distcp. Since a file could have been 
> replaced or renamed to, the file name and size combination is no t reliable, 
> but the combination of file id and size is unique.
> 3. id based protocol support (e.g., NFS)
> 4. to make the pluggable block placement policy use fileid instead of 
> filename (HDFS-385).

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


  1   2   3   4   5   >