[jira] [Commented] (HDFS-3178) Add states for journal synchronization in journal daemon

2012-04-04 Thread Suresh Srinivas (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3178?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13246058#comment-13246058
 ] 

Suresh Srinivas commented on HDFS-3178:
---

Nicholas, if you are going to check for all states in StateMachine, you might 
as well move that code into JournalService. Rest of the code looks good.

 Add states for journal synchronization in journal daemon
 

 Key: HDFS-3178
 URL: https://issues.apache.org/jira/browse/HDFS-3178
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: ha, name-node
Reporter: Tsz Wo (Nicholas), SZE
Assignee: Tsz Wo (Nicholas), SZE
 Attachments: h3178_20120403_svn_mv.patch, svn_mv.sh


 Journal in a new daemon has to be synchronized to the current transaction.  
 It requires new states such as WaitingForRoll, Syncing and Synced.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3192) Active NN should exit when it has not received a getServiceStatus() rpc from ZKFC for timeout secs

2012-04-04 Thread Suresh Srinivas (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3192?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13246059#comment-13246059
 ] 

Suresh Srinivas commented on HDFS-3192:
---

Hari, agree with Aaron that this should not be a subtask of HDFS-3092.

 Active NN should exit when it has not received a getServiceStatus() rpc from 
 ZKFC for timeout secs
 --

 Key: HDFS-3192
 URL: https://issues.apache.org/jira/browse/HDFS-3192
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: ha, name-node
Reporter: Hari Mankude
Assignee: Hari Mankude



--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Created] (HDFS-3193) HDFS journal logging improvement

2012-04-04 Thread Suresh Srinivas (Created) (JIRA)
HDFS journal logging improvement


 Key: HDFS-3193
 URL: https://issues.apache.org/jira/browse/HDFS-3193
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: name-node
Reporter: Suresh Srinivas


I want to use this jira to discuss doing HDFS editlog logging before namespace 
update. Couple of bugs (though have been solved with work arounds) HDFS-2815, 
HDFS-3191 are related to this. Also currently an update to namespace becomes 
visible to other readers, even though it is not persisted in editlog.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HDFS-3183) Add JournalManager implementation to JournalDaemons for storing edits

2012-04-04 Thread Suresh Srinivas (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3183?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Suresh Srinivas updated HDFS-3183:
--

Description: 
The JournalManager is used in HA configuration and uses the following journal 
daemons on:
- Local namenode
- Other namenode
- A configured JournalDaemon target from the configuration

  was:
The JournalManager is used in HA configuration and uses the following journal 
targets:
- local namenode
- Other namenode
- A configured JournalDaemon target from configuration

Summary: Add JournalManager implementation to JournalDaemons for 
storing edits  (was: Add JournalManager implementation to use local namenode, 
remote namenode and a configured JournalDaemon for storing editlogs)

 Add JournalManager implementation to JournalDaemons for storing edits
 -

 Key: HDFS-3183
 URL: https://issues.apache.org/jira/browse/HDFS-3183
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: ha, name-node
Reporter: Suresh Srinivas

 The JournalManager is used in HA configuration and uses the following journal 
 daemons on:
 - Local namenode
 - Other namenode
 - A configured JournalDaemon target from the configuration

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HDFS-3084) FenceMethod.tryFence() and ShellCommandFencer should pass namenodeId as well as host:port

2012-04-04 Thread Todd Lipcon (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3084?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Todd Lipcon updated HDFS-3084:
--

Attachment: hdfs-3084.txt

This patch, in conjunction with HADOOP-8007, provides the necessary facilities:

- the configuration used to instantiate the NodeFencer is now the configuration 
of the _target_ node, rather than the _fencing_ node. This means that all of 
the configs like RPC address have already had the nameservice and namenodeid 
chopped off
- in addition to the variables described in HADOOP-8007, we add 
$target_namenodeid and $target_nameserviceid to the fencing environment
- added getters for namenode id and nameservice id to NNHAServiceTarget. If a 
custom fencing implementation is HDFS specific, it can downcast to the 
implementation class to access these getters. Alternatively, it may call 
{{getFencingParameters()}} and then use the resulting string map to get at this 
info.

The patch depends on HADOOP-8007. I'll submit it to QA once that other patch is 
committed.

 FenceMethod.tryFence() and ShellCommandFencer should pass namenodeId as well 
 as host:port
 -

 Key: HDFS-3084
 URL: https://issues.apache.org/jira/browse/HDFS-3084
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: ha
Affects Versions: 0.24.0, 0.23.3
Reporter: Philip Zeyliger
Assignee: Todd Lipcon
 Attachments: hdfs-3084.txt


 The FenceMethod interface passes along the host:port of the NN that needs to 
 be fenced.  That's great for the common case.  However, it's likely necessary 
 to have extra configuration parameters for fencing, and these are typically 
 keyed off the nameserviceId.namenodeId (if, for nothing else, consistency 
 with all the other parameters that are keyed off of namespaceId.namenodeId).  
 Obviously this can be backed out from the host:port, but it's inconvenient, 
 and requires iterating through all the configs.
 The shell interface exhibits the same issue: host:port is great for most 
 fencers, but if you need extra configs (like the host:port of the power 
 supply unit), those are harder to pipe through without the namenodeId.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3084) FenceMethod.tryFence() and ShellCommandFencer should pass namenodeId as well as host:port

2012-04-04 Thread Aaron T. Myers (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3084?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13246087#comment-13246087
 ] 

Aaron T. Myers commented on HDFS-3084:
--

+1, the patch looks good to me. Good stuff, Todd.

 FenceMethod.tryFence() and ShellCommandFencer should pass namenodeId as well 
 as host:port
 -

 Key: HDFS-3084
 URL: https://issues.apache.org/jira/browse/HDFS-3084
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: ha
Affects Versions: 0.24.0, 0.23.3
Reporter: Philip Zeyliger
Assignee: Todd Lipcon
 Attachments: hdfs-3084.txt


 The FenceMethod interface passes along the host:port of the NN that needs to 
 be fenced.  That's great for the common case.  However, it's likely necessary 
 to have extra configuration parameters for fencing, and these are typically 
 keyed off the nameserviceId.namenodeId (if, for nothing else, consistency 
 with all the other parameters that are keyed off of namespaceId.namenodeId).  
 Obviously this can be backed out from the host:port, but it's inconvenient, 
 and requires iterating through all the configs.
 The shell interface exhibits the same issue: host:port is great for most 
 fencers, but if you need extra configs (like the host:port of the power 
 supply unit), those are harder to pipe through without the namenodeId.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3084) FenceMethod.tryFence() and ShellCommandFencer should pass namenodeId as well as host:port

2012-04-04 Thread Eli Collins (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3084?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13246086#comment-13246086
 ] 

Eli Collins commented on HDFS-3084:
---

+1 looks good

 FenceMethod.tryFence() and ShellCommandFencer should pass namenodeId as well 
 as host:port
 -

 Key: HDFS-3084
 URL: https://issues.apache.org/jira/browse/HDFS-3084
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: ha
Affects Versions: 0.24.0, 0.23.3
Reporter: Philip Zeyliger
Assignee: Todd Lipcon
 Attachments: hdfs-3084.txt


 The FenceMethod interface passes along the host:port of the NN that needs to 
 be fenced.  That's great for the common case.  However, it's likely necessary 
 to have extra configuration parameters for fencing, and these are typically 
 keyed off the nameserviceId.namenodeId (if, for nothing else, consistency 
 with all the other parameters that are keyed off of namespaceId.namenodeId).  
 Obviously this can be backed out from the host:port, but it's inconvenient, 
 and requires iterating through all the configs.
 The shell interface exhibits the same issue: host:port is great for most 
 fencers, but if you need extra configs (like the host:port of the power 
 supply unit), those are harder to pipe through without the namenodeId.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HDFS-3102) Add CLI tool to initialize the shared-edits dir

2012-04-04 Thread Aaron T. Myers (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3102?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Aaron T. Myers updated HDFS-3102:
-

Attachment: HDFS-3102.patch

Here's a patch rebased on trunk.

 Add CLI tool to initialize the shared-edits dir
 ---

 Key: HDFS-3102
 URL: https://issues.apache.org/jira/browse/HDFS-3102
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: ha, name-node
Affects Versions: 0.24.0, 2.0.0
Reporter: Todd Lipcon
Assignee: Aaron T. Myers
 Attachments: HDFS-3102.patch, HDFS-3102.patch


 Currently in order to make a non-HA NN HA, you need to initialize the shared 
 edits dir. This can be done manually by cping directories around. It would be 
 preferable to add a namenode -initializeSharedEdits command to achieve this 
 same effect.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HDFS-3084) FenceMethod.tryFence() and ShellCommandFencer should pass namenodeId as well as host:port

2012-04-04 Thread Todd Lipcon (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3084?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Todd Lipcon updated HDFS-3084:
--

Target Version/s: 0.24.0, 0.23.3  (was: 0.23.3, 0.24.0)
  Status: Patch Available  (was: Open)

 FenceMethod.tryFence() and ShellCommandFencer should pass namenodeId as well 
 as host:port
 -

 Key: HDFS-3084
 URL: https://issues.apache.org/jira/browse/HDFS-3084
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: ha
Affects Versions: 0.24.0, 0.23.3
Reporter: Philip Zeyliger
Assignee: Todd Lipcon
 Attachments: hdfs-3084.txt


 The FenceMethod interface passes along the host:port of the NN that needs to 
 be fenced.  That's great for the common case.  However, it's likely necessary 
 to have extra configuration parameters for fencing, and these are typically 
 keyed off the nameserviceId.namenodeId (if, for nothing else, consistency 
 with all the other parameters that are keyed off of namespaceId.namenodeId).  
 Obviously this can be backed out from the host:port, but it's inconvenient, 
 and requires iterating through all the configs.
 The shell interface exhibits the same issue: host:port is great for most 
 fencers, but if you need extra configs (like the host:port of the power 
 supply unit), those are harder to pipe through without the namenodeId.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3102) Add CLI tool to initialize the shared-edits dir

2012-04-04 Thread Hadoop QA (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3102?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13246129#comment-13246129
 ] 

Hadoop QA commented on HDFS-3102:
-

-1 overall.  Here are the results of testing the latest attachment 
  http://issues.apache.org/jira/secure/attachment/12521291/HDFS-3102.patch
  against trunk revision .

+1 @author.  The patch does not contain any @author tags.

+1 tests included.  The patch appears to include 6 new or modified tests.

+1 javadoc.  The javadoc tool did not generate any warning messages.

+1 javac.  The applied patch does not increase the total number of javac 
compiler warnings.

+1 eclipse:eclipse.  The patch built with eclipse:eclipse.

+1 findbugs.  The patch does not introduce any new Findbugs (version 1.3.9) 
warnings.

+1 release audit.  The applied patch does not increase the total number of 
release audit warnings.

-1 core tests.  The patch failed these unit tests:
  
org.apache.hadoop.hdfs.server.namenode.ha.TestPipelinesFailover

+1 contrib tests.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-HDFS-Build/2181//testReport/
Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/2181//console

This message is automatically generated.

 Add CLI tool to initialize the shared-edits dir
 ---

 Key: HDFS-3102
 URL: https://issues.apache.org/jira/browse/HDFS-3102
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: ha, name-node
Affects Versions: 0.24.0, 2.0.0
Reporter: Todd Lipcon
Assignee: Aaron T. Myers
 Attachments: HDFS-3102.patch, HDFS-3102.patch


 Currently in order to make a non-HA NN HA, you need to initialize the shared 
 edits dir. This can be done manually by cping directories around. It would be 
 preferable to add a namenode -initializeSharedEdits command to achieve this 
 same effect.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Created] (HDFS-3194) Continuous block scanning at DN side

2012-04-04 Thread suja s (Created) (JIRA)
Continuous block scanning at DN side


 Key: HDFS-3194
 URL: https://issues.apache.org/jira/browse/HDFS-3194
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 1.0.3
Reporter: suja s
Priority: Minor
 Fix For: 1.0.3


Block scanning interval by default should be taken as 21 days(3 weeks) and each 
block scanning should happen once in 21 days.
Here the block is being scanned continuosly.

2012-04-03 10:44:47,056 INFO 
org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner: Verification 
succeeded for 
BP-241703115-xx.xx.xx.55-1333086229434:blk_-2666054955039014473_1003
2012-04-03 10:45:02,064 INFO 
org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner: Verification 
succeeded for 
BP-241703115-xx.xx.xx.55-1333086229434:blk_-2666054955039014473_1003
2012-04-03 10:45:17,071 INFO 
org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner: Verification 
succeeded for 
BP-241703115-xx.xx.xx.55-1333086229434:blk_-2666054955039014473_1003
2012-04-03 10:45:32,079 INFO 
org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner: Verification 
succeeded for BP-241703115-xx.xx.xx.55-1333086229434:blk_-2666054955039014473

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3084) FenceMethod.tryFence() and ShellCommandFencer should pass namenodeId as well as host:port

2012-04-04 Thread Hadoop QA (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3084?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13246143#comment-13246143
 ] 

Hadoop QA commented on HDFS-3084:
-

+1 overall.  Here are the results of testing the latest attachment 
  http://issues.apache.org/jira/secure/attachment/12521282/hdfs-3084.txt
  against trunk revision .

+1 @author.  The patch does not contain any @author tags.

+1 tests included.  The patch appears to include 6 new or modified tests.

+1 javadoc.  The javadoc tool did not generate any warning messages.

+1 javac.  The applied patch does not increase the total number of javac 
compiler warnings.

+1 eclipse:eclipse.  The patch built with eclipse:eclipse.

+1 findbugs.  The patch does not introduce any new Findbugs (version 1.3.9) 
warnings.

+1 release audit.  The applied patch does not increase the total number of 
release audit warnings.

+1 core tests.  The patch passed unit tests in .

+1 contrib tests.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-HDFS-Build/2182//testReport/
Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/2182//console

This message is automatically generated.

 FenceMethod.tryFence() and ShellCommandFencer should pass namenodeId as well 
 as host:port
 -

 Key: HDFS-3084
 URL: https://issues.apache.org/jira/browse/HDFS-3084
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: ha
Affects Versions: 0.24.0, 0.23.3
Reporter: Philip Zeyliger
Assignee: Todd Lipcon
 Attachments: hdfs-3084.txt


 The FenceMethod interface passes along the host:port of the NN that needs to 
 be fenced.  That's great for the common case.  However, it's likely necessary 
 to have extra configuration parameters for fencing, and these are typically 
 keyed off the nameserviceId.namenodeId (if, for nothing else, consistency 
 with all the other parameters that are keyed off of namespaceId.namenodeId).  
 Obviously this can be backed out from the host:port, but it's inconvenient, 
 and requires iterating through all the configs.
 The shell interface exhibits the same issue: host:port is great for most 
 fencers, but if you need extra configs (like the host:port of the power 
 supply unit), those are harder to pipe through without the namenodeId.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3175) When the disk space is available back,Namenode resource monitor can automatically take off safemode.

2012-04-04 Thread liaowenrui (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3175?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13246150#comment-13246150
 ] 

liaowenrui commented on HDFS-3175:
--

hi Aaron,

I think that automatically leave safemode mechanism is better than manual.and 
It should be optimized to automatical approach.

Is it right?

 When the disk space is available back,Namenode resource monitor can 
 automatically take off safemode.
 

 Key: HDFS-3175
 URL: https://issues.apache.org/jira/browse/HDFS-3175
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: name-node
Affects Versions: 0.24.0, 2.0.0
Reporter: liaowenrui
 Attachments: HDFS-3175.patch, HDFS-3175.patch, HDFS-3175.patch, 
 testcase




--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3175) When the disk space is available back,Namenode resource monitor can automatically take off safemode.

2012-04-04 Thread liaowenrui (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3175?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13246164#comment-13246164
 ] 

liaowenrui commented on HDFS-3175:
--

Hi Aaron, I have seen your opinion in other issue HDFS-2914

https://issues.apache.org/jira/browse/HDFS-2914?focusedCommentId=13203135page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-13203135

This can solve that issue as well.

 When the disk space is available back,Namenode resource monitor can 
 automatically take off safemode.
 

 Key: HDFS-3175
 URL: https://issues.apache.org/jira/browse/HDFS-3175
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: name-node
Affects Versions: 0.24.0, 2.0.0
Reporter: liaowenrui
 Attachments: HDFS-3175.patch, HDFS-3175.patch, HDFS-3175.patch, 
 testcase




--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HDFS-2964) No notice in any logs if dfs.datanode.du.reserved is greater than available disk space

2012-04-04 Thread madhukara phatak (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-2964?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

madhukara phatak updated HDFS-2964:
---

Attachment: HDFS-2964.patch

Added a warning message in FsVolumeImpl.java where it reads the latest 
capacity. I m creating the patch first time so not sure its the correct place 
to add the warning message.

 No notice in any logs if dfs.datanode.du.reserved is greater than available 
 disk space
 --

 Key: HDFS-2964
 URL: https://issues.apache.org/jira/browse/HDFS-2964
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: data-node, name-node
Affects Versions: 0.20.2
Reporter: Robert J Berger
Priority: Minor
 Attachments: HDFS-2964.patch


 We spent a long time tracking down why a test hdfs cluster seemed to be 
 running fine, but would not allow the mapred system to come up complaining 
 that could only be replicated to 0 nodes, instead of 1.
 There were no namenode or datanode errors in any of the logs. hadoop fsck 
 said everything was good. At first glance dfsadmin -report looked good. It 
 wasn't until I realized that there was 0 Capacity available that we poked 
 around and found 
 https://groups.google.com/a/cloudera.org/group/scm-users/msg/a4252d6623adbc2d 
 which mentioned that the reserverd space might be greater than the disk 
 space available. And we did find that our dfs.datanode.du.reserved was indeed 
 higher than our actual since we were only testing a small cluster.
 It seems that there should be some warning or error in the logs that say that.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-2964) No notice in any logs if dfs.datanode.du.reserved is greater than available disk space

2012-04-04 Thread Harsh J (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2964?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13246193#comment-13246193
 ] 

Harsh J commented on HDFS-2964:
---

Hello Madhukara,

Please attach a proper diff that carries only the changes you have done. Your 
current diff is a total rewrite of file. Ensure you use unix newlines in your 
editor/save-actions/etc., in case that is what led to this.

 No notice in any logs if dfs.datanode.du.reserved is greater than available 
 disk space
 --

 Key: HDFS-2964
 URL: https://issues.apache.org/jira/browse/HDFS-2964
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: data-node, name-node
Affects Versions: 0.20.2
Reporter: Robert J Berger
Priority: Minor
 Attachments: HDFS-2964.patch


 We spent a long time tracking down why a test hdfs cluster seemed to be 
 running fine, but would not allow the mapred system to come up complaining 
 that could only be replicated to 0 nodes, instead of 1.
 There were no namenode or datanode errors in any of the logs. hadoop fsck 
 said everything was good. At first glance dfsadmin -report looked good. It 
 wasn't until I realized that there was 0 Capacity available that we poked 
 around and found 
 https://groups.google.com/a/cloudera.org/group/scm-users/msg/a4252d6623adbc2d 
 which mentioned that the reserverd space might be greater than the disk 
 space available. And we did find that our dfs.datanode.du.reserved was indeed 
 higher than our actual since we were only testing a small cluster.
 It seems that there should be some warning or error in the logs that say that.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3194) Continuous block scanning at DN side

2012-04-04 Thread Harsh J (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3194?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13246199#comment-13246199
 ] 

Harsh J commented on HDFS-3194:
---

org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner does not exist in 
branch-1. It is a branch-2 feature (attached to Federation). What version are 
you really using and reporting for? Please fix the version field to reflect the 
right version…

 Continuous block scanning at DN side
 

 Key: HDFS-3194
 URL: https://issues.apache.org/jira/browse/HDFS-3194
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 1.0.3
Reporter: suja s
Priority: Minor
 Fix For: 1.0.3


 Block scanning interval by default should be taken as 21 days(3 weeks) and 
 each block scanning should happen once in 21 days.
 Here the block is being scanned continuosly.
 2012-04-03 10:44:47,056 INFO 
 org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner: Verification 
 succeeded for 
 BP-241703115-xx.xx.xx.55-1333086229434:blk_-2666054955039014473_1003
 2012-04-03 10:45:02,064 INFO 
 org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner: Verification 
 succeeded for 
 BP-241703115-xx.xx.xx.55-1333086229434:blk_-2666054955039014473_1003
 2012-04-03 10:45:17,071 INFO 
 org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner: Verification 
 succeeded for 
 BP-241703115-xx.xx.xx.55-1333086229434:blk_-2666054955039014473_1003
 2012-04-03 10:45:32,079 INFO 
 org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner: Verification 
 succeeded for BP-241703115-xx.xx.xx.55-1333086229434:blk_-2666054955039014473

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HDFS-2964) No notice in any logs if dfs.datanode.du.reserved is greater than available disk space

2012-04-04 Thread madhukara phatak (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-2964?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

madhukara phatak updated HDFS-2964:
---

Attachment: HDFS-2964-1.patch

Fixed the patch file to have only changed code and converted new lines into UNIX

 No notice in any logs if dfs.datanode.du.reserved is greater than available 
 disk space
 --

 Key: HDFS-2964
 URL: https://issues.apache.org/jira/browse/HDFS-2964
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: data-node, name-node
Affects Versions: 0.20.2
Reporter: Robert J Berger
Priority: Minor
 Attachments: HDFS-2964-1.patch, HDFS-2964.patch


 We spent a long time tracking down why a test hdfs cluster seemed to be 
 running fine, but would not allow the mapred system to come up complaining 
 that could only be replicated to 0 nodes, instead of 1.
 There were no namenode or datanode errors in any of the logs. hadoop fsck 
 said everything was good. At first glance dfsadmin -report looked good. It 
 wasn't until I realized that there was 0 Capacity available that we poked 
 around and found 
 https://groups.google.com/a/cloudera.org/group/scm-users/msg/a4252d6623adbc2d 
 which mentioned that the reserverd space might be greater than the disk 
 space available. And we did find that our dfs.datanode.du.reserved was indeed 
 higher than our actual since we were only testing a small cluster.
 It seems that there should be some warning or error in the logs that say that.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3176) JsonUtil should not parse the MD5MD5CRC32FileChecksum bytes on its own.

2012-04-04 Thread Hudson (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3176?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13246209#comment-13246209
 ] 

Hudson commented on HDFS-3176:
--

Integrated in Hadoop-Hdfs-0.23-Build #218 (See 
[https://builds.apache.org/job/Hadoop-Hdfs-0.23-Build/218/])
svn merge -c 1309114 from trunk for HDFS-3176. Use 
MD5MD5CRC32FileChecksum.readFields() in JsonUtil. (Revision 1309115)

 Result = SUCCESS
szetszwo : 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1309115
Files : 
* /hadoop/common/branches/branch-0.23/hadoop-hdfs-project/hadoop-hdfs
* 
/hadoop/common/branches/branch-0.23/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* 
/hadoop/common/branches/branch-0.23/hadoop-hdfs-project/hadoop-hdfs/src/main/java
* 
/hadoop/common/branches/branch-0.23/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/web/JsonUtil.java


 JsonUtil should not parse the MD5MD5CRC32FileChecksum bytes on its own.
 ---

 Key: HDFS-3176
 URL: https://issues.apache.org/jira/browse/HDFS-3176
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: hdfs client
Affects Versions: 0.23.0, 1.0.1
Reporter: Kihwal Lee
Assignee: Kihwal Lee
 Fix For: 1.1.0, 0.23.3, 2.0.0, 3.0.0

 Attachments: hdfs-3176-branch-1.patch, hdfs-3176.patch


 Currently JsonUtil used by webhdfs parses MD5MD5CRC32FileChecksum binary 
 bytes on its own and contructs a MD5MD5CRC32FileChecksum. It should instead 
 call MD5MD5CRC32FileChecksum.readFields().

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3166) Hftp connections do not have a timeout

2012-04-04 Thread Hudson (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3166?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13246210#comment-13246210
 ] 

Hudson commented on HDFS-3166:
--

Integrated in Hadoop-Hdfs-0.23-Build #218 (See 
[https://builds.apache.org/job/Hadoop-Hdfs-0.23-Build/218/])
svn merge -c 1309103 from trunk for HDFS-3166. Add timeout to Hftp 
connections. (Revision 1309106)

 Result = SUCCESS
szetszwo : 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1309106
Files : 
* /hadoop/common/branches/branch-0.23/hadoop-hdfs-project/hadoop-hdfs
* 
/hadoop/common/branches/branch-0.23/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* 
/hadoop/common/branches/branch-0.23/hadoop-hdfs-project/hadoop-hdfs/src/main/java
* 
/hadoop/common/branches/branch-0.23/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/HftpFileSystem.java
* 
/hadoop/common/branches/branch-0.23/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/HsftpFileSystem.java
* 
/hadoop/common/branches/branch-0.23/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/tools/DelegationTokenFetcher.java
* 
/hadoop/common/branches/branch-0.23/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/web/URLUtils.java
* 
/hadoop/common/branches/branch-0.23/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestHftpURLTimeouts.java


 Hftp connections do not have a timeout
 --

 Key: HDFS-3166
 URL: https://issues.apache.org/jira/browse/HDFS-3166
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: hdfs client
Affects Versions: 0.23.2, 0.23.3, 2.0.0, 3.0.0
Reporter: Daryn Sharp
Assignee: Daryn Sharp
Priority: Critical
 Fix For: 0.23.3, 2.0.0, 3.0.0

 Attachments: HADOOP-8221.branch-1.patch, HADOOP-8221.patch, 
 HADOOP-8221.patch, HDFS-3166.patch, HDFS-3166.patch


 Hftp connections do not have read timeouts.  This leads to indefinitely hung 
 sockets when there is a network outage during which time the remote host 
 closed the socket.
 This may also affect WebHdfs, etc.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3176) JsonUtil should not parse the MD5MD5CRC32FileChecksum bytes on its own.

2012-04-04 Thread Hudson (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3176?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13246230#comment-13246230
 ] 

Hudson commented on HDFS-3176:
--

Integrated in Hadoop-Hdfs-trunk #1005 (See 
[https://builds.apache.org/job/Hadoop-Hdfs-trunk/1005/])
HDFS-3176. Use MD5MD5CRC32FileChecksum.readFields() in JsonUtil .  
Contributed by Kihwal Lee (Revision 1309114)

 Result = FAILURE
szetszwo : 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1309114
Files : 
* /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/web/JsonUtil.java


 JsonUtil should not parse the MD5MD5CRC32FileChecksum bytes on its own.
 ---

 Key: HDFS-3176
 URL: https://issues.apache.org/jira/browse/HDFS-3176
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: hdfs client
Affects Versions: 0.23.0, 1.0.1
Reporter: Kihwal Lee
Assignee: Kihwal Lee
 Fix For: 1.1.0, 0.23.3, 2.0.0, 3.0.0

 Attachments: hdfs-3176-branch-1.patch, hdfs-3176.patch


 Currently JsonUtil used by webhdfs parses MD5MD5CRC32FileChecksum binary 
 bytes on its own and contructs a MD5MD5CRC32FileChecksum. It should instead 
 call MD5MD5CRC32FileChecksum.readFields().

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3000) Add a public API for setting quotas

2012-04-04 Thread Hudson (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3000?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13246232#comment-13246232
 ] 

Hudson commented on HDFS-3000:
--

Integrated in Hadoop-Hdfs-trunk #1005 (See 
[https://builds.apache.org/job/Hadoop-Hdfs-trunk/1005/])
HDFS-3000. Add a public API for setting quotas. Contributed by Aaron T. 
Myers. (Revision 1309227)

 Result = FAILURE
atm : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1309227
Files : 
* /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/client
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/client/HdfsAdmin.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestHdfsAdmin.java


 Add a public API for setting quotas
 ---

 Key: HDFS-3000
 URL: https://issues.apache.org/jira/browse/HDFS-3000
 Project: Hadoop HDFS
  Issue Type: New Feature
  Components: hdfs client
Affects Versions: 2.0.0
Reporter: Aaron T. Myers
Assignee: Aaron T. Myers
 Attachments: HDFS-3000.patch, HDFS-3000.patch, HDFS-3000.patch, 
 HDFS-3000.patch, HDFS-3000.patch


 Currently one can set the quota of a file or directory from the command line, 
 but if a user wants to set it programmatically, they need to use 
 DistributedFileSystem, which is annotated InterfaceAudience.Private.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3168) Clean up FSNamesystem and BlockManager

2012-04-04 Thread Hudson (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3168?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13246231#comment-13246231
 ] 

Hudson commented on HDFS-3168:
--

Integrated in Hadoop-Hdfs-trunk #1005 (See 
[https://builds.apache.org/job/Hadoop-Hdfs-trunk/1005/])
HDFS-3168. Remove unnecessary throw IOException and change fields to 
final in FSNamesystem and BlockManager. (Revision 1309218)

 Result = FAILURE
szetszwo : 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1309218
Files : 
* /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/BlockManager.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSNamesystem.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/NameNode.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/ha/EditLogTailer.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/blockmanagement/BlockManagerTestUtil.java


 Clean up FSNamesystem and BlockManager
 --

 Key: HDFS-3168
 URL: https://issues.apache.org/jira/browse/HDFS-3168
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: name-node
Reporter: Tsz Wo (Nicholas), SZE
Assignee: Tsz Wo (Nicholas), SZE
 Fix For: 0.24.0, 0.23.3

 Attachments: h3168_20120330.patch, h3168_20120402.patch, 
 h3168_20120403.patch




--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3187) Upgrade guava to 11.0.2

2012-04-04 Thread Hudson (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3187?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13246233#comment-13246233
 ] 

Hudson commented on HDFS-3187:
--

Integrated in Hadoop-Hdfs-trunk #1005 (See 
[https://builds.apache.org/job/Hadoop-Hdfs-trunk/1005/])
HDFS-3187. Upgrade guava to 11.0.2. Contributed by Todd Lipcon. (Revision 
1309181)

 Result = FAILURE
todd : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1309181
Files : 
* /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/MiniDFSCluster.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/FSImageTestUtil.java
* /hadoop/common/trunk/hadoop-project/pom.xml


 Upgrade guava to 11.0.2
 ---

 Key: HDFS-3187
 URL: https://issues.apache.org/jira/browse/HDFS-3187
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: build
Affects Versions: 2.0.0
Reporter: Todd Lipcon
Assignee: Todd Lipcon
Priority: Minor
 Fix For: 2.0.0

 Attachments: hdfs-3187.txt


 Guava r11 includes some nice features which we'd like to use in the 
 implementation of HDFS-3077. In particular, 
 {{MoreExecutors.listeningDecorator}} allows a normal {{ExecutorService}} to 
 be turned into a {{ListeningExecutorService}}, so that tasks can be submitted 
 to it and then wrapped as {{ListenableFuture}}s.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3166) Hftp connections do not have a timeout

2012-04-04 Thread Hudson (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3166?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13246234#comment-13246234
 ] 

Hudson commented on HDFS-3166:
--

Integrated in Hadoop-Hdfs-trunk #1005 (See 
[https://builds.apache.org/job/Hadoop-Hdfs-trunk/1005/])
HDFS-3166. Add timeout to Hftp connections.  Contributed by Daryn Sharp 
(Revision 1309103)

 Result = FAILURE
szetszwo : 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1309103
Files : 
* /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/HftpFileSystem.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/HsftpFileSystem.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/tools/DelegationTokenFetcher.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/web/URLUtils.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestHftpURLTimeouts.java


 Hftp connections do not have a timeout
 --

 Key: HDFS-3166
 URL: https://issues.apache.org/jira/browse/HDFS-3166
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: hdfs client
Affects Versions: 0.23.2, 0.23.3, 2.0.0, 3.0.0
Reporter: Daryn Sharp
Assignee: Daryn Sharp
Priority: Critical
 Fix For: 0.23.3, 2.0.0, 3.0.0

 Attachments: HADOOP-8221.branch-1.patch, HADOOP-8221.patch, 
 HADOOP-8221.patch, HDFS-3166.patch, HDFS-3166.patch


 Hftp connections do not have read timeouts.  This leads to indefinitely hung 
 sockets when there is a network outage during which time the remote host 
 closed the socket.
 This may also affect WebHdfs, etc.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-1362) Provide volume management functionality for DataNode

2012-04-04 Thread Uma Maheswara Rao G (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-1362?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13246251#comment-13246251
 ] 

Uma Maheswara Rao G commented on HDFS-1362:
---

Any update on this? - Thanks

 Provide volume management functionality for DataNode
 

 Key: HDFS-1362
 URL: https://issues.apache.org/jira/browse/HDFS-1362
 Project: Hadoop HDFS
  Issue Type: New Feature
  Components: data-node
Affects Versions: 0.23.0
Reporter: Wang Xu
Assignee: Wang Xu
 Fix For: 0.24.0

 Attachments: DataNode Volume Refreshment in HDFS-1362.pdf, 
 HDFS-1362.4_w7001.txt, HDFS-1362.5.patch, HDFS-1362.6.patch, 
 HDFS-1362.7.patch, HDFS-1362.8.patch, HDFS-1362.txt, 
 Provide_volume_management_for_DN_v1.pdf


 The current management unit in Hadoop is a node, i.e. if a node failed, it 
 will be kicked out and all the data on the node will be replicated.
 As almost all SATA controller support hotplug, we add a new command line 
 interface to datanode, thus it can list, add or remove a volume online, which 
 means we can change a disk without node decommission. Moreover, if the failed 
 disk still readable and the node has enouth space, it can migrate data on the 
 disks to other disks in the same node.
 A more detailed design document will be attached.
 The original version in our lab is implemented against 0.20 datanode 
 directly, and is it better to implemented it in contrib? Or any other 
 suggestion?

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3122) Block recovery with closeFile flag true can race with blockReport. Due to this blocks are getting marked as corrupt.

2012-04-04 Thread amith (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3122?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13246256#comment-13246256
 ] 

amith commented on HDFS-3122:
-

Hi Folks

I thought of  some sample solution for this problem please give your inputs to 
the same.

When any blockrecovery happen, commitBlockSynchronization is called we can 
store the old and new generation stamp for this block in a map which is like 
historyMap = new HashMapString, ArrayListLong(). Here ArrayList is of size 
2 which contain old and new GenerationStamp. 

For every recovery this map is updated with the block and Generation Stamps. 

Consider the scenario when BlockReport has arrived @ NN and delayed. 
Now if the any BlockRecovery completed (historyMap will have the entry of old 
and new Generation Stamps). 
Now the Blockreport processing started. Here 

{code}
case RWR: 
  if (!storedBlock.isComplete()) { 
return null; // not corrupt 
  } else if (storedBlock.getGenerationStamp() != iblk.getGenerationStamp() 

!historyMap.get(iblk.getBlockId()).get(0) != 
iblk.getGenerationStamp() )) { 
return new BlockToMarkCorrupt(storedBlock, 
reported  + reportedState +  replica with genstamp  + 
iblk.getGenerationStamp() +  does not match COMPLETE block's  + 
genstamp in block map  + storedBlock.getGenerationStamp()); 
  } else { // COMPLETE block, same genstamp
{code}

Here we are checking like 
if (Block GenerationStamp from BlockMap != BlockReport's Block GenerationStamp 
and (blockGenerationStamp is newly changed due to recovery then check the) 
old GenerationStamp is not equal BlockReports Block GenerationStamp) then { // 
mark block as corrupt the block } 

Map is populated in CommitBlockSynchronization and cleared when BlockReport is 
processed for this block with new generationstamp

 



 Block recovery with closeFile flag true can race with blockReport. Due to 
 this blocks are getting marked as corrupt.
 

 Key: HDFS-3122
 URL: https://issues.apache.org/jira/browse/HDFS-3122
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: data-node, name-node
Affects Versions: 0.23.0, 0.24.0
Reporter: Uma Maheswara Rao G
Assignee: Uma Maheswara Rao G
Priority: Critical
 Attachments: blockCorrupt.txt


 *Block Report* can *race* with *Block Recovery* with closeFile flag true.
  Block report generated just before block recovery at DN side and due to N/W 
 problems, block report got delayed to NN. 
 After this, recovery success and generation stamp modifies to new one. 
 And primary DN invokes the commitBlockSynchronization and block got updated 
 in NN side. Also block got marked as complete, since the closeFile flag was 
 true. Updated with new genstamp.
 Now blockReport started processing at NN side. This particular block from RBW 
 (when it generated the BR at DN), and file was completed at NN side.
 Finally block will be marked as corrupt because of genstamp mismatch.
 {code}
  case RWR:
   if (!storedBlock.isComplete()) {
 return null; // not corrupt
   } else if (storedBlock.getGenerationStamp() != 
 iblk.getGenerationStamp()) {
 return new BlockToMarkCorrupt(storedBlock,
 reported  + reportedState +  replica with genstamp  +
 iblk.getGenerationStamp() +  does not match COMPLETE block's  +
 genstamp in block map  + storedBlock.getGenerationStamp());
   } else { // COMPLETE block, same genstamp
 {code}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3194) Continuous block scanning at DN side

2012-04-04 Thread Daryn Sharp (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3194?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13246268#comment-13246268
 ] 

Daryn Sharp commented on HDFS-3194:
---

I've seen this behavior in branch-2  branch-3.  Last year, I think Matt Foley 
fixed this or a similar issue with excessive block scanning.  This may be a 
regression.

 Continuous block scanning at DN side
 

 Key: HDFS-3194
 URL: https://issues.apache.org/jira/browse/HDFS-3194
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 1.0.3
Reporter: suja s
Priority: Minor
 Fix For: 1.0.3


 Block scanning interval by default should be taken as 21 days(3 weeks) and 
 each block scanning should happen once in 21 days.
 Here the block is being scanned continuosly.
 2012-04-03 10:44:47,056 INFO 
 org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner: Verification 
 succeeded for 
 BP-241703115-xx.xx.xx.55-1333086229434:blk_-2666054955039014473_1003
 2012-04-03 10:45:02,064 INFO 
 org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner: Verification 
 succeeded for 
 BP-241703115-xx.xx.xx.55-1333086229434:blk_-2666054955039014473_1003
 2012-04-03 10:45:17,071 INFO 
 org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner: Verification 
 succeeded for 
 BP-241703115-xx.xx.xx.55-1333086229434:blk_-2666054955039014473_1003
 2012-04-03 10:45:32,079 INFO 
 org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner: Verification 
 succeeded for BP-241703115-xx.xx.xx.55-1333086229434:blk_-2666054955039014473

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3176) JsonUtil should not parse the MD5MD5CRC32FileChecksum bytes on its own.

2012-04-04 Thread Hudson (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3176?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13246274#comment-13246274
 ] 

Hudson commented on HDFS-3176:
--

Integrated in Hadoop-Mapreduce-trunk #1040 (See 
[https://builds.apache.org/job/Hadoop-Mapreduce-trunk/1040/])
HDFS-3176. Use MD5MD5CRC32FileChecksum.readFields() in JsonUtil .  
Contributed by Kihwal Lee (Revision 1309114)

 Result = FAILURE
szetszwo : 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1309114
Files : 
* /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/web/JsonUtil.java


 JsonUtil should not parse the MD5MD5CRC32FileChecksum bytes on its own.
 ---

 Key: HDFS-3176
 URL: https://issues.apache.org/jira/browse/HDFS-3176
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: hdfs client
Affects Versions: 0.23.0, 1.0.1
Reporter: Kihwal Lee
Assignee: Kihwal Lee
 Fix For: 1.1.0, 0.23.3, 2.0.0, 3.0.0

 Attachments: hdfs-3176-branch-1.patch, hdfs-3176.patch


 Currently JsonUtil used by webhdfs parses MD5MD5CRC32FileChecksum binary 
 bytes on its own and contructs a MD5MD5CRC32FileChecksum. It should instead 
 call MD5MD5CRC32FileChecksum.readFields().

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3168) Clean up FSNamesystem and BlockManager

2012-04-04 Thread Hudson (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3168?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13246275#comment-13246275
 ] 

Hudson commented on HDFS-3168:
--

Integrated in Hadoop-Mapreduce-trunk #1040 (See 
[https://builds.apache.org/job/Hadoop-Mapreduce-trunk/1040/])
HDFS-3168. Remove unnecessary throw IOException and change fields to 
final in FSNamesystem and BlockManager. (Revision 1309218)

 Result = FAILURE
szetszwo : 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1309218
Files : 
* /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/BlockManager.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSNamesystem.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/NameNode.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/ha/EditLogTailer.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/blockmanagement/BlockManagerTestUtil.java


 Clean up FSNamesystem and BlockManager
 --

 Key: HDFS-3168
 URL: https://issues.apache.org/jira/browse/HDFS-3168
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: name-node
Reporter: Tsz Wo (Nicholas), SZE
Assignee: Tsz Wo (Nicholas), SZE
 Fix For: 0.24.0, 0.23.3

 Attachments: h3168_20120330.patch, h3168_20120402.patch, 
 h3168_20120403.patch




--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3000) Add a public API for setting quotas

2012-04-04 Thread Hudson (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3000?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13246276#comment-13246276
 ] 

Hudson commented on HDFS-3000:
--

Integrated in Hadoop-Mapreduce-trunk #1040 (See 
[https://builds.apache.org/job/Hadoop-Mapreduce-trunk/1040/])
HDFS-3000. Add a public API for setting quotas. Contributed by Aaron T. 
Myers. (Revision 1309227)

 Result = FAILURE
atm : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1309227
Files : 
* /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/client
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/client/HdfsAdmin.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestHdfsAdmin.java


 Add a public API for setting quotas
 ---

 Key: HDFS-3000
 URL: https://issues.apache.org/jira/browse/HDFS-3000
 Project: Hadoop HDFS
  Issue Type: New Feature
  Components: hdfs client
Affects Versions: 2.0.0
Reporter: Aaron T. Myers
Assignee: Aaron T. Myers
 Attachments: HDFS-3000.patch, HDFS-3000.patch, HDFS-3000.patch, 
 HDFS-3000.patch, HDFS-3000.patch


 Currently one can set the quota of a file or directory from the command line, 
 but if a user wants to set it programmatically, they need to use 
 DistributedFileSystem, which is annotated InterfaceAudience.Private.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3166) Hftp connections do not have a timeout

2012-04-04 Thread Hudson (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3166?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13246278#comment-13246278
 ] 

Hudson commented on HDFS-3166:
--

Integrated in Hadoop-Mapreduce-trunk #1040 (See 
[https://builds.apache.org/job/Hadoop-Mapreduce-trunk/1040/])
HDFS-3166. Add timeout to Hftp connections.  Contributed by Daryn Sharp 
(Revision 1309103)

 Result = FAILURE
szetszwo : 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1309103
Files : 
* /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/HftpFileSystem.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/HsftpFileSystem.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/tools/DelegationTokenFetcher.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/web/URLUtils.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestHftpURLTimeouts.java


 Hftp connections do not have a timeout
 --

 Key: HDFS-3166
 URL: https://issues.apache.org/jira/browse/HDFS-3166
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: hdfs client
Affects Versions: 0.23.2, 0.23.3, 2.0.0, 3.0.0
Reporter: Daryn Sharp
Assignee: Daryn Sharp
Priority: Critical
 Fix For: 0.23.3, 2.0.0, 3.0.0

 Attachments: HADOOP-8221.branch-1.patch, HADOOP-8221.patch, 
 HADOOP-8221.patch, HDFS-3166.patch, HDFS-3166.patch


 Hftp connections do not have read timeouts.  This leads to indefinitely hung 
 sockets when there is a network outage during which time the remote host 
 closed the socket.
 This may also affect WebHdfs, etc.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3187) Upgrade guava to 11.0.2

2012-04-04 Thread Hudson (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3187?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13246277#comment-13246277
 ] 

Hudson commented on HDFS-3187:
--

Integrated in Hadoop-Mapreduce-trunk #1040 (See 
[https://builds.apache.org/job/Hadoop-Mapreduce-trunk/1040/])
HDFS-3187. Upgrade guava to 11.0.2. Contributed by Todd Lipcon. (Revision 
1309181)

 Result = FAILURE
todd : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1309181
Files : 
* /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/MiniDFSCluster.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/FSImageTestUtil.java
* /hadoop/common/trunk/hadoop-project/pom.xml


 Upgrade guava to 11.0.2
 ---

 Key: HDFS-3187
 URL: https://issues.apache.org/jira/browse/HDFS-3187
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: build
Affects Versions: 2.0.0
Reporter: Todd Lipcon
Assignee: Todd Lipcon
Priority: Minor
 Fix For: 2.0.0

 Attachments: hdfs-3187.txt


 Guava r11 includes some nice features which we'd like to use in the 
 implementation of HDFS-3077. In particular, 
 {{MoreExecutors.listeningDecorator}} allows a normal {{ExecutorService}} to 
 be turned into a {{ListeningExecutorService}}, so that tasks can be submitted 
 to it and then wrapped as {{ListenableFuture}}s.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3084) FenceMethod.tryFence() and ShellCommandFencer should pass namenodeId as well as host:port

2012-04-04 Thread Daryn Sharp (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3084?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13246316#comment-13246316
 ] 

Daryn Sharp commented on HDFS-3084:
---

I haven't been following all the fencing jiras, so qq, are tokens and/or their 
services involved?

 FenceMethod.tryFence() and ShellCommandFencer should pass namenodeId as well 
 as host:port
 -

 Key: HDFS-3084
 URL: https://issues.apache.org/jira/browse/HDFS-3084
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: ha
Affects Versions: 0.24.0, 0.23.3
Reporter: Philip Zeyliger
Assignee: Todd Lipcon
 Attachments: hdfs-3084.txt


 The FenceMethod interface passes along the host:port of the NN that needs to 
 be fenced.  That's great for the common case.  However, it's likely necessary 
 to have extra configuration parameters for fencing, and these are typically 
 keyed off the nameserviceId.namenodeId (if, for nothing else, consistency 
 with all the other parameters that are keyed off of namespaceId.namenodeId).  
 Obviously this can be backed out from the host:port, but it's inconvenient, 
 and requires iterating through all the configs.
 The shell interface exhibits the same issue: host:port is great for most 
 fencers, but if you need extra configs (like the host:port of the power 
 supply unit), those are harder to pipe through without the namenodeId.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HDFS-3109) Remove hsqldb exclusions from pom.xml

2012-04-04 Thread Thomas Graves (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3109?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Thomas Graves updated HDFS-3109:


Target Version/s: 0.23.3  (was: 0.23.2)

 Remove hsqldb exclusions from pom.xml
 -

 Key: HDFS-3109
 URL: https://issues.apache.org/jira/browse/HDFS-3109
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 0.23.2
Reporter: Ravi Prakash
Assignee: Ravi Prakash
 Attachments: HDFS-3109.patch


 Related to MAPREDUCE-3621

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Created] (HDFS-3195) Compilation error in FSNamesystem

2012-04-04 Thread Jason Lowe (Created) (JIRA)
Compilation error in FSNamesystem
-

 Key: HDFS-3195
 URL: https://issues.apache.org/jira/browse/HDFS-3195
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: build
Affects Versions: 2.0.0
Reporter: Jason Lowe
Priority: Blocker


Build fails when trying to build branch-2:

[ERROR] 
/hadoop/src/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSNamesystem.java:[2741,32]
 unreported exception java.io.IOException; must be caught or declared to be 
thrown


--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3195) Compilation error in FSNamesystem

2012-04-04 Thread Jason Lowe (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3195?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13246381#comment-13246381
 ] 

Jason Lowe commented on HDFS-3195:
--

Appears to have been caused by HDFS-3168 per [this 
comment|https://issues.apache.org/jira/browse/HDFS-3168?focusedCommentId=13245996page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-13245996].

 Compilation error in FSNamesystem
 -

 Key: HDFS-3195
 URL: https://issues.apache.org/jira/browse/HDFS-3195
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: build
Affects Versions: 2.0.0
Reporter: Jason Lowe
Priority: Blocker

 Build fails when trying to build branch-2:
 [ERROR] 
 /hadoop/src/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSNamesystem.java:[2741,32]
  unreported exception java.io.IOException; must be caught or declared to be 
 thrown

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3092) Enable journal protocol based editlog streaming for standby namenode

2012-04-04 Thread Flavio Junqueira (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3092?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13246385#comment-13246385
 ] 

Flavio Junqueira commented on HDFS-3092:


bq. Is there a way to turn off the striping even if Quorom size (Q) is less 
than Ensemble size (E)?
We like the idea that each Journal file contains ALL entries.
Our default config: Q is 2 and set of JDs is 3 (roughly equivalent to E).

Right now the write set is the same as the ack set, since we haven't had the 
use case you suggest so far. We have anticipated the need of different ways of 
scheduling reads and writes, and it is fairly simple to make it cover your use 
case. Ivan had actually suggested this separation between ack sets and write 
sets previously to deal with slow disks. 

I have tried to reflect this discussion in BOOKKEEPER-208 if you're interested. 
 

 Enable journal protocol based editlog streaming for standby namenode
 

 Key: HDFS-3092
 URL: https://issues.apache.org/jira/browse/HDFS-3092
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: ha, name-node
Affects Versions: 0.24.0, 0.23.3
Reporter: Suresh Srinivas
Assignee: Suresh Srinivas
 Attachments: MultipleSharedJournals.pdf, MultipleSharedJournals.pdf


 Currently standby namenode relies on reading shared editlogs to stay current 
 with the active namenode, for namespace changes. BackupNode used streaming 
 edits from active namenode for doing the same. This jira is to explore using 
 journal protocol based editlog streams for the standby namenode. A daemon in 
 standby will get the editlogs from the active and write it to local edits. To 
 begin with, the existing standby mechanism of reading from a file, will 
 continue to be used, instead of from shared edits, from the local edits.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Assigned] (HDFS-3195) Compilation error in FSNamesystem

2012-04-04 Thread Eli Collins (Assigned) (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3195?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Eli Collins reassigned HDFS-3195:
-

Assignee: Eli Collins

 Compilation error in FSNamesystem
 -

 Key: HDFS-3195
 URL: https://issues.apache.org/jira/browse/HDFS-3195
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: build
Affects Versions: 2.0.0
Reporter: Jason Lowe
Assignee: Eli Collins
Priority: Blocker

 Build fails when trying to build branch-2:
 [ERROR] 
 /hadoop/src/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSNamesystem.java:[2741,32]
  unreported exception java.io.IOException; must be caught or declared to be 
 thrown

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HDFS-3195) Compilation error in FSNamesystem

2012-04-04 Thread Eli Collins (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3195?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Eli Collins updated HDFS-3195:
--

Attachment: hdfs-3195.txt

Patch attached. 

The throws clause was removed in HDFS-2301 and it looks like HDFS-3168 
re-introduced it, which is kind of troublesome. I'll ping Nicholas on HDFS-3168 
to double check that he didn't accidentally revert other changes.

 Compilation error in FSNamesystem
 -

 Key: HDFS-3195
 URL: https://issues.apache.org/jira/browse/HDFS-3195
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: build
Affects Versions: 2.0.0
Reporter: Jason Lowe
Assignee: Eli Collins
Priority: Blocker
 Attachments: hdfs-3195.txt


 Build fails when trying to build branch-2:
 [ERROR] 
 /hadoop/src/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSNamesystem.java:[2741,32]
  unreported exception java.io.IOException; must be caught or declared to be 
 thrown

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HDFS-3165) HDFS Balancer scripts are refering to wrong path of hadoop-daemon.sh

2012-04-04 Thread amith (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3165?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

amith updated HDFS-3165:


Status: Open  (was: Patch Available)

 HDFS Balancer scripts are refering to wrong path of hadoop-daemon.sh
 

 Key: HDFS-3165
 URL: https://issues.apache.org/jira/browse/HDFS-3165
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: balancer
Affects Versions: 1.0.1
 Environment: HDFS Balancer
Reporter: amith
Priority: Minor
 Attachments: HDFS-3165.patch, HDFS-3165.patch

   Original Estimate: 1m
  Remaining Estimate: 1m

 HDFS Balancer scripts are refering to wrong path of hadoop-daemon.sh
 HOST-xx-xx-xx-xx:/home/amith/V1R2/namenode1/sbin # ./start-balancer.sh
 ./start-balancer.sh: line 27: 
 /home/amith/V1R2/namenode1/bin/hadoop-daemon.sh: No such file or directory
 currently hadoop-daemon.sh script is in sbin and not bin.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HDFS-3165) HDFS Balancer scripts are refering to wrong path of hadoop-daemon.sh

2012-04-04 Thread amith (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3165?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

amith updated HDFS-3165:


Attachment: HDFS-3165.patch

 HDFS Balancer scripts are refering to wrong path of hadoop-daemon.sh
 

 Key: HDFS-3165
 URL: https://issues.apache.org/jira/browse/HDFS-3165
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: balancer
Affects Versions: 1.0.1
 Environment: HDFS Balancer
Reporter: amith
Priority: Minor
 Attachments: HDFS-3165.patch, HDFS-3165.patch

   Original Estimate: 1m
  Remaining Estimate: 1m

 HDFS Balancer scripts are refering to wrong path of hadoop-daemon.sh
 HOST-xx-xx-xx-xx:/home/amith/V1R2/namenode1/sbin # ./start-balancer.sh
 ./start-balancer.sh: line 27: 
 /home/amith/V1R2/namenode1/bin/hadoop-daemon.sh: No such file or directory
 currently hadoop-daemon.sh script is in sbin and not bin.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Assigned] (HDFS-3195) Compilation error in FSNamesystem

2012-04-04 Thread Jason Lowe (Assigned) (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3195?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jason Lowe reassigned HDFS-3195:


Assignee: Tsz Wo (Nicholas), SZE  (was: Eli Collins)

Nicholas, could you take a look at this?  I think HDFS-3168 needs to be backed 
out of branch-2 or HDFS-2564 should be pulled into branch-2 to fix the build.

 Compilation error in FSNamesystem
 -

 Key: HDFS-3195
 URL: https://issues.apache.org/jira/browse/HDFS-3195
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: build
Affects Versions: 2.0.0
Reporter: Jason Lowe
Assignee: Tsz Wo (Nicholas), SZE
Priority: Blocker
 Attachments: hdfs-3195.txt


 Build fails when trying to build branch-2:
 [ERROR] 
 /hadoop/src/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSNamesystem.java:[2741,32]
  unreported exception java.io.IOException; must be caught or declared to be 
 thrown

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Assigned] (HDFS-3195) Compilation error in FSNamesystem

2012-04-04 Thread Jason Lowe (Assigned) (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3195?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jason Lowe reassigned HDFS-3195:


Assignee: Eli Collins  (was: Tsz Wo (Nicholas), SZE)

Whoops, comment race, sorry back to Eli.

 Compilation error in FSNamesystem
 -

 Key: HDFS-3195
 URL: https://issues.apache.org/jira/browse/HDFS-3195
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: build
Affects Versions: 2.0.0
Reporter: Jason Lowe
Assignee: Eli Collins
Priority: Blocker
 Attachments: hdfs-3195.txt


 Build fails when trying to build branch-2:
 [ERROR] 
 /hadoop/src/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSNamesystem.java:[2741,32]
  unreported exception java.io.IOException; must be caught or declared to be 
 thrown

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3195) Compilation error in FSNamesystem

2012-04-04 Thread Eli Collins (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3195?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13246397#comment-13246397
 ] 

Eli Collins commented on HDFS-3195:
---

Actually, it looks like HDFS-3168 removed the throws from reassignLease, which 
should have been done when  HDFS-2301 (part of HA) was merged to trunk.

 Compilation error in FSNamesystem
 -

 Key: HDFS-3195
 URL: https://issues.apache.org/jira/browse/HDFS-3195
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: build
Affects Versions: 2.0.0
Reporter: Jason Lowe
Assignee: Eli Collins
Priority: Blocker
 Attachments: hdfs-3195.txt


 Build fails when trying to build branch-2:
 [ERROR] 
 /hadoop/src/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSNamesystem.java:[2741,32]
  unreported exception java.io.IOException; must be caught or declared to be 
 thrown

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3165) HDFS Balancer scripts are refering to wrong path of hadoop-daemon.sh

2012-04-04 Thread amith (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3165?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13246396#comment-13246396
 ] 

amith commented on HDFS-3165:
-

Sorry for the delay in providing the correct patch.
please accept the new patch which is attached newly :)



 HDFS Balancer scripts are refering to wrong path of hadoop-daemon.sh
 

 Key: HDFS-3165
 URL: https://issues.apache.org/jira/browse/HDFS-3165
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: balancer
Affects Versions: 1.0.1
 Environment: HDFS Balancer
Reporter: amith
Priority: Minor
 Attachments: HDFS-3165.patch, HDFS-3165.patch

   Original Estimate: 1m
  Remaining Estimate: 1m

 HDFS Balancer scripts are refering to wrong path of hadoop-daemon.sh
 HOST-xx-xx-xx-xx:/home/amith/V1R2/namenode1/sbin # ./start-balancer.sh
 ./start-balancer.sh: line 27: 
 /home/amith/V1R2/namenode1/bin/hadoop-daemon.sh: No such file or directory
 currently hadoop-daemon.sh script is in sbin and not bin.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3102) Add CLI tool to initialize the shared-edits dir

2012-04-04 Thread Aaron T. Myers (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3102?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13246419#comment-13246419
 ] 

Aaron T. Myers commented on HDFS-3102:
--

The test failure appears to be unrelated. The test passes locally for me, and 
failed on the Jenkins slave with a java.io.IOException: Too many open files; 
error.

 Add CLI tool to initialize the shared-edits dir
 ---

 Key: HDFS-3102
 URL: https://issues.apache.org/jira/browse/HDFS-3102
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: ha, name-node
Affects Versions: 0.24.0, 2.0.0
Reporter: Todd Lipcon
Assignee: Aaron T. Myers
 Attachments: HDFS-3102.patch, HDFS-3102.patch


 Currently in order to make a non-HA NN HA, you need to initialize the shared 
 edits dir. This can be done manually by cping directories around. It would be 
 preferable to add a namenode -initializeSharedEdits command to achieve this 
 same effect.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3195) Compilation error in FSNamesystem

2012-04-04 Thread Eli Collins (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3195?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13246422#comment-13246422
 ] 

Eli Collins commented on HDFS-3195:
---

ATM had a better idea, fixing this by merging HDFS-2564 to branch-2, I'll do 
that and close that out when I've committed and verified the build is fixed.

 Compilation error in FSNamesystem
 -

 Key: HDFS-3195
 URL: https://issues.apache.org/jira/browse/HDFS-3195
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: build
Affects Versions: 2.0.0
Reporter: Jason Lowe
Assignee: Eli Collins
Priority: Blocker
 Attachments: hdfs-3195.txt


 Build fails when trying to build branch-2:
 [ERROR] 
 /hadoop/src/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSNamesystem.java:[2741,32]
  unreported exception java.io.IOException; must be caught or declared to be 
 thrown

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HDFS-2564) Cleanup unnecessary exceptions thrown and unnecessary casts

2012-04-04 Thread Eli Collins (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-2564?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Eli Collins updated HDFS-2564:
--

 Target Version/s:   (was: 0.23.1)
Affects Version/s: (was: 0.24.0)
Fix Version/s: (was: 3.0.0)
   2.0.0

I've merged this to branch-2

 Cleanup unnecessary exceptions thrown and unnecessary casts
 ---

 Key: HDFS-2564
 URL: https://issues.apache.org/jira/browse/HDFS-2564
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: data-node, hdfs client, name-node
Reporter: Hari Mankude
Assignee: Hari Mankude
Priority: Minor
 Fix For: 2.0.0

 Attachments: HDFS-2564-1.txt, hadoop-2564.trunk.patch, 
 hadoop-2564.trunk.patch, hadoop-2564.trunk.patch


 Cleaning up some of the java files with unnecessary exceptions and casts.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Resolved] (HDFS-3195) Compilation error in FSNamesystem

2012-04-04 Thread Eli Collins (Resolved) (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3195?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Eli Collins resolved HDFS-3195.
---

Resolution: Fixed

I've merged HDFS-2564 to branch-2, which fixes this issue.

 Compilation error in FSNamesystem
 -

 Key: HDFS-3195
 URL: https://issues.apache.org/jira/browse/HDFS-3195
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: build
Affects Versions: 2.0.0
Reporter: Jason Lowe
Assignee: Eli Collins
Priority: Blocker
 Attachments: hdfs-3195.txt


 Build fails when trying to build branch-2:
 [ERROR] 
 /hadoop/src/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSNamesystem.java:[2741,32]
  unreported exception java.io.IOException; must be caught or declared to be 
 thrown

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3084) FenceMethod.tryFence() and ShellCommandFencer should pass namenodeId as well as host:port

2012-04-04 Thread Todd Lipcon (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3084?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13246457#comment-13246457
 ] 

Todd Lipcon commented on HDFS-3084:
---

No, the scripts and failover controllers use keytab-based or straight user 
credentials.

 FenceMethod.tryFence() and ShellCommandFencer should pass namenodeId as well 
 as host:port
 -

 Key: HDFS-3084
 URL: https://issues.apache.org/jira/browse/HDFS-3084
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: ha
Affects Versions: 0.24.0, 0.23.3
Reporter: Philip Zeyliger
Assignee: Todd Lipcon
 Attachments: hdfs-3084.txt


 The FenceMethod interface passes along the host:port of the NN that needs to 
 be fenced.  That's great for the common case.  However, it's likely necessary 
 to have extra configuration parameters for fencing, and these are typically 
 keyed off the nameserviceId.namenodeId (if, for nothing else, consistency 
 with all the other parameters that are keyed off of namespaceId.namenodeId).  
 Obviously this can be backed out from the host:port, but it's inconvenient, 
 and requires iterating through all the configs.
 The shell interface exhibits the same issue: host:port is great for most 
 fencers, but if you need extra configs (like the host:port of the power 
 supply unit), those are harder to pipe through without the namenodeId.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-2564) Cleanup unnecessary exceptions thrown and unnecessary casts

2012-04-04 Thread Hudson (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2564?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13246458#comment-13246458
 ] 

Hudson commented on HDFS-2564:
--

Integrated in Hadoop-Hdfs-trunk-Commit #2069 (See 
[https://builds.apache.org/job/Hadoop-Hdfs-trunk-Commit/2069/])
Updated CHANGES.txt to reflect HDFS-2564 merge (Revision 1309488)

 Result = SUCCESS
eli : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1309488
Files : 
* /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt


 Cleanup unnecessary exceptions thrown and unnecessary casts
 ---

 Key: HDFS-2564
 URL: https://issues.apache.org/jira/browse/HDFS-2564
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: data-node, hdfs client, name-node
Reporter: Hari Mankude
Assignee: Hari Mankude
Priority: Minor
 Fix For: 2.0.0

 Attachments: HDFS-2564-1.txt, hadoop-2564.trunk.patch, 
 hadoop-2564.trunk.patch, hadoop-2564.trunk.patch


 Cleaning up some of the java files with unnecessary exceptions and casts.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-2564) Cleanup unnecessary exceptions thrown and unnecessary casts

2012-04-04 Thread Hudson (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2564?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13246460#comment-13246460
 ] 

Hudson commented on HDFS-2564:
--

Integrated in Hadoop-Common-trunk-Commit #1994 (See 
[https://builds.apache.org/job/Hadoop-Common-trunk-Commit/1994/])
Updated CHANGES.txt to reflect HDFS-2564 merge (Revision 1309488)

 Result = SUCCESS
eli : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1309488
Files : 
* /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt


 Cleanup unnecessary exceptions thrown and unnecessary casts
 ---

 Key: HDFS-2564
 URL: https://issues.apache.org/jira/browse/HDFS-2564
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: data-node, hdfs client, name-node
Reporter: Hari Mankude
Assignee: Hari Mankude
Priority: Minor
 Fix For: 2.0.0

 Attachments: HDFS-2564-1.txt, hadoop-2564.trunk.patch, 
 hadoop-2564.trunk.patch, hadoop-2564.trunk.patch


 Cleaning up some of the java files with unnecessary exceptions and casts.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HDFS-3000) Add a public API for setting quotas

2012-04-04 Thread Aaron T. Myers (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3000?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Aaron T. Myers updated HDFS-3000:
-

   Resolution: Fixed
Fix Version/s: 2.0.0
 Hadoop Flags: Reviewed
   Status: Resolved  (was: Patch Available)

I've just committed this to trunk and branch-2.

Thanks a lot for the reviews.

 Add a public API for setting quotas
 ---

 Key: HDFS-3000
 URL: https://issues.apache.org/jira/browse/HDFS-3000
 Project: Hadoop HDFS
  Issue Type: New Feature
  Components: hdfs client
Affects Versions: 2.0.0
Reporter: Aaron T. Myers
Assignee: Aaron T. Myers
 Fix For: 2.0.0

 Attachments: HDFS-3000.patch, HDFS-3000.patch, HDFS-3000.patch, 
 HDFS-3000.patch, HDFS-3000.patch


 Currently one can set the quota of a file or directory from the command line, 
 but if a user wants to set it programmatically, they need to use 
 DistributedFileSystem, which is annotated InterfaceAudience.Private.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HDFS-3050) rework OEV to share more code with the NameNode

2012-04-04 Thread Eli Collins (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3050?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Eli Collins updated HDFS-3050:
--

Target Version/s: 2.0.0

 rework OEV to share more code with the NameNode
 ---

 Key: HDFS-3050
 URL: https://issues.apache.org/jira/browse/HDFS-3050
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: name-node
Reporter: Colin Patrick McCabe
Assignee: Colin Patrick McCabe
Priority: Minor
 Attachments: HDFS-3050.006.patch, HDFS-3050.007.patch, 
 HDFS-3050.008.patch, HDFS-3050.009.patch, HDFS-3050.010.patch, 
 HDFS-3050.011.patch, HDFS-3050.012.patch, HDFS-3050.014.patch, 
 HDFS-3050.015.patch, HDFS-3050.016.patch, HDFS-3050.017.patch, 
 HDFS-3050.018.patch


 Current, OEV (the offline edits viewer) re-implements all of the opcode 
 parsing logic found in the NameNode.  This duplicated code creates a 
 maintenance burden for us.
 OEV should be refactored to simply use the normal EditLog parsing code, 
 rather than rolling its own.  By using the existing FSEditLogLoader code to 
 load edits in OEV, we can avoid having to update two places when the format 
 changes.
 We should not put opcode checksums into the XML, because they are a 
 serialization detail, not related to what the data is what we're storing.  
 This will also make it possible to modify the XML file and translate this 
 modified file back to a binary edits log file.
 Finally, this changes introduces --fix-txids.  When OEV is passed this flag, 
 it will close gaps in the transaction log by modifying the sequence numbers.  
 This is useful if you want to modify the edit log XML (say, by removing a 
 transaction), and transform the modified XML back into a valid binary edit 
 log file.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3102) Add CLI tool to initialize the shared-edits dir

2012-04-04 Thread Eli Collins (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3102?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13246501#comment-13246501
 ] 

Eli Collins commented on HDFS-3102:
---

Two minor comments, otherwise looks excellent

- Since initializeSharedEdits swallows most IOEs and returns a boolean I'd 
update the throws javadoc to indicate an IOE isn't throw for most error cases. 
Or perhaps swallow the IOE from unlockAll and return false tehre as well
- Good to add a test that the force option works, specifically w/o specifying 
it we don't blow away your existing shared edits dirs. The code obviously looks 
correct but this would be  a bad bug to introduce so worth checking


 Add CLI tool to initialize the shared-edits dir
 ---

 Key: HDFS-3102
 URL: https://issues.apache.org/jira/browse/HDFS-3102
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: ha, name-node
Affects Versions: 0.24.0, 2.0.0
Reporter: Todd Lipcon
Assignee: Aaron T. Myers
 Attachments: HDFS-3102.patch, HDFS-3102.patch


 Currently in order to make a non-HA NN HA, you need to initialize the shared 
 edits dir. This can be done manually by cping directories around. It would be 
 preferable to add a namenode -initializeSharedEdits command to achieve this 
 same effect.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-2564) Cleanup unnecessary exceptions thrown and unnecessary casts

2012-04-04 Thread Hudson (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2564?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13246502#comment-13246502
 ] 

Hudson commented on HDFS-2564:
--

Integrated in Hadoop-Mapreduce-trunk-Commit #2006 (See 
[https://builds.apache.org/job/Hadoop-Mapreduce-trunk-Commit/2006/])
Updated CHANGES.txt to reflect HDFS-2564 merge (Revision 1309488)

 Result = SUCCESS
eli : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1309488
Files : 
* /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt


 Cleanup unnecessary exceptions thrown and unnecessary casts
 ---

 Key: HDFS-2564
 URL: https://issues.apache.org/jira/browse/HDFS-2564
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: data-node, hdfs client, name-node
Reporter: Hari Mankude
Assignee: Hari Mankude
Priority: Minor
 Fix For: 2.0.0

 Attachments: HDFS-2564-1.txt, hadoop-2564.trunk.patch, 
 hadoop-2564.trunk.patch, hadoop-2564.trunk.patch


 Cleaning up some of the java files with unnecessary exceptions and casts.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HDFS-3102) Add CLI tool to initialize the shared-edits dir

2012-04-04 Thread Aaron T. Myers (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3102?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Aaron T. Myers updated HDFS-3102:
-

Attachment: HDFS-3102.patch

Here's an updated patch which incorporates Eli's feedback.

bq. Since initializeSharedEdits swallows most IOEs and returns a boolean I'd 
update the throws javadoc to indicate an IOE isn't throw for most error cases. 
Or perhaps swallow the IOE from unlockAll and return false tehre as well

Good thinking. I made it catch/log/return in the case of an IOE from unlockAll, 
and removed the throws declaration.

bq. Good to add a test that the force option works, specifically w/o specifying 
it we don't blow away your existing shared edits dirs. The code obviously looks 
correct but this would be a bad bug to introduce so worth checking

Note that the force option isn't actually exposed to the user - it's only used 
in testing.

Regardless, you're right that it's useful to add a test that we won't overwrite 
directories without confirmation. I've add a little test to do so.

 Add CLI tool to initialize the shared-edits dir
 ---

 Key: HDFS-3102
 URL: https://issues.apache.org/jira/browse/HDFS-3102
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: ha, name-node
Affects Versions: 0.24.0, 2.0.0
Reporter: Todd Lipcon
Assignee: Aaron T. Myers
 Attachments: HDFS-3102.patch, HDFS-3102.patch, HDFS-3102.patch


 Currently in order to make a non-HA NN HA, you need to initialize the shared 
 edits dir. This can be done manually by cping directories around. It would be 
 preferable to add a namenode -initializeSharedEdits command to achieve this 
 same effect.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3102) Add CLI tool to initialize the shared-edits dir

2012-04-04 Thread Eli Collins (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3102?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13246542#comment-13246542
 ] 

Eli Collins commented on HDFS-3102:
---

+1 looks good

 Add CLI tool to initialize the shared-edits dir
 ---

 Key: HDFS-3102
 URL: https://issues.apache.org/jira/browse/HDFS-3102
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: ha, name-node
Affects Versions: 0.24.0, 2.0.0
Reporter: Todd Lipcon
Assignee: Aaron T. Myers
 Attachments: HDFS-3102.patch, HDFS-3102.patch, HDFS-3102.patch


 Currently in order to make a non-HA NN HA, you need to initialize the shared 
 edits dir. This can be done manually by cping directories around. It would be 
 preferable to add a namenode -initializeSharedEdits command to achieve this 
 same effect.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HDFS-3050) rework OEV to share more code with the NameNode

2012-04-04 Thread Colin Patrick McCabe (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3050?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Colin Patrick McCabe updated HDFS-3050:
---

Attachment: HDFS-3050.019.patch

* fix error handling bug that could lead to open files getting leaked 
(theoretically)

* suppress javadoc warnings resulting from com.sun.* API use

 rework OEV to share more code with the NameNode
 ---

 Key: HDFS-3050
 URL: https://issues.apache.org/jira/browse/HDFS-3050
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: name-node
Reporter: Colin Patrick McCabe
Assignee: Colin Patrick McCabe
Priority: Minor
 Attachments: HDFS-3050.006.patch, HDFS-3050.007.patch, 
 HDFS-3050.008.patch, HDFS-3050.009.patch, HDFS-3050.010.patch, 
 HDFS-3050.011.patch, HDFS-3050.012.patch, HDFS-3050.014.patch, 
 HDFS-3050.015.patch, HDFS-3050.016.patch, HDFS-3050.017.patch, 
 HDFS-3050.018.patch, HDFS-3050.019.patch


 Current, OEV (the offline edits viewer) re-implements all of the opcode 
 parsing logic found in the NameNode.  This duplicated code creates a 
 maintenance burden for us.
 OEV should be refactored to simply use the normal EditLog parsing code, 
 rather than rolling its own.  By using the existing FSEditLogLoader code to 
 load edits in OEV, we can avoid having to update two places when the format 
 changes.
 We should not put opcode checksums into the XML, because they are a 
 serialization detail, not related to what the data is what we're storing.  
 This will also make it possible to modify the XML file and translate this 
 modified file back to a binary edits log file.
 Finally, this changes introduces --fix-txids.  When OEV is passed this flag, 
 it will close gaps in the transaction log by modifying the sequence numbers.  
 This is useful if you want to modify the edit log XML (say, by removing a 
 transaction), and transform the modified XML back into a valid binary edit 
 log file.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3195) Compilation error in FSNamesystem

2012-04-04 Thread Tsz Wo (Nicholas), SZE (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3195?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13246564#comment-13246564
 ] 

Tsz Wo (Nicholas), SZE commented on HDFS-3195:
--

Very sorry that I did not check branch-2 when merging HDFS-3168 last night and 
thanks for fixing it!

 Compilation error in FSNamesystem
 -

 Key: HDFS-3195
 URL: https://issues.apache.org/jira/browse/HDFS-3195
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: build
Affects Versions: 2.0.0
Reporter: Jason Lowe
Assignee: Eli Collins
Priority: Blocker
 Attachments: hdfs-3195.txt


 Build fails when trying to build branch-2:
 [ERROR] 
 /hadoop/src/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSNamesystem.java:[2741,32]
  unreported exception java.io.IOException; must be caught or declared to be 
 thrown

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3168) Clean up FSNamesystem and BlockManager

2012-04-04 Thread Tsz Wo (Nicholas), SZE (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3168?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13246572#comment-13246572
 ] 

Tsz Wo (Nicholas), SZE commented on HDFS-3168:
--

 Also, not a big deal, but in the future please get a +1 from a committer 
 before committing a patch.

Aaron, it is nothing to do with it.  Any contributor could review code.  It was 
a merging problem.

 Clean up FSNamesystem and BlockManager
 --

 Key: HDFS-3168
 URL: https://issues.apache.org/jira/browse/HDFS-3168
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: name-node
Reporter: Tsz Wo (Nicholas), SZE
Assignee: Tsz Wo (Nicholas), SZE
 Fix For: 0.24.0, 0.23.3

 Attachments: h3168_20120330.patch, h3168_20120402.patch, 
 h3168_20120403.patch




--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HDFS-3084) FenceMethod.tryFence() and ShellCommandFencer should pass namenodeId as well as host:port

2012-04-04 Thread Todd Lipcon (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3084?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Todd Lipcon updated HDFS-3084:
--

  Resolution: Fixed
   Fix Version/s: 2.0.0
Target Version/s: 2.0.0  (was: 0.23.3, 0.24.0)
Hadoop Flags: Reviewed
  Status: Resolved  (was: Patch Available)

Committed to branch-2 and trunk, thanks for reviews.

 FenceMethod.tryFence() and ShellCommandFencer should pass namenodeId as well 
 as host:port
 -

 Key: HDFS-3084
 URL: https://issues.apache.org/jira/browse/HDFS-3084
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: ha
Affects Versions: 0.24.0, 0.23.3
Reporter: Philip Zeyliger
Assignee: Todd Lipcon
 Fix For: 2.0.0

 Attachments: hdfs-3084.txt


 The FenceMethod interface passes along the host:port of the NN that needs to 
 be fenced.  That's great for the common case.  However, it's likely necessary 
 to have extra configuration parameters for fencing, and these are typically 
 keyed off the nameserviceId.namenodeId (if, for nothing else, consistency 
 with all the other parameters that are keyed off of namespaceId.namenodeId).  
 Obviously this can be backed out from the host:port, but it's inconvenient, 
 and requires iterating through all the configs.
 The shell interface exhibits the same issue: host:port is great for most 
 fencers, but if you need extra configs (like the host:port of the power 
 supply unit), those are harder to pipe through without the namenodeId.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3084) FenceMethod.tryFence() and ShellCommandFencer should pass namenodeId as well as host:port

2012-04-04 Thread Hudson (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3084?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13246589#comment-13246589
 ] 

Hudson commented on HDFS-3084:
--

Integrated in Hadoop-Hdfs-trunk-Commit #2070 (See 
[https://builds.apache.org/job/Hadoop-Hdfs-trunk-Commit/2070/])
HDFS-3084. FenceMethod.tryFence() and ShellCommandFencer should pass 
namenodeId as well as host:port. Contributed by Todd Lipcon. (Revision 1309526)

 Result = SUCCESS
todd : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1309526
Files : 
* /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/DFSUtil.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/tools/NNHAServiceTarget.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestDFSUtil.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/tools/TestDFSHAAdminMiniCluster.java


 FenceMethod.tryFence() and ShellCommandFencer should pass namenodeId as well 
 as host:port
 -

 Key: HDFS-3084
 URL: https://issues.apache.org/jira/browse/HDFS-3084
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: ha
Affects Versions: 0.24.0, 0.23.3
Reporter: Philip Zeyliger
Assignee: Todd Lipcon
 Fix For: 2.0.0

 Attachments: hdfs-3084.txt


 The FenceMethod interface passes along the host:port of the NN that needs to 
 be fenced.  That's great for the common case.  However, it's likely necessary 
 to have extra configuration parameters for fencing, and these are typically 
 keyed off the nameserviceId.namenodeId (if, for nothing else, consistency 
 with all the other parameters that are keyed off of namespaceId.namenodeId).  
 Obviously this can be backed out from the host:port, but it's inconvenient, 
 and requires iterating through all the configs.
 The shell interface exhibits the same issue: host:port is great for most 
 fencers, but if you need extra configs (like the host:port of the power 
 supply unit), those are harder to pipe through without the namenodeId.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3084) FenceMethod.tryFence() and ShellCommandFencer should pass namenodeId as well as host:port

2012-04-04 Thread Hudson (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3084?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13246591#comment-13246591
 ] 

Hudson commented on HDFS-3084:
--

Integrated in Hadoop-Common-trunk-Commit #1995 (See 
[https://builds.apache.org/job/Hadoop-Common-trunk-Commit/1995/])
HDFS-3084. FenceMethod.tryFence() and ShellCommandFencer should pass 
namenodeId as well as host:port. Contributed by Todd Lipcon. (Revision 1309526)

 Result = SUCCESS
todd : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1309526
Files : 
* /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/DFSUtil.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/tools/NNHAServiceTarget.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestDFSUtil.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/tools/TestDFSHAAdminMiniCluster.java


 FenceMethod.tryFence() and ShellCommandFencer should pass namenodeId as well 
 as host:port
 -

 Key: HDFS-3084
 URL: https://issues.apache.org/jira/browse/HDFS-3084
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: ha
Affects Versions: 0.24.0, 0.23.3
Reporter: Philip Zeyliger
Assignee: Todd Lipcon
 Fix For: 2.0.0

 Attachments: hdfs-3084.txt


 The FenceMethod interface passes along the host:port of the NN that needs to 
 be fenced.  That's great for the common case.  However, it's likely necessary 
 to have extra configuration parameters for fencing, and these are typically 
 keyed off the nameserviceId.namenodeId (if, for nothing else, consistency 
 with all the other parameters that are keyed off of namespaceId.namenodeId).  
 Obviously this can be backed out from the host:port, but it's inconvenient, 
 and requires iterating through all the configs.
 The shell interface exhibits the same issue: host:port is great for most 
 fencers, but if you need extra configs (like the host:port of the power 
 supply unit), those are harder to pipe through without the namenodeId.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3084) FenceMethod.tryFence() and ShellCommandFencer should pass namenodeId as well as host:port

2012-04-04 Thread Hudson (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3084?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13246596#comment-13246596
 ] 

Hudson commented on HDFS-3084:
--

Integrated in Hadoop-Mapreduce-trunk-Commit #2007 (See 
[https://builds.apache.org/job/Hadoop-Mapreduce-trunk-Commit/2007/])
HDFS-3084. FenceMethod.tryFence() and ShellCommandFencer should pass 
namenodeId as well as host:port. Contributed by Todd Lipcon. (Revision 1309526)

 Result = SUCCESS
todd : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1309526
Files : 
* /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/DFSUtil.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/tools/NNHAServiceTarget.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestDFSUtil.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/tools/TestDFSHAAdminMiniCluster.java


 FenceMethod.tryFence() and ShellCommandFencer should pass namenodeId as well 
 as host:port
 -

 Key: HDFS-3084
 URL: https://issues.apache.org/jira/browse/HDFS-3084
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: ha
Affects Versions: 0.24.0, 0.23.3
Reporter: Philip Zeyliger
Assignee: Todd Lipcon
 Fix For: 2.0.0

 Attachments: hdfs-3084.txt


 The FenceMethod interface passes along the host:port of the NN that needs to 
 be fenced.  That's great for the common case.  However, it's likely necessary 
 to have extra configuration parameters for fencing, and these are typically 
 keyed off the nameserviceId.namenodeId (if, for nothing else, consistency 
 with all the other parameters that are keyed off of namespaceId.namenodeId).  
 Obviously this can be backed out from the host:port, but it's inconvenient, 
 and requires iterating through all the configs.
 The shell interface exhibits the same issue: host:port is great for most 
 fencers, but if you need extra configs (like the host:port of the power 
 supply unit), those are harder to pipe through without the namenodeId.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3168) Clean up FSNamesystem and BlockManager

2012-04-04 Thread Todd Lipcon (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3168?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13246600#comment-13246600
 ] 

Todd Lipcon commented on HDFS-3168:
---

bq. Aaron, it is nothing to do with it. Any contributor could review code. It 
was a merging problem.

Is that the case? I have no opinion on this particular patch and whether a 
different reviewer might have seen the issue. But I thought you had to get a 
committer +1 to commit things...

 Clean up FSNamesystem and BlockManager
 --

 Key: HDFS-3168
 URL: https://issues.apache.org/jira/browse/HDFS-3168
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: name-node
Reporter: Tsz Wo (Nicholas), SZE
Assignee: Tsz Wo (Nicholas), SZE
 Fix For: 0.24.0, 0.23.3

 Attachments: h3168_20120330.patch, h3168_20120402.patch, 
 h3168_20120403.patch




--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3168) Clean up FSNamesystem and BlockManager

2012-04-04 Thread Tsz Wo (Nicholas), SZE (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3168?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13246610#comment-13246610
 ] 

Tsz Wo (Nicholas), SZE commented on HDFS-3168:
--

Could the committer be the same as the contributor?

 Clean up FSNamesystem and BlockManager
 --

 Key: HDFS-3168
 URL: https://issues.apache.org/jira/browse/HDFS-3168
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: name-node
Reporter: Tsz Wo (Nicholas), SZE
Assignee: Tsz Wo (Nicholas), SZE
 Fix For: 0.24.0, 0.23.3

 Attachments: h3168_20120330.patch, h3168_20120402.patch, 
 h3168_20120403.patch




--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3102) Add CLI tool to initialize the shared-edits dir

2012-04-04 Thread Hadoop QA (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3102?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13246624#comment-13246624
 ] 

Hadoop QA commented on HDFS-3102:
-

+1 overall.  Here are the results of testing the latest attachment 
  http://issues.apache.org/jira/secure/attachment/12521349/HDFS-3102.patch
  against trunk revision .

+1 @author.  The patch does not contain any @author tags.

+1 tests included.  The patch appears to include 6 new or modified tests.

+1 javadoc.  The javadoc tool did not generate any warning messages.

+1 javac.  The applied patch does not increase the total number of javac 
compiler warnings.

+1 eclipse:eclipse.  The patch built with eclipse:eclipse.

+1 findbugs.  The patch does not introduce any new Findbugs (version 1.3.9) 
warnings.

+1 release audit.  The applied patch does not increase the total number of 
release audit warnings.

+1 core tests.  The patch passed unit tests in .

+1 contrib tests.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-HDFS-Build/2183//testReport/
Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/2183//console

This message is automatically generated.

 Add CLI tool to initialize the shared-edits dir
 ---

 Key: HDFS-3102
 URL: https://issues.apache.org/jira/browse/HDFS-3102
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: ha, name-node
Affects Versions: 0.24.0, 2.0.0
Reporter: Todd Lipcon
Assignee: Aaron T. Myers
 Attachments: HDFS-3102.patch, HDFS-3102.patch, HDFS-3102.patch


 Currently in order to make a non-HA NN HA, you need to initialize the shared 
 edits dir. This can be done manually by cping directories around. It would be 
 preferable to add a namenode -initializeSharedEdits command to achieve this 
 same effect.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HDFS-3178) Add states for journal synchronization in journal daemon

2012-04-04 Thread Tsz Wo (Nicholas), SZE (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3178?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tsz Wo (Nicholas), SZE updated HDFS-3178:
-

Attachment: h3178_20120404_svn_mv.patch

The file should be 3178_20120404_svn_mv.patch, not h3179_20120403.patch.

 Add states for journal synchronization in journal daemon
 

 Key: HDFS-3178
 URL: https://issues.apache.org/jira/browse/HDFS-3178
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: ha, name-node
Reporter: Tsz Wo (Nicholas), SZE
Assignee: Tsz Wo (Nicholas), SZE
 Attachments: h3178_20120403_svn_mv.patch, 
 h3178_20120404_svn_mv.patch, svn_mv.sh


 Journal in a new daemon has to be synchronized to the current transaction.  
 It requires new states such as WaitingForRoll, Syncing and Synced.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HDFS-3178) Add states for journal synchronization in journal daemon

2012-04-04 Thread Tsz Wo (Nicholas), SZE (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3178?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tsz Wo (Nicholas), SZE updated HDFS-3178:
-

Attachment: h3179_20120403.patch

h3179_20120403.patch:
- renames StateMachine to StateHandler
- moves JournalListener to a standalone interface.

(The patch did not include removing the old files so that it could be applied 
cleanly after svn_mv.sh.)

 Add states for journal synchronization in journal daemon
 

 Key: HDFS-3178
 URL: https://issues.apache.org/jira/browse/HDFS-3178
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: ha, name-node
Reporter: Tsz Wo (Nicholas), SZE
Assignee: Tsz Wo (Nicholas), SZE
 Attachments: h3178_20120403_svn_mv.patch, 
 h3178_20120404_svn_mv.patch, svn_mv.sh


 Journal in a new daemon has to be synchronized to the current transaction.  
 It requires new states such as WaitingForRoll, Syncing and Synced.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HDFS-3178) Add states for journal synchronization in journal daemon

2012-04-04 Thread Tsz Wo (Nicholas), SZE (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3178?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tsz Wo (Nicholas), SZE updated HDFS-3178:
-

Attachment: (was: h3179_20120403.patch)

 Add states for journal synchronization in journal daemon
 

 Key: HDFS-3178
 URL: https://issues.apache.org/jira/browse/HDFS-3178
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: ha, name-node
Reporter: Tsz Wo (Nicholas), SZE
Assignee: Tsz Wo (Nicholas), SZE
 Attachments: h3178_20120403_svn_mv.patch, 
 h3178_20120404_svn_mv.patch, svn_mv.sh


 Journal in a new daemon has to be synchronized to the current transaction.  
 It requires new states such as WaitingForRoll, Syncing and Synced.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3168) Clean up FSNamesystem and BlockManager

2012-04-04 Thread Todd Lipcon (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3168?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13246651#comment-13246651
 ] 

Todd Lipcon commented on HDFS-3168:
---

By my understanding of our policies, the committer who provides the +1 has to 
be someone separate than the patch author. On branches I'm fine being lax here, 
since we need three +1s to merge a branch, but on trunk, I think it merits a 
discussion if there is disagreement on what our policies are.

 Clean up FSNamesystem and BlockManager
 --

 Key: HDFS-3168
 URL: https://issues.apache.org/jira/browse/HDFS-3168
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: name-node
Reporter: Tsz Wo (Nicholas), SZE
Assignee: Tsz Wo (Nicholas), SZE
 Fix For: 0.24.0, 0.23.3

 Attachments: h3168_20120330.patch, h3168_20120402.patch, 
 h3168_20120403.patch




--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HDFS-3178) Add states for journal synchronization in journal daemon

2012-04-04 Thread Tsz Wo (Nicholas), SZE (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3178?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tsz Wo (Nicholas), SZE updated HDFS-3178:
-

Attachment: h3178_20120404.patch

h3178_20120404.patch: same patch for Jenkins.

 Add states for journal synchronization in journal daemon
 

 Key: HDFS-3178
 URL: https://issues.apache.org/jira/browse/HDFS-3178
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: ha, name-node
Reporter: Tsz Wo (Nicholas), SZE
Assignee: Tsz Wo (Nicholas), SZE
 Attachments: h3178_20120403_svn_mv.patch, h3178_20120404.patch, 
 h3178_20120404_svn_mv.patch, svn_mv.sh


 Journal in a new daemon has to be synchronized to the current transaction.  
 It requires new states such as WaitingForRoll, Syncing and Synced.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HDFS-3178) Add states for journal synchronization in journal daemon

2012-04-04 Thread Tsz Wo (Nicholas), SZE (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3178?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tsz Wo (Nicholas), SZE updated HDFS-3178:
-

Status: Patch Available  (was: Open)

 Add states for journal synchronization in journal daemon
 

 Key: HDFS-3178
 URL: https://issues.apache.org/jira/browse/HDFS-3178
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: ha, name-node
Reporter: Tsz Wo (Nicholas), SZE
Assignee: Tsz Wo (Nicholas), SZE
 Attachments: h3178_20120403_svn_mv.patch, h3178_20120404.patch, 
 h3178_20120404_svn_mv.patch, svn_mv.sh


 Journal in a new daemon has to be synchronized to the current transaction.  
 It requires new states such as WaitingForRoll, Syncing and Synced.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3168) Clean up FSNamesystem and BlockManager

2012-04-04 Thread Tsz Wo (Nicholas), SZE (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3168?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13246657#comment-13246657
 ] 

Tsz Wo (Nicholas), SZE commented on HDFS-3168:
--

 By my understanding of our policies, the committer who provides the +1 has to 
 be someone separate than the patch author. ...

I think this is not true.  After a committer provided a patch and a 
non-committer reviewed it, the same committer could commit the patch.

 Clean up FSNamesystem and BlockManager
 --

 Key: HDFS-3168
 URL: https://issues.apache.org/jira/browse/HDFS-3168
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: name-node
Reporter: Tsz Wo (Nicholas), SZE
Assignee: Tsz Wo (Nicholas), SZE
 Fix For: 0.24.0, 0.23.3

 Attachments: h3168_20120330.patch, h3168_20120402.patch, 
 h3168_20120403.patch




--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3192) Active NN should exit when it has not received a getServiceStatus() rpc from ZKFC for timeout secs

2012-04-04 Thread Hari Mankude (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3192?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13246661#comment-13246661
 ] 

Hari Mankude commented on HDFS-3192:


bq.Why?

I think it's an advantage that the FC may die and come back, or that you may 
start the FCs after the NNs.

Well, if FC restarts the health monitoring within the timeout period, then NN 
will not die. However, if FC is having a gc pause or is not restarting, then NN 
should die. This is the first level of protection where in if NN is healthy, it 
can stonith itself.



 Active NN should exit when it has not received a getServiceStatus() rpc from 
 ZKFC for timeout secs
 --

 Key: HDFS-3192
 URL: https://issues.apache.org/jira/browse/HDFS-3192
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: ha, name-node
Reporter: Hari Mankude
Assignee: Hari Mankude



--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Created] (HDFS-3196) Implement JournalListener for writing journal to local disk

2012-04-04 Thread Tsz Wo (Nicholas), SZE (Created) (JIRA)
Implement JournalListener for writing journal to local disk
---

 Key: HDFS-3196
 URL: https://issues.apache.org/jira/browse/HDFS-3196
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Tsz Wo (Nicholas), SZE
Assignee: Tsz Wo (Nicholas), SZE




--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3192) Active NN should exit when it has not received a getServiceStatus() rpc from ZKFC for timeout secs

2012-04-04 Thread Todd Lipcon (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3192?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13246671#comment-13246671
 ] 

Todd Lipcon commented on HDFS-3192:
---

Why add multiple stonith paths, given we need external stonith anyway? It just 
adds to the complexity by increasing the number of scenarios we have to debug, 
etc.

That is to say: if the ZKFC dies, then it will lose its lock, and the other 
node will stonith this one when it takes over. What's the benefit of having it 
abort itself at the same time? In fact, it seems to be detrimental, because if 
it stays up, the other node can do a graceful transitionToStandby() call rather 
than having to do something more drastic like a full abort.

 Active NN should exit when it has not received a getServiceStatus() rpc from 
 ZKFC for timeout secs
 --

 Key: HDFS-3192
 URL: https://issues.apache.org/jira/browse/HDFS-3192
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: ha, name-node
Reporter: Hari Mankude
Assignee: Hari Mankude



--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3050) rework OEV to share more code with the NameNode

2012-04-04 Thread Hadoop QA (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3050?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13246675#comment-13246675
 ] 

Hadoop QA commented on HDFS-3050:
-

-1 overall.  Here are the results of testing the latest attachment 
  http://issues.apache.org/jira/secure/attachment/12521352/HDFS-3050.019.patch
  against trunk revision .

+1 @author.  The patch does not contain any @author tags.

+1 tests included.  The patch appears to include 9 new or modified tests.

-1 javadoc.  The javadoc tool appears to have generated 2 warning messages.

-1 javac.  The applied patch generated 20 javac compiler warnings (more 
than the trunk's current 14 warnings).

+1 eclipse:eclipse.  The patch built with eclipse:eclipse.

+1 findbugs.  The patch does not introduce any new Findbugs (version 1.3.9) 
warnings.

+1 release audit.  The applied patch does not increase the total number of 
release audit warnings.

-1 core tests.  The patch failed these unit tests:
  
org.apache.hadoop.hdfs.server.namenode.TestValidateConfigurationSettings

+1 contrib tests.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-HDFS-Build/2184//testReport/
Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/2184//console

This message is automatically generated.

 rework OEV to share more code with the NameNode
 ---

 Key: HDFS-3050
 URL: https://issues.apache.org/jira/browse/HDFS-3050
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: name-node
Reporter: Colin Patrick McCabe
Assignee: Colin Patrick McCabe
Priority: Minor
 Attachments: HDFS-3050.006.patch, HDFS-3050.007.patch, 
 HDFS-3050.008.patch, HDFS-3050.009.patch, HDFS-3050.010.patch, 
 HDFS-3050.011.patch, HDFS-3050.012.patch, HDFS-3050.014.patch, 
 HDFS-3050.015.patch, HDFS-3050.016.patch, HDFS-3050.017.patch, 
 HDFS-3050.018.patch, HDFS-3050.019.patch


 Current, OEV (the offline edits viewer) re-implements all of the opcode 
 parsing logic found in the NameNode.  This duplicated code creates a 
 maintenance burden for us.
 OEV should be refactored to simply use the normal EditLog parsing code, 
 rather than rolling its own.  By using the existing FSEditLogLoader code to 
 load edits in OEV, we can avoid having to update two places when the format 
 changes.
 We should not put opcode checksums into the XML, because they are a 
 serialization detail, not related to what the data is what we're storing.  
 This will also make it possible to modify the XML file and translate this 
 modified file back to a binary edits log file.
 Finally, this changes introduces --fix-txids.  When OEV is passed this flag, 
 it will close gaps in the transaction log by modifying the sequence numbers.  
 This is useful if you want to modify the edit log XML (say, by removing a 
 transaction), and transform the modified XML back into a valid binary edit 
 log file.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-2185) HA: HDFS portion of ZK-based FailoverController

2012-04-04 Thread Mingjie Lai (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2185?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13246676#comment-13246676
 ] 

Mingjie Lai commented on HDFS-2185:
---

(still posting comments here since the design doc is attached here)

@todd

Thanks for adding the manual failover section 2.7 in the design doc. 

However I have some questions for what you described in 2.7.2:
- HAAdmin makes an RPC failoverToYou() to the target ZKFC
- target ZKFC sends an RPC concedeLock() to the currently active ZKFC.
- the active sends a transitionToStandby() RPC to its local node

IMO the chain of RPCs is quite complicated, not easy to debug and troubleshoot 
in operation. Because you're trying to resolve the 2 problems, auto and manual 
failover, at one place -- ZKFC. 

How about seperate the 2 cases:
- add commands at haadmin to start/stop autofailover
- stop-autofailover requests all ZKFC to exitElection
- start-autofailover requests all ZKFC to enterElction
- haadmin is responsible for handle manual failover (as current implementation)
- admins can only perform manual failover when autofailover is stopped
- can be used to specify one particular active NN

Pros:
- existing manual fo code can be kept mostly
- although new RPC is added to ZKFC but we don't need them to talk to each 
other. the manual failover logic is all handled at client -- haadmin. 
- easier to extend to the case of multiple standby NNs

cons:
- administrator needs to explicitly start/stop autofailover, in addition to 
ZKFC process. 


 HA: HDFS portion of ZK-based FailoverController
 ---

 Key: HDFS-2185
 URL: https://issues.apache.org/jira/browse/HDFS-2185
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: auto-failover, ha
Affects Versions: 0.24.0, 0.23.3
Reporter: Eli Collins
Assignee: Todd Lipcon
 Fix For: Auto failover (HDFS-3042)

 Attachments: Failover_Controller.jpg, hdfs-2185.txt, hdfs-2185.txt, 
 hdfs-2185.txt, hdfs-2185.txt, hdfs-2185.txt, zkfc-design.pdf, 
 zkfc-design.pdf, zkfc-design.pdf, zkfc-design.pdf, zkfc-design.tex


 This jira is for a ZK-based FailoverController daemon. The FailoverController 
 is a separate daemon from the NN that does the following:
 * Initiates leader election (via ZK) when necessary
 * Performs health monitoring (aka failure detection)
 * Performs fail-over (standby to active and active to standby transitions)
 * Heartbeats to ensure the liveness
 It should have the same/similar interface as the Linux HA RM to aid 
 pluggability.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HDFS-3102) Add CLI tool to initialize the shared-edits dir

2012-04-04 Thread Aaron T. Myers (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3102?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Aaron T. Myers updated HDFS-3102:
-

Target Version/s: 0.24.0, 2.0.0  (was: 2.0.0, 0.24.0)
  Issue Type: New Feature  (was: Improvement)

 Add CLI tool to initialize the shared-edits dir
 ---

 Key: HDFS-3102
 URL: https://issues.apache.org/jira/browse/HDFS-3102
 Project: Hadoop HDFS
  Issue Type: New Feature
  Components: ha, name-node
Affects Versions: 0.24.0, 2.0.0
Reporter: Todd Lipcon
Assignee: Aaron T. Myers
 Attachments: HDFS-3102.patch, HDFS-3102.patch, HDFS-3102.patch


 Currently in order to make a non-HA NN HA, you need to initialize the shared 
 edits dir. This can be done manually by cping directories around. It would be 
 preferable to add a namenode -initializeSharedEdits command to achieve this 
 same effect.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3192) Active NN should exit when it has not received a getServiceStatus() rpc from ZKFC for timeout secs

2012-04-04 Thread Hari Mankude (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3192?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13246687#comment-13246687
 ] 

Hari Mankude commented on HDFS-3192:


bq. Why add multiple stonith paths, given we need external stonith anyway? It 
just adds to the complexity by increasing the number of scenarios we have to 
debug, etc.

I thought we are not going to have external stonith using special devices and 
that is mainly the reason why we are going through hoops to implement fencing 
in journal daemons. 

bq. That is to say: if the ZKFC dies, then it will lose its lock, and the other 
node will stonith this one when it takes over. What's the benefit of having it 
abort itself at the same time? In fact, it seems to be detrimental, because if 
it stays up, the other node can do a graceful transitionToStandby() call rather 
than having to do something more drastic like a full abort.

I disagree about two items here.

1. Why is the behaviour different from what happens when zkfc loses the 
ephemeral node? Currently zkfc when it loses the ephemeral node will shutdown 
the active NN. Similarly if active NN does not hear from zkfc, it implies that 
zkfc is dead, going through gc pause essentially resulting in loss of ephemeral 
node. 

2.  If active NN loses quorum, it has to shutdown. There is no way to do a 
transitionToStandby() especially since the log is updated after NN metadata is 
updated and there is no way to roll back the last update. This is just one of 
the issues that we are aware of where a rollback would be necessary. There 
might be other situations where rollback is required. In fact, one of the most 
of the difficult APIs to implement correctly would be transitionToStandby() 
from active state. 

 Active NN should exit when it has not received a getServiceStatus() rpc from 
 ZKFC for timeout secs
 --

 Key: HDFS-3192
 URL: https://issues.apache.org/jira/browse/HDFS-3192
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: ha, name-node
Reporter: Hari Mankude
Assignee: Hari Mankude



--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HDFS-3102) Add CLI tool to initialize the shared-edits dir

2012-04-04 Thread Aaron T. Myers (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3102?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Aaron T. Myers updated HDFS-3102:
-

  Resolution: Fixed
   Fix Version/s: 2.0.0
Target Version/s: 0.24.0, 2.0.0  (was: 2.0.0, 0.24.0)
Hadoop Flags: Reviewed
  Status: Resolved  (was: Patch Available)

Thanks a lot for the reviews, Eli. I've just committed this to trunk and 
branch-2.

 Add CLI tool to initialize the shared-edits dir
 ---

 Key: HDFS-3102
 URL: https://issues.apache.org/jira/browse/HDFS-3102
 Project: Hadoop HDFS
  Issue Type: New Feature
  Components: ha, name-node
Affects Versions: 0.24.0, 2.0.0
Reporter: Todd Lipcon
Assignee: Aaron T. Myers
 Fix For: 2.0.0

 Attachments: HDFS-3102.patch, HDFS-3102.patch, HDFS-3102.patch


 Currently in order to make a non-HA NN HA, you need to initialize the shared 
 edits dir. This can be done manually by cping directories around. It would be 
 preferable to add a namenode -initializeSharedEdits command to achieve this 
 same effect.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3102) Add CLI tool to initialize the shared-edits dir

2012-04-04 Thread Hudson (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3102?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13246696#comment-13246696
 ] 

Hudson commented on HDFS-3102:
--

Integrated in Hadoop-Hdfs-trunk-Commit #2072 (See 
[https://builds.apache.org/job/Hadoop-Hdfs-trunk-Commit/2072/])
HDFS-3102. Add CLI tool to initialize the shared-edits dir. Contributed by 
Aaron T. Myers. (Revision 1309580)

 Result = SUCCESS
atm : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1309580
Files : 
* /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/common/HdfsServerConstants.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSNamesystem.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/NameNode.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/ha/TestBootstrapStandby.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/ha/TestInitializeSharedEdits.java


 Add CLI tool to initialize the shared-edits dir
 ---

 Key: HDFS-3102
 URL: https://issues.apache.org/jira/browse/HDFS-3102
 Project: Hadoop HDFS
  Issue Type: New Feature
  Components: ha, name-node
Affects Versions: 0.24.0, 2.0.0
Reporter: Todd Lipcon
Assignee: Aaron T. Myers
 Fix For: 2.0.0

 Attachments: HDFS-3102.patch, HDFS-3102.patch, HDFS-3102.patch


 Currently in order to make a non-HA NN HA, you need to initialize the shared 
 edits dir. This can be done manually by cping directories around. It would be 
 preferable to add a namenode -initializeSharedEdits command to achieve this 
 same effect.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3102) Add CLI tool to initialize the shared-edits dir

2012-04-04 Thread Hudson (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3102?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13246700#comment-13246700
 ] 

Hudson commented on HDFS-3102:
--

Integrated in Hadoop-Common-trunk-Commit #1997 (See 
[https://builds.apache.org/job/Hadoop-Common-trunk-Commit/1997/])
HDFS-3102. Add CLI tool to initialize the shared-edits dir. Contributed by 
Aaron T. Myers. (Revision 1309580)

 Result = SUCCESS
atm : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1309580
Files : 
* /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/common/HdfsServerConstants.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSNamesystem.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/NameNode.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/ha/TestBootstrapStandby.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/ha/TestInitializeSharedEdits.java


 Add CLI tool to initialize the shared-edits dir
 ---

 Key: HDFS-3102
 URL: https://issues.apache.org/jira/browse/HDFS-3102
 Project: Hadoop HDFS
  Issue Type: New Feature
  Components: ha, name-node
Affects Versions: 0.24.0, 2.0.0
Reporter: Todd Lipcon
Assignee: Aaron T. Myers
 Fix For: 2.0.0

 Attachments: HDFS-3102.patch, HDFS-3102.patch, HDFS-3102.patch


 Currently in order to make a non-HA NN HA, you need to initialize the shared 
 edits dir. This can be done manually by cping directories around. It would be 
 preferable to add a namenode -initializeSharedEdits command to achieve this 
 same effect.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3102) Add CLI tool to initialize the shared-edits dir

2012-04-04 Thread Hudson (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3102?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13246709#comment-13246709
 ] 

Hudson commented on HDFS-3102:
--

Integrated in Hadoop-Mapreduce-trunk-Commit #2009 (See 
[https://builds.apache.org/job/Hadoop-Mapreduce-trunk-Commit/2009/])
HDFS-3102. Add CLI tool to initialize the shared-edits dir. Contributed by 
Aaron T. Myers. (Revision 1309580)

 Result = SUCCESS
atm : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1309580
Files : 
* /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/common/HdfsServerConstants.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSNamesystem.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/NameNode.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/ha/TestBootstrapStandby.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/ha/TestInitializeSharedEdits.java


 Add CLI tool to initialize the shared-edits dir
 ---

 Key: HDFS-3102
 URL: https://issues.apache.org/jira/browse/HDFS-3102
 Project: Hadoop HDFS
  Issue Type: New Feature
  Components: ha, name-node
Affects Versions: 0.24.0, 2.0.0
Reporter: Todd Lipcon
Assignee: Aaron T. Myers
 Fix For: 2.0.0

 Attachments: HDFS-3102.patch, HDFS-3102.patch, HDFS-3102.patch


 Currently in order to make a non-HA NN HA, you need to initialize the shared 
 edits dir. This can be done manually by cping directories around. It would be 
 preferable to add a namenode -initializeSharedEdits command to achieve this 
 same effect.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3192) Active NN should exit when it has not received a getServiceStatus() rpc from ZKFC for timeout secs

2012-04-04 Thread Todd Lipcon (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3192?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13246718#comment-13246718
 ] 

Todd Lipcon commented on HDFS-3192:
---

bq. I thought we are not going to have external stonith using special devices 
and that is mainly the reason why we are going through hoops to implement 
fencing in journal daemons.

In the current design, which uses a filer, we *require* external stonith 
devices. There is no correct way of doing it without either stonith or storage 
fencing.

The proposal with the journal-daemon based fencing is essentailly the same as 
storage fencing - just that we do it with our own software storage instead of a 
NAS/SAN.

bq. Why is the behaviour different from what happens when zkfc loses the 
ephemeral node? Currently zkfc when it loses the ephemeral node will shutdown 
the active NN

No, it doesn't - it will transition it to standby. But, as I commented 
elsewhere, this is redundant, because the _new_ active is actually going to 
fence it anyway before taking over.

bq. Similarly if active NN does not hear from zkfc, it implies that zkfc is 
dead, going through gc pause essentially resulting in loss of ephemeral node.

But this can reduce uptime. For example, imagine an administrator accidentally 
changes the ACL on zookeeper. This causes both ZKFCs to get an authentication 
error and crash at the same time. With your design, both NNs will then commit 
suicide. With the existing implementation, the system will continue to run in 
its existing state -- i.e no new failovers will occur, but whoever is active 
will remain active.

bq.  If active NN loses quorum, it has to shutdown

Yes, it has to shut down _before_ it does any edits, or it has to be fenced by 
the next active. Notification of session loss is asynchronous. The same is true 
of your proposal. In either case it can take arbitrarily long before it 
notices that it should not be active. So we still require that the new active 
fence it before it becomes active. So, this proposal doesn't solve any problems.

bq. In fact, one of the most of the difficult APIs to implement correctly would 
be transitionToStandby() from active state.

We already have that implemented. It syncs any existing edits, and then stops 
allowing new ones. We allow failover from one node to another without aborting, 
so long as it's graceful. This is perfectly correct. If we need to do a 
non-graceful failover, we fence the node by STONITH or by disallowing further 
access to the edit logs (which indirectly causes the node to abort, since 
logSync() fails).

It seems you're trying to solve problems we've already solved.

 Active NN should exit when it has not received a getServiceStatus() rpc from 
 ZKFC for timeout secs
 --

 Key: HDFS-3192
 URL: https://issues.apache.org/jira/browse/HDFS-3192
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: ha, name-node
Reporter: Hari Mankude
Assignee: Hari Mankude



--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3168) Clean up FSNamesystem and BlockManager

2012-04-04 Thread Aaron T. Myers (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3168?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13246720#comment-13246720
 ] 

Aaron T. Myers commented on HDFS-3168:
--

bq. Aaron, it is nothing to do with it. Any contributor could review code. It 
was a merging problem.

I definitely didn't mean to imply that this particular compilation issue would 
have been caught by a different reviewer. I'm confident I would not have caught 
this particular issue myself. :)

bq. I think this is not true. After a committer provided a patch and a 
non-committer reviewed it, the same committer could commit the patch.

I think the bylaws are a little ambiguous on the subject. Per the bylaws:

{quote}
*Code Change*
A change made to a codebase of the project and committed by a committer. This 
includes source code, documentation, website content, etc. Lazy consensus of 
active committers, but with a minimum of one +1. The code can be committed 
after the first +1, unless the code change represents a merge from a branch, in 
which case three +1s are required.
{quote}

This would seem to imply that a review by a non-committer contributor is 
non-binding. It does not, however, clear up the issue of whether or not a 
committer can provide a +1 of their own patch.

FWIW, my understanding is the same as Todd's on this subject.

 Clean up FSNamesystem and BlockManager
 --

 Key: HDFS-3168
 URL: https://issues.apache.org/jira/browse/HDFS-3168
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: name-node
Reporter: Tsz Wo (Nicholas), SZE
Assignee: Tsz Wo (Nicholas), SZE
 Fix For: 0.24.0, 0.23.3

 Attachments: h3168_20120330.patch, h3168_20120402.patch, 
 h3168_20120403.patch




--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3121) hdfs tests for HADOOP-8014

2012-04-04 Thread Suresh Srinivas (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3121?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13246724#comment-13246724
 ] 

Suresh Srinivas commented on HDFS-3121:
---

Comments:
# Not very clear THis test is test the fix.. Instead of describing running 
into serialization problem etc. can you just describe what the test is. Also 
please move the description as class level javadoc.
# minor: defaultBLockSize - defaultBlockSize 
# Please ensure line lengths are with in 80 chars limit
# testGetDefaultBlockSize() indententation at try is not correct. I am also 
not clear about the comment createFile... What has createFile got to do with 
the test?
# Please add a brief description of what the test is testing.
# Please consider: When expecting an exception in tests, you can move fail() 
within try clause, after the method that you expect exception from. This avoids 
also return from catch blocks. 


 hdfs tests for HADOOP-8014
 --

 Key: HDFS-3121
 URL: https://issues.apache.org/jira/browse/HDFS-3121
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 0.23.2, 0.23.3
Reporter: John George
Assignee: John George
 Attachments: hdfs-3121.patch, hdfs-3121.patch, hdfs-3121.patch


 This JIRA is to write tests for viewing quota using viewfs.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3178) Add states for journal synchronization in journal daemon

2012-04-04 Thread Hadoop QA (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3178?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13246725#comment-13246725
 ] 

Hadoop QA commented on HDFS-3178:
-

+1 overall.  Here are the results of testing the latest attachment 
  http://issues.apache.org/jira/secure/attachment/12521374/h3178_20120404.patch
  against trunk revision .

+1 @author.  The patch does not contain any @author tags.

+1 tests included.  The patch appears to include 6 new or modified tests.

+1 javadoc.  The javadoc tool did not generate any warning messages.

+1 javac.  The applied patch does not increase the total number of javac 
compiler warnings.

+1 eclipse:eclipse.  The patch built with eclipse:eclipse.

+1 findbugs.  The patch does not introduce any new Findbugs (version 1.3.9) 
warnings.

+1 release audit.  The applied patch does not increase the total number of 
release audit warnings.

+1 core tests.  The patch passed unit tests in .

+1 contrib tests.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-HDFS-Build/2185//testReport/
Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/2185//console

This message is automatically generated.

 Add states for journal synchronization in journal daemon
 

 Key: HDFS-3178
 URL: https://issues.apache.org/jira/browse/HDFS-3178
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: ha, name-node
Reporter: Tsz Wo (Nicholas), SZE
Assignee: Tsz Wo (Nicholas), SZE
 Attachments: h3178_20120403_svn_mv.patch, h3178_20120404.patch, 
 h3178_20120404_svn_mv.patch, svn_mv.sh


 Journal in a new daemon has to be synchronized to the current transaction.  
 It requires new states such as WaitingForRoll, Syncing and Synced.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-2185) HA: HDFS portion of ZK-based FailoverController

2012-04-04 Thread Todd Lipcon (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2185?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13246729#comment-13246729
 ] 

Todd Lipcon commented on HDFS-2185:
---

Hi Mingjie. Thanks for taking a look.

The idea for the chain of RPCs is from talking with some folks here who work on 
Hadoop deployment. Their opinion was the following: currently, most of the 
Hadoop client tools are too thick. For example, in the current manual 
failover implementation, the fencing is run on the admin client. This means 
that you have to run the haadmin command from a machine that has access to all 
of the necessary fencing scripts, key files, etc. That's a little bizarre -- 
you would expect to configure these kinds of things only on the central 
location, not on the client.

So, we decided that it makes sense to push the management of the whole failover 
process into the FCs themselves, and just use a single RPC to kick off the 
whole failover process. This keeps the client thin.

As for your proposed alternative, here are a few thoughts:

bq. existing manual fo code can be kept mostly
We actually share much of the code already. But, the problem with using the 
existing code exactly as is, is that the failover controllers always expect to 
have complete control over the system. If the state of the NNs changes 
underneath the ZKFC, then the state in ZK will become inconsistent with the 
actual state of the system, and it's very easy to get into split brain 
scenarios. So, the idea is that, when auto-failover is enabled, *all* decisions 
must be made by ZKFCs. That way we can make sure the ZK state doesn't get out 
of sync.

bq. although new RPC is added to ZKFC but we don't need them to talk to each 
other. the manual failover logic is all handled at client – haadmin.
As noted above I think this is a con, not a pro, because it requires 
configuring fencing scripts at the client, and likely requiring that the client 
have read-write access to ZK

bq. easier to extend to the case of multiple standby NNs

I think the extension path to multiple standby is actually equally easy with 
both approaches. The solution in the ZKFC-managed implementation is to add a 
new znode like PreferredActive and have nodes avoid becoming active unless 
they're listed as preferred. The target node of the failover can just set 
itself to be preferred before asking the other node to cede the lock.


Some other advantages that I probably didn't explain well in the design doc:
- this design is fault tolerant. If the target node crashes in the middle of 
the process, then the old active will automatically regain the active state 
after its rejoin timeout elapses. With a client-managed setup, a well-meaning 
admin may ^C the process in the middle and leave the system with no active at 
all.
- no need to introduce disable/enable to auto-failover. Just having both 
nodes quit the election wouldn't work, since one would end up quitting before 
the other, causing a blip where an unnecessary (random) failover occurred. We 
could carefully orchestrate the order of quitting, so the active quits last, 
but I think it still gets complicated.

 HA: HDFS portion of ZK-based FailoverController
 ---

 Key: HDFS-2185
 URL: https://issues.apache.org/jira/browse/HDFS-2185
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: auto-failover, ha
Affects Versions: 0.24.0, 0.23.3
Reporter: Eli Collins
Assignee: Todd Lipcon
 Fix For: Auto failover (HDFS-3042)

 Attachments: Failover_Controller.jpg, hdfs-2185.txt, hdfs-2185.txt, 
 hdfs-2185.txt, hdfs-2185.txt, hdfs-2185.txt, zkfc-design.pdf, 
 zkfc-design.pdf, zkfc-design.pdf, zkfc-design.pdf, zkfc-design.tex


 This jira is for a ZK-based FailoverController daemon. The FailoverController 
 is a separate daemon from the NN that does the following:
 * Initiates leader election (via ZK) when necessary
 * Performs health monitoring (aka failure detection)
 * Performs fail-over (standby to active and active to standby transitions)
 * Heartbeats to ensure the liveness
 It should have the same/similar interface as the Linux HA RM to aid 
 pluggability.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3178) Add states for journal synchronization in journal daemon

2012-04-04 Thread Todd Lipcon (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3178?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13246731#comment-13246731
 ] 

Todd Lipcon commented on HDFS-3178:
---

Hi Nicholas. Could you please add some javadoc to the state enum values 
explaining the purpose of each state, and what the transitions are between 
them? Or augment the design doc for HDFS-3092 with this state machine, and 
reference it from the code?

 Add states for journal synchronization in journal daemon
 

 Key: HDFS-3178
 URL: https://issues.apache.org/jira/browse/HDFS-3178
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: ha, name-node
Reporter: Tsz Wo (Nicholas), SZE
Assignee: Tsz Wo (Nicholas), SZE
 Attachments: h3178_20120403_svn_mv.patch, h3178_20120404.patch, 
 h3178_20120404_svn_mv.patch, svn_mv.sh


 Journal in a new daemon has to be synchronized to the current transaction.  
 It requires new states such as WaitingForRoll, Syncing and Synced.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3168) Clean up FSNamesystem and BlockManager

2012-04-04 Thread Todd Lipcon (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3168?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13246734#comment-13246734
 ] 

Todd Lipcon commented on HDFS-3168:
---

Let's ask the dev list. I'll start a thread.

 Clean up FSNamesystem and BlockManager
 --

 Key: HDFS-3168
 URL: https://issues.apache.org/jira/browse/HDFS-3168
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: name-node
Reporter: Tsz Wo (Nicholas), SZE
Assignee: Tsz Wo (Nicholas), SZE
 Fix For: 0.24.0, 0.23.3

 Attachments: h3168_20120330.patch, h3168_20120402.patch, 
 h3168_20120403.patch




--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3192) Active NN should exit when it has not received a getServiceStatus() rpc from ZKFC for timeout secs

2012-04-04 Thread Hari Mankude (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3192?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13246752#comment-13246752
 ] 

Hari Mankude commented on HDFS-3192:


bq.   bq. Why is the behaviour different from what happens when zkfc loses the 
ephemeral node? Currently zkfc when it loses the ephemeral node will shutdown 
the active NN

bq. No, it doesn't - it will transition it to standby. But, as I commented 
elsewhere, this is redundant, because the new active is actually going to fence 
it anyway before taking over.

Well this is incorrect behaviour. It does not handle the situation that I 
mentioned earlier about requiring rollbacks. The transition to standby will 
have result in old active having incorrect in-memory state. The only way around 
this is to shutdown the active. The reason is that as soon zkfc on NN1 has lost 
the ephemeral znode, it is possible that zkfc on NN2 has taken over the znode 
and NN2 has started fencing the journals. There is no way to gracefully 
coordinate this with NN1. This would result in NN1 getting quorum loss which in 
turn could leave the in-memory state in NN1 in an inconsistent shape. Do you 
agree again that in-memory state of NN1 is inconsistent with the editlogs?

bq. Similarly if active NN does not hear from zkfc, it implies that zkfc is 
dead, going through gc pause essentially resulting in loss of ephemeral node.

bq. But this can reduce uptime. For example, imagine an administrator 
accidentally changes the ACL on zookeeper. This causes both ZKFCs to get an 
authentication error and crash at the same time. With your design, both NNs 
will then commit suicide. With the existing implementation, the system will 
continue to run in its existing state – i.e no new failovers will occur, but 
whoever is active will remain active.

Firstly, how often does some change ACLs in zookeeper? Secondly, why is ZKFC 
dying when this happens? ZKFC must be more robust than NN. NN is a resource 
that is controlled by ZKFC. We should make zkfc more robust to handle zookeeper 
acl changes if this is a common occurance.

bq. If active NN loses quorum, it has to shutdown

bq. Yes, it has to shut down before it does any edits, or it has to be fenced 
by the next active. Notification of session loss is asynchronous. The same is 
true of your proposal. In either case it can take arbitrarily long before it 
notices that it should not be active. So we still require that the new active 
fence it before it becomes active. So, this proposal doesn't solve any problems.

My proposal was not meant to handle active NN losing quorum. My proposal is 
shutdown NN when ZKFC has died or is in a gc pause. 

My comment was with regards to earlier comment regarding doing a 
transitionToStandby(). Do you agree that active NN has invalid in-memory state 
and cannot go through transitionToStandby() when it loses quorum? There seems 
to be two solutions. 
1. Implement rollback for various types of editlog entries and then do 
transitionToStandby() OR
2. Shutdown NN when it loses quorum
Does this sound right?









 Active NN should exit when it has not received a getServiceStatus() rpc from 
 ZKFC for timeout secs
 --

 Key: HDFS-3192
 URL: https://issues.apache.org/jira/browse/HDFS-3192
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: ha, name-node
Reporter: Hari Mankude
Assignee: Hari Mankude



--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Created] (HDFS-3197) Duplicate (and incorrect) class comments in a few tests

2012-04-04 Thread Aaron T. Myers (Created) (JIRA)
Duplicate (and incorrect) class comments in a few tests
---

 Key: HDFS-3197
 URL: https://issues.apache.org/jira/browse/HDFS-3197
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: test
Affects Versions: 0.24.0
Reporter: Aaron T. Myers
Priority: Trivial


Somewhat hilariously, TestFileCreationClient, TestDatanodeDeath, 
TestReplaceDatanodeOnFailure, and TestDatanodeRegistration all have the 
following class comment:

{noformat}
/**
 * This class tests that a file need not be closed before its
 * data can be read by another client.
 */
{noformat}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3192) Active NN should exit when it has not received a getServiceStatus() rpc from ZKFC for timeout secs

2012-04-04 Thread Todd Lipcon (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3192?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13246778#comment-13246778
 ] 

Todd Lipcon commented on HDFS-3192:
---

I think there is confusion here over the terminology loses quorum.

I agree completely about the following: if any NN fails to sync its edit logs, 
it needs to abort. This is already what it does today - no changes necessary.
If the edit log happens to be implemented using a quorum protocol (HDFS-3092 or 
HDFS-3077 for example), then that behavior should be maintained. The 
JournalManager implementation needs to throw an exception in response to 
logSync(). That will cause the NN to abort.

That's all that's necessary for correctness - an NN won't ack success to a 
write unless it successfully syncs it, and will abort rather than rollback, 
since we have no rollback capability.

In the above sense, loses quorum really means loses write access to the edit 
logs.


If instead you're talking about loses quorum as loses its ZK session, then 
no abort is necessary, because it may still be able to write to its edits. So 
long as it's getting success back from editLog.logSync(), then the edits are 
being persisted. It is the responsibility of the next active to fence access to 
the shared edits. It may do so in one of two ways:
1) Edits fencing: ensure that the next write to the edits mechanism throws IOE. 
In the case of FileJournalManager on NAS, this is done via an RPC to the NAS 
system to fence the given export.
2) STONITH: ensure that the next write fails because power has been yanked from 
the machine.

Alternatively, the new active may first try a graceful transition:
3) Gracefully ask the prior active to stop writing. The prior active flushes 
anything buffered, successfully syncs, and then enters standby mode.


Notably, self-stonith upon losing the ZK lease is not an option, because it 
may take arbitrarily long before it notices. EG:
1) NN1 writing to edits log
2) ZKFC1 loses lease, but doesn't know about it yet
3) ZKFC2 gets lease
4) NN2 becomes active, starts writing logs
5) NN1 writes some edits. World explodes.
6) ZKFC1 gets asynchronous notification from ZK that it lots its session. 
Anything you do at this point is _too late_.

Before step 4, NN2 must use a fencing mechanism. *Regardless* of whatever steps 
NN1 or ZKFC1 might take in step 6.


 Active NN should exit when it has not received a getServiceStatus() rpc from 
 ZKFC for timeout secs
 --

 Key: HDFS-3192
 URL: https://issues.apache.org/jira/browse/HDFS-3192
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: ha, name-node
Reporter: Hari Mankude
Assignee: Hari Mankude



--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3050) rework OEV to share more code with the NameNode

2012-04-04 Thread Colin Patrick McCabe (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3050?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13246784#comment-13246784
 ] 

Colin Patrick McCabe commented on HDFS-3050:


The TestValidateConfigurationSettings is another random jenkins failure (port 
in use). 

 rework OEV to share more code with the NameNode
 ---

 Key: HDFS-3050
 URL: https://issues.apache.org/jira/browse/HDFS-3050
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: name-node
Reporter: Colin Patrick McCabe
Assignee: Colin Patrick McCabe
Priority: Minor
 Attachments: HDFS-3050.006.patch, HDFS-3050.007.patch, 
 HDFS-3050.008.patch, HDFS-3050.009.patch, HDFS-3050.010.patch, 
 HDFS-3050.011.patch, HDFS-3050.012.patch, HDFS-3050.014.patch, 
 HDFS-3050.015.patch, HDFS-3050.016.patch, HDFS-3050.017.patch, 
 HDFS-3050.018.patch, HDFS-3050.019.patch


 Current, OEV (the offline edits viewer) re-implements all of the opcode 
 parsing logic found in the NameNode.  This duplicated code creates a 
 maintenance burden for us.
 OEV should be refactored to simply use the normal EditLog parsing code, 
 rather than rolling its own.  By using the existing FSEditLogLoader code to 
 load edits in OEV, we can avoid having to update two places when the format 
 changes.
 We should not put opcode checksums into the XML, because they are a 
 serialization detail, not related to what the data is what we're storing.  
 This will also make it possible to modify the XML file and translate this 
 modified file back to a binary edits log file.
 Finally, this changes introduces --fix-txids.  When OEV is passed this flag, 
 it will close gaps in the transaction log by modifying the sequence numbers.  
 This is useful if you want to modify the edit log XML (say, by removing a 
 transaction), and transform the modified XML back into a valid binary edit 
 log file.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HDFS-3050) rework OEV to share more code with the NameNode

2012-04-04 Thread Colin Patrick McCabe (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3050?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Colin Patrick McCabe updated HDFS-3050:
---

Attachment: HDFS-3050.020.patch

* update OK_JAVADOC_WARNINGS

 rework OEV to share more code with the NameNode
 ---

 Key: HDFS-3050
 URL: https://issues.apache.org/jira/browse/HDFS-3050
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: name-node
Reporter: Colin Patrick McCabe
Assignee: Colin Patrick McCabe
Priority: Minor
 Attachments: HDFS-3050.006.patch, HDFS-3050.007.patch, 
 HDFS-3050.008.patch, HDFS-3050.009.patch, HDFS-3050.010.patch, 
 HDFS-3050.011.patch, HDFS-3050.012.patch, HDFS-3050.014.patch, 
 HDFS-3050.015.patch, HDFS-3050.016.patch, HDFS-3050.017.patch, 
 HDFS-3050.018.patch, HDFS-3050.019.patch, HDFS-3050.020.patch


 Current, OEV (the offline edits viewer) re-implements all of the opcode 
 parsing logic found in the NameNode.  This duplicated code creates a 
 maintenance burden for us.
 OEV should be refactored to simply use the normal EditLog parsing code, 
 rather than rolling its own.  By using the existing FSEditLogLoader code to 
 load edits in OEV, we can avoid having to update two places when the format 
 changes.
 We should not put opcode checksums into the XML, because they are a 
 serialization detail, not related to what the data is what we're storing.  
 This will also make it possible to modify the XML file and translate this 
 modified file back to a binary edits log file.
 Finally, this changes introduces --fix-txids.  When OEV is passed this flag, 
 it will close gaps in the transaction log by modifying the sequence numbers.  
 This is useful if you want to modify the edit log XML (say, by removing a 
 transaction), and transform the modified XML back into a valid binary edit 
 log file.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-2696) fuse-dfs build problems

2012-04-04 Thread Eli Collins (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2696?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13246794#comment-13246794
 ] 

Eli Collins commented on HDFS-2696:
---

Hey Bruno,
Can you generate one patch that applies against trunk so I can review?

Thanks! 

 fuse-dfs build problems
 ---

 Key: HDFS-2696
 URL: https://issues.apache.org/jira/browse/HDFS-2696
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: contrib/fuse-dfs
Affects Versions: 0.23.1
 Environment: linux
Reporter: Petru Dimulescu
  Labels: bigtop, build, maven, patch
 Fix For: 0.23.3

 Attachments: HDFS-2696-plus.patch, dfsfuse-build.patch.zip


 fuse-dfs has some compilation problems on the 0.23 and 0.24(head) branches. 
 Some details on how to compile fuse-dfs on 0.23 is here : 
 http://mail-archives.apache.org/mod_mbox/hadoop-hdfs-user/201112.mbox/%3c4ee23195.7030...@gmail.com%3E
  and a wiki page follows. 
 The attached patch removes some problems (an inexistnant 
 ivy/library.properties, detecting jni.h location).

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




  1   2   >