[jira] [Updated] (HDFS-3049) During the normal loading NN startup process, fall back on a different EditLog if we see one that is corrupt

2012-04-13 Thread Colin Patrick McCabe (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3049?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Colin Patrick McCabe updated HDFS-3049:
---

 Target Version/s: 2.0.0
Affects Version/s: 0.23.0
Fix Version/s: (was: 0.24.0)

> During the normal loading NN startup process, fall back on a different 
> EditLog if we see one that is corrupt
> 
>
> Key: HDFS-3049
> URL: https://issues.apache.org/jira/browse/HDFS-3049
> Project: Hadoop HDFS
>  Issue Type: New Feature
>  Components: name-node
>Affects Versions: 0.23.0
>Reporter: Colin Patrick McCabe
>Assignee: Colin Patrick McCabe
>Priority: Minor
> Attachments: HDFS-3049.001.patch
>
>
> During the NameNode startup process, we load an image, and then apply edit 
> logs to it until we believe that we have all the latest changes.  
> Unfortunately, if there is an I/O error while reading any of these files, in 
> most cases, we simply abort the startup process.  We should try harder to 
> locate a readable edit log and/or image file.
> *There are three main use cases for this feature:*
> 1. If the operating system does not honor fsync (usually due to a 
> misconfiguration), a file may end up in an inconsistent state.
> 2. In certain older releases where we did not use fallocate() or similar to 
> pre-reserve blocks, a disk full condition may cause a truncated log in one 
> edit directory.
> 3. There may be a bug in HDFS which results in some of the data directories 
> receiving corrupt data, but not all.  This is the least likely use case.
> *Proposed changes to normal NN startup*
> * We should try a different FSImage if we can't load the first one we try.
> * We should examine other FSEditLogs if we can't load the first one(s) we try.
> * We should fail if we can't find EditLogs that would bring us up to what we 
> believe is the latest transaction ID.
> Proposed changes to recovery mode NN startup:
> we should list out all the available storage directories and allow the 
> operator to select which one he wants to use.
> Something like this:
> {code}
> Multiple storage directories found.
> 1. /foo/bar
> edits__curent__XYZ  size:213421345   md5:2345345
> image  size:213421345   md5:2345345
> 2. /foo/baz
> edits__curent__XYZ  size:213421345   md5:2345345345
> image  size:213421345   md5:2345345
> Which one would you like to use? (1/2)
> {code}
> As usual in recovery mode, we want to be flexible about error handling.  In 
> this case, this means that we should NOT fail if we can't find EditLogs that 
> would bring us up to what we believe is the latest transaction ID.
> *Not addressed by this feature*
> This feature will not address the case where an attempt to access the 
> NameNode name directory or directories hangs because of an I/O error.  This 
> may happen, for example, when trying to load an image from a hard-mounted NFS 
> directory, when the NFS server has gone away.  Just as now, the operator will 
> have to notice this problem and take steps to correct it.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HDFS-3134) harden edit log loader against malformed or malicious input

2012-04-13 Thread Colin Patrick McCabe (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3134?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Colin Patrick McCabe updated HDFS-3134:
---

Target Version/s: 2.0.0
   Fix Version/s: (was: 2.0.0)

> harden edit log loader against malformed or malicious input
> ---
>
> Key: HDFS-3134
> URL: https://issues.apache.org/jira/browse/HDFS-3134
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: name-node
>Affects Versions: 0.23.0
>Reporter: Colin Patrick McCabe
>Assignee: Colin Patrick McCabe
> Attachments: HDFS-3134.001.patch, HDFS-3134.002.patch, 
> HDFS-3134.003.patch, HDFS-3134.004.patch, HDFS-3134.005.patch
>
>
> Currently, the edit log loader does not handle bad or malicious input 
> sensibly.
> We can often cause OutOfMemory exceptions, null pointer exceptions, or other 
> unchecked exceptions to be thrown by feeding the edit log loader bad input.  
> In some environments, an out of memory error can cause the JVM process to be 
> terminated.
> It's clear that we want these exceptions to be thrown as IOException instead 
> of as unchecked exceptions.  We also want to avoid out of memory situations.
> The main task here is to put a sensible upper limit on the lengths of arrays 
> and strings we allocate on command.  The other task is to try to avoid 
> creating unchecked exceptions (by dereferencing potentially-NULL pointers, 
> for example).  Instead, we should verify ahead of time and give a more 
> sensible error message that reflects the problem with the input.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HDFS-3280) DFSOutputStream.sync should not be synchronized

2012-04-13 Thread Todd Lipcon (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3280?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Todd Lipcon updated HDFS-3280:
--

   Resolution: Fixed
Fix Version/s: 2.0.0
 Hadoop Flags: Reviewed
   Status: Resolved  (was: Patch Available)

Committed to branch-2 (as noted above, this problem isn't present in trunk 
since sync() was removed).

Thanks also to Matteo Bertozzi and Jon Hsieh for their help tracking this issue 
down.

> DFSOutputStream.sync should not be synchronized
> ---
>
> Key: HDFS-3280
> URL: https://issues.apache.org/jira/browse/HDFS-3280
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: hdfs client
>Affects Versions: 2.0.0
>Reporter: Todd Lipcon
>Assignee: Todd Lipcon
>Priority: Critical
> Fix For: 2.0.0
>
> Attachments: hdfs-3280.txt
>
>
> HDFS-895 added an optimization to make hflush() much faster by 
> unsynchronizing it. But, we forgot to un-synchronize the deprecated 
> {{sync()}} wrapper method. This makes the HBase WAL really slow on 0.23+ 
> since it doesn't take advantage of HDFS-895 anymore.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3280) DFSOutputStream.sync should not be synchronized

2012-04-13 Thread Todd Lipcon (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3280?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13253951#comment-13253951
 ] 

Todd Lipcon commented on HDFS-3280:
---

bq. Ah, so this explains what you guys thought might be an interaction with 
Nagle?

Yep, turned out to be much simpler :)

The patch failed on Hudson due to HDFS-3034 having removed the deprecated 
method. I'll commit this based on Aaron's +1 and based on my manual stress 
testing using HBase's HLog class, which uses this method.

No unit tests since it's hard to unit test for performance, and the hflush 
equivalent is already tested by TestMultithreadedHflush

> DFSOutputStream.sync should not be synchronized
> ---
>
> Key: HDFS-3280
> URL: https://issues.apache.org/jira/browse/HDFS-3280
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: hdfs client
>Affects Versions: 2.0.0
>Reporter: Todd Lipcon
>Assignee: Todd Lipcon
>Priority: Critical
> Attachments: hdfs-3280.txt
>
>
> HDFS-895 added an optimization to make hflush() much faster by 
> unsynchronizing it. But, we forgot to un-synchronize the deprecated 
> {{sync()}} wrapper method. This makes the HBase WAL really slow on 0.23+ 
> since it doesn't take advantage of HDFS-895 anymore.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3280) DFSOutputStream.sync should not be synchronized

2012-04-13 Thread Hadoop QA (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3280?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13253948#comment-13253948
 ] 

Hadoop QA commented on HDFS-3280:
-

-1 overall.  Here are the results of testing the latest attachment 
  http://issues.apache.org/jira/secure/attachment/12522653/hdfs-3280.txt
  against trunk revision .

+1 @author.  The patch does not contain any @author tags.

-1 tests included.  The patch doesn't appear to include any new or modified 
tests.
Please justify why no new tests are needed for this 
patch.
Also please list what manual steps were performed to 
verify this patch.

-1 patch.  The patch command could not apply the patch.

Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/2277//console

This message is automatically generated.

> DFSOutputStream.sync should not be synchronized
> ---
>
> Key: HDFS-3280
> URL: https://issues.apache.org/jira/browse/HDFS-3280
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: hdfs client
>Affects Versions: 2.0.0
>Reporter: Todd Lipcon
>Assignee: Todd Lipcon
>Priority: Critical
> Attachments: hdfs-3280.txt
>
>
> HDFS-895 added an optimization to make hflush() much faster by 
> unsynchronizing it. But, we forgot to un-synchronize the deprecated 
> {{sync()}} wrapper method. This makes the HBase WAL really slow on 0.23+ 
> since it doesn't take advantage of HDFS-895 anymore.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3280) DFSOutputStream.sync should not be synchronized

2012-04-13 Thread Aaron T. Myers (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3280?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13253946#comment-13253946
 ] 

Aaron T. Myers commented on HDFS-3280:
--

+1 pending Jenkins. 

> DFSOutputStream.sync should not be synchronized
> ---
>
> Key: HDFS-3280
> URL: https://issues.apache.org/jira/browse/HDFS-3280
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: hdfs client
>Affects Versions: 2.0.0
>Reporter: Todd Lipcon
>Assignee: Todd Lipcon
>Priority: Critical
> Attachments: hdfs-3280.txt
>
>
> HDFS-895 added an optimization to make hflush() much faster by 
> unsynchronizing it. But, we forgot to un-synchronize the deprecated 
> {{sync()}} wrapper method. This makes the HBase WAL really slow on 0.23+ 
> since it doesn't take advantage of HDFS-895 anymore.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3049) During the normal loading NN startup process, fall back on a different EditLog if we see one that is corrupt

2012-04-13 Thread Hadoop QA (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3049?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13253945#comment-13253945
 ] 

Hadoop QA commented on HDFS-3049:
-

-1 overall.  Here are the results of testing the latest attachment 
  http://issues.apache.org/jira/secure/attachment/12522637/HDFS-3049.001.patch
  against trunk revision .

+1 @author.  The patch does not contain any @author tags.

+1 tests included.  The patch appears to include 3 new or modified test 
files.

+1 javadoc.  The javadoc tool did not generate any warning messages.

+1 javac.  The applied patch does not increase the total number of javac 
compiler warnings.

+1 eclipse:eclipse.  The patch built with eclipse:eclipse.

+1 findbugs.  The patch does not introduce any new Findbugs (version 1.3.9) 
warnings.

+1 release audit.  The applied patch does not increase the total number of 
release audit warnings.

-1 core tests.  The patch failed these unit tests:
  org.apache.hadoop.hdfs.server.namenode.TestFSEditLogLoader
  org.apache.hadoop.hdfs.server.namenode.TestGenericJournalConf

+1 contrib tests.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-HDFS-Build/2276//testReport/
Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/2276//console

This message is automatically generated.

> During the normal loading NN startup process, fall back on a different 
> EditLog if we see one that is corrupt
> 
>
> Key: HDFS-3049
> URL: https://issues.apache.org/jira/browse/HDFS-3049
> Project: Hadoop HDFS
>  Issue Type: New Feature
>  Components: name-node
>Reporter: Colin Patrick McCabe
>Assignee: Colin Patrick McCabe
>Priority: Minor
> Fix For: 0.24.0
>
> Attachments: HDFS-3049.001.patch
>
>
> During the NameNode startup process, we load an image, and then apply edit 
> logs to it until we believe that we have all the latest changes.  
> Unfortunately, if there is an I/O error while reading any of these files, in 
> most cases, we simply abort the startup process.  We should try harder to 
> locate a readable edit log and/or image file.
> *There are three main use cases for this feature:*
> 1. If the operating system does not honor fsync (usually due to a 
> misconfiguration), a file may end up in an inconsistent state.
> 2. In certain older releases where we did not use fallocate() or similar to 
> pre-reserve blocks, a disk full condition may cause a truncated log in one 
> edit directory.
> 3. There may be a bug in HDFS which results in some of the data directories 
> receiving corrupt data, but not all.  This is the least likely use case.
> *Proposed changes to normal NN startup*
> * We should try a different FSImage if we can't load the first one we try.
> * We should examine other FSEditLogs if we can't load the first one(s) we try.
> * We should fail if we can't find EditLogs that would bring us up to what we 
> believe is the latest transaction ID.
> Proposed changes to recovery mode NN startup:
> we should list out all the available storage directories and allow the 
> operator to select which one he wants to use.
> Something like this:
> {code}
> Multiple storage directories found.
> 1. /foo/bar
> edits__curent__XYZ  size:213421345   md5:2345345
> image  size:213421345   md5:2345345
> 2. /foo/baz
> edits__curent__XYZ  size:213421345   md5:2345345345
> image  size:213421345   md5:2345345
> Which one would you like to use? (1/2)
> {code}
> As usual in recovery mode, we want to be flexible about error handling.  In 
> this case, this means that we should NOT fail if we can't find EditLogs that 
> would bring us up to what we believe is the latest transaction ID.
> *Not addressed by this feature*
> This feature will not address the case where an attempt to access the 
> NameNode name directory or directories hangs because of an I/O error.  This 
> may happen, for example, when trying to load an image from a hard-mounted NFS 
> directory, when the NFS server has gone away.  Just as now, the operator will 
> have to notice this problem and take steps to correct it.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Created] (HDFS-3281) TestParallelRead times out with Clover enabled

2012-04-13 Thread Henry Robinson (Created) (JIRA)
TestParallelRead times out with Clover enabled
--

 Key: HDFS-3281
 URL: https://issues.apache.org/jira/browse/HDFS-3281
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 3.0.0
Reporter: Henry Robinson
Assignee: Henry Robinson


Although it passes in a respectable time in pre-commit builds, when Clover is 
enabled TestParallelRead timesout (so runs > 15 minutes). 

I haven't been able to reproduce this on my development machines - the test 
runs in ~3 minutes and I don't think there's a factor of five difference in 
speed between this and the Jenkins box. 

I'm continuing to investigate the cause. If we can't find anything in the short 
term, we should temporarily disable this test to let the build go back to 
passing. 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3280) DFSOutputStream.sync should not be synchronized

2012-04-13 Thread Andrew Purtell (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3280?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13253942#comment-13253942
 ] 

Andrew Purtell commented on HDFS-3280:
--

Ah, so this explains what you guys thought might be an interaction with Nagle?

> DFSOutputStream.sync should not be synchronized
> ---
>
> Key: HDFS-3280
> URL: https://issues.apache.org/jira/browse/HDFS-3280
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: hdfs client
>Affects Versions: 2.0.0
>Reporter: Todd Lipcon
>Assignee: Todd Lipcon
>Priority: Critical
> Attachments: hdfs-3280.txt
>
>
> HDFS-895 added an optimization to make hflush() much faster by 
> unsynchronizing it. But, we forgot to un-synchronize the deprecated 
> {{sync()}} wrapper method. This makes the HBase WAL really slow on 0.23+ 
> since it doesn't take advantage of HDFS-895 anymore.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HDFS-3280) DFSOutputStream.sync should not be synchronized

2012-04-13 Thread Todd Lipcon (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3280?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Todd Lipcon updated HDFS-3280:
--

Status: Patch Available  (was: Open)

> DFSOutputStream.sync should not be synchronized
> ---
>
> Key: HDFS-3280
> URL: https://issues.apache.org/jira/browse/HDFS-3280
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: hdfs client
>Affects Versions: 2.0.0
>Reporter: Todd Lipcon
>Assignee: Todd Lipcon
>Priority: Critical
> Attachments: hdfs-3280.txt
>
>
> HDFS-895 added an optimization to make hflush() much faster by 
> unsynchronizing it. But, we forgot to un-synchronize the deprecated 
> {{sync()}} wrapper method. This makes the HBase WAL really slow on 0.23+ 
> since it doesn't take advantage of HDFS-895 anymore.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3280) DFSOutputStream.sync should not be synchronized

2012-04-13 Thread Todd Lipcon (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3280?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13253941#comment-13253941
 ] 

Todd Lipcon commented on HDFS-3280:
---

Verified that this increased my benchmark performance by a factor of two.

> DFSOutputStream.sync should not be synchronized
> ---
>
> Key: HDFS-3280
> URL: https://issues.apache.org/jira/browse/HDFS-3280
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: hdfs client
>Affects Versions: 2.0.0
>Reporter: Todd Lipcon
>Assignee: Todd Lipcon
>Priority: Critical
> Attachments: hdfs-3280.txt
>
>
> HDFS-895 added an optimization to make hflush() much faster by 
> unsynchronizing it. But, we forgot to un-synchronize the deprecated 
> {{sync()}} wrapper method. This makes the HBase WAL really slow on 0.23+ 
> since it doesn't take advantage of HDFS-895 anymore.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HDFS-3280) DFSOutputStream.sync should not be synchronized

2012-04-13 Thread Todd Lipcon (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3280?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Todd Lipcon updated HDFS-3280:
--

Attachment: hdfs-3280.txt

Trivial patch should fix the issue. I'll verify with an HLog benchmark on a 
cluster as well.

This is definitely correct since the method is just a one-line call to another 
method which is unsynchronized.

> DFSOutputStream.sync should not be synchronized
> ---
>
> Key: HDFS-3280
> URL: https://issues.apache.org/jira/browse/HDFS-3280
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: hdfs client
>Affects Versions: 2.0.0
>Reporter: Todd Lipcon
>Assignee: Todd Lipcon
>Priority: Critical
> Attachments: hdfs-3280.txt
>
>
> HDFS-895 added an optimization to make hflush() much faster by 
> unsynchronizing it. But, we forgot to un-synchronize the deprecated 
> {{sync()}} wrapper method. This makes the HBase WAL really slow on 0.23+ 
> since it doesn't take advantage of HDFS-895 anymore.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Created] (HDFS-3280) DFSOutputStream.sync should not be synchronized

2012-04-13 Thread Todd Lipcon (Created) (JIRA)
DFSOutputStream.sync should not be synchronized
---

 Key: HDFS-3280
 URL: https://issues.apache.org/jira/browse/HDFS-3280
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: hdfs client
Affects Versions: 2.0.0
Reporter: Todd Lipcon
Assignee: Todd Lipcon
Priority: Critical


HDFS-895 added an optimization to make hflush() much faster by unsynchronizing 
it. But, we forgot to un-synchronize the deprecated {{sync()}} wrapper method. 
This makes the HBase WAL really slow on 0.23+ since it doesn't take advantage 
of HDFS-895 anymore.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-2708) Stats for the # of blocks per DN

2012-04-13 Thread Hudson (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2708?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13253937#comment-13253937
 ] 

Hudson commented on HDFS-2708:
--

Integrated in Hadoop-Mapreduce-trunk-Commit #2088 (See 
[https://builds.apache.org/job/Hadoop-Mapreduce-trunk-Commit/2088/])
HDFS-2708. Stats for the # of blocks per DN. Contributed by Aaron T. Myers. 
(Revision 1326039)

 Result = FAILURE
atm : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1326039
Files : 
* /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSNamesystem.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestNameNodeMXBean.java


> Stats for the # of blocks per DN
> 
>
> Key: HDFS-2708
> URL: https://issues.apache.org/jira/browse/HDFS-2708
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: data-node, name-node
>Affects Versions: 2.0.0
>Reporter: Eli Collins
>Assignee: Aaron T. Myers
>Priority: Minor
> Fix For: 2.0.0
>
> Attachments: HDFS-2708.patch
>
>
> It would be useful for tools to be able to retrieve the total number of 
> blocks on each datanode.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-2708) Stats for the # of blocks per DN

2012-04-13 Thread Hudson (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2708?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13253932#comment-13253932
 ] 

Hudson commented on HDFS-2708:
--

Integrated in Hadoop-Hdfs-trunk-Commit #2147 (See 
[https://builds.apache.org/job/Hadoop-Hdfs-trunk-Commit/2147/])
HDFS-2708. Stats for the # of blocks per DN. Contributed by Aaron T. Myers. 
(Revision 1326039)

 Result = SUCCESS
atm : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1326039
Files : 
* /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSNamesystem.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestNameNodeMXBean.java


> Stats for the # of blocks per DN
> 
>
> Key: HDFS-2708
> URL: https://issues.apache.org/jira/browse/HDFS-2708
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: data-node, name-node
>Affects Versions: 2.0.0
>Reporter: Eli Collins
>Assignee: Aaron T. Myers
>Priority: Minor
> Fix For: 2.0.0
>
> Attachments: HDFS-2708.patch
>
>
> It would be useful for tools to be able to retrieve the total number of 
> blocks on each datanode.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-2708) Stats for the # of blocks per DN

2012-04-13 Thread Hudson (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2708?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13253933#comment-13253933
 ] 

Hudson commented on HDFS-2708:
--

Integrated in Hadoop-Common-trunk-Commit #2074 (See 
[https://builds.apache.org/job/Hadoop-Common-trunk-Commit/2074/])
HDFS-2708. Stats for the # of blocks per DN. Contributed by Aaron T. Myers. 
(Revision 1326039)

 Result = SUCCESS
atm : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1326039
Files : 
* /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSNamesystem.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestNameNodeMXBean.java


> Stats for the # of blocks per DN
> 
>
> Key: HDFS-2708
> URL: https://issues.apache.org/jira/browse/HDFS-2708
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: data-node, name-node
>Affects Versions: 2.0.0
>Reporter: Eli Collins
>Assignee: Aaron T. Myers
>Priority: Minor
> Fix For: 2.0.0
>
> Attachments: HDFS-2708.patch
>
>
> It would be useful for tools to be able to retrieve the total number of 
> blocks on each datanode.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HDFS-2708) Stats for the # of blocks per DN

2012-04-13 Thread Aaron T. Myers (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-2708?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Aaron T. Myers updated HDFS-2708:
-

   Resolution: Fixed
Fix Version/s: 2.0.0
 Hadoop Flags: Reviewed
   Status: Resolved  (was: Patch Available)

Thanks a lot for the review, Todd. I've just committed this to trunk and 
branch-2.

> Stats for the # of blocks per DN
> 
>
> Key: HDFS-2708
> URL: https://issues.apache.org/jira/browse/HDFS-2708
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: data-node, name-node
>Affects Versions: 2.0.0
>Reporter: Eli Collins
>Assignee: Aaron T. Myers
>Priority: Minor
> Fix For: 2.0.0
>
> Attachments: HDFS-2708.patch
>
>
> It would be useful for tools to be able to retrieve the total number of 
> blocks on each datanode.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-2708) Stats for the # of blocks per DN

2012-04-13 Thread Hadoop QA (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2708?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13253923#comment-13253923
 ] 

Hadoop QA commented on HDFS-2708:
-

+1 overall.  Here are the results of testing the latest attachment 
  http://issues.apache.org/jira/secure/attachment/12522641/HDFS-2708.patch
  against trunk revision .

+1 @author.  The patch does not contain any @author tags.

+1 tests included.  The patch appears to include 1 new or modified test 
files.

+1 javadoc.  The javadoc tool did not generate any warning messages.

+1 javac.  The applied patch does not increase the total number of javac 
compiler warnings.

+1 eclipse:eclipse.  The patch built with eclipse:eclipse.

+1 findbugs.  The patch does not introduce any new Findbugs (version 1.3.9) 
warnings.

+1 release audit.  The applied patch does not increase the total number of 
release audit warnings.

+1 core tests.  The patch passed unit tests in .

+1 contrib tests.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-HDFS-Build/2275//testReport/
Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/2275//console

This message is automatically generated.

> Stats for the # of blocks per DN
> 
>
> Key: HDFS-2708
> URL: https://issues.apache.org/jira/browse/HDFS-2708
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: data-node, name-node
>Affects Versions: 2.0.0
>Reporter: Eli Collins
>Assignee: Aaron T. Myers
>Priority: Minor
> Attachments: HDFS-2708.patch
>
>
> It would be useful for tools to be able to retrieve the total number of 
> blocks on each datanode.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HDFS-3049) During the normal loading NN startup process, fall back on a different EditLog if we see one that is corrupt

2012-04-13 Thread Colin Patrick McCabe (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3049?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Colin Patrick McCabe updated HDFS-3049:
---

Status: Patch Available  (was: Open)

> During the normal loading NN startup process, fall back on a different 
> EditLog if we see one that is corrupt
> 
>
> Key: HDFS-3049
> URL: https://issues.apache.org/jira/browse/HDFS-3049
> Project: Hadoop HDFS
>  Issue Type: New Feature
>  Components: name-node
>Reporter: Colin Patrick McCabe
>Assignee: Colin Patrick McCabe
>Priority: Minor
> Fix For: 0.24.0
>
> Attachments: HDFS-3049.001.patch
>
>
> During the NameNode startup process, we load an image, and then apply edit 
> logs to it until we believe that we have all the latest changes.  
> Unfortunately, if there is an I/O error while reading any of these files, in 
> most cases, we simply abort the startup process.  We should try harder to 
> locate a readable edit log and/or image file.
> *There are three main use cases for this feature:*
> 1. If the operating system does not honor fsync (usually due to a 
> misconfiguration), a file may end up in an inconsistent state.
> 2. In certain older releases where we did not use fallocate() or similar to 
> pre-reserve blocks, a disk full condition may cause a truncated log in one 
> edit directory.
> 3. There may be a bug in HDFS which results in some of the data directories 
> receiving corrupt data, but not all.  This is the least likely use case.
> *Proposed changes to normal NN startup*
> * We should try a different FSImage if we can't load the first one we try.
> * We should examine other FSEditLogs if we can't load the first one(s) we try.
> * We should fail if we can't find EditLogs that would bring us up to what we 
> believe is the latest transaction ID.
> Proposed changes to recovery mode NN startup:
> we should list out all the available storage directories and allow the 
> operator to select which one he wants to use.
> Something like this:
> {code}
> Multiple storage directories found.
> 1. /foo/bar
> edits__curent__XYZ  size:213421345   md5:2345345
> image  size:213421345   md5:2345345
> 2. /foo/baz
> edits__curent__XYZ  size:213421345   md5:2345345345
> image  size:213421345   md5:2345345
> Which one would you like to use? (1/2)
> {code}
> As usual in recovery mode, we want to be flexible about error handling.  In 
> this case, this means that we should NOT fail if we can't find EditLogs that 
> would bring us up to what we believe is the latest transaction ID.
> *Not addressed by this feature*
> This feature will not address the case where an attempt to access the 
> NameNode name directory or directories hangs because of an I/O error.  This 
> may happen, for example, when trying to load an image from a hard-mounted NFS 
> directory, when the NFS server has gone away.  Just as now, the operator will 
> have to notice this problem and take steps to correct it.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HDFS-3134) harden edit log loader against malformed or malicious input

2012-04-13 Thread Colin Patrick McCabe (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3134?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Colin Patrick McCabe updated HDFS-3134:
---

Affects Version/s: 0.23.0
Fix Version/s: 2.0.0

> harden edit log loader against malformed or malicious input
> ---
>
> Key: HDFS-3134
> URL: https://issues.apache.org/jira/browse/HDFS-3134
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: name-node
>Affects Versions: 0.23.0
>Reporter: Colin Patrick McCabe
>Assignee: Colin Patrick McCabe
> Fix For: 2.0.0
>
> Attachments: HDFS-3134.001.patch, HDFS-3134.002.patch, 
> HDFS-3134.003.patch, HDFS-3134.004.patch, HDFS-3134.005.patch
>
>
> Currently, the edit log loader does not handle bad or malicious input 
> sensibly.
> We can often cause OutOfMemory exceptions, null pointer exceptions, or other 
> unchecked exceptions to be thrown by feeding the edit log loader bad input.  
> In some environments, an out of memory error can cause the JVM process to be 
> terminated.
> It's clear that we want these exceptions to be thrown as IOException instead 
> of as unchecked exceptions.  We also want to avoid out of memory situations.
> The main task here is to put a sensible upper limit on the lengths of arrays 
> and strings we allocate on command.  The other task is to try to avoid 
> creating unchecked exceptions (by dereferencing potentially-NULL pointers, 
> for example).  Instead, we should verify ahead of time and give a more 
> sensible error message that reflects the problem with the input.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-2708) Stats for the # of blocks per DN

2012-04-13 Thread Todd Lipcon (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2708?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13253903#comment-13253903
 ] 

Todd Lipcon commented on HDFS-2708:
---

+1 pending jenkins

> Stats for the # of blocks per DN
> 
>
> Key: HDFS-2708
> URL: https://issues.apache.org/jira/browse/HDFS-2708
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: data-node, name-node
>Affects Versions: 2.0.0
>Reporter: Eli Collins
>Assignee: Aaron T. Myers
>Priority: Minor
> Attachments: HDFS-2708.patch
>
>
> It would be useful for tools to be able to retrieve the total number of 
> blocks on each datanode.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HDFS-2708) Stats for the # of blocks per DN

2012-04-13 Thread Aaron T. Myers (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-2708?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Aaron T. Myers updated HDFS-2708:
-

Attachment: HDFS-2708.patch

Here's a patch for trunk which adds a "numBlocks" field to each DN in the 
"LiveNodes" field of the NameNodeInfo JMX interface.

> Stats for the # of blocks per DN
> 
>
> Key: HDFS-2708
> URL: https://issues.apache.org/jira/browse/HDFS-2708
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: data-node, name-node
>Affects Versions: 2.0.0
>Reporter: Eli Collins
>Assignee: Aaron T. Myers
>Priority: Minor
> Attachments: HDFS-2708.patch
>
>
> It would be useful for tools to be able to retrieve the total number of 
> blocks on each datanode.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HDFS-2708) Stats for the # of blocks per DN

2012-04-13 Thread Aaron T. Myers (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-2708?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Aaron T. Myers updated HDFS-2708:
-

Status: Patch Available  (was: Open)

> Stats for the # of blocks per DN
> 
>
> Key: HDFS-2708
> URL: https://issues.apache.org/jira/browse/HDFS-2708
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: data-node, name-node
>Affects Versions: 2.0.0
>Reporter: Eli Collins
>Assignee: Aaron T. Myers
>Priority: Minor
> Attachments: HDFS-2708.patch
>
>
> It would be useful for tools to be able to retrieve the total number of 
> blocks on each datanode.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3273) Refactor BackupImage and FSEditLog

2012-04-13 Thread Aaron T. Myers (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3273?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13253897#comment-13253897
 ] 

Aaron T. Myers commented on HDFS-3273:
--

Hi Nicholas, could you please add some motivation for this change? Perhaps in 
the "Description" field of this JIRA? Thanks a lot.

> Refactor BackupImage and FSEditLog
> --
>
> Key: HDFS-3273
> URL: https://issues.apache.org/jira/browse/HDFS-3273
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: ha, name-node
>Reporter: Tsz Wo (Nicholas), SZE
>Assignee: Tsz Wo (Nicholas), SZE
> Fix For: 3.0.0
>
> Attachments: h3273_20120413.patch
>
>


--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Created] (HDFS-3279) One of the FSEditLog constructor should be moved to TestEditLog

2012-04-13 Thread Tsz Wo (Nicholas), SZE (Created) (JIRA)
One of the FSEditLog constructor should be moved to TestEditLog
---

 Key: HDFS-3279
 URL: https://issues.apache.org/jira/browse/HDFS-3279
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: name-node
Reporter: Tsz Wo (Nicholas), SZE
Assignee: Arpit Gupta
Priority: Minor


The FSEditLog constructor with @VisibleForTesting is used only in TestEditLog.  
It could be simply declared as a static method in TestEditLog.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HDFS-2708) Stats for the # of blocks per DN

2012-04-13 Thread Aaron T. Myers (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-2708?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Aaron T. Myers updated HDFS-2708:
-

  Description: It would be useful for tools to be able to retrieve the 
total number of blocks on each datanode.  (was: It would be useful for tools to 
be able to retrieve the total # blocks in the file system (and also display eg 
via dfsadmin report, this is currently only available via FSNamesystemMetrics, 
so would add to ClientProtocol#getStats?) and the total number of blocks on 
each datanode (via DataNodeInfo).)
 Target Version/s: 2.0.0  (was: 1.1.0, 0.23.1)
Affects Version/s: 2.0.0
 Assignee: Aaron T. Myers
  Summary: Stats for the # of blocks per DN  (was: Stats for the 
total # blocks and blocks per DN)

Let's separate the original description of this JIRA into multiple distinct 
issues. In this JIRA, let's just add support for getting the number of blocks 
on each DN via the FSNamesystem metrics, since that interface already has 
support for retrieving the total number of blocks in the file system.

If we really do want to add other interfaces to retrieve both of these metrics, 
let's do that in a separate JIRA.

> Stats for the # of blocks per DN
> 
>
> Key: HDFS-2708
> URL: https://issues.apache.org/jira/browse/HDFS-2708
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: data-node, name-node
>Affects Versions: 2.0.0
>Reporter: Eli Collins
>Assignee: Aaron T. Myers
>Priority: Minor
>
> It would be useful for tools to be able to retrieve the total number of 
> blocks on each datanode.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-2765) TestNameEditsConfigs is incorrectly swallowing IOE

2012-04-13 Thread Hudson (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2765?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13253888#comment-13253888
 ] 

Hudson commented on HDFS-2765:
--

Integrated in Hadoop-Mapreduce-trunk-Commit #2087 (See 
[https://builds.apache.org/job/Hadoop-Mapreduce-trunk-Commit/2087/])
Move CHANGES.txt message for HDFS-2765 to be under 2.0.0 instead of trunk. 
(Revision 1326013)

 Result = FAILURE
atm : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1326013
Files : 
* /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt


> TestNameEditsConfigs is incorrectly swallowing IOE
> --
>
> Key: HDFS-2765
> URL: https://issues.apache.org/jira/browse/HDFS-2765
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: test
>Affects Versions: 2.0.0
>Reporter: Aaron T. Myers
>Assignee: Aaron T. Myers
> Fix For: 2.0.0
>
> Attachments: HDFS-2765.patch
>
>
> The final portion of this test case is swallowing an IOE and in so doing 
> appearing to succeed, although it should not be succeeding as-written.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3273) Refactor BackupImage and FSEditLog

2012-04-13 Thread Hudson (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3273?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13253889#comment-13253889
 ] 

Hudson commented on HDFS-3273:
--

Integrated in Hadoop-Mapreduce-trunk-Commit #2087 (See 
[https://builds.apache.org/job/Hadoop-Mapreduce-trunk-Commit/2087/])
HDFS-3273. Refactor BackupImage and FSEditLog, and rename 
JournalListener.rollLogs(..) to startLogSegment(..). (Revision 1326016)

 Result = FAILURE
szetszwo : 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1326016
Files : 
* /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/journalservice/JournalListener.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/journalservice/JournalService.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/BackupImage.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/BackupNode.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSEditLog.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSImage.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/journalservice/TestJournalService.java


> Refactor BackupImage and FSEditLog
> --
>
> Key: HDFS-3273
> URL: https://issues.apache.org/jira/browse/HDFS-3273
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: ha, name-node
>Reporter: Tsz Wo (Nicholas), SZE
>Assignee: Tsz Wo (Nicholas), SZE
> Fix For: 3.0.0
>
> Attachments: h3273_20120413.patch
>
>


--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3042) Automatic failover support for NN HA

2012-04-13 Thread Todd Lipcon (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3042?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13253883#comment-13253883
 ] 

Todd Lipcon commented on HDFS-3042:
---

Sanjay: this is being done in a branch... it's in branches/HDFS-3042 in SVN.

> Automatic failover support for NN HA
> 
>
> Key: HDFS-3042
> URL: https://issues.apache.org/jira/browse/HDFS-3042
> Project: Hadoop HDFS
>  Issue Type: New Feature
>  Components: auto-failover, ha
>Reporter: Todd Lipcon
>Assignee: Todd Lipcon
>
> HDFS-1623 was the umbrella task for implementation of NN HA capabilities. 
> However, it only focused on manually-triggered failover.
> Given that the HDFS-1623 branch will be merged shortly, I'm opening this JIRA 
> to consolidate/track subtasks for automatic failover support and related 
> improvements.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3273) Refactor BackupImage and FSEditLog

2012-04-13 Thread Hudson (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3273?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13253878#comment-13253878
 ] 

Hudson commented on HDFS-3273:
--

Integrated in Hadoop-Common-trunk-Commit #2073 (See 
[https://builds.apache.org/job/Hadoop-Common-trunk-Commit/2073/])
HDFS-3273. Refactor BackupImage and FSEditLog, and rename 
JournalListener.rollLogs(..) to startLogSegment(..). (Revision 1326016)

 Result = SUCCESS
szetszwo : 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1326016
Files : 
* /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/journalservice/JournalListener.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/journalservice/JournalService.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/BackupImage.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/BackupNode.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSEditLog.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSImage.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/journalservice/TestJournalService.java


> Refactor BackupImage and FSEditLog
> --
>
> Key: HDFS-3273
> URL: https://issues.apache.org/jira/browse/HDFS-3273
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: ha, name-node
>Reporter: Tsz Wo (Nicholas), SZE
>Assignee: Tsz Wo (Nicholas), SZE
> Fix For: 3.0.0
>
> Attachments: h3273_20120413.patch
>
>


--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-2765) TestNameEditsConfigs is incorrectly swallowing IOE

2012-04-13 Thread Hudson (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2765?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13253877#comment-13253877
 ] 

Hudson commented on HDFS-2765:
--

Integrated in Hadoop-Common-trunk-Commit #2073 (See 
[https://builds.apache.org/job/Hadoop-Common-trunk-Commit/2073/])
Move CHANGES.txt message for HDFS-2765 to be under 2.0.0 instead of trunk. 
(Revision 1326013)

 Result = SUCCESS
atm : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1326013
Files : 
* /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt


> TestNameEditsConfigs is incorrectly swallowing IOE
> --
>
> Key: HDFS-2765
> URL: https://issues.apache.org/jira/browse/HDFS-2765
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: test
>Affects Versions: 2.0.0
>Reporter: Aaron T. Myers
>Assignee: Aaron T. Myers
> Fix For: 2.0.0
>
> Attachments: HDFS-2765.patch
>
>
> The final portion of this test case is swallowing an IOE and in so doing 
> appearing to succeed, although it should not be succeeding as-written.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-2765) TestNameEditsConfigs is incorrectly swallowing IOE

2012-04-13 Thread Hudson (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2765?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13253875#comment-13253875
 ] 

Hudson commented on HDFS-2765:
--

Integrated in Hadoop-Hdfs-trunk-Commit #2146 (See 
[https://builds.apache.org/job/Hadoop-Hdfs-trunk-Commit/2146/])
Move CHANGES.txt message for HDFS-2765 to be under 2.0.0 instead of trunk. 
(Revision 1326013)

 Result = SUCCESS
atm : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1326013
Files : 
* /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt


> TestNameEditsConfigs is incorrectly swallowing IOE
> --
>
> Key: HDFS-2765
> URL: https://issues.apache.org/jira/browse/HDFS-2765
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: test
>Affects Versions: 2.0.0
>Reporter: Aaron T. Myers
>Assignee: Aaron T. Myers
> Fix For: 2.0.0
>
> Attachments: HDFS-2765.patch
>
>
> The final portion of this test case is swallowing an IOE and in so doing 
> appearing to succeed, although it should not be succeeding as-written.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3273) Refactor BackupImage and FSEditLog

2012-04-13 Thread Hudson (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3273?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13253876#comment-13253876
 ] 

Hudson commented on HDFS-3273:
--

Integrated in Hadoop-Hdfs-trunk-Commit #2146 (See 
[https://builds.apache.org/job/Hadoop-Hdfs-trunk-Commit/2146/])
HDFS-3273. Refactor BackupImage and FSEditLog, and rename 
JournalListener.rollLogs(..) to startLogSegment(..). (Revision 1326016)

 Result = SUCCESS
szetszwo : 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1326016
Files : 
* /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/journalservice/JournalListener.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/journalservice/JournalService.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/BackupImage.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/BackupNode.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSEditLog.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSImage.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/journalservice/TestJournalService.java


> Refactor BackupImage and FSEditLog
> --
>
> Key: HDFS-3273
> URL: https://issues.apache.org/jira/browse/HDFS-3273
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: ha, name-node
>Reporter: Tsz Wo (Nicholas), SZE
>Assignee: Tsz Wo (Nicholas), SZE
> Fix For: 3.0.0
>
> Attachments: h3273_20120413.patch
>
>


--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HDFS-3273) Refactor BackupImage and FSEditLog

2012-04-13 Thread Tsz Wo (Nicholas), SZE (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3273?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tsz Wo (Nicholas), SZE updated HDFS-3273:
-

   Resolution: Fixed
Fix Version/s: 3.0.0
 Hadoop Flags: Reviewed
   Status: Resolved  (was: Patch Available)

Suresh, thanks for reviewing this.

I have committed this.

> Refactor BackupImage and FSEditLog
> --
>
> Key: HDFS-3273
> URL: https://issues.apache.org/jira/browse/HDFS-3273
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: ha, name-node
>Reporter: Tsz Wo (Nicholas), SZE
>Assignee: Tsz Wo (Nicholas), SZE
> Fix For: 3.0.0
>
> Attachments: h3273_20120413.patch
>
>


--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Created] (HDFS-3278) Umbrella Jira for HDFS-HA Phase 2

2012-04-13 Thread Sanjay Radia (Created) (JIRA)
Umbrella Jira for HDFS-HA Phase 2
-

 Key: HDFS-3278
 URL: https://issues.apache.org/jira/browse/HDFS-3278
 Project: Hadoop HDFS
  Issue Type: New Feature
Reporter: Sanjay Radia
Assignee: Sanjay Radia


HDFS-1623 gives a high level architecture and design for hot automatic failover 
of the NN. Branch HDFS-1623 was merged into trunk for tactical reasons even 
though the work for HA was not complete, Branch HDFS-1623 contained mechanisms 
for keeping a standby Hot (ie read from shared journal), dual block reports, 
fencing of DNs, Zookeeper library for leader election etc. This Umbrella jira 
covers the remaining work for HA and will link all the jiras for the remaining 
work. Unlike HDFS-1623 no single branch will be created - work will proceed in 
parallel branches.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Resolved] (HDFS-2765) TestNameEditsConfigs is incorrectly swallowing IOE

2012-04-13 Thread Aaron T. Myers (Resolved) (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-2765?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Aaron T. Myers resolved HDFS-2765.
--

   Resolution: Fixed
Fix Version/s: 2.0.0

I've back-ported this fix to branch-2.

> TestNameEditsConfigs is incorrectly swallowing IOE
> --
>
> Key: HDFS-2765
> URL: https://issues.apache.org/jira/browse/HDFS-2765
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: test
>Affects Versions: 2.0.0
>Reporter: Aaron T. Myers
>Assignee: Aaron T. Myers
> Fix For: 2.0.0
>
> Attachments: HDFS-2765.patch
>
>
> The final portion of this test case is swallowing an IOE and in so doing 
> appearing to succeed, although it should not be succeeding as-written.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3273) Refactor BackupImage and FSEditLog

2012-04-13 Thread Suresh Srinivas (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3273?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13253869#comment-13253869
 ] 

Suresh Srinivas commented on HDFS-3273:
---

+1 for the patch.

> Refactor BackupImage and FSEditLog
> --
>
> Key: HDFS-3273
> URL: https://issues.apache.org/jira/browse/HDFS-3273
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: ha, name-node
>Reporter: Tsz Wo (Nicholas), SZE
>Assignee: Tsz Wo (Nicholas), SZE
> Attachments: h3273_20120413.patch
>
>


--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3042) Automatic failover support for NN HA

2012-04-13 Thread Sanjay Radia (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3042?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13253867#comment-13253867
 ] 

Sanjay Radia commented on HDFS-3042:


Todd, I would have preferred to have all these done in branch as with the rest 
of the HA work. 
There is work that remains and this is best done in a branch.

> Automatic failover support for NN HA
> 
>
> Key: HDFS-3042
> URL: https://issues.apache.org/jira/browse/HDFS-3042
> Project: Hadoop HDFS
>  Issue Type: New Feature
>  Components: auto-failover, ha
>Reporter: Todd Lipcon
>Assignee: Todd Lipcon
>
> HDFS-1623 was the umbrella task for implementation of NN HA capabilities. 
> However, it only focused on manually-triggered failover.
> Given that the HDFS-1623 branch will be merged shortly, I'm opening this JIRA 
> to consolidate/track subtasks for automatic failover support and related 
> improvements.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Resolved] (HDFS-3274) Create a new journal_edits_dir key to support journal nodes

2012-04-13 Thread Tsz Wo (Nicholas), SZE (Resolved) (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3274?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tsz Wo (Nicholas), SZE resolved HDFS-3274.
--

   Resolution: Fixed
Fix Version/s: Shared journals (HDFS-3092)
 Hadoop Flags: Reviewed

I have committed this.  Thanks, Hari!

> Create a new journal_edits_dir key to support journal nodes 
> 
>
> Key: HDFS-3274
> URL: https://issues.apache.org/jira/browse/HDFS-3274
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: ha, name-node
>Reporter: Hari Mankude
>Assignee: Hari Mankude
> Fix For: Shared journals (HDFS-3092)
>
> Attachments: hdfs-3274-1.patch, hdfs-3274.patch
>
>


--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HDFS-3049) During the normal loading NN startup process, fall back on a different EditLog if we see one that is corrupt

2012-04-13 Thread Colin Patrick McCabe (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3049?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Colin Patrick McCabe updated HDFS-3049:
---

Summary: During the normal loading NN startup process, fall back on a 
different EditLog if we see one that is corrupt  (was: During the normal 
loading NN startup process, fall back on a different image or EditLog if we see 
one that is corrupt)

remove references to FSImage (there is now a separate JIRA for that)

> During the normal loading NN startup process, fall back on a different 
> EditLog if we see one that is corrupt
> 
>
> Key: HDFS-3049
> URL: https://issues.apache.org/jira/browse/HDFS-3049
> Project: Hadoop HDFS
>  Issue Type: New Feature
>  Components: name-node
>Reporter: Colin Patrick McCabe
>Assignee: Colin Patrick McCabe
>Priority: Minor
> Fix For: 0.24.0
>
> Attachments: HDFS-3049.001.patch
>
>
> During the NameNode startup process, we load an image, and then apply edit 
> logs to it until we believe that we have all the latest changes.  
> Unfortunately, if there is an I/O error while reading any of these files, in 
> most cases, we simply abort the startup process.  We should try harder to 
> locate a readable edit log and/or image file.
> *There are three main use cases for this feature:*
> 1. If the operating system does not honor fsync (usually due to a 
> misconfiguration), a file may end up in an inconsistent state.
> 2. In certain older releases where we did not use fallocate() or similar to 
> pre-reserve blocks, a disk full condition may cause a truncated log in one 
> edit directory.
> 3. There may be a bug in HDFS which results in some of the data directories 
> receiving corrupt data, but not all.  This is the least likely use case.
> *Proposed changes to normal NN startup*
> * We should try a different FSImage if we can't load the first one we try.
> * We should examine other FSEditLogs if we can't load the first one(s) we try.
> * We should fail if we can't find EditLogs that would bring us up to what we 
> believe is the latest transaction ID.
> Proposed changes to recovery mode NN startup:
> we should list out all the available storage directories and allow the 
> operator to select which one he wants to use.
> Something like this:
> {code}
> Multiple storage directories found.
> 1. /foo/bar
> edits__curent__XYZ  size:213421345   md5:2345345
> image  size:213421345   md5:2345345
> 2. /foo/baz
> edits__curent__XYZ  size:213421345   md5:2345345345
> image  size:213421345   md5:2345345
> Which one would you like to use? (1/2)
> {code}
> As usual in recovery mode, we want to be flexible about error handling.  In 
> this case, this means that we should NOT fail if we can't find EditLogs that 
> would bring us up to what we believe is the latest transaction ID.
> *Not addressed by this feature*
> This feature will not address the case where an attempt to access the 
> NameNode name directory or directories hangs because of an I/O error.  This 
> may happen, for example, when trying to load an image from a hard-mounted NFS 
> directory, when the NFS server has gone away.  Just as now, the operator will 
> have to notice this problem and take steps to correct it.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Created] (HDFS-3277) fail over to loading a different FSImage if the first one we try to load is corrupt

2012-04-13 Thread Colin Patrick McCabe (Created) (JIRA)
fail over to loading a different FSImage if the first one we try to load is 
corrupt
---

 Key: HDFS-3277
 URL: https://issues.apache.org/jira/browse/HDFS-3277
 Project: Hadoop HDFS
  Issue Type: Bug
Reporter: Colin Patrick McCabe
Assignee: Colin Patrick McCabe


Most users store multiple copies of the FSImage in order to prevent 
catastrophic data loss if a hard disk fails.  However, our image loading code 
is currently not set up to start reading another FSImage if loading the first 
one does not succeed.  We should add this capability.

We should also be sure to remove the FSImage directory that failed from the 
list of FSImage directories to write to, in the way we normally do when a write 
(as opopsed to read) fails.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HDFS-3049) During the normal loading NN startup process, fall back on a different image or EditLog if we see one that is corrupt

2012-04-13 Thread Colin Patrick McCabe (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3049?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Colin Patrick McCabe updated HDFS-3049:
---

Attachment: HDFS-3049.001.patch

* implement edit log failover (no image stuff in here)

> During the normal loading NN startup process, fall back on a different image 
> or EditLog if we see one that is corrupt
> -
>
> Key: HDFS-3049
> URL: https://issues.apache.org/jira/browse/HDFS-3049
> Project: Hadoop HDFS
>  Issue Type: New Feature
>  Components: name-node
>Reporter: Colin Patrick McCabe
>Assignee: Colin Patrick McCabe
>Priority: Minor
> Fix For: 0.24.0
>
> Attachments: HDFS-3049.001.patch
>
>
> During the NameNode startup process, we load an image, and then apply edit 
> logs to it until we believe that we have all the latest changes.  
> Unfortunately, if there is an I/O error while reading any of these files, in 
> most cases, we simply abort the startup process.  We should try harder to 
> locate a readable edit log and/or image file.
> *There are three main use cases for this feature:*
> 1. If the operating system does not honor fsync (usually due to a 
> misconfiguration), a file may end up in an inconsistent state.
> 2. In certain older releases where we did not use fallocate() or similar to 
> pre-reserve blocks, a disk full condition may cause a truncated log in one 
> edit directory.
> 3. There may be a bug in HDFS which results in some of the data directories 
> receiving corrupt data, but not all.  This is the least likely use case.
> *Proposed changes to normal NN startup*
> * We should try a different FSImage if we can't load the first one we try.
> * We should examine other FSEditLogs if we can't load the first one(s) we try.
> * We should fail if we can't find EditLogs that would bring us up to what we 
> believe is the latest transaction ID.
> Proposed changes to recovery mode NN startup:
> we should list out all the available storage directories and allow the 
> operator to select which one he wants to use.
> Something like this:
> {code}
> Multiple storage directories found.
> 1. /foo/bar
> edits__curent__XYZ  size:213421345   md5:2345345
> image  size:213421345   md5:2345345
> 2. /foo/baz
> edits__curent__XYZ  size:213421345   md5:2345345345
> image  size:213421345   md5:2345345
> Which one would you like to use? (1/2)
> {code}
> As usual in recovery mode, we want to be flexible about error handling.  In 
> this case, this means that we should NOT fail if we can't find EditLogs that 
> would bring us up to what we believe is the latest transaction ID.
> *Not addressed by this feature*
> This feature will not address the case where an attempt to access the 
> NameNode name directory or directories hangs because of an I/O error.  This 
> may happen, for example, when trying to load an image from a hard-mounted NFS 
> directory, when the NFS server has gone away.  Just as now, the operator will 
> have to notice this problem and take steps to correct it.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Reopened] (HDFS-2765) TestNameEditsConfigs is incorrectly swallowing IOE

2012-04-13 Thread Aaron T. Myers (Reopened) (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-2765?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Aaron T. Myers reopened HDFS-2765:
--


Reopening for back-port to branch-2.

> TestNameEditsConfigs is incorrectly swallowing IOE
> --
>
> Key: HDFS-2765
> URL: https://issues.apache.org/jira/browse/HDFS-2765
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: test
>Affects Versions: 2.0.0
>Reporter: Aaron T. Myers
>Assignee: Aaron T. Myers
> Attachments: HDFS-2765.patch
>
>
> The final portion of this test case is swallowing an IOE and in so doing 
> appearing to succeed, although it should not be succeeding as-written.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HDFS-2765) TestNameEditsConfigs is incorrectly swallowing IOE

2012-04-13 Thread Aaron T. Myers (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-2765?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Aaron T. Myers updated HDFS-2765:
-

 Target Version/s: 2.0.0  (was: 0.24.0)
Affects Version/s: (was: 0.24.0)
   2.0.0
Fix Version/s: (was: 3.0.0)

> TestNameEditsConfigs is incorrectly swallowing IOE
> --
>
> Key: HDFS-2765
> URL: https://issues.apache.org/jira/browse/HDFS-2765
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: test
>Affects Versions: 2.0.0
>Reporter: Aaron T. Myers
>Assignee: Aaron T. Myers
> Attachments: HDFS-2765.patch
>
>
> The final portion of this test case is swallowing an IOE and in so doing 
> appearing to succeed, although it should not be succeeding as-written.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HDFS-3274) Create a new journal_edits_dir key to support journal nodes

2012-04-13 Thread Hari Mankude (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3274?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hari Mankude updated HDFS-3274:
---

Attachment: hdfs-3274-1.patch

> Create a new journal_edits_dir key to support journal nodes 
> 
>
> Key: HDFS-3274
> URL: https://issues.apache.org/jira/browse/HDFS-3274
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: ha, name-node
>Reporter: Hari Mankude
>Assignee: Hari Mankude
> Attachments: hdfs-3274-1.patch, hdfs-3274.patch
>
>


--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3274) Create a new journal_edits_dir key to support journal nodes

2012-04-13 Thread Hari Mankude (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3274?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13253851#comment-13253851
 ] 

Hari Mankude commented on HDFS-3274:


@Brandon
done, removed the invalid comment
@Nicholas
done. created javadoc.

Will update the patch shortly.

> Create a new journal_edits_dir key to support journal nodes 
> 
>
> Key: HDFS-3274
> URL: https://issues.apache.org/jira/browse/HDFS-3274
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: ha, name-node
>Reporter: Hari Mankude
>Assignee: Hari Mankude
> Attachments: hdfs-3274.patch
>
>


--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3274) Create a new journal_edits_dir key to support journal nodes

2012-04-13 Thread Tsz Wo (Nicholas), SZE (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3274?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13253848#comment-13253848
 ] 

Tsz Wo (Nicholas), SZE commented on HDFS-3274:
--

+1 patch looks good.

Please change initializeJournalStorage(..) private and the comment above to 
javadoc.

> Create a new journal_edits_dir key to support journal nodes 
> 
>
> Key: HDFS-3274
> URL: https://issues.apache.org/jira/browse/HDFS-3274
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: ha, name-node
>Reporter: Hari Mankude
>Assignee: Hari Mankude
> Attachments: hdfs-3274.patch
>
>


--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Created] (HDFS-3276) initializeSharedEdits should have a -nonInteractive flag

2012-04-13 Thread Vinithra Varadharajan (Created) (JIRA)
initializeSharedEdits should have a -nonInteractive flag


 Key: HDFS-3276
 URL: https://issues.apache.org/jira/browse/HDFS-3276
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: ha, name-node
Affects Versions: 2.0.0
Reporter: Vinithra Varadharajan
Priority: Minor


Similar to format and bootstrapStandby, would be nice to have -nonInteractive 
as an option on initializeSharedEdits

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HDFS-3275) Format command overwrites contents of non-empty shared edits dir if name dirs are empty without any prompting

2012-04-13 Thread Aaron T. Myers (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3275?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Aaron T. Myers updated HDFS-3275:
-

Description: 
To reproduce:
# Configure a NameNode with namedirs and a shared edits dir, all of which are 
empty.
# Run hdfs namenode -format. Namedirs and shared edits dir gets populated.
# Delete the contents of the namedirs. Leave the shared edits dir as is. Check 
the timestamps of the shared edits dir contents.
# Run format again. The namedirs as well as the shared edits dir get formatted. 
The shared edits dir's contents have been replaced without any prompting.

  was:
To reproduce:
1) Configure a NameNode with namedirs and a shared edits dir, all of which are 
empty.
2) Run hdfs namenode -format. Namedirs and shared edits dir gets populated.
2) Delete the contents of the namedirs. Leave the shared edits dir as is. Check 
the timestamps of the shared edits dir contents.
3) Run format again. The namedirs as well as the shared edits dir get 
formatted. The shared edits dir's contents have been replaced without any 
prompting.


Updating the numbering of the steps. (The original numbering had two step 2's.)

> Format command overwrites contents of non-empty shared edits dir if name dirs 
> are empty without any prompting
> -
>
> Key: HDFS-3275
> URL: https://issues.apache.org/jira/browse/HDFS-3275
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: ha, name-node
>Affects Versions: 2.0.0
>Reporter: Vinithra Varadharajan
>
> To reproduce:
> # Configure a NameNode with namedirs and a shared edits dir, all of which are 
> empty.
> # Run hdfs namenode -format. Namedirs and shared edits dir gets populated.
> # Delete the contents of the namedirs. Leave the shared edits dir as is. 
> Check the timestamps of the shared edits dir contents.
> # Run format again. The namedirs as well as the shared edits dir get 
> formatted. The shared edits dir's contents have been replaced without any 
> prompting.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Created] (HDFS-3275) Format command overwrites contents of non-empty shared edits dir if name dirs are empty without any prompting

2012-04-13 Thread Vinithra Varadharajan (Created) (JIRA)
Format command overwrites contents of non-empty shared edits dir if name dirs 
are empty without any prompting
-

 Key: HDFS-3275
 URL: https://issues.apache.org/jira/browse/HDFS-3275
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: ha, name-node
Affects Versions: 2.0.0
Reporter: Vinithra Varadharajan


To reproduce:
1) Configure a NameNode with namedirs and a shared edits dir, all of which are 
empty.
2) Run hdfs namenode -format. Namedirs and shared edits dir gets populated.
2) Delete the contents of the namedirs. Leave the shared edits dir as is. Check 
the timestamps of the shared edits dir contents.
3) Run format again. The namedirs as well as the shared edits dir get 
formatted. The shared edits dir's contents have been replaced without any 
prompting.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Resolved] (HDFS-3261) TestHASafeMode fails on HDFS-3042 branch

2012-04-13 Thread Todd Lipcon (Resolved) (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3261?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Todd Lipcon resolved HDFS-3261.
---

   Resolution: Fixed
Fix Version/s: Auto failover (HDFS-3042)
 Hadoop Flags: Reviewed

Committed this yesterday, just forgot to resolve

> TestHASafeMode fails on HDFS-3042 branch
> 
>
> Key: HDFS-3261
> URL: https://issues.apache.org/jira/browse/HDFS-3261
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: test
>Affects Versions: Auto failover (HDFS-3042)
>Reporter: Todd Lipcon
>Assignee: Todd Lipcon
>Priority: Trivial
> Fix For: Auto failover (HDFS-3042)
>
> Attachments: hdfs-3261.txt
>
>
> TestHASafeMode started failing on the HDFS-3042 branch after the commit of 
> HADOOP-8247. The reason is that testEnterSafeModeInANNShouldNotThrowNPE 
> restarts the active node, and then tries to make an RPC to it right after 
> restarting. The RPC picks up a cached connection to the old (restarted) NN, 
> which causes an EOFException. This was just due to a test change that was 
> made in HADOOP-8247, not due to any change made by the actual patch.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-2799) Trim fs.checkpoint.dir values

2012-04-13 Thread Hudson (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2799?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13253816#comment-13253816
 ] 

Hudson commented on HDFS-2799:
--

Integrated in Hadoop-Mapreduce-trunk-Commit #2085 (See 
[https://builds.apache.org/job/Hadoop-Mapreduce-trunk-Commit/2085/])
HDFS-2799. Trim fs.checkpoint.dir values. Contributed by Amith D K 
(Revision 1325963)

 Result = ABORTED
eli : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1325963
Files : 
* /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSImage.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestNameEditsConfigs.java


> Trim fs.checkpoint.dir values
> -
>
> Key: HDFS-2799
> URL: https://issues.apache.org/jira/browse/HDFS-2799
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: name-node
>Affects Versions: 0.23.0
>Reporter: Eli Collins
>Assignee: amith
>  Labels: noob
> Fix For: 2.0.0
>
> Attachments: HDFS-2799.patch, HDFS-2799.patch
>
>
> fs.checkpoint.dir values need to be trimmed like dfs.name.dir and 
> dfs.data.dir values so eg the following works. This currently results in the 
> directory "HADOOP_HOME/?/home/eli/hadoop/dirs3/dfs/chkpoint1" being created.
> {noformat}
>   
> fs.checkpoint.dir
>  
> /home/eli/hadoop/dirs3/dfs/chkpoint1,
> /home/eli/hadoop/dirs3/dfs/chkpoint2
>  
>   
> {noformat}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3271) src/fuse_users.c: use re-entrant versions of getpwuid, getgid, etc

2012-04-13 Thread Colin Patrick McCabe (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3271?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13253812#comment-13253812
 ] 

Colin Patrick McCabe commented on HDFS-3271:


I guess we could make FUSE respect that configuration option as well.

> src/fuse_users.c: use re-entrant versions of getpwuid, getgid, etc
> --
>
> Key: HDFS-3271
> URL: https://issues.apache.org/jira/browse/HDFS-3271
> Project: Hadoop HDFS
>  Issue Type: Improvement
>Reporter: Colin Patrick McCabe
>Assignee: Colin Patrick McCabe
>Priority: Minor
>
> Use the re-entrant versions of these functions rather than using locking

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3274) Create a new journal_edits_dir key to support journal nodes

2012-04-13 Thread Brandon Li (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3274?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13253815#comment-13253815
 ] 

Brandon Li commented on HDFS-3274:
--

The patch looks good to me. Please remove the following comments in method 
formatStorage() sine they are irrelevant now: "For now, use the namespacedirs 
and edits dir to save the journal info..."

> Create a new journal_edits_dir key to support journal nodes 
> 
>
> Key: HDFS-3274
> URL: https://issues.apache.org/jira/browse/HDFS-3274
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: ha, name-node
>Reporter: Hari Mankude
>Assignee: Hari Mankude
> Attachments: hdfs-3274.patch
>
>


--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HDFS-3134) harden edit log loader against malformed or malicious input

2012-04-13 Thread Colin Patrick McCabe (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3134?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Colin Patrick McCabe updated HDFS-3134:
---

Attachment: HDFS-3134.005.patch

new patch with the common stuff split out.  Requires HADOOP-8275

> harden edit log loader against malformed or malicious input
> ---
>
> Key: HDFS-3134
> URL: https://issues.apache.org/jira/browse/HDFS-3134
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: name-node
>Reporter: Colin Patrick McCabe
>Assignee: Colin Patrick McCabe
> Attachments: HDFS-3134.001.patch, HDFS-3134.002.patch, 
> HDFS-3134.003.patch, HDFS-3134.004.patch, HDFS-3134.005.patch
>
>
> Currently, the edit log loader does not handle bad or malicious input 
> sensibly.
> We can often cause OutOfMemory exceptions, null pointer exceptions, or other 
> unchecked exceptions to be thrown by feeding the edit log loader bad input.  
> In some environments, an out of memory error can cause the JVM process to be 
> terminated.
> It's clear that we want these exceptions to be thrown as IOException instead 
> of as unchecked exceptions.  We also want to avoid out of memory situations.
> The main task here is to put a sensible upper limit on the lengths of arrays 
> and strings we allocate on command.  The other task is to try to avoid 
> creating unchecked exceptions (by dereferencing potentially-NULL pointers, 
> for example).  Instead, we should verify ahead of time and give a more 
> sensible error message that reflects the problem with the input.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HDFS-3134) harden edit log loader against malformed or malicious input

2012-04-13 Thread Colin Patrick McCabe (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3134?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Colin Patrick McCabe updated HDFS-3134:
---

Status: Open  (was: Patch Available)

> harden edit log loader against malformed or malicious input
> ---
>
> Key: HDFS-3134
> URL: https://issues.apache.org/jira/browse/HDFS-3134
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: name-node
>Reporter: Colin Patrick McCabe
>Assignee: Colin Patrick McCabe
> Attachments: HDFS-3134.001.patch, HDFS-3134.002.patch, 
> HDFS-3134.003.patch, HDFS-3134.004.patch
>
>
> Currently, the edit log loader does not handle bad or malicious input 
> sensibly.
> We can often cause OutOfMemory exceptions, null pointer exceptions, or other 
> unchecked exceptions to be thrown by feeding the edit log loader bad input.  
> In some environments, an out of memory error can cause the JVM process to be 
> terminated.
> It's clear that we want these exceptions to be thrown as IOException instead 
> of as unchecked exceptions.  We also want to avoid out of memory situations.
> The main task here is to put a sensible upper limit on the lengths of arrays 
> and strings we allocate on command.  The other task is to try to avoid 
> creating unchecked exceptions (by dereferencing potentially-NULL pointers, 
> for example).  Instead, we should verify ahead of time and give a more 
> sensible error message that reflects the problem with the input.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-2799) Trim fs.checkpoint.dir values

2012-04-13 Thread Hudson (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2799?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13253780#comment-13253780
 ] 

Hudson commented on HDFS-2799:
--

Integrated in Hadoop-Hdfs-trunk-Commit #2144 (See 
[https://builds.apache.org/job/Hadoop-Hdfs-trunk-Commit/2144/])
HDFS-2799. Trim fs.checkpoint.dir values. Contributed by Amith D K 
(Revision 1325963)

 Result = SUCCESS
eli : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1325963
Files : 
* /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSImage.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestNameEditsConfigs.java


> Trim fs.checkpoint.dir values
> -
>
> Key: HDFS-2799
> URL: https://issues.apache.org/jira/browse/HDFS-2799
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: name-node
>Affects Versions: 0.23.0
>Reporter: Eli Collins
>Assignee: amith
>  Labels: noob
> Fix For: 2.0.0
>
> Attachments: HDFS-2799.patch, HDFS-2799.patch
>
>
> fs.checkpoint.dir values need to be trimmed like dfs.name.dir and 
> dfs.data.dir values so eg the following works. This currently results in the 
> directory "HADOOP_HOME/?/home/eli/hadoop/dirs3/dfs/chkpoint1" being created.
> {noformat}
>   
> fs.checkpoint.dir
>  
> /home/eli/hadoop/dirs3/dfs/chkpoint1,
> /home/eli/hadoop/dirs3/dfs/chkpoint2
>  
>   
> {noformat}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-2799) Trim fs.checkpoint.dir values

2012-04-13 Thread Hudson (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2799?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13253777#comment-13253777
 ] 

Hudson commented on HDFS-2799:
--

Integrated in Hadoop-Common-trunk-Commit #2071 (See 
[https://builds.apache.org/job/Hadoop-Common-trunk-Commit/2071/])
HDFS-2799. Trim fs.checkpoint.dir values. Contributed by Amith D K 
(Revision 1325963)

 Result = SUCCESS
eli : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1325963
Files : 
* /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSImage.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestNameEditsConfigs.java


> Trim fs.checkpoint.dir values
> -
>
> Key: HDFS-2799
> URL: https://issues.apache.org/jira/browse/HDFS-2799
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: name-node
>Affects Versions: 0.23.0
>Reporter: Eli Collins
>Assignee: amith
>  Labels: noob
> Fix For: 2.0.0
>
> Attachments: HDFS-2799.patch, HDFS-2799.patch
>
>
> fs.checkpoint.dir values need to be trimmed like dfs.name.dir and 
> dfs.data.dir values so eg the following works. This currently results in the 
> directory "HADOOP_HOME/?/home/eli/hadoop/dirs3/dfs/chkpoint1" being created.
> {noformat}
>   
> fs.checkpoint.dir
>  
> /home/eli/hadoop/dirs3/dfs/chkpoint1,
> /home/eli/hadoop/dirs3/dfs/chkpoint2
>  
>   
> {noformat}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HDFS-2799) Trim fs.checkpoint.dir values

2012-04-13 Thread Eli Collins (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-2799?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Eli Collins updated HDFS-2799:
--

  Resolution: Fixed
   Fix Version/s: 2.0.0
Target Version/s:   (was: 0.23.1)
Hadoop Flags: Reviewed
  Status: Resolved  (was: Patch Available)

I've committed this and merged to branch-2. Thanks Amith!

> Trim fs.checkpoint.dir values
> -
>
> Key: HDFS-2799
> URL: https://issues.apache.org/jira/browse/HDFS-2799
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: name-node
>Affects Versions: 0.23.0
>Reporter: Eli Collins
>Assignee: amith
>  Labels: noob
> Fix For: 2.0.0
>
> Attachments: HDFS-2799.patch, HDFS-2799.patch
>
>
> fs.checkpoint.dir values need to be trimmed like dfs.name.dir and 
> dfs.data.dir values so eg the following works. This currently results in the 
> directory "HADOOP_HOME/?/home/eli/hadoop/dirs3/dfs/chkpoint1" being created.
> {noformat}
>   
> fs.checkpoint.dir
>  
> /home/eli/hadoop/dirs3/dfs/chkpoint1,
> /home/eli/hadoop/dirs3/dfs/chkpoint2
>  
>   
> {noformat}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-2799) Trim fs.checkpoint.dir values

2012-04-13 Thread Eli Collins (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2799?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13253758#comment-13253758
 ] 

Eli Collins commented on HDFS-2799:
---

+1 looks good. Two small nits: the first line of the test javadoc was indented 
incorrectly and the test name had a spelling mistake ("trimm"), I'll fix those 
and commit.

> Trim fs.checkpoint.dir values
> -
>
> Key: HDFS-2799
> URL: https://issues.apache.org/jira/browse/HDFS-2799
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: name-node
>Affects Versions: 0.23.0
>Reporter: Eli Collins
>Assignee: amith
>  Labels: noob
> Attachments: HDFS-2799.patch, HDFS-2799.patch
>
>
> fs.checkpoint.dir values need to be trimmed like dfs.name.dir and 
> dfs.data.dir values so eg the following works. This currently results in the 
> directory "HADOOP_HOME/?/home/eli/hadoop/dirs3/dfs/chkpoint1" being created.
> {noformat}
>   
> fs.checkpoint.dir
>  
> /home/eli/hadoop/dirs3/dfs/chkpoint1,
> /home/eli/hadoop/dirs3/dfs/chkpoint2
>  
>   
> {noformat}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Assigned] (HDFS-2799) Trim fs.checkpoint.dir values

2012-04-13 Thread Eli Collins (Assigned) (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-2799?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Eli Collins reassigned HDFS-2799:
-

Assignee: amith

> Trim fs.checkpoint.dir values
> -
>
> Key: HDFS-2799
> URL: https://issues.apache.org/jira/browse/HDFS-2799
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: name-node
>Affects Versions: 0.23.0
>Reporter: Eli Collins
>Assignee: amith
>  Labels: noob
> Attachments: HDFS-2799.patch, HDFS-2799.patch
>
>
> fs.checkpoint.dir values need to be trimmed like dfs.name.dir and 
> dfs.data.dir values so eg the following works. This currently results in the 
> directory "HADOOP_HOME/?/home/eli/hadoop/dirs3/dfs/chkpoint1" being created.
> {noformat}
>   
> fs.checkpoint.dir
>  
> /home/eli/hadoop/dirs3/dfs/chkpoint1,
> /home/eli/hadoop/dirs3/dfs/chkpoint2
>  
>   
> {noformat}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HDFS-3274) Create a new journal_edits_dir key to support journal nodes

2012-04-13 Thread Hari Mankude (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3274?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hari Mankude updated HDFS-3274:
---

Attachment: hdfs-3274.patch

> Create a new journal_edits_dir key to support journal nodes 
> 
>
> Key: HDFS-3274
> URL: https://issues.apache.org/jira/browse/HDFS-3274
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: ha, name-node
>Reporter: Hari Mankude
>Assignee: Hari Mankude
> Attachments: hdfs-3274.patch
>
>


--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HDFS-3274) Create a new journal_edits_dir key to support journal nodes

2012-04-13 Thread Hari Mankude (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3274?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hari Mankude updated HDFS-3274:
---

Attachment: (was: hdfs-3274.patch)

> Create a new journal_edits_dir key to support journal nodes 
> 
>
> Key: HDFS-3274
> URL: https://issues.apache.org/jira/browse/HDFS-3274
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: ha, name-node
>Reporter: Hari Mankude
>Assignee: Hari Mankude
> Attachments: hdfs-3274.patch
>
>


--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3196) Implement JournalListener for writing journal to local disk

2012-04-13 Thread Tsz Wo (Nicholas), SZE (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3196?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13253755#comment-13253755
 ] 

Tsz Wo (Nicholas), SZE commented on HDFS-3196:
--

Hari's patch in HDFS-3274 has editlog initialization.  We should simply use it 
in JournalDiskWriter.

> Implement JournalListener for writing journal to local disk
> ---
>
> Key: HDFS-3196
> URL: https://issues.apache.org/jira/browse/HDFS-3196
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: ha, name-node
>Reporter: Tsz Wo (Nicholas), SZE
>Assignee: Tsz Wo (Nicholas), SZE
> Attachments: h3196_20120412.patch
>
>


--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HDFS-3274) Create a new journal_edits_dir key to support journal nodes

2012-04-13 Thread Hari Mankude (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3274?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hari Mankude updated HDFS-3274:
---

Attachment: hdfs-3274.patch

> Create a new journal_edits_dir key to support journal nodes 
> 
>
> Key: HDFS-3274
> URL: https://issues.apache.org/jira/browse/HDFS-3274
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: ha, name-node
>Reporter: Hari Mankude
>Assignee: Hari Mankude
> Attachments: hdfs-3274.patch
>
>


--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3273) Refactor BackupImage and FSEditLog

2012-04-13 Thread Hadoop QA (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3273?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13253745#comment-13253745
 ] 

Hadoop QA commented on HDFS-3273:
-

+1 overall.  Here are the results of testing the latest attachment 
  http://issues.apache.org/jira/secure/attachment/12522608/h3273_20120413.patch
  against trunk revision .

+1 @author.  The patch does not contain any @author tags.

+1 tests included.  The patch appears to include 1 new or modified test 
files.

+1 javadoc.  The javadoc tool did not generate any warning messages.

+1 javac.  The applied patch does not increase the total number of javac 
compiler warnings.

+1 eclipse:eclipse.  The patch built with eclipse:eclipse.

+1 findbugs.  The patch does not introduce any new Findbugs (version 1.3.9) 
warnings.

+1 release audit.  The applied patch does not increase the total number of 
release audit warnings.

+1 core tests.  The patch passed unit tests in .

+1 contrib tests.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-HDFS-Build/2273//testReport/
Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/2273//console

This message is automatically generated.

> Refactor BackupImage and FSEditLog
> --
>
> Key: HDFS-3273
> URL: https://issues.apache.org/jira/browse/HDFS-3273
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: ha, name-node
>Reporter: Tsz Wo (Nicholas), SZE
>Assignee: Tsz Wo (Nicholas), SZE
> Attachments: h3273_20120413.patch
>
>


--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-2704) NameNodeResouceChecker#checkAvailableResources should check for inodes

2012-04-13 Thread Aaron T. Myers (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2704?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13253742#comment-13253742
 ] 

Aaron T. Myers commented on HDFS-2704:
--

bq. having a corrupted edit log because we tried to roll them when there were 
no inodes available is a very bad thing.

Certainly true, though it seems unlikely that this would result in a truly 
corrupt edit log that the NN couldn't restart from, since running out of inodes 
would presumably result in no file being created at all, which isn't an invalid 
edit log.

bq. NameNode also does create local image/edits files upon every checkpointing 
and such - so the inode monitoring is surely relevant.

+1

> NameNodeResouceChecker#checkAvailableResources should check for inodes
> --
>
> Key: HDFS-2704
> URL: https://issues.apache.org/jira/browse/HDFS-2704
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: name-node
>Affects Versions: 0.24.0
>Reporter: Eli Collins
>Assignee: SreeHari
>
> NameNodeResouceChecker#checkAvailableResources currently just checks for free 
> space. However inodes are also a file system resource that needs to be 
> available (you can run out of inodes but have free space).

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3201) Implement Journal reader for reading transactions from local edits log or remote Journal node

2012-04-13 Thread Hadoop QA (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3201?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13253730#comment-13253730
 ] 

Hadoop QA commented on HDFS-3201:
-

-1 overall.  Here are the results of testing the latest attachment 
  
http://issues.apache.org/jira/secure/attachment/12522598/HDFS-3201.HDFS-3092.patch
  against trunk revision .

+1 @author.  The patch does not contain any @author tags.

+1 tests included.  The patch appears to include 1 new or modified test 
files.

-1 patch.  The patch command could not apply the patch.

Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/2274//console

This message is automatically generated.

> Implement Journal reader for reading transactions from local edits log or 
> remote Journal node
> -
>
> Key: HDFS-3201
> URL: https://issues.apache.org/jira/browse/HDFS-3201
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: ha, name-node
>Reporter: Brandon Li
>Assignee: Brandon Li
> Attachments: HDFS-3201.HDFS-3092.patch
>
>


--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HDFS-3201) Implement Journal reader for reading transactions from local edits log or remote Journal node

2012-04-13 Thread Brandon Li (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3201?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Brandon Li updated HDFS-3201:
-

Status: Patch Available  (was: Open)

> Implement Journal reader for reading transactions from local edits log or 
> remote Journal node
> -
>
> Key: HDFS-3201
> URL: https://issues.apache.org/jira/browse/HDFS-3201
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: ha, name-node
>Reporter: Brandon Li
>Assignee: Brandon Li
> Attachments: HDFS-3201.HDFS-3092.patch
>
>


--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HDFS-3136) Multiple SLF4J binding warning

2012-04-13 Thread Tsz Wo (Nicholas), SZE (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3136?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tsz Wo (Nicholas), SZE updated HDFS-3136:
-

Fix Version/s: 0.23.3

Merged to 0.23.

> Multiple SLF4J binding warning
> --
>
> Key: HDFS-3136
> URL: https://issues.apache.org/jira/browse/HDFS-3136
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: build
>Affects Versions: 0.23.0
>Reporter: Jason Lowe
>Assignee: Jason Lowe
> Fix For: 0.23.3, 2.0.0, 3.0.0
>
> Attachments: HDFS-3136.patch
>
>
> This is the HDFS portion of HADOOP-8005.  HDFS no longer depends upon slf4j, 
> so removing it from the assembly will eliminate the HDFS-portion of the 
> multiple SLF4J warnings.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Created] (HDFS-3274) Create a new journal_edits_dir key to support journal nodes

2012-04-13 Thread Hari Mankude (Created) (JIRA)
Create a new journal_edits_dir key to support journal nodes 


 Key: HDFS-3274
 URL: https://issues.apache.org/jira/browse/HDFS-3274
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Hari Mankude
Assignee: Hari Mankude




--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3196) Implement JournalListener for writing journal to local disk

2012-04-13 Thread Tsz Wo (Nicholas), SZE (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3196?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13253704#comment-13253704
 ] 

Tsz Wo (Nicholas), SZE commented on HDFS-3196:
--

I created HDFS-3273 for refactoring.  Then, I am going to work on editlog 
initialization here.

> Implement JournalListener for writing journal to local disk
> ---
>
> Key: HDFS-3196
> URL: https://issues.apache.org/jira/browse/HDFS-3196
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: ha, name-node
>Reporter: Tsz Wo (Nicholas), SZE
>Assignee: Tsz Wo (Nicholas), SZE
> Attachments: h3196_20120412.patch
>
>


--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HDFS-3273) Refactor BackupImage and FSEditLog

2012-04-13 Thread Tsz Wo (Nicholas), SZE (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3273?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tsz Wo (Nicholas), SZE updated HDFS-3273:
-

Status: Patch Available  (was: Open)

> Refactor BackupImage and FSEditLog
> --
>
> Key: HDFS-3273
> URL: https://issues.apache.org/jira/browse/HDFS-3273
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: ha, name-node
>Reporter: Tsz Wo (Nicholas), SZE
>Assignee: Tsz Wo (Nicholas), SZE
> Attachments: h3273_20120413.patch
>
>


--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HDFS-3273) Refactor BackupImage and FSEditLog

2012-04-13 Thread Tsz Wo (Nicholas), SZE (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3273?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tsz Wo (Nicholas), SZE updated HDFS-3273:
-

Attachment: h3273_20120413.patch

h3273_20120413.patch:
- moves some code from BackupImage to FSEditLog;
- renames JournalListener.rollLogs(..) to startLogSegment(..).

> Refactor BackupImage and FSEditLog
> --
>
> Key: HDFS-3273
> URL: https://issues.apache.org/jira/browse/HDFS-3273
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: ha, name-node
>Reporter: Tsz Wo (Nicholas), SZE
>Assignee: Tsz Wo (Nicholas), SZE
> Attachments: h3273_20120413.patch
>
>


--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Created] (HDFS-3273) Refactor BackupImage and FSEditLog

2012-04-13 Thread Tsz Wo (Nicholas), SZE (Created) (JIRA)
Refactor BackupImage and FSEditLog
--

 Key: HDFS-3273
 URL: https://issues.apache.org/jira/browse/HDFS-3273
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Tsz Wo (Nicholas), SZE
Assignee: Tsz Wo (Nicholas), SZE




--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Created] (HDFS-3272) Make it possible to state MIME type for a webhdfs OPEN operation's result

2012-04-13 Thread Steve Loughran (Created) (JIRA)
Make it possible to state MIME type for a webhdfs OPEN operation's result
-

 Key: HDFS-3272
 URL: https://issues.apache.org/jira/browse/HDFS-3272
 Project: Hadoop HDFS
  Issue Type: Improvement
Affects Versions: 1.0.1
Reporter: Steve Loughran
Priority: Minor


when you do a GET from the browser with webhdfs, you get the file, but it comes 
over as a binary as the browser doesn't know what type it is. Having a mime 
mapping table and such like would be one solution, but another is simply to add 
a {{mime}} query parameter that would provide a string to be reflected back to 
the caller as the Content-Type header in the HTTP response.

e.g.
{code}
http://ranier:50070/webhdfs/v1/results/Debounce/part-r-0.csv?op=open&mime=text/csv
 
{code}
would generate a 307 redirect to the datanode, with the 
{code}
http://dn1:50075/webhdfs/v1/results/Debounce/part-r-0.csv?op=open&mime=text/csv
 
{code}
which would then generate the result
{code}
200 OK
Content-Type:text/csv

GATE4,eb8bd736445f415e18886ba037f84829,55000,2007-01-14,14:01:54,
GATE4,ec58edcce1049fa665446dc1fa690638,8030803000,2007-01-14,13:52:31,
...
{code}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HDFS-3201) Implement Journal reader for reading transactions from local edits log or remote Journal node

2012-04-13 Thread Brandon Li (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3201?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Brandon Li updated HDFS-3201:
-

Attachment: HDFS-3201.HDFS-3092.patch

The current plan is to use http server to transfer finalized edit log segments.

This patch includes the following changes:
1. create a http server for journal node
2. have a new get image severlet to be used only by journal node
3. have a test to test file copy between journal nodes

> Implement Journal reader for reading transactions from local edits log or 
> remote Journal node
> -
>
> Key: HDFS-3201
> URL: https://issues.apache.org/jira/browse/HDFS-3201
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: ha, name-node
>Reporter: Brandon Li
>Assignee: Brandon Li
> Attachments: HDFS-3201.HDFS-3092.patch
>
>


--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3194) Continuous block scanning at DN side

2012-04-13 Thread Hadoop QA (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3194?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13253579#comment-13253579
 ] 

Hadoop QA commented on HDFS-3194:
-

-1 overall.  Here are the results of testing the latest attachment 
  http://issues.apache.org/jira/secure/attachment/12522585/HDFS-3194.patch
  against trunk revision .

+1 @author.  The patch does not contain any @author tags.

+1 tests included.  The patch appears to include 1 new or modified test 
files.

+1 javadoc.  The javadoc tool did not generate any warning messages.

+1 javac.  The applied patch does not increase the total number of javac 
compiler warnings.

+1 eclipse:eclipse.  The patch built with eclipse:eclipse.

-1 findbugs.  The patch appears to introduce 2 new Findbugs (version 1.3.9) 
warnings.

+1 release audit.  The applied patch does not increase the total number of 
release audit warnings.

-1 core tests.  The patch failed these unit tests:
  org.apache.hadoop.hdfs.TestSetrepDecreasing
  
org.apache.hadoop.hdfs.server.blockmanagement.TestOverReplicatedBlocks

+1 contrib tests.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-HDFS-Build/2272//testReport/
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HDFS-Build/2272//artifact/trunk/hadoop-hdfs-project/patchprocess/newPatchFindbugsWarningshadoop-hdfs.html
Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/2272//console

This message is automatically generated.

> Continuous block scanning at DN side
> 
>
> Key: HDFS-3194
> URL: https://issues.apache.org/jira/browse/HDFS-3194
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 3.0.0
>Reporter: suja s
>Priority: Minor
> Fix For: 3.0.0
>
> Attachments: HDFS-3194.patch
>
>
> Block scanning interval by default should be taken as 21 days(3 weeks) and 
> each block scanning should happen once in 21 days.
> Here the block is being scanned continuosly.
> 2012-04-03 10:44:47,056 INFO 
> org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner: Verification 
> succeeded for 
> BP-241703115-xx.xx.xx.55-1333086229434:blk_-2666054955039014473_1003
> 2012-04-03 10:45:02,064 INFO 
> org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner: Verification 
> succeeded for 
> BP-241703115-xx.xx.xx.55-1333086229434:blk_-2666054955039014473_1003
> 2012-04-03 10:45:17,071 INFO 
> org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner: Verification 
> succeeded for 
> BP-241703115-xx.xx.xx.55-1333086229434:blk_-2666054955039014473_1003
> 2012-04-03 10:45:32,079 INFO 
> org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner: Verification 
> succeeded for BP-241703115-xx.xx.xx.55-1333086229434:blk_-2666054955039014473

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3271) src/fuse_users.c: use re-entrant versions of getpwuid, getgid, etc

2012-04-13 Thread Todd Lipcon (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3271?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13253576#comment-13253576
 ] 

Todd Lipcon commented on HDFS-3271:
---

See HADOOP-7156. the reentrant functions are broken on some OSes so we switched 
to locking and non-reentrant calls.

> src/fuse_users.c: use re-entrant versions of getpwuid, getgid, etc
> --
>
> Key: HDFS-3271
> URL: https://issues.apache.org/jira/browse/HDFS-3271
> Project: Hadoop HDFS
>  Issue Type: Improvement
>Reporter: Colin Patrick McCabe
>Assignee: Colin Patrick McCabe
>Priority: Minor
>
> Use the re-entrant versions of these functions rather than using locking

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-2799) Trim fs.checkpoint.dir values

2012-04-13 Thread Hadoop QA (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2799?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13253544#comment-13253544
 ] 

Hadoop QA commented on HDFS-2799:
-

+1 overall.  Here are the results of testing the latest attachment 
  http://issues.apache.org/jira/secure/attachment/12522584/HDFS-2799.patch
  against trunk revision .

+1 @author.  The patch does not contain any @author tags.

+1 tests included.  The patch appears to include 1 new or modified test 
files.

+1 javadoc.  The javadoc tool did not generate any warning messages.

+1 javac.  The applied patch does not increase the total number of javac 
compiler warnings.

+1 eclipse:eclipse.  The patch built with eclipse:eclipse.

+1 findbugs.  The patch does not introduce any new Findbugs (version 1.3.9) 
warnings.

+1 release audit.  The applied patch does not increase the total number of 
release audit warnings.

+1 core tests.  The patch passed unit tests in .

+1 contrib tests.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-HDFS-Build/2271//testReport/
Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/2271//console

This message is automatically generated.

> Trim fs.checkpoint.dir values
> -
>
> Key: HDFS-2799
> URL: https://issues.apache.org/jira/browse/HDFS-2799
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: name-node
>Affects Versions: 0.23.0
>Reporter: Eli Collins
>  Labels: noob
> Attachments: HDFS-2799.patch, HDFS-2799.patch
>
>
> fs.checkpoint.dir values need to be trimmed like dfs.name.dir and 
> dfs.data.dir values so eg the following works. This currently results in the 
> directory "HADOOP_HOME/?/home/eli/hadoop/dirs3/dfs/chkpoint1" being created.
> {noformat}
>   
> fs.checkpoint.dir
>  
> /home/eli/hadoop/dirs3/dfs/chkpoint1,
> /home/eli/hadoop/dirs3/dfs/chkpoint2
>  
>   
> {noformat}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Created] (HDFS-3271) src/fuse_users.c: use re-entrant versions of getpwuid, getgid, etc

2012-04-13 Thread Colin Patrick McCabe (Created) (JIRA)
src/fuse_users.c: use re-entrant versions of getpwuid, getgid, etc
--

 Key: HDFS-3271
 URL: https://issues.apache.org/jira/browse/HDFS-3271
 Project: Hadoop HDFS
  Issue Type: Improvement
Reporter: Colin Patrick McCabe
Priority: Minor


Use the re-entrant versions of these functions rather than using locking

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Assigned] (HDFS-3271) src/fuse_users.c: use re-entrant versions of getpwuid, getgid, etc

2012-04-13 Thread Colin Patrick McCabe (Assigned) (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3271?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Colin Patrick McCabe reassigned HDFS-3271:
--

Assignee: Colin Patrick McCabe

> src/fuse_users.c: use re-entrant versions of getpwuid, getgid, etc
> --
>
> Key: HDFS-3271
> URL: https://issues.apache.org/jira/browse/HDFS-3271
> Project: Hadoop HDFS
>  Issue Type: Improvement
>Reporter: Colin Patrick McCabe
>Assignee: Colin Patrick McCabe
>Priority: Minor
>
> Use the re-entrant versions of these functions rather than using locking

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Created] (HDFS-3270) run valgrind on fuse-dfs, fix any memory leaks

2012-04-13 Thread Colin Patrick McCabe (Created) (JIRA)
run valgrind on fuse-dfs, fix any memory leaks
--

 Key: HDFS-3270
 URL: https://issues.apache.org/jira/browse/HDFS-3270
 Project: Hadoop HDFS
  Issue Type: Improvement
Reporter: Colin Patrick McCabe
Assignee: Colin Patrick McCabe
Priority: Minor


run valgrind on fuse-dfs, fix any memory leaks

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-2704) NameNodeResouceChecker#checkAvailableResources should check for inodes

2012-04-13 Thread Harsh J (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2704?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13253501#comment-13253501
 ] 

Harsh J commented on HDFS-2704:
---

NameNode also does create local image/edits files upon every checkpointing and 
such - so the inode monitoring is surely relevant.

> NameNodeResouceChecker#checkAvailableResources should check for inodes
> --
>
> Key: HDFS-2704
> URL: https://issues.apache.org/jira/browse/HDFS-2704
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: name-node
>Affects Versions: 0.24.0
>Reporter: Eli Collins
>Assignee: SreeHari
>
> NameNodeResouceChecker#checkAvailableResources currently just checks for free 
> space. However inodes are also a file system resource that needs to be 
> available (you can run out of inodes but have free space).

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3194) Continuous block scanning at DN side

2012-04-13 Thread amith (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3194?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13253497#comment-13253497
 ] 

amith commented on HDFS-3194:
-

Hi Suja,

I have done analysis on the reported defect and came up with a patch.

DataBlockScanner thread will scan for the new blocks for every 5 seconds(new 
period start), but in the scan method of BlockPoolSliceScanner finally we are 
rolling the logs(irrespective of entries in logs).

When a new period starts we are loading the number of entries from the logs and 
non scanned blocks are scanned.
Since there is log roll for every 5 sec after 10 seconds the previously scanned 
entries are lost

causing the block scan for the next period. Thus we can see the time difference 
is 15 sec for every consecutive scan.

In the provided patch I have introduced log roll only after certain number of 
entries in the logs are reached(which is similar to 20.x version HDFS 
behaviour), care is taken not to break log interface which already exist.

> Continuous block scanning at DN side
> 
>
> Key: HDFS-3194
> URL: https://issues.apache.org/jira/browse/HDFS-3194
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 3.0.0
>Reporter: suja s
>Priority: Minor
> Fix For: 3.0.0
>
> Attachments: HDFS-3194.patch
>
>
> Block scanning interval by default should be taken as 21 days(3 weeks) and 
> each block scanning should happen once in 21 days.
> Here the block is being scanned continuosly.
> 2012-04-03 10:44:47,056 INFO 
> org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner: Verification 
> succeeded for 
> BP-241703115-xx.xx.xx.55-1333086229434:blk_-2666054955039014473_1003
> 2012-04-03 10:45:02,064 INFO 
> org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner: Verification 
> succeeded for 
> BP-241703115-xx.xx.xx.55-1333086229434:blk_-2666054955039014473_1003
> 2012-04-03 10:45:17,071 INFO 
> org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner: Verification 
> succeeded for 
> BP-241703115-xx.xx.xx.55-1333086229434:blk_-2666054955039014473_1003
> 2012-04-03 10:45:32,079 INFO 
> org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner: Verification 
> succeeded for BP-241703115-xx.xx.xx.55-1333086229434:blk_-2666054955039014473

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3194) Continuous block scanning at DN side

2012-04-13 Thread amith (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3194?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13253496#comment-13253496
 ] 

amith commented on HDFS-3194:
-

Hi Suja,

I have done analysis on the reported defect and came up with a patch.

DataBlockScanner thread will scan for the new blocks for every 5 seconds(new 
period start), but in the scan method of BlockPoolSliceScanner finally we are 
rolling the logs(irrespective of entries in logs).

When a new period starts we are loading the number of entries from the logs and 
non scanned blocks are scanned.
Since there is log roll for every 5 sec after 10 seconds the previously scanned 
entries are lost

causing the block scan for the next period. Thus we can see the time difference 
is 15 sec for every consecutive scan.

In the provided patch I have introduced log roll only after certain number of 
entries in the logs are reached(which is similar to 20.x version HDFS 
behaviour), care is taken not to break log interface which already exist.

> Continuous block scanning at DN side
> 
>
> Key: HDFS-3194
> URL: https://issues.apache.org/jira/browse/HDFS-3194
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 3.0.0
>Reporter: suja s
>Priority: Minor
> Fix For: 3.0.0
>
> Attachments: HDFS-3194.patch
>
>
> Block scanning interval by default should be taken as 21 days(3 weeks) and 
> each block scanning should happen once in 21 days.
> Here the block is being scanned continuosly.
> 2012-04-03 10:44:47,056 INFO 
> org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner: Verification 
> succeeded for 
> BP-241703115-xx.xx.xx.55-1333086229434:blk_-2666054955039014473_1003
> 2012-04-03 10:45:02,064 INFO 
> org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner: Verification 
> succeeded for 
> BP-241703115-xx.xx.xx.55-1333086229434:blk_-2666054955039014473_1003
> 2012-04-03 10:45:17,071 INFO 
> org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner: Verification 
> succeeded for 
> BP-241703115-xx.xx.xx.55-1333086229434:blk_-2666054955039014473_1003
> 2012-04-03 10:45:32,079 INFO 
> org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner: Verification 
> succeeded for BP-241703115-xx.xx.xx.55-1333086229434:blk_-2666054955039014473

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-2799) Trim fs.checkpoint.dir values

2012-04-13 Thread amith (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2799?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13253484#comment-13253484
 ] 

amith commented on HDFS-2799:
-

Hi Harsh/Eli

I have provided a patch can u please review the same

> Trim fs.checkpoint.dir values
> -
>
> Key: HDFS-2799
> URL: https://issues.apache.org/jira/browse/HDFS-2799
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: name-node
>Affects Versions: 0.23.0
>Reporter: Eli Collins
>  Labels: noob
> Attachments: HDFS-2799.patch, HDFS-2799.patch
>
>
> fs.checkpoint.dir values need to be trimmed like dfs.name.dir and 
> dfs.data.dir values so eg the following works. This currently results in the 
> directory "HADOOP_HOME/?/home/eli/hadoop/dirs3/dfs/chkpoint1" being created.
> {noformat}
>   
> fs.checkpoint.dir
>  
> /home/eli/hadoop/dirs3/dfs/chkpoint1,
> /home/eli/hadoop/dirs3/dfs/chkpoint2
>  
>   
> {noformat}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HDFS-3194) Continuous block scanning at DN side

2012-04-13 Thread amith (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3194?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

amith updated HDFS-3194:


Status: Patch Available  (was: Open)

> Continuous block scanning at DN side
> 
>
> Key: HDFS-3194
> URL: https://issues.apache.org/jira/browse/HDFS-3194
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 3.0.0
>Reporter: suja s
>Priority: Minor
> Fix For: 3.0.0
>
> Attachments: HDFS-3194.patch
>
>
> Block scanning interval by default should be taken as 21 days(3 weeks) and 
> each block scanning should happen once in 21 days.
> Here the block is being scanned continuosly.
> 2012-04-03 10:44:47,056 INFO 
> org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner: Verification 
> succeeded for 
> BP-241703115-xx.xx.xx.55-1333086229434:blk_-2666054955039014473_1003
> 2012-04-03 10:45:02,064 INFO 
> org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner: Verification 
> succeeded for 
> BP-241703115-xx.xx.xx.55-1333086229434:blk_-2666054955039014473_1003
> 2012-04-03 10:45:17,071 INFO 
> org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner: Verification 
> succeeded for 
> BP-241703115-xx.xx.xx.55-1333086229434:blk_-2666054955039014473_1003
> 2012-04-03 10:45:32,079 INFO 
> org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner: Verification 
> succeeded for BP-241703115-xx.xx.xx.55-1333086229434:blk_-2666054955039014473

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HDFS-3194) Continuous block scanning at DN side

2012-04-13 Thread amith (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3194?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

amith updated HDFS-3194:


Attachment: HDFS-3194.patch

> Continuous block scanning at DN side
> 
>
> Key: HDFS-3194
> URL: https://issues.apache.org/jira/browse/HDFS-3194
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 3.0.0
>Reporter: suja s
>Priority: Minor
> Fix For: 3.0.0
>
> Attachments: HDFS-3194.patch
>
>
> Block scanning interval by default should be taken as 21 days(3 weeks) and 
> each block scanning should happen once in 21 days.
> Here the block is being scanned continuosly.
> 2012-04-03 10:44:47,056 INFO 
> org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner: Verification 
> succeeded for 
> BP-241703115-xx.xx.xx.55-1333086229434:blk_-2666054955039014473_1003
> 2012-04-03 10:45:02,064 INFO 
> org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner: Verification 
> succeeded for 
> BP-241703115-xx.xx.xx.55-1333086229434:blk_-2666054955039014473_1003
> 2012-04-03 10:45:17,071 INFO 
> org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner: Verification 
> succeeded for 
> BP-241703115-xx.xx.xx.55-1333086229434:blk_-2666054955039014473_1003
> 2012-04-03 10:45:32,079 INFO 
> org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner: Verification 
> succeeded for BP-241703115-xx.xx.xx.55-1333086229434:blk_-2666054955039014473

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HDFS-2799) Trim fs.checkpoint.dir values

2012-04-13 Thread amith (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-2799?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

amith updated HDFS-2799:


Attachment: HDFS-2799.patch

> Trim fs.checkpoint.dir values
> -
>
> Key: HDFS-2799
> URL: https://issues.apache.org/jira/browse/HDFS-2799
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: name-node
>Affects Versions: 0.23.0
>Reporter: Eli Collins
>  Labels: noob
> Attachments: HDFS-2799.patch, HDFS-2799.patch
>
>
> fs.checkpoint.dir values need to be trimmed like dfs.name.dir and 
> dfs.data.dir values so eg the following works. This currently results in the 
> directory "HADOOP_HOME/?/home/eli/hadoop/dirs3/dfs/chkpoint1" being created.
> {noformat}
>   
> fs.checkpoint.dir
>  
> /home/eli/hadoop/dirs3/dfs/chkpoint1,
> /home/eli/hadoop/dirs3/dfs/chkpoint2
>  
>   
> {noformat}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-2704) NameNodeResouceChecker#checkAvailableResources should check for inodes

2012-04-13 Thread Robert Joseph Evans (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2704?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13253392#comment-13253392
 ] 

Robert Joseph Evans commented on HDFS-2704:
---

I personally would like to see it go into safemode, just because you never know 
what in the system will start to have problems when you run out of inodes, or 
any other resource for that matter.  I would hope most standard things would 
fail gracefully, but you never know and having a corrupted edit log because we 
tried to roll them when there were no inodes available is a very bad thing.

> NameNodeResouceChecker#checkAvailableResources should check for inodes
> --
>
> Key: HDFS-2704
> URL: https://issues.apache.org/jira/browse/HDFS-2704
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: name-node
>Affects Versions: 0.24.0
>Reporter: Eli Collins
>Assignee: SreeHari
>
> NameNodeResouceChecker#checkAvailableResources currently just checks for free 
> space. However inodes are also a file system resource that needs to be 
> available (you can run out of inodes but have free space).

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3256) HDFS considers blocks under-replicated if topology script is configured with only 1 rack

2012-04-13 Thread Hudson (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3256?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13253387#comment-13253387
 ] 

Hudson commented on HDFS-3256:
--

Integrated in Hadoop-Mapreduce-trunk #1048 (See 
[https://builds.apache.org/job/Hadoop-Mapreduce-trunk/1048/])
HDFS-3256. HDFS considers blocks under-replicated if topology script is 
configured with only 1 rack. Contributed by Aaron T. Myers. (Revision 1325531)

 Result = SUCCESS
atm : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1325531
Files : 
* /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/BlockManager.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/DatanodeManager.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/blockmanagement/TestBlockManager.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/blockmanagement/TestBlocksWithNotEnoughRacks.java


> HDFS considers blocks under-replicated if topology script is configured with 
> only 1 rack
> 
>
> Key: HDFS-3256
> URL: https://issues.apache.org/jira/browse/HDFS-3256
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 2.0.0
>Reporter: Aaron T. Myers
>Assignee: Aaron T. Myers
> Fix For: 2.0.0
>
> Attachments: HDFS-3256.patch, HDFS-3256.patch, HDFS-3256.patch, 
> HDFS-3256.patch, HDFS-3256.patch, HDFS-3256.patch
>
>
> HDFS treats the mere presence of a topology script being configured as 
> evidence that there are multiple racks. If there is in fact only a single 
> rack, the NN will try to place the blocks on at least two racks, and thus 
> blocks will be considered to be under-replicated.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3255) HA DFS returns wrong token service

2012-04-13 Thread Hudson (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3255?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13253388#comment-13253388
 ] 

Hudson commented on HDFS-3255:
--

Integrated in Hadoop-Mapreduce-trunk #1048 (See 
[https://builds.apache.org/job/Hadoop-Mapreduce-trunk/1048/])
HDFS-3255. HA DFS returns wrong token service. Contributed by Daryn Sharp. 
(Revision 1325414)

 Result = SUCCESS
todd : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1325414
Files : 
* /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/DistributedFileSystem.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/ha/TestDelegationTokensWithHA.java


> HA DFS returns wrong token service
> --
>
> Key: HDFS-3255
> URL: https://issues.apache.org/jira/browse/HDFS-3255
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: ha, hdfs client
>Affects Versions: 2.0.0
>Reporter: Daryn Sharp
>Assignee: Daryn Sharp
>Priority: Critical
> Fix For: 2.0.0
>
> Attachments: HDFS-3255.patch
>
>
> {{fs.getCanonicalService()}} must be equal to 
> {{fs.getDelegationToken(renewer).getService()}}.  When HA is enabled, the DFS 
> token's service is a logical uri, but {{dfs.getCanonicalService()}} is only 
> returning the hostname of the logical uri.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3259) NameNode#initializeSharedEdits should populate shared edits dir with edit log segments

2012-04-13 Thread Hudson (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3259?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13253384#comment-13253384
 ] 

Hudson commented on HDFS-3259:
--

Integrated in Hadoop-Mapreduce-trunk #1048 (See 
[https://builds.apache.org/job/Hadoop-Mapreduce-trunk/1048/])
HDFS-3259. NameNode#initializeSharedEdits should populate shared edits dir 
with edit log segments. Contributed by Aaron T. Myers. (Revision 1325518)

 Result = SUCCESS
atm : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1325518
Files : 
* /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSEditLog.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FileJournalManager.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/NameNode.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/ha/TestInitializeSharedEdits.java


> NameNode#initializeSharedEdits should populate shared edits dir with edit log 
> segments
> --
>
> Key: HDFS-3259
> URL: https://issues.apache.org/jira/browse/HDFS-3259
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: ha, name-node
>Affects Versions: 2.0.0
>Reporter: Aaron T. Myers
>Assignee: Aaron T. Myers
> Fix For: 2.0.0
>
> Attachments: HDFS-3259.patch, HDFS-3259.patch, HDFS-3259.patch, 
> HDFS-3259.patch, HDFS-3259.patch
>
>
> Currently initializeSharedEdits formats the shared dir so that subsequent 
> edit log segments will be written there. However, it would be nice to 
> automatically populate this dir with edit log segments with transactions 
> going back to the last fsimage.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3255) HA DFS returns wrong token service

2012-04-13 Thread Hudson (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3255?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13253348#comment-13253348
 ] 

Hudson commented on HDFS-3255:
--

Integrated in Hadoop-Hdfs-trunk #1013 (See 
[https://builds.apache.org/job/Hadoop-Hdfs-trunk/1013/])
HDFS-3255. HA DFS returns wrong token service. Contributed by Daryn Sharp. 
(Revision 1325414)

 Result = FAILURE
todd : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1325414
Files : 
* /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/DistributedFileSystem.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/ha/TestDelegationTokensWithHA.java


> HA DFS returns wrong token service
> --
>
> Key: HDFS-3255
> URL: https://issues.apache.org/jira/browse/HDFS-3255
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: ha, hdfs client
>Affects Versions: 2.0.0
>Reporter: Daryn Sharp
>Assignee: Daryn Sharp
>Priority: Critical
> Fix For: 2.0.0
>
> Attachments: HDFS-3255.patch
>
>
> {{fs.getCanonicalService()}} must be equal to 
> {{fs.getDelegationToken(renewer).getService()}}.  When HA is enabled, the DFS 
> token's service is a logical uri, but {{dfs.getCanonicalService()}} is only 
> returning the hostname of the logical uri.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3256) HDFS considers blocks under-replicated if topology script is configured with only 1 rack

2012-04-13 Thread Hudson (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3256?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13253347#comment-13253347
 ] 

Hudson commented on HDFS-3256:
--

Integrated in Hadoop-Hdfs-trunk #1013 (See 
[https://builds.apache.org/job/Hadoop-Hdfs-trunk/1013/])
HDFS-3256. HDFS considers blocks under-replicated if topology script is 
configured with only 1 rack. Contributed by Aaron T. Myers. (Revision 1325531)

 Result = FAILURE
atm : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1325531
Files : 
* /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/BlockManager.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/DatanodeManager.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/blockmanagement/TestBlockManager.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/blockmanagement/TestBlocksWithNotEnoughRacks.java


> HDFS considers blocks under-replicated if topology script is configured with 
> only 1 rack
> 
>
> Key: HDFS-3256
> URL: https://issues.apache.org/jira/browse/HDFS-3256
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 2.0.0
>Reporter: Aaron T. Myers
>Assignee: Aaron T. Myers
> Fix For: 2.0.0
>
> Attachments: HDFS-3256.patch, HDFS-3256.patch, HDFS-3256.patch, 
> HDFS-3256.patch, HDFS-3256.patch, HDFS-3256.patch
>
>
> HDFS treats the mere presence of a topology script being configured as 
> evidence that there are multiple racks. If there is in fact only a single 
> rack, the NN will try to place the blocks on at least two racks, and thus 
> blocks will be considered to be under-replicated.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-3259) NameNode#initializeSharedEdits should populate shared edits dir with edit log segments

2012-04-13 Thread Hudson (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3259?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13253344#comment-13253344
 ] 

Hudson commented on HDFS-3259:
--

Integrated in Hadoop-Hdfs-trunk #1013 (See 
[https://builds.apache.org/job/Hadoop-Hdfs-trunk/1013/])
HDFS-3259. NameNode#initializeSharedEdits should populate shared edits dir 
with edit log segments. Contributed by Aaron T. Myers. (Revision 1325518)

 Result = FAILURE
atm : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1325518
Files : 
* /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSEditLog.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FileJournalManager.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/NameNode.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/ha/TestInitializeSharedEdits.java


> NameNode#initializeSharedEdits should populate shared edits dir with edit log 
> segments
> --
>
> Key: HDFS-3259
> URL: https://issues.apache.org/jira/browse/HDFS-3259
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: ha, name-node
>Affects Versions: 2.0.0
>Reporter: Aaron T. Myers
>Assignee: Aaron T. Myers
> Fix For: 2.0.0
>
> Attachments: HDFS-3259.patch, HDFS-3259.patch, HDFS-3259.patch, 
> HDFS-3259.patch, HDFS-3259.patch
>
>
> Currently initializeSharedEdits formats the shared dir so that subsequent 
> edit log segments will be written there. However, it would be nice to 
> automatically populate this dir with edit log segments with transactions 
> going back to the last fsimage.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HDFS-119) logSync() may block NameNode forever.

2012-04-13 Thread Konstantin Shvachko (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-119?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Konstantin Shvachko updated HDFS-119:
-

  Component/s: name-node
Fix Version/s: 1.1.0

> logSync() may block NameNode forever.
> -
>
> Key: HDFS-119
> URL: https://issues.apache.org/jira/browse/HDFS-119
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: name-node
>Reporter: Konstantin Shvachko
>Assignee: Suresh Srinivas
> Fix For: 0.21.0, 1.1.0
>
> Attachments: HDFS-119-branch-1.0.patch, HDFS-119-branch-1.0.patch, 
> HDFS-119.patch, HDFS-119.patch
>
>
> # {{FSEditLog.logSync()}} first waits until {{isSyncRunning}} is false and 
> then performs syncing to file streams by calling 
> {{EditLogOutputStream.flush()}}.
> If an exception is thrown after {{isSyncRunning}} is set to {{true}} all 
> threads will always wait on this condition.
> An {{IOException}} may be thrown by {{EditLogOutputStream.setReadyToFlush()}} 
> or a {{RuntimeException}} may be thrown by {{EditLogOutputStream.flush()}} or 
> by {{processIOError()}}.
> # The loop that calls {{eStream.flush()}} for multiple 
> {{EditLogOutputStream}}-s is not synchronized, which means that another 
> thread may encounter an error and modify {{editStreams}} by say calling 
> {{processIOError()}}. Then the iterating process in {{logSync()}} will break 
> with {{IndexOutOfBoundException}}.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-119) logSync() may block NameNode forever.

2012-04-13 Thread Konstantin Shvachko (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-119?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13253229#comment-13253229
 ] 

Konstantin Shvachko commented on HDFS-119:
--

Committed to branch 1.

> logSync() may block NameNode forever.
> -
>
> Key: HDFS-119
> URL: https://issues.apache.org/jira/browse/HDFS-119
> Project: Hadoop HDFS
>  Issue Type: Bug
>Reporter: Konstantin Shvachko
>Assignee: Suresh Srinivas
> Fix For: 0.21.0
>
> Attachments: HDFS-119-branch-1.0.patch, HDFS-119-branch-1.0.patch, 
> HDFS-119.patch, HDFS-119.patch
>
>
> # {{FSEditLog.logSync()}} first waits until {{isSyncRunning}} is false and 
> then performs syncing to file streams by calling 
> {{EditLogOutputStream.flush()}}.
> If an exception is thrown after {{isSyncRunning}} is set to {{true}} all 
> threads will always wait on this condition.
> An {{IOException}} may be thrown by {{EditLogOutputStream.setReadyToFlush()}} 
> or a {{RuntimeException}} may be thrown by {{EditLogOutputStream.flush()}} or 
> by {{processIOError()}}.
> # The loop that calls {{eStream.flush()}} for multiple 
> {{EditLogOutputStream}}-s is not synchronized, which means that another 
> thread may encounter an error and modify {{editStreams}} by say calling 
> {{processIOError()}}. Then the iterating process in {{logSync()}} will break 
> with {{IndexOutOfBoundException}}.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HDFS-119) logSync() may block NameNode forever.

2012-04-13 Thread Konstantin Shvachko (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-119?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Konstantin Shvachko updated HDFS-119:
-

Attachment: HDFS-119-branch-1.0.patch

Updated patch to reflect current branch state. No code changes, just line 
numbers.
Plus the code style change (indentation), suggested by Brandon.

> logSync() may block NameNode forever.
> -
>
> Key: HDFS-119
> URL: https://issues.apache.org/jira/browse/HDFS-119
> Project: Hadoop HDFS
>  Issue Type: Bug
>Reporter: Konstantin Shvachko
>Assignee: Suresh Srinivas
> Fix For: 0.21.0
>
> Attachments: HDFS-119-branch-1.0.patch, HDFS-119-branch-1.0.patch, 
> HDFS-119.patch, HDFS-119.patch
>
>
> # {{FSEditLog.logSync()}} first waits until {{isSyncRunning}} is false and 
> then performs syncing to file streams by calling 
> {{EditLogOutputStream.flush()}}.
> If an exception is thrown after {{isSyncRunning}} is set to {{true}} all 
> threads will always wait on this condition.
> An {{IOException}} may be thrown by {{EditLogOutputStream.setReadyToFlush()}} 
> or a {{RuntimeException}} may be thrown by {{EditLogOutputStream.flush()}} or 
> by {{processIOError()}}.
> # The loop that calls {{eStream.flush()}} for multiple 
> {{EditLogOutputStream}}-s is not synchronized, which means that another 
> thread may encounter an error and modify {{editStreams}} by say calling 
> {{processIOError()}}. Then the iterating process in {{logSync()}} will break 
> with {{IndexOutOfBoundException}}.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




  1   2   >