[jira] [Commented] (HDFS-2747) HA: entering safe mode after starting SBN can NPE

2012-01-12 Thread Uma Maheswara Rao G (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2747?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13185446#comment-13185446
 ] 

Uma Maheswara Rao G commented on HDFS-2747:
---

Yes, Todd. Before that, I am thinking, can we skip logSyncAll invocation from 
enterSafeMode call when numTransactions is zero in FSEditlog?
do you think any other impacts with this? Presetly i couldnot think of any 
scenario where it can impact.

> HA: entering safe mode after starting SBN can NPE
> -
>
> Key: HDFS-2747
> URL: https://issues.apache.org/jira/browse/HDFS-2747
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: ha
>Affects Versions: HA branch (HDFS-1623)
>Reporter: Eli Collins
>Assignee: Uma Maheswara Rao G
>
> Entering Safemode on the primary after while it's already in safemode after 
> the SBN is started results in an NPE: 
> {noformat}
> hadoop-0.24.0-SNAPSHOT $ ./bin/hdfs dfsadmin -safemode get
> Safe mode is ON
> hadoop-0.24.0-SNAPSHOT $ ./bin/hdfs dfsadmin -safemode enter
> safemode: java.lang.NullPointerException
> {noformat}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-2700) TestDataNodeMultipleRegistrations is failing in trunk

2012-01-12 Thread Uma Maheswara Rao G (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2700?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13185435#comment-13185435
 ] 

Uma Maheswara Rao G commented on HDFS-2700:
---

Sure, i will do that from next time. :-)

> TestDataNodeMultipleRegistrations is failing in trunk
> -
>
> Key: HDFS-2700
> URL: https://issues.apache.org/jira/browse/HDFS-2700
> Project: Hadoop HDFS
>  Issue Type: Bug
>Reporter: Uma Maheswara Rao G
>Assignee: Uma Maheswara Rao G
> Fix For: 0.24.0
>
> Attachments: HDFS-2700.patch
>
>
> TestDataNodeMultipleRegistrations  is failing from last couple of builds
> https://builds.apache.org/job/PreCommit-HDFS-Build/lastCompletedBuild/testReport/

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HDFS-2335) DataNodeCluster and NNStorage always pull fresh entropy

2012-01-12 Thread Tsz Wo (Nicholas), SZE (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-2335?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tsz Wo (Nicholas), SZE updated HDFS-2335:
-

Fix Version/s: 0.23.1
   0.24.0

Added "Fix Version/s".

> DataNodeCluster and NNStorage always pull fresh entropy
> ---
>
> Key: HDFS-2335
> URL: https://issues.apache.org/jira/browse/HDFS-2335
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: data-node, name-node
>Affects Versions: 0.23.0, 0.24.0, 1.0.0
>Reporter: Eli Collins
>Assignee: Uma Maheswara Rao G
> Fix For: 0.24.0, 0.23.1
>
> Attachments: HDFS-2335.patch, HDFS-2335.patch
>
>
> Jira for giving DataNodeCluster and NNStorage the same treatment as 
> HDFS-1835. They're not truly cryptographic uses as well. We should also 
> factor this out to a utility method, seems like the three uses are slightly 
> different, eg one uses DFSUtil.getRandom and the other creates a new Random 
> object.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-2700) TestDataNodeMultipleRegistrations is failing in trunk

2012-01-12 Thread Tsz Wo (Nicholas), SZE (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2700?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13185431#comment-13185431
 ] 

Tsz Wo (Nicholas), SZE commented on HDFS-2700:
--

{quote}
TestDataNodeMultipleRegistrations is failing from last couple of builds
https://builds.apache.org/job/PreCommit-HDFS-Build/lastCompletedBuild/testReport/
{quote}
It may be better to use the build number(s) instead of using the link with 
lastCompletedBuild next time since the "last" build keeps changing from time to 
time.  :)

> TestDataNodeMultipleRegistrations is failing in trunk
> -
>
> Key: HDFS-2700
> URL: https://issues.apache.org/jira/browse/HDFS-2700
> Project: Hadoop HDFS
>  Issue Type: Bug
>Reporter: Uma Maheswara Rao G
>Assignee: Uma Maheswara Rao G
> Fix For: 0.24.0
>
> Attachments: HDFS-2700.patch
>
>
> TestDataNodeMultipleRegistrations  is failing from last couple of builds
> https://builds.apache.org/job/PreCommit-HDFS-Build/lastCompletedBuild/testReport/

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-2731) Autopopulate standby name dirs if they're empty

2012-01-12 Thread Sanjay Radia (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2731?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13185417#comment-13185417
 ] 

Sanjay Radia commented on HDFS-2731:


Can we generalize this: on startup, a NN (PNN or SBN) should make all dirs 
consistent - i.e. copy any missing images and logs.


> Autopopulate standby name dirs if they're empty
> ---
>
> Key: HDFS-2731
> URL: https://issues.apache.org/jira/browse/HDFS-2731
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: ha
>Affects Versions: HA branch (HDFS-1623)
>Reporter: Eli Collins
>Assignee: Eli Collins
>
> To setup a SBN we currently format the primary then manually copy the name 
> dirs to the SBN. The SBN should do this automatically. Specifically, on NN 
> startup, if HA with a shared edits dir is configured and populated, if the 
> SBN has empty name dirs it should downloads the image and log from the 
> primary (as an optimization it could copy the logs from the shared dir). If 
> the other NN is still in standby then it should fail to start as it does 
> currently.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-2731) Autopopulate standby name dirs if they're empty

2012-01-12 Thread Sanjay Radia (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2731?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13185416#comment-13185416
 ] 

Sanjay Radia commented on HDFS-2731:


>if the SBN has empty name dirs it should downloads the image and log from the 
>primary (as an optimization it could copy the logs from the shared dir).
I am missing this:  both Image and Edits should be on shared dirs (ie no need 
to copy image from primary - it should be available in shared dir).



> Autopopulate standby name dirs if they're empty
> ---
>
> Key: HDFS-2731
> URL: https://issues.apache.org/jira/browse/HDFS-2731
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: ha
>Affects Versions: HA branch (HDFS-1623)
>Reporter: Eli Collins
>Assignee: Eli Collins
>
> To setup a SBN we currently format the primary then manually copy the name 
> dirs to the SBN. The SBN should do this automatically. Specifically, on NN 
> startup, if HA with a shared edits dir is configured and populated, if the 
> SBN has empty name dirs it should downloads the image and log from the 
> primary (as an optimization it could copy the logs from the shared dir). If 
> the other NN is still in standby then it should fail to start as it does 
> currently.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-2768) BackupNode stop can not close proxy connections because it is not a proxy instance.

2012-01-12 Thread Uma Maheswara Rao G (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2768?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13185408#comment-13185408
 ] 

Uma Maheswara Rao G commented on HDFS-2768:
---

Eli, I think in BPOfferSevice also we used 
DatanodeProtocolClientSideTranslatorPB directly instead of DatanodeProtocol. In 
that class many places already specialized to 
DatanodeProtocolClientSideTranslatorPB.
Do you think we can change that as well to DatanodeProtocol along with it ? 
(or) is it ok, as  **TranslatorPBs are mainly delegating calls to actual 
proxies? 

> BackupNode stop can not close proxy connections because it is not a proxy 
> instance.
> ---
>
> Key: HDFS-2768
> URL: https://issues.apache.org/jira/browse/HDFS-2768
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: name-node
>Affects Versions: 0.24.0
>Reporter: Uma Maheswara Rao G
>Assignee: Uma Maheswara Rao G
> Attachments: HDFS-2768.patch
>
>
> Observe this from BackupNode tests:
> java.lang.IllegalArgumentException: not a proxy instance
>   at java.lang.reflect.Proxy.getInvocationHandler(Unknown Source)
>   at org.apache.hadoop.ipc.RPC.stopProxy(RPC.java:557)
>   at 
> org.apache.hadoop.hdfs.server.namenode.BackupNode.stop(BackupNode.java:194)
>   at 
> org.apache.hadoop.hdfs.server.namenode.TestBackupNode.testCheckpoint(TestBackupNode.java:355)
>   at 
> org.apache.hadoop.hdfs.server.namenode.TestBackupNode.testBackupNode(TestBackupNode.java:241)
>   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>   at sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
>   at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
>   at java.lang.reflect.Method.invoke(Unknown Source)
>   at junit.framework.TestCase.runTest(TestCase.java:168)
>   at junit.framework.TestCase.runBare(TestCase.java:134)
>   at junit.framework.TestResult$1.protect(TestResult.java:110)

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-2768) BackupNode stop can not close proxy connections because it is not a proxy instance.

2012-01-12 Thread Uma Maheswara Rao G (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2768?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13185400#comment-13185400
 ] 

Uma Maheswara Rao G commented on HDFS-2768:
---

Thanks a lot Eli, for the review.
Here NamenodeProtocolTranslatorPB  are delegating the calls to it and actual 
proxy maintained in NamenodeProtocolTranslatorPB class. And also 
NamenodeProtocol is not a closeable.

infact, the otherway could be to add the instanceOf closeable checks in close 
and keep the NamenodeProtocol as it is.

> BackupNode stop can not close proxy connections because it is not a proxy 
> instance.
> ---
>
> Key: HDFS-2768
> URL: https://issues.apache.org/jira/browse/HDFS-2768
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: name-node
>Affects Versions: 0.24.0
>Reporter: Uma Maheswara Rao G
>Assignee: Uma Maheswara Rao G
> Attachments: HDFS-2768.patch
>
>
> Observe this from BackupNode tests:
> java.lang.IllegalArgumentException: not a proxy instance
>   at java.lang.reflect.Proxy.getInvocationHandler(Unknown Source)
>   at org.apache.hadoop.ipc.RPC.stopProxy(RPC.java:557)
>   at 
> org.apache.hadoop.hdfs.server.namenode.BackupNode.stop(BackupNode.java:194)
>   at 
> org.apache.hadoop.hdfs.server.namenode.TestBackupNode.testCheckpoint(TestBackupNode.java:355)
>   at 
> org.apache.hadoop.hdfs.server.namenode.TestBackupNode.testBackupNode(TestBackupNode.java:241)
>   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>   at sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
>   at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
>   at java.lang.reflect.Method.invoke(Unknown Source)
>   at junit.framework.TestCase.runTest(TestCase.java:168)
>   at junit.framework.TestCase.runBare(TestCase.java:134)
>   at junit.framework.TestResult$1.protect(TestResult.java:110)

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HDFS-2787) HDFS build failing in 0.22

2012-01-12 Thread Tsz Wo (Nicholas), SZE (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-2787?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tsz Wo (Nicholas), SZE updated HDFS-2787:
-

Summary: HDFS build failing in 0.22  (was: HDFS build failing)

Revised the summary.

> HDFS build failing in 0.22
> --
>
> Key: HDFS-2787
> URL: https://issues.apache.org/jira/browse/HDFS-2787
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 0.22.0
> Environment: Ubuntu 11.04 / Ant 1.8.1
>Reporter: Plamen Jeliazkov
>
> On a fresh install of Hadoop 0.22.0, I can get common to build fine, but when 
> I try to run "ant eclipse" or "ant jar" on hdfs I get this:
> [ivy:resolve]   
> http://repo1.maven.org/maven2/org/apache/hadoop/hadoop-common/0.22.0-SNAPSHOT/hadoop-common-0.22.0-SNAPSHOT.jar
> [ivy:resolve] ::
> [ivy:resolve] ::  UNRESOLVED DEPENDENCIES ::
> [ivy:resolve] ::
> [ivy:resolve] :: 
> org.apache.hadoop#hadoop-common;0.22.0-SNAPSHOT: not found
> [ivy:resolve] ::
> [ivy:resolve] 
> [ivy:resolve] :: USE VERBOSE OR DEBUG MESSAGE LEVEL FOR MORE DETAILS
> BUILD FAILED
> /home/hadoop/Desktop/hadoop-0.22.0/hdfs/build.xml:1814: impossible to resolve 
> dependencies:
>   resolve failed - see output for details
> From my own investigation of the repo, I can see there is no 
> "0.22.0-SNAPSHOT" folder,
> there is instead a "0.22.0" folder, and inside of it there is 
> hadoop-common-0.22.0.jar (without the -SNAPSHOT).

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-2782) HA: Support multiple shared edits dirs

2012-01-12 Thread Sanjay Radia (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2782?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13185359#comment-13185359
 ] 

Sanjay Radia commented on HDFS-2782:


2 questions:
- If you have multiple dirs within the same NFS server, are you expecting them 
to fail independently?
- For multiple NFS servers, will the standby be expected to read from any/all 
of them?

> HA: Support multiple shared edits dirs
> --
>
> Key: HDFS-2782
> URL: https://issues.apache.org/jira/browse/HDFS-2782
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: ha
>Affects Versions: HA branch (HDFS-1623)
>Reporter: Eli Collins
>Assignee: Eli Collins
>
> Supporting multiple shared dirs will improve availability (eg see HDFS-2769). 
> You may want to use multiple shared dirs on a single filer (eg for better 
> fault isolation) or because you want to use multiple filers/mounts. Per 
> HDFS-2752 (and HDFS-2735) we need to do things like use the JournalSet in 
> EditLogTailer and add tests.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-2681) Add ZK client for leader election

2012-01-12 Thread Suresh Srinivas (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2681?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13185358#comment-13185358
 ] 

Suresh Srinivas commented on HDFS-2681:
---

Some comments for the first version of the patch:
# General - How is multithreaded use for this library handled?
# ActiveStandbyElector
#* Add javadoc to the class on how to use the class. Also callback interface 
can be described in more detail on when the call back is made and perhaps some 
description of what is expected of the app. notifyError() particularly needs 
better documentation on when to expect this callback.
#* Please document the enumerations.
#* Constructor should check for null - at least for call back passed. Otherwise 
you will get null pointer exception.
#* joinElection() you may want to copy the byte[] data passed or at least 
document that the data[] must not be changed by the caller.
#* #getNewZooKeeper() seems unnecessary and can be removed. Creation of 
ZooKeeper() can be moved to createConnection() it self.
#* Make member variable that are initialized only once in the constructor final.
#* activeData could be better name for appData.
#* Please check if all the params are documented in methods. For example 
constructor is missing one of the params in the doc. Same is true with 
exceptions thrown.
#* quitElection() should not check zkClient non null, as terminateConnection 
already checks it.
#* getActiveData() - how about not throwing KeeperException? Also 
ActiveNotFoundException should wrap the exception caught from ZK.

I have not complete the review. These are some prelimiary comments


> Add ZK client for leader election
> -
>
> Key: HDFS-2681
> URL: https://issues.apache.org/jira/browse/HDFS-2681
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: ha
>Affects Versions: HA branch (HDFS-1623)
>Reporter: Suresh Srinivas
>Assignee: Bikas Saha
> Fix For: HA branch (HDFS-1623)
>
> Attachments: HDFS-2681.HDFS-1623.patch, HDFS-2681.HDFS-1623.patch, 
> Zookeeper based Leader Election and Monitoring Library.pdf
>
>
> ZKClient needs to support the following capabilities:
> # Ability to create a znode for co-ordinating leader election.
> # Ability to monitor and receive call backs when active znode status changes.
> # Ability to get information about the active node.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HDFS-2681) Add ZK client for leader election

2012-01-12 Thread Bikas Saha (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-2681?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Bikas Saha updated HDFS-2681:
-

Attachment: HDFS-2681.HDFS-1623.patch

New patch with inconsistent names fixed across comments and code.

> Add ZK client for leader election
> -
>
> Key: HDFS-2681
> URL: https://issues.apache.org/jira/browse/HDFS-2681
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: ha
>Affects Versions: HA branch (HDFS-1623)
>Reporter: Suresh Srinivas
>Assignee: Bikas Saha
> Fix For: HA branch (HDFS-1623)
>
> Attachments: HDFS-2681.HDFS-1623.patch, HDFS-2681.HDFS-1623.patch, 
> Zookeeper based Leader Election and Monitoring Library.pdf
>
>
> ZKClient needs to support the following capabilities:
> # Ability to create a znode for co-ordinating leader election.
> # Ability to monitor and receive call backs when active znode status changes.
> # Ability to get information about the active node.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-2681) Add ZK client for leader election

2012-01-12 Thread Bikas Saha (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2681?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13185312#comment-13185312
 ] 

Bikas Saha commented on HDFS-2681:
--

please ignore inconsistencies in comments while I fix them.

> Add ZK client for leader election
> -
>
> Key: HDFS-2681
> URL: https://issues.apache.org/jira/browse/HDFS-2681
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: ha
>Affects Versions: HA branch (HDFS-1623)
>Reporter: Suresh Srinivas
>Assignee: Bikas Saha
> Fix For: HA branch (HDFS-1623)
>
> Attachments: HDFS-2681.HDFS-1623.patch, Zookeeper based Leader 
> Election and Monitoring Library.pdf
>
>
> ZKClient needs to support the following capabilities:
> # Ability to create a znode for co-ordinating leader election.
> # Ability to monitor and receive call backs when active znode status changes.
> # Ability to get information about the active node.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Resolved] (HDFS-2787) HDFS build failing

2012-01-12 Thread Plamen Jeliazkov (Resolved) (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-2787?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Plamen Jeliazkov resolved HDFS-2787.


Resolution: Not A Problem

Don't trust the release mirrors for 0.22.0. Download it straight from Jenkins 
or the branch.

> HDFS build failing
> --
>
> Key: HDFS-2787
> URL: https://issues.apache.org/jira/browse/HDFS-2787
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 0.22.0
> Environment: Ubuntu 11.04 / Ant 1.8.1
>Reporter: Plamen Jeliazkov
>
> On a fresh install of Hadoop 0.22.0, I can get common to build fine, but when 
> I try to run "ant eclipse" or "ant jar" on hdfs I get this:
> [ivy:resolve]   
> http://repo1.maven.org/maven2/org/apache/hadoop/hadoop-common/0.22.0-SNAPSHOT/hadoop-common-0.22.0-SNAPSHOT.jar
> [ivy:resolve] ::
> [ivy:resolve] ::  UNRESOLVED DEPENDENCIES ::
> [ivy:resolve] ::
> [ivy:resolve] :: 
> org.apache.hadoop#hadoop-common;0.22.0-SNAPSHOT: not found
> [ivy:resolve] ::
> [ivy:resolve] 
> [ivy:resolve] :: USE VERBOSE OR DEBUG MESSAGE LEVEL FOR MORE DETAILS
> BUILD FAILED
> /home/hadoop/Desktop/hadoop-0.22.0/hdfs/build.xml:1814: impossible to resolve 
> dependencies:
>   resolve failed - see output for details
> From my own investigation of the repo, I can see there is no 
> "0.22.0-SNAPSHOT" folder,
> there is instead a "0.22.0" folder, and inside of it there is 
> hadoop-common-0.22.0.jar (without the -SNAPSHOT).

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Created] (HDFS-2787) HDFS build failing

2012-01-12 Thread Plamen Jeliazkov (Created) (JIRA)
HDFS build failing
--

 Key: HDFS-2787
 URL: https://issues.apache.org/jira/browse/HDFS-2787
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 0.22.0
 Environment: Ubuntu 11.04 / Ant 1.8.1
Reporter: Plamen Jeliazkov


On a fresh install of Hadoop 0.22.0, I can get common to build fine, but when I 
try to run "ant eclipse" or "ant jar" on hdfs I get this:

[ivy:resolve] 
http://repo1.maven.org/maven2/org/apache/hadoop/hadoop-common/0.22.0-SNAPSHOT/hadoop-common-0.22.0-SNAPSHOT.jar
[ivy:resolve]   ::
[ivy:resolve]   ::  UNRESOLVED DEPENDENCIES ::
[ivy:resolve]   ::
[ivy:resolve]   :: org.apache.hadoop#hadoop-common;0.22.0-SNAPSHOT: not 
found
[ivy:resolve]   ::
[ivy:resolve] 
[ivy:resolve] :: USE VERBOSE OR DEBUG MESSAGE LEVEL FOR MORE DETAILS

BUILD FAILED
/home/hadoop/Desktop/hadoop-0.22.0/hdfs/build.xml:1814: impossible to resolve 
dependencies:
resolve failed - see output for details


>From my own investigation of the repo, I can see there is no "0.22.0-SNAPSHOT" 
>folder,
there is instead a "0.22.0" folder, and inside of it there is 
hadoop-common-0.22.0.jar (without the -SNAPSHOT).

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-2777) When copying a file out of HDFS, modifying it, and uploading it back into HDFS, the put fails due to a CRC mismatch

2012-01-12 Thread Kihwal Lee (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2777?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13185163#comment-13185163
 ] 

Kihwal Lee commented on HDFS-2777:
--

Actually ignoreCRC is for disabling checksum verification. CRC copying needs to 
be disabled by default when -crc is not specified.


> When copying a file out of HDFS, modifying it, and uploading it back into 
> HDFS, the put fails due to a CRC mismatch
> ---
>
> Key: HDFS-2777
> URL: https://issues.apache.org/jira/browse/HDFS-2777
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: hdfs client
>Affects Versions: 0.23.1
> Environment: KR at Yahoo
>Reporter: Kevin J. Price
>
> Performing an hdfs -get on a file, modifying the file, and using hdfs -put to 
> place the file back into HDFS results in a checksum error.  It seems that the 
> problem is a .crc file being generated locally from the -get command even 
> though the -crc option was NOT specified.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-2777) When copying a file out of HDFS, modifying it, and uploading it back into HDFS, the put fails due to a CRC mismatch

2012-01-12 Thread Kihwal Lee (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2777?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13185154#comment-13185154
 ] 

Kihwal Lee commented on HDFS-2777:
--

It is due to the change in the default behavior. FsShell for some reason now 
creates CRC files (..crc) in the local file system by default. 
Unaware of existence of the .filename.crc, users can modify files stored in the 
local file system and make the CRC invalid. The default behavior needs to be 
reverted back to ingnoreCrc.

> When copying a file out of HDFS, modifying it, and uploading it back into 
> HDFS, the put fails due to a CRC mismatch
> ---
>
> Key: HDFS-2777
> URL: https://issues.apache.org/jira/browse/HDFS-2777
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: hdfs client
>Affects Versions: 0.23.1
> Environment: KR at Yahoo
>Reporter: Kevin J. Price
>
> Performing an hdfs -get on a file, modifying the file, and using hdfs -put to 
> place the file back into HDFS results in a checksum error.  It seems that the 
> problem is a .crc file being generated locally from the -get command even 
> though the -crc option was NOT specified.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HDFS-2681) Add ZK client for leader election

2012-01-12 Thread Bikas Saha (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-2681?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Bikas Saha updated HDFS-2681:
-

Attachment: HDFS-2681.HDFS-1623.patch
Zookeeper based Leader Election and Monitoring Library.pdf

Hi,
I have submitted a short design document for this library and a patch that 
implements the functionality. Both are up for review and comments.
Thanks

> Add ZK client for leader election
> -
>
> Key: HDFS-2681
> URL: https://issues.apache.org/jira/browse/HDFS-2681
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: ha
>Affects Versions: HA branch (HDFS-1623)
>Reporter: Suresh Srinivas
>Assignee: Bikas Saha
> Fix For: HA branch (HDFS-1623)
>
> Attachments: HDFS-2681.HDFS-1623.patch, Zookeeper based Leader 
> Election and Monitoring Library.pdf
>
>
> ZKClient needs to support the following capabilities:
> # Ability to create a znode for co-ordinating leader election.
> # Ability to monitor and receive call backs when active znode status changes.
> # Ability to get information about the active node.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-2742) HA: observed dataloss in replication stress test

2012-01-12 Thread Hari Mankude (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2742?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13185133#comment-13185133
 ] 

Hari Mankude commented on HDFS-2742:


This is a good idea. It keeps the log stream separate from the datanode based 
block msgs. We would need to handle "lost" msgs. For example, primary might 
have received a block finalized msg and the same msg was not delivered to 
standby. So, the block rbw msg on the standby queue will have to be purged 
after a timeout.

> HA: observed dataloss in replication stress test
> 
>
> Key: HDFS-2742
> URL: https://issues.apache.org/jira/browse/HDFS-2742
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: data-node, ha, name-node
>Affects Versions: HA branch (HDFS-1623)
>Reporter: Todd Lipcon
>Assignee: Todd Lipcon
>Priority: Blocker
> Attachments: hdfs-2742.txt, log-colorized.txt
>
>
> The replication stress test case failed over the weekend since one of the 
> replicas went missing. Still diagnosing the issue, but it seems like the 
> chain of events was something like:
> - a block report was generated on one of the nodes while the block was being 
> written - thus the block report listed the block as RBW
> - when the standby replayed this queued message, it was replayed after the 
> file was marked complete. Thus it marked this replica as corrupt
> - it asked the DN holding the corrupt replica to delete it. And, I think, 
> removed it from the block map at this time.
> - That DN then did another block report before receiving the deletion. This 
> caused it to be re-added to the block map, since it was "FINALIZED" now.
> - Replication was lowered on the file, and it counted the above replica as 
> non-corrupt, and asked for the other replicas to be deleted.
> - All replicas were lost.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-2768) BackupNode stop can not close proxy connections because it is not a proxy instance.

2012-01-12 Thread Eli Collins (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2768?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13185129#comment-13185129
 ] 

Eli Collins commented on HDFS-2768:
---

The type of backupNode and namenode should be NamenodeProtocol not 
NamenodeProtocolTranslatorPB right? Otherwise looks great.

> BackupNode stop can not close proxy connections because it is not a proxy 
> instance.
> ---
>
> Key: HDFS-2768
> URL: https://issues.apache.org/jira/browse/HDFS-2768
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: name-node
>Affects Versions: 0.24.0
>Reporter: Uma Maheswara Rao G
>Assignee: Uma Maheswara Rao G
> Attachments: HDFS-2768.patch
>
>
> Observe this from BackupNode tests:
> java.lang.IllegalArgumentException: not a proxy instance
>   at java.lang.reflect.Proxy.getInvocationHandler(Unknown Source)
>   at org.apache.hadoop.ipc.RPC.stopProxy(RPC.java:557)
>   at 
> org.apache.hadoop.hdfs.server.namenode.BackupNode.stop(BackupNode.java:194)
>   at 
> org.apache.hadoop.hdfs.server.namenode.TestBackupNode.testCheckpoint(TestBackupNode.java:355)
>   at 
> org.apache.hadoop.hdfs.server.namenode.TestBackupNode.testBackupNode(TestBackupNode.java:241)
>   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>   at sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
>   at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
>   at java.lang.reflect.Method.invoke(Unknown Source)
>   at junit.framework.TestCase.runTest(TestCase.java:168)
>   at junit.framework.TestCase.runBare(TestCase.java:134)
>   at junit.framework.TestResult$1.protect(TestResult.java:110)

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-2601) Proposal to store edits and checkpoints inside HDFS itself for namenode HA

2012-01-12 Thread Hari Mankude (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2601?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13185122#comment-13185122
 ] 

Hari Mankude commented on HDFS-2601:


Karthik,

One of things that is not explicitly mentioned is whether edit log 
writes/checkpoints are written synchronously or asynchronously to HDFS. A 
synchronous write of edit log entries might have write latency issues as 
mentioned by Sanjay. 

Asynchronous writes are also acceptable and HDFS write latency issues are not 
very critical with async writes. The disadvantage of asynchronously writing 
edit log is that the edit log in HDFS will lag the actual log and could result 
in loss of data if HDFS based image has to be used. 

However, having more than one copy of fsimage/edit logs in HDFS improves the 
recoverability of HDFS immensely (without total loss of data) in case of 
catastrophic NN failures. In fact, storing multiple checkpoints with the 
associated edit logs also will help in rollback situations if there is metadata 
corruption at NN.


> Proposal to store edits and checkpoints inside HDFS itself for namenode HA
> --
>
> Key: HDFS-2601
> URL: https://issues.apache.org/jira/browse/HDFS-2601
> Project: Hadoop HDFS
>  Issue Type: New Feature
>  Components: name-node
>Reporter: Karthik Ranganathan
>
> Would have liked to make this a "brainstorming" JIRA but couldn't find the 
> option for some reason.
> I have talked to a quite a few people about this proposal at Facebook 
> internally (HDFS folks like Hairong and Dhruba, as well as HBase folks 
> interested in this feature), and wanted to broaden the audience.
> At the core of the HA feature, we need 2 things:
> A. the secondary NN (or avatar stand-by or whatever we call it) needs to read 
> all the fsedits and fsimage data written by the primary NN
> B. Once the stand-by has taken over, the old NN should not be allowed to make 
> any edits
> The basic idea is as follows (there are some variants, we can hone in on the 
> details if we like the general approach):
> 1. The write path for fsedits and fsimage: 
> 1.1 The NN uses a dfs client to write fsedits and fsimage. These will be 
> regular hdfs files written using the write pipeline.
> 1.2 Let us say the fsimage and fsedits files are written to a well-known 
> location in the local HDFS itself (say /.META or some such location)
> 1.3 The create files and add blocks to files in this path are not written to 
> fsimage or fsedits. The location of the blocks for the files in this location 
> are known to all namenodes - primary and standby - somehow (some 
> possibilities here - write these block ids to zk or use reserved block ids or 
> write some meta-data into the blocks itself and store the blocks in a well 
> known location on all the datanodes)
> 1.4 If the replication factor on the write pipeline decreases, we close the 
> block immediately and allow NN to re-replicate to bring up the replication 
> factor. We continue writing to a new block
> 2. The read path on a NN failure
> 2.1 Since the new NN "knows" the location of the blocks for the fsedits and 
> fsimage (again the same possibilities as mentioned above), there is nothing 
> to do to determine this
> 2.2 It can read the files it needs using the HDFS client itself
> 3. Fencing - if a NN is unresponsive, a new NN takes over, old NN should not 
> be allowed to perform any action
> 3.1 Use HDFS lease recovery for the fsedits and fsimage files - the new NN 
> will close all these files baing written to by the old NN (and hence all the 
> blocks)
> 3.2 The new NN (avatar NN) will write its address into ZK to let everyone 
> know its the master
> 3.3 The new NN now gets the lease for these files and starts writing into the 
> fsedits and fsimage
> 3.4 The old NN cannot write into the file as the block it was writing to was 
> closed and it does not have the lease. If it needs to re-open these files, it 
> needs to check zk to see it is indeed the current master, if not it should 
> exit.
> 4. Misc considerations:
> 4.1 If needed, we can specify favored nodes to place the blocks for this data 
> in specific set of nodes (say we want to use a different set of RAIDed nodes, 
> etc). 
> 4.2 Since we wont record the entries for /.META in fsedits and fsimage, a 
> "hadoop dfs -ls /" command wont show the files. This is probably ok, and can 
> be fixed if not.
> 4.3 If we have 256MB block sizes, then 20GB fsimage file would need 80 block 
> ids - the NN would need only these 80 block ids to read all the fsedits data. 
> The fsimage data is even lesser. This is very tractable using a variety of 
> the techniques (the possibilities mentioned above).
> The advantage is that we are re-using the existing HDFS client (with some 
> enhancements of 

[jira] [Commented] (HDFS-2742) HA: observed dataloss in replication stress test

2012-01-12 Thread Suresh Srinivas (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2742?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13185118#comment-13185118
 ] 

Suresh Srinivas commented on HDFS-2742:
---

Here is an idea to handle this problem:
On standby for rbw replicas, we do nothing and just put then in a queue when 
received in block report. At this point two things can happen:
# Block received comes from datanode indicating finalization of block. We use 
this to just remove entry from rbw queue.
# Block received never comes, then lease recovery has to handle it. We can use 
rbw to populate the block if it does not exist, before lease recovery.

> HA: observed dataloss in replication stress test
> 
>
> Key: HDFS-2742
> URL: https://issues.apache.org/jira/browse/HDFS-2742
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: data-node, ha, name-node
>Affects Versions: HA branch (HDFS-1623)
>Reporter: Todd Lipcon
>Assignee: Todd Lipcon
>Priority: Blocker
> Attachments: hdfs-2742.txt, log-colorized.txt
>
>
> The replication stress test case failed over the weekend since one of the 
> replicas went missing. Still diagnosing the issue, but it seems like the 
> chain of events was something like:
> - a block report was generated on one of the nodes while the block was being 
> written - thus the block report listed the block as RBW
> - when the standby replayed this queued message, it was replayed after the 
> file was marked complete. Thus it marked this replica as corrupt
> - it asked the DN holding the corrupt replica to delete it. And, I think, 
> removed it from the block map at this time.
> - That DN then did another block report before receiving the deletion. This 
> caused it to be re-added to the block map, since it was "FINALIZED" now.
> - Replication was lowered on the file, and it counted the above replica as 
> non-corrupt, and asked for the other replicas to be deleted.
> - All replicas were lost.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-2747) HA: entering safe mode after starting SBN can NPE

2012-01-12 Thread Todd Lipcon (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2747?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13185076#comment-13185076
 ] 

Todd Lipcon commented on HDFS-2747:
---

Ah, that makes sense. Good sleuthing. Planning to provide a patch?

> HA: entering safe mode after starting SBN can NPE
> -
>
> Key: HDFS-2747
> URL: https://issues.apache.org/jira/browse/HDFS-2747
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: ha
>Affects Versions: HA branch (HDFS-1623)
>Reporter: Eli Collins
>Assignee: Uma Maheswara Rao G
>
> Entering Safemode on the primary after while it's already in safemode after 
> the SBN is started results in an NPE: 
> {noformat}
> hadoop-0.24.0-SNAPSHOT $ ./bin/hdfs dfsadmin -safemode get
> Safe mode is ON
> hadoop-0.24.0-SNAPSHOT $ ./bin/hdfs dfsadmin -safemode enter
> safemode: java.lang.NullPointerException
> {noformat}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HDFS-2652) Port token service changes from 205

2012-01-12 Thread Daryn Sharp (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-2652?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Daryn Sharp updated HDFS-2652:
--

Target Version/s: 0.24.0, 0.23.1  (was: 0.23.1, 0.24.0)
  Issue Type: Improvement  (was: New Feature)

> Port token service changes from 205
> ---
>
> Key: HDFS-2652
> URL: https://issues.apache.org/jira/browse/HDFS-2652
> Project: Hadoop HDFS
>  Issue Type: Improvement
>Affects Versions: 0.24.0, 0.23.1
>Reporter: Daryn Sharp
>Assignee: Daryn Sharp
>
> Need to merge the 205 token bug fixes and the feature to enable 
> hostname-based tokens.  See jiras linked to HADOOP-7808 for more details.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Created] (HDFS-2786) Fix host-based token incompatibilities in DFSUtil

2012-01-12 Thread Daryn Sharp (Created) (JIRA)
Fix host-based token incompatibilities in DFSUtil
-

 Key: HDFS-2786
 URL: https://issues.apache.org/jira/browse/HDFS-2786
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: name-node, security
Affects Versions: 0.24.0, 0.23.1
Reporter: Daryn Sharp


DFSUtil introduces new static methods that duplicate functionality in NetUtils. 
 These new methods lack the logic necessary for host-based tokens to work.  
After speaking with Suresh, the approach being taken is:
* DFSUtil.getSocketAddress will be removed.  Callers will be reverted to using 
the NetUtils version.
* DFSUtil.getDFSClient will changed to take accept a uri/host:port string 
instead of an InetSocketAddress.  The method will internal call 
NetUtils.createSocketAddr. This alleviates the callers from being required to 
call NetUtils.createSocketAddr and reduce the opportunity for error that will 
break host-based tokens.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Created] (HDFS-2785) Update webhdfs and httpfs for host-based token support

2012-01-12 Thread Daryn Sharp (Created) (JIRA)
Update webhdfs and httpfs for host-based token support
--

 Key: HDFS-2785
 URL: https://issues.apache.org/jira/browse/HDFS-2785
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: name-node, security
Affects Versions: 0.24.0, 0.23.1
Reporter: Daryn Sharp
Assignee: Robert Joseph Evans


Need to port 205 tokens into these filesystems.  Will mainly involve ensuring 
code duplicated from hftp is updated accordingly.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Created] (HDFS-2784) Update hftp and hdfs for host-based token support

2012-01-12 Thread Daryn Sharp (Created) (JIRA)
Update hftp and hdfs for host-based token support
-

 Key: HDFS-2784
 URL: https://issues.apache.org/jira/browse/HDFS-2784
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: hdfs client, name-node, security
Affects Versions: 0.24.0, 0.23.1
Reporter: Daryn Sharp
Assignee: Kihwal Lee


Need to port 205 token changes and update any new related code dealing with 
tokens in these filesystems.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Assigned] (HDFS-2754) HA: enable dfs.namenode.name.dir.restore if HA is enabled

2012-01-12 Thread Uma Maheswara Rao G (Assigned) (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-2754?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Uma Maheswara Rao G reassigned HDFS-2754:
-

Assignee: Uma Maheswara Rao G

> HA: enable dfs.namenode.name.dir.restore if HA is enabled
> -
>
> Key: HDFS-2754
> URL: https://issues.apache.org/jira/browse/HDFS-2754
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: ha, name-node
>Affects Versions: HA branch (HDFS-1623)
>Reporter: Eli Collins
>Assignee: Uma Maheswara Rao G
>
> If HA is enabled it seems like we should always try to restore failed name 
> dirs. Let's auto-enable name dir restoration if HA is enabled, at least for 
> shared edits dirs.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-2747) HA: entering safe mode after starting SBN can NPE

2012-01-12 Thread Uma Maheswara Rao G (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2747?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13185016#comment-13185016
 ] 

Uma Maheswara Rao G commented on HDFS-2747:
---

Here the reason for NullPoniterException is, initially when we are entering 
into safemode, editLogStream will be null in FSEditlog because there is no log 
segment started yet.

The reason for success of second time safemode entering is, The same synctxtid 
already visited once with previous safemode trail. So it will assume that this 
transaction was already flushed and will return. There won't be any exception 
in logSyncAll call and will enter into safemode successfully again.

java.lang.NullPointerException
at 
org.apache.hadoop.hdfs.server.namenode.FSEditLog.logSync(FSEditLog.java:524)
at 
org.apache.hadoop.hdfs.server.namenode.FSEditLog.logSyncAll(FSEditLog.java:450)
at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.enterSafeMode(FSNamesystem.java:3780)
at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.setSafeMode(FSNamesystem.java:3648)
at 
org.apache.hadoop.hdfs.server.namenode.FSImageTestUtil.enterSafemode(FSImageTestUtil.java:510)





> HA: entering safe mode after starting SBN can NPE
> -
>
> Key: HDFS-2747
> URL: https://issues.apache.org/jira/browse/HDFS-2747
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: ha
>Affects Versions: HA branch (HDFS-1623)
>Reporter: Eli Collins
>Assignee: Uma Maheswara Rao G
>
> Entering Safemode on the primary after while it's already in safemode after 
> the SBN is started results in an NPE: 
> {noformat}
> hadoop-0.24.0-SNAPSHOT $ ./bin/hdfs dfsadmin -safemode get
> Safe mode is ON
> hadoop-0.24.0-SNAPSHOT $ ./bin/hdfs dfsadmin -safemode enter
> safemode: java.lang.NullPointerException
> {noformat}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-69) Improve dfsadmin command line help

2012-01-12 Thread Hudson (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-69?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13184940#comment-13184940
 ] 

Hudson commented on HDFS-69:


Integrated in Hadoop-Mapreduce-trunk #956 (See 
[https://builds.apache.org/job/Hadoop-Mapreduce-trunk/956/])
HDFS-69. Improve the 'dfsadmin' commandline help. (harsh)

harsh : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1230398
Files : 
* /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/tools/DFSAdmin.java


> Improve dfsadmin command line help 
> ---
>
> Key: HDFS-69
> URL: https://issues.apache.org/jira/browse/HDFS-69
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 1.0.0
>Reporter: Ravi Phulari
>Assignee: Harsh J
>Priority: Minor
> Fix For: 0.23.1
>
> Attachments: HDFS-69.patch
>
>
> Enhance dfsadmin command line help informing "A quota of one forces a 
> directory to remain empty" 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-69) Improve dfsadmin command line help

2012-01-12 Thread Hudson (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-69?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13184930#comment-13184930
 ] 

Hudson commented on HDFS-69:


Integrated in Hadoop-Mapreduce-0.23-Build #158 (See 
[https://builds.apache.org/job/Hadoop-Mapreduce-0.23-Build/158/])
merge HDFS-69 (harsh)

harsh : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1230414
Files : 
* /hadoop/common/branches/branch-0.23
* /hadoop/common/branches/branch-0.23/hadoop-common-project
* /hadoop/common/branches/branch-0.23/hadoop-common-project/hadoop-auth
* /hadoop/common/branches/branch-0.23/hadoop-common-project/hadoop-common
* 
/hadoop/common/branches/branch-0.23/hadoop-common-project/hadoop-common/CHANGES.txt
* 
/hadoop/common/branches/branch-0.23/hadoop-common-project/hadoop-common/src/main/docs
* 
/hadoop/common/branches/branch-0.23/hadoop-common-project/hadoop-common/src/main/java
* 
/hadoop/common/branches/branch-0.23/hadoop-common-project/hadoop-common/src/test/core
* /hadoop/common/branches/branch-0.23/hadoop-hdfs-project/hadoop-hdfs
* 
/hadoop/common/branches/branch-0.23/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* 
/hadoop/common/branches/branch-0.23/hadoop-hdfs-project/hadoop-hdfs/src/main/java
* 
/hadoop/common/branches/branch-0.23/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/tools/DFSAdmin.java
* 
/hadoop/common/branches/branch-0.23/hadoop-hdfs-project/hadoop-hdfs/src/main/native
* 
/hadoop/common/branches/branch-0.23/hadoop-hdfs-project/hadoop-hdfs/src/main/webapps/datanode
* 
/hadoop/common/branches/branch-0.23/hadoop-hdfs-project/hadoop-hdfs/src/main/webapps/hdfs
* 
/hadoop/common/branches/branch-0.23/hadoop-hdfs-project/hadoop-hdfs/src/main/webapps/secondary
* 
/hadoop/common/branches/branch-0.23/hadoop-hdfs-project/hadoop-hdfs/src/test/hdfs
* /hadoop/common/branches/branch-0.23/hadoop-mapreduce-project
* /hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/.gitignore
* /hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/CHANGES.txt
* /hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/bin
* /hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/conf
* 
/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/resources/mapred-default.xml
* 
/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/hadoop-mapreduce-examples
* 
/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/hadoop-yarn/hadoop-yarn-site
* 
/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/hadoop-yarn/hadoop-yarn-site/src/site/apt
* /hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/src/c++
* /hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/src/contrib
* 
/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/src/contrib/block_forensics
* 
/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/src/contrib/build-contrib.xml
* 
/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/src/contrib/build.xml
* 
/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/src/contrib/data_join
* 
/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/src/contrib/eclipse-plugin
* /hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/src/contrib/index
* 
/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/src/contrib/vaidya
* /hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/src/examples
* /hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/src/java
* /hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/src/test/mapred
* 
/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/src/test/mapred/org/apache/hadoop/fs
* 
/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/src/test/mapred/org/apache/hadoop/hdfs
* 
/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/src/test/mapred/org/apache/hadoop/io/FileBench.java
* 
/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/src/test/mapred/org/apache/hadoop/io/TestSequenceFileMergeProgress.java
* 
/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/src/test/mapred/org/apache/hadoop/ipc
* 
/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/src/test/mapred/org/apache/hadoop/security/authorize/TestServiceLevelAuthorization.java
* 
/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/src/test/mapred/org/apache/hadoop/test/MapredTestDriver.java
* /hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/src/webapps/job
* /hadoop/common/branches/branch-0.23/hadoop-project
* /hadoop/common/branches/branch-0.23/hadoop-project/src/site


> Improve dfsadmin command line help 
> ---
>
> Key: HDFS-69
> URL: https://issues.apache.org/jira/browse/HDFS-69
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 1.0.0
>Rep

[jira] [Commented] (HDFS-69) Improve dfsadmin command line help

2012-01-12 Thread Hudson (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-69?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13184907#comment-13184907
 ] 

Hudson commented on HDFS-69:


Integrated in Hadoop-Hdfs-0.23-Build #136 (See 
[https://builds.apache.org/job/Hadoop-Hdfs-0.23-Build/136/])
merge HDFS-69 (harsh)

harsh : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1230414
Files : 
* /hadoop/common/branches/branch-0.23
* /hadoop/common/branches/branch-0.23/hadoop-common-project
* /hadoop/common/branches/branch-0.23/hadoop-common-project/hadoop-auth
* /hadoop/common/branches/branch-0.23/hadoop-common-project/hadoop-common
* 
/hadoop/common/branches/branch-0.23/hadoop-common-project/hadoop-common/CHANGES.txt
* 
/hadoop/common/branches/branch-0.23/hadoop-common-project/hadoop-common/src/main/docs
* 
/hadoop/common/branches/branch-0.23/hadoop-common-project/hadoop-common/src/main/java
* 
/hadoop/common/branches/branch-0.23/hadoop-common-project/hadoop-common/src/test/core
* /hadoop/common/branches/branch-0.23/hadoop-hdfs-project/hadoop-hdfs
* 
/hadoop/common/branches/branch-0.23/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* 
/hadoop/common/branches/branch-0.23/hadoop-hdfs-project/hadoop-hdfs/src/main/java
* 
/hadoop/common/branches/branch-0.23/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/tools/DFSAdmin.java
* 
/hadoop/common/branches/branch-0.23/hadoop-hdfs-project/hadoop-hdfs/src/main/native
* 
/hadoop/common/branches/branch-0.23/hadoop-hdfs-project/hadoop-hdfs/src/main/webapps/datanode
* 
/hadoop/common/branches/branch-0.23/hadoop-hdfs-project/hadoop-hdfs/src/main/webapps/hdfs
* 
/hadoop/common/branches/branch-0.23/hadoop-hdfs-project/hadoop-hdfs/src/main/webapps/secondary
* 
/hadoop/common/branches/branch-0.23/hadoop-hdfs-project/hadoop-hdfs/src/test/hdfs
* /hadoop/common/branches/branch-0.23/hadoop-mapreduce-project
* /hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/.gitignore
* /hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/CHANGES.txt
* /hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/bin
* /hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/conf
* 
/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/resources/mapred-default.xml
* 
/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/hadoop-mapreduce-examples
* 
/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/hadoop-yarn/hadoop-yarn-site
* 
/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/hadoop-yarn/hadoop-yarn-site/src/site/apt
* /hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/src/c++
* /hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/src/contrib
* 
/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/src/contrib/block_forensics
* 
/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/src/contrib/build-contrib.xml
* 
/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/src/contrib/build.xml
* 
/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/src/contrib/data_join
* 
/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/src/contrib/eclipse-plugin
* /hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/src/contrib/index
* 
/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/src/contrib/vaidya
* /hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/src/examples
* /hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/src/java
* /hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/src/test/mapred
* 
/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/src/test/mapred/org/apache/hadoop/fs
* 
/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/src/test/mapred/org/apache/hadoop/hdfs
* 
/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/src/test/mapred/org/apache/hadoop/io/FileBench.java
* 
/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/src/test/mapred/org/apache/hadoop/io/TestSequenceFileMergeProgress.java
* 
/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/src/test/mapred/org/apache/hadoop/ipc
* 
/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/src/test/mapred/org/apache/hadoop/security/authorize/TestServiceLevelAuthorization.java
* 
/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/src/test/mapred/org/apache/hadoop/test/MapredTestDriver.java
* /hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/src/webapps/job
* /hadoop/common/branches/branch-0.23/hadoop-project
* /hadoop/common/branches/branch-0.23/hadoop-project/src/site


> Improve dfsadmin command line help 
> ---
>
> Key: HDFS-69
> URL: https://issues.apache.org/jira/browse/HDFS-69
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 1.0.0
>Reporter: Rav

[jira] [Commented] (HDFS-69) Improve dfsadmin command line help

2012-01-12 Thread Hudson (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-69?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13184894#comment-13184894
 ] 

Hudson commented on HDFS-69:


Integrated in Hadoop-Hdfs-trunk #923 (See 
[https://builds.apache.org/job/Hadoop-Hdfs-trunk/923/])
HDFS-69. Improve the 'dfsadmin' commandline help. (harsh)

harsh : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1230398
Files : 
* /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/tools/DFSAdmin.java


> Improve dfsadmin command line help 
> ---
>
> Key: HDFS-69
> URL: https://issues.apache.org/jira/browse/HDFS-69
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 1.0.0
>Reporter: Ravi Phulari
>Assignee: Harsh J
>Priority: Minor
> Fix For: 0.23.1
>
> Attachments: HDFS-69.patch
>
>
> Enhance dfsadmin command line help informing "A quota of one forces a 
> directory to remain empty" 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira