[jira] [Created] (HADOOP-12351) Can't run test-patch with start-build-env.sh

2015-08-24 Thread Jakob Homan (JIRA)
Jakob Homan created HADOOP-12351:


 Summary: Can't run test-patch with start-build-env.sh
 Key: HADOOP-12351
 URL: https://issues.apache.org/jira/browse/HADOOP-12351
 Project: Hadoop Common
  Issue Type: Bug
  Components: test, yetus
Reporter: Jakob Homan
Priority: Minor


The Docker instance started by start-build-env.sh drops the user into a root 
directory wherein other directories (say ~/download) are not accessible, so one 
cannot pull patches to test into it.  The dev-support/test-patch.sh script 
requires a clean environment with no extra files, such as the patch to test.  
Between these two restrictions, one can't use the docker env to run test-patch. 
 We should either -v a volume where patches can be stashed, or allow 
test-patch.sh to ignore patch files' existences.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Resolved] (HADOOP-12068) Remove @author tags

2015-06-05 Thread Jakob Homan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-12068?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jakob Homan resolved HADOOP-12068.
--
Resolution: Invalid

Wrong project.

> Remove @author tags
> ---
>
> Key: HADOOP-12068
> URL: https://issues.apache.org/jira/browse/HADOOP-12068
> Project: Hadoop Common
>  Issue Type: Bug
>        Reporter: Jakob Homan
>
> Apache generally frowns on @author tags (as they imply ownership of the code 
> by a single individual, see 
> https://blogs.oracle.com/ahe/entry/coding_conventions_and_attribution [best 
> asf link is not working, annoyingly] and 
> https://cwiki.apache.org/confluence/display/AVRO/How+To+Contribute as an 
> example)



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HADOOP-12068) Remove @author tags

2015-06-05 Thread Jakob Homan (JIRA)
Jakob Homan created HADOOP-12068:


 Summary: Remove @author tags
 Key: HADOOP-12068
 URL: https://issues.apache.org/jira/browse/HADOOP-12068
 Project: Hadoop Common
  Issue Type: Bug
Reporter: Jakob Homan


Apache generally frowns on @author tags (as they imply ownership of the code by 
a single individual, see 
https://blogs.oracle.com/ahe/entry/coding_conventions_and_attribution [best asf 
link is not working, annoyingly] and 
https://cwiki.apache.org/confluence/display/AVRO/How+To+Contribute as an 
example)



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HADOOP-12016) Typo in FileSystem:: listStatusIterator

2015-05-21 Thread Jakob Homan (JIRA)
Jakob Homan created HADOOP-12016:


 Summary: Typo in FileSystem:: listStatusIterator
 Key: HADOOP-12016
 URL: https://issues.apache.org/jira/browse/HADOOP-12016
 Project: Hadoop Common
  Issue Type: Bug
  Components: fs
Reporter: Jakob Homan
Priority: Trivial


{code}  public RemoteIterator listStatusIterator(final Path p)
  throws FileNotFoundException, IOException {
return new RemoteIterator() {
//...
  throw new NoSuchElementException("No more entry in " + p);
}
return stats[i++];
  }{code}
Should be 'no more entries'



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HADOOP-12009) FileSystemContractBaseTest:testListStatus should not assume listStatus returns sorted results

2015-05-20 Thread Jakob Homan (JIRA)
Jakob Homan created HADOOP-12009:


 Summary: FileSystemContractBaseTest:testListStatus should not 
assume listStatus returns sorted results
 Key: HADOOP-12009
 URL: https://issues.apache.org/jira/browse/HADOOP-12009
 Project: Hadoop Common
  Issue Type: Improvement
  Components: test
Reporter: Jakob Homan
Priority: Minor


FileSystem.listStatus does not guarantee that implementations will return 
sorted entries:
{code}  /**
   * List the statuses of the files/directories in the given path if the path is
   * a directory.
   * 
   * @param f given path
   * @return the statuses of the files/directories in the given patch
   * @throws FileNotFoundException when the path does not exist;
   * IOException see specific implementation
   */
  public abstract FileStatus[] listStatus(Path f) throws FileNotFoundException, 
 IOException;{code}
However, FileSystemContractBaseTest, expects the elements to come back sorted:
{code}Path[] testDirs = { path("/test/hadoop/a"),
path("/test/hadoop/b"),
path("/test/hadoop/c/1"), };
   
// ...

paths = fs.listStatus(path("/test/hadoop"));
assertEquals(3, paths.length);
assertEquals(path("/test/hadoop/a"), paths[0].getPath());
assertEquals(path("/test/hadoop/b"), paths[1].getPath());
assertEquals(path("/test/hadoop/c"), paths[2].getPath());{code}
We should pass this test as long as all the paths are there, regardless of 
their ordering.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


Re: [VOTE] Change by-laws on release votes: 5 days instead of 7

2014-06-24 Thread Jakob Homan
+1 (binding)


On Tue, Jun 24, 2014 at 10:33 AM, Zhijie Shen  wrote:

> +1 (non-binding)
>
>
> On Wed, Jun 25, 2014 at 1:26 AM, Aaron T. Myers  wrote:
>
> > +1 (binding)
> >
> > --
> > Aaron T. Myers
> > Software Engineer, Cloudera
> >
> >
> > On Tue, Jun 24, 2014 at 1:53 AM, Arun C Murthy 
> > wrote:
> >
> > > Folks,
> > >
> > >  As discussed, I'd like to call a vote on changing our by-laws to
> change
> > > release votes from 7 days to 5.
> > >
> > >  I've attached the change to by-laws I'm proposing.
> > >
> > >  Please vote, the vote will the usual period of 7 days.
> > >
> > > thanks,
> > > Arun
> > >
> > > 
> > >
> > > [main]$ svn diff
> > > Index: author/src/documentation/content/xdocs/bylaws.xml
> > > ===
> > > --- author/src/documentation/content/xdocs/bylaws.xml   (revision
> > 1605015)
> > > +++ author/src/documentation/content/xdocs/bylaws.xml   (working copy)
> > > @@ -344,7 +344,16 @@
> > >  Votes are open for a period of 7 days to allow all active
> > >  voters time to consider the vote. Votes relating to code
> > >  changes are not subject to a strict timetable but should be
> > > -made as timely as possible.
> > > +made as timely as possible.
> > > +
> > > + 
> > > +  Product Release - Vote Timeframe
> > > +   Release votes, alone, run for a period of 5 days. All
> > other
> > > + votes are subject to the above timeframe of 7 days.
> > > + 
> > > +   
> > > +   
> > > +
> > > 
> > > 
> > >  
> > > --
> > > CONFIDENTIALITY NOTICE
> > > NOTICE: This message is intended for the use of the individual or
> entity
> > to
> > > which it is addressed and may contain information that is confidential,
> > > privileged and exempt from disclosure under applicable law. If the
> reader
> > > of this message is not the intended recipient, you are hereby notified
> > that
> > > any printing, copying, dissemination, distribution, disclosure or
> > > forwarding of this communication is strictly prohibited. If you have
> > > received this communication in error, please contact the sender
> > immediately
> > > and delete it from your system. Thank You.
> > >
> >
>
>
>
> --
> Zhijie Shen
> Hortonworks Inc.
> http://hortonworks.com/
>
> --
> CONFIDENTIALITY NOTICE
> NOTICE: This message is intended for the use of the individual or entity to
> which it is addressed and may contain information that is confidential,
> privileged and exempt from disclosure under applicable law. If the reader
> of this message is not the intended recipient, you are hereby notified that
> any printing, copying, dissemination, distribution, disclosure or
> forwarding of this communication is strictly prohibited. If you have
> received this communication in error, please contact the sender immediately
> and delete it from your system. Thank You.
>


HADOOP JIRA near full of contributors, error message not helpful

2014-02-14 Thread Jakob Homan
FYI, ran into trouble adding a new contributor in the HADOOP common project
(https://issues.apache.org/jira/browse/INFRA-7293).  Looks like we've hit
the limit for contributors, but the error message is not very helpful.
Suresh has cleaned up some, but would be good to keep in mind in the future
when adding people.

-Jakob


[jira] [Resolved] (HADOOP-8322) Log entry for successful auth has misspelling

2012-05-03 Thread Jakob Homan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-8322?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jakob Homan resolved HADOOP-8322.
-

Resolution: Duplicate

> Log entry for successful auth has misspelling
> -
>
> Key: HADOOP-8322
> URL: https://issues.apache.org/jira/browse/HADOOP-8322
> Project: Hadoop Common
>  Issue Type: Improvement
>        Reporter: Jakob Homan
>Assignee: Boris Shkolnik
>Priority: Minor
>  Labels: newbie
>
> Server.java:
> {code}  private static final String AUTH_SUCCESSFULL_FOR = "Auth successfull 
> for "; {code}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Created] (HADOOP-8322) Log entry for successful auth has misspelling

2012-04-26 Thread Jakob Homan (JIRA)
Jakob Homan created HADOOP-8322:
---

 Summary: Log entry for successful auth has misspelling
 Key: HADOOP-8322
 URL: https://issues.apache.org/jira/browse/HADOOP-8322
 Project: Hadoop Common
  Issue Type: Improvement
Reporter: Jakob Homan
Priority: Minor


Server.java:
{code}  private static final String AUTH_SUCCESSFULL_FOR = "Auth successfull 
for "; {code}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Created] (HADOOP-8229) DistCp doesn't handle non-existent paths correctly

2012-03-29 Thread Jakob Homan (Created) (JIRA)
DistCp doesn't handle non-existent paths correctly
--

 Key: HADOOP-8229
 URL: https://issues.apache.org/jira/browse/HADOOP-8229
 Project: Hadoop Common
  Issue Type: Bug
Affects Versions: 1.0.0, 0.20.2
Reporter: Jakob Homan


Assume /user/jhoman/blork doesn't exist:
{noformat}[tardis]$ hadoop distcp 'hftp://nn:50070/user/jhoman/blork' /tmp/plork
12/03/29 22:04:33 INFO tools.DistCp: 
srcPaths=[hftp://nn:50070/user/jhoman/blork]
12/03/29 22:04:33 INFO tools.DistCp: destPath=/tmp/plork
[Fatal Error] :1:173: XML document structures must start and end within the 
same entity.
With failures, global counters are inaccurate; consider running with -i
Copy failed: java.io.IOException: invalid xml directory content
at 
org.apache.hadoop.hdfs.HftpFileSystem$LsParser.fetchList(HftpFileSystem.java:427)
at 
org.apache.hadoop.hdfs.HftpFileSystem$LsParser.getFileStatus(HftpFileSystem.java:432)
at 
org.apache.hadoop.hdfs.HftpFileSystem.getFileStatus(HftpFileSystem.java:461)
at org.apache.hadoop.fs.FileSystem.exists(FileSystem.java:768)
at org.apache.hadoop.tools.DistCp.checkSrcPath(DistCp.java:636)
at org.apache.hadoop.tools.DistCp.copy(DistCp.java:656)
at org.apache.hadoop.tools.DistCp.run(DistCp.java:881)
at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:65)
at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:79)
at org.apache.hadoop.tools.DistCp.main(DistCp.java:908)
Caused by: org.xml.sax.SAXParseException: XML document structures must start 
and end within the same entity.
at 
com.sun.org.apache.xerces.internal.parsers.AbstractSAXParser.parse(AbstractSAXParser.java:1231)
at 
org.apache.hadoop.hdfs.HftpFileSystem$LsParser.fetchList(HftpFileSystem.java:421)
... 9 more{noformat}
This is because the ListPathsServlet hits an NPE when it calls 
nnproxy.getFileInfo(path); on the non-existent path and just bails, leaving the 
resulting XML unformed.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Created] (HADOOP-7927) Can't build packages 205+ on OSX

2011-12-15 Thread Jakob Homan (Created) (JIRA)
Can't build packages 205+ on OSX


 Key: HADOOP-7927
 URL: https://issues.apache.org/jira/browse/HADOOP-7927
 Project: Hadoop Common
  Issue Type: Bug
Affects Versions: 0.20.205.0, 1.0.0
Reporter: Jakob Homan


Currently the ant build script tries to reference the native directories, which 
are not built on OSX, breaking the build:
{noformat}bin-package:
[mkdir] Created dir: 
/Users/jhoman/repos/hadoop-common/build/hadoop-0.20.205.1
[mkdir] Created dir: 
/Users/jhoman/repos/hadoop-common/build/hadoop-0.20.205.1/bin
[mkdir] Created dir: 
/Users/jhoman/repos/hadoop-common/build/hadoop-0.20.205.1/etc/hadoop
[mkdir] Created dir: 
/Users/jhoman/repos/hadoop-common/build/hadoop-0.20.205.1/lib
[mkdir] Created dir: 
/Users/jhoman/repos/hadoop-common/build/hadoop-0.20.205.1/libexec
[mkdir] Created dir: 
/Users/jhoman/repos/hadoop-common/build/hadoop-0.20.205.1/sbin
[mkdir] Created dir: 
/Users/jhoman/repos/hadoop-common/build/hadoop-0.20.205.1/share/hadoop/contrib
[mkdir] Created dir: 
/Users/jhoman/repos/hadoop-common/build/hadoop-0.20.205.1/share/hadoop/webapps
[mkdir] Created dir: 
/Users/jhoman/repos/hadoop-common/build/hadoop-0.20.205.1/share/hadoop/templates/conf
 [copy] Copying 11 files to 
/Users/jhoman/repos/hadoop-common/build/hadoop-0.20.205.1/share/hadoop/templates/conf
 [copy] Copying 39 files to 
/Users/jhoman/repos/hadoop-common/build/hadoop-0.20.205.1/share/hadoop/lib
 [copy] Copying 15 files to 
/Users/jhoman/repos/hadoop-common/build/hadoop-0.20.205.1/share/hadoop/lib

BUILD FAILED
/Users/jhoman/repos/hadoop-common/build.xml:1611: 
/Users/jhoman/repos/hadoop-common/build/hadoop-0.20.205.1/native does not exist.
{noformat}

Once one fixes this, one discovers the build is also trying to build the linux 
task controller, regardless of whether not the native flag is set, which also 
fails.  

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




Re: [VOTE] Should we release 0.20.204.0rc2?

2011-08-18 Thread Jakob Homan
>> This vote is still running with no votes other than mine.

The vote was started 9 days ago and, if it kept running after Allen's
vote, it would have ended three days ago with the result of 1-0 to
release.  As such, let's call this release 2Owen3.


[jira] [Resolved] (HADOOP-7458) Namenode not get started! FSNamesystem initialization failed. java.io.FileNotFoundException

2011-07-14 Thread Jakob Homan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-7458?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jakob Homan resolved HADOOP-7458.
-

Resolution: Invalid

This appears to be an issue of file corruption.  The best forum for resolving 
it is the mailing list, rather than the issue tracker.  Please write to the 
hdfs users' list.  Closing.

> Namenode not get started! FSNamesystem initialization failed. 
> java.io.FileNotFoundException
> ---
>
> Key: HADOOP-7458
> URL: https://issues.apache.org/jira/browse/HADOOP-7458
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs
>Affects Versions: 0.20.2
> Environment: CentOS release 5.5 (Final), 18 node Cluster 
>Reporter: Sakthivel Murugasamy
>Priority: Blocker
>  Labels: hadoop
>   Original Estimate: 48h
>  Remaining Estimate: 48h
>
> 2011-07-13 12:04:12,967 ERROR 
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: FSNamesystem 
> initialization failed.
> java.io.FileNotFoundException: File does not exist: 
> /opt/data/tmp/mapred/system/job_201107041958_0120/j^@^@^@^@^@^@
> at 
> org.apache.hadoop.hdfs.server.namenode.FSDirectory.unprotectedSetPermission(FSDirectory.java:544)
> at 
> org.apache.hadoop.hdfs.server.namenode.FSEditLog.loadFSEdits(FSEditLog.java:724)
> at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.loadFSEdits(FSImage.java:992)
> at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.loadFSImage(FSImage.java:812)
> at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:364)
> at 
> org.apache.hadoop.hdfs.server.namenode.FSDirectory.loadFSImage(FSDirectory.java:87)
> at 
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.initialize(FSNamesystem.java:311)
> at 
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.(FSNamesystem.java:292)
> at 
> org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:201)
> at 
> org.apache.hadoop.hdfs.server.namenode.NameNode.(NameNode.java:279)
> at 
> org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:956)
> at 
> org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:965)
> 2011-07-13 12:04:13,006 ERROR 
> org.apache.hadoop.hdfs.server.namenode.NameNode: 
> java.io.FileNotFoundException: File does not exist: 
> /opt/data/tmp/mapred/system/job_201107041958_0120/j^@^@^@^@^@^@
> at 
> org.apache.hadoop.hdfs.server.namenode.FSDirectory.unprotectedSetPermission(FSDirectory.java:544)
> at 
> org.apache.hadoop.hdfs.server.namenode.FSEditLog.loadFSEdits(FSEditLog.java:724)
> at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.loadFSEdits(FSImage.java:992)
> at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.loadFSImage(FSImage.java:812)
> at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:364)
> at 
> org.apache.hadoop.hdfs.server.namenode.FSDirectory.loadFSImage(FSDirectory.java:87)
> at 
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.initialize(FSNamesystem.java:311)
> at 
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.(FSNamesystem.java:292)
> at 
> org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:201)
> at 
> org.apache.hadoop.hdfs.server.namenode.NameNode.(NameNode.java:279)
> at 
> org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:956)
> at 
> org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:965)
> In the path /opt/data/tmp/mapred, "system/" folder itself is not available

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Created] (HADOOP-7457) Remove out-of-date Chinese language documentation

2011-07-11 Thread Jakob Homan (JIRA)
Remove out-of-date Chinese language documentation
-

 Key: HADOOP-7457
 URL: https://issues.apache.org/jira/browse/HADOOP-7457
 Project: Hadoop Common
  Issue Type: Improvement
  Components: documentation
Reporter: Jakob Homan


The Chinese language documentation haven't been updated (other than copyright 
years and svn moves) since their original contribution several years ago.  
Worse than no docs is out-of-date, wrong docs.  We should delete them from the 
source tree.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




Re: [VOTE] Release candidate 0.20.203.0-rc0

2011-05-03 Thread Jakob Homan
Tested the RC on a single node cluster, kicked the tires.  Looks good.
 +1 on its release.

Regardless of how the RC got here, we only get benefit from releasing
it.  It represents a huge chunk of work from our contributors,
provides needed features for our users and moves us one step closer to
making regular releases again.

-Jakob


On Tue, May 3, 2011 at 11:33 AM, Konstantin Boudnik  wrote:
> Yup, exactly right - it has been reverted in the trunk as well. Thanks
> for digging this up, Koji!
>
> On Tue, May 3, 2011 at 11:22, Koji Noguchi  wrote:
 except
 HADOOP-6386 and HADOOP-6428.
>>> causes a rolling port side effect in TT
>>>
>> I remember bugging Cos and Rob to revert HADOOP-6386.
>> https://issues.apache.org/jira/browse/HADOOP-6760?focusedCommentId=12867342&;
>> page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#commen
>> t-12867342
>>
>> Koji
>>
>> On 5/2/11 9:43 PM, "Konstantin Boudnik"  wrote:
>>
>>> On Mon, May 2, 2011 at 16:56, Arun C Murthy  wrote:
>>>

 On May 2, 2011, at 3:01 PM, Arun C Murthy wrote:


> On May 2, 2011, at 12:31 PM, Tom White wrote:
>
>> I just did a quick search, and these are the JIRAs that are in 0.20.2
>> but appear not to be in 0.20.203.0.
>>
>
> Thanks Tom.
>
> I did a quick analysis:
>
> # Remaining for 0.20.203
>  * HADOOP-5611
>  * HADOOP-5612
>  * HADOOP-5623
>  * HDFS-596
>  * HDFS-723
>  * HDFS-732
>  * HDFS-579
>  * MAPREDUCE-1070
>  * HADOOP-6315
>  * MAPREDUCE-1163
>  * HADOOP-5759
>  * HADOOP-6269
>  * HADOOP-6386
>  * HADOOP-6428
>

 Owen, Suresh and I have committed everything on this list except
 HADOOP-6386 and HADOOP-6428. Not sure which of the two are
 relevant/necessary, I'll check with Cos.  Other than that hadoop-0.20.203
 now a superset of hadoop-0.20.2.

>>>
>>> I have looked somewhat more into these two JIRAs and if I remember correctly
>>> this fix causes a rolling port side effect in TT and it has been reverted in
>>> 0.20.200 (Y! Fred? release) because Ops weren't happy about this (I am sure
>>> you can check internal Git to cross-verify my recollection).
>>>
>>> Considering above, these might be better left outside of the release and,
>>> perhaps, they should be reverted in trunk as well.
>>>
>>> Cos
>>>
>>>
 thanks,
 Arun

>>
>>
>


Re: VOTE: Committing HADOOP-6949 to 0.22 branch

2011-03-29 Thread Jakob Homan
+1
n.b. that the vote lost hdfs and common dev at some point.  I've added
them back.

On Tue, Mar 29, 2011 at 9:18 AM, Amit Sangroya  wrote:
> +1
>
> On Tue, Mar 29, 2011 at 6:04 PM, Stephen Boesch  wrote:
>> +1
>>
>> 2011/3/29 Doug Cutting 
>>
>>> +1
>>>
>>> I don't think this creates an incompatibility.  It changes the RPC wire
>>> format, but we already require that clients and servers run identical
>>> builds.  No application that ran with a prior version of Hadoop would be
>>> broken by this change when it upgrades to this version of Hadoop.
>>>
>>> Doug
>>>
>>> On 03/28/2011 09:39 PM, Konstantin Shvachko wrote:
>>> > HADOOP-6949 introduced a very important optimization to the RPC layer.
>>> Based
>>> > on the benchmarks presented in HDFS-1583 this provides an order of
>>> magnitude
>>> > improvement of current RPC implementation.
>>> > RPC is a common component of Hadoop projects. Many of them should benefit
>>> > from this change. But since this is an incompatible change it requires a
>>> > vote to be included into a previous branch.
>>> > Please vote for inclusion of this change into branch 0.22.
>>> >
>>> > +1 from me.
>>> >
>>> > Thanks,
>>> > --Konstantin
>>> >
>>>
>>
>


[jira] Resolved: (HADOOP-6236) clean-up Trash.moveToTrash() code. Should throw exception instead of returning false

2011-03-09 Thread Jakob Homan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-6236?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jakob Homan resolved HADOOP-6236.
-

Resolution: Duplicate

Resolving as duplicate of HADOOP-6345

> clean-up Trash.moveToTrash() code. Should throw exception instead of 
> returning false
> 
>
> Key: HADOOP-6236
> URL: https://issues.apache.org/jira/browse/HADOOP-6236
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs
>Reporter: Boris Shkolnik
>
> In case of errors it should throw an exception instead of returning false.
> One valid case is when interval is 0 (trash disabled), but that could be 
> moved to a separate method isEnabled().

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] Reopened: (HADOOP-6463) Uninformative error message for some sort of file permission problem from the hadoop command

2011-02-19 Thread Jakob Homan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-6463?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jakob Homan reopened HADOOP-6463:
-


> Uninformative error message for some sort of file permission problem from the 
> hadoop command
> 
>
> Key: HADOOP-6463
> URL: https://issues.apache.org/jira/browse/HADOOP-6463
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: util
>Affects Versions: 0.20.1
>Reporter: Benson Margulies
> Attachments: HADOOP-6463.patch
>
>
> It would be better if this error message would include the pathname for which 
> permission was denied.
> benson@cn0:/u1/benson/hadoop-test$ /u1/hadoop/bin/hadoop jar 
> /u1/hadoop/hadoop-0.20.1-examples.jar wordcount /users/benson/gutenberg 
> /users/benson/gutenberg-output
> Exception in thread "main" java.io.IOException: Permission denied
>   at java.io.UnixFileSystem.createFileExclusively(Native Method)
>   at java.io.File.checkAndCreate(File.java:1716)
>   at java.io.File.createTempFile(File.java:1804)
>   at org.apache.hadoop.util.RunJar.main(RunJar.java:115)

-- 
This message is automatically generated by JIRA.
-
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] Resolved: (HADOOP-6463) Uninformative error message for some sort of file permission problem from the hadoop command

2011-02-19 Thread Jakob Homan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-6463?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jakob Homan resolved HADOOP-6463.
-

Resolution: Won't Fix

> Uninformative error message for some sort of file permission problem from the 
> hadoop command
> 
>
> Key: HADOOP-6463
> URL: https://issues.apache.org/jira/browse/HADOOP-6463
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: util
>Affects Versions: 0.20.1
>Reporter: Benson Margulies
> Attachments: HADOOP-6463.patch
>
>
> It would be better if this error message would include the pathname for which 
> permission was denied.
> benson@cn0:/u1/benson/hadoop-test$ /u1/hadoop/bin/hadoop jar 
> /u1/hadoop/hadoop-0.20.1-examples.jar wordcount /users/benson/gutenberg 
> /users/benson/gutenberg-output
> Exception in thread "main" java.io.IOException: Permission denied
>   at java.io.UnixFileSystem.createFileExclusively(Native Method)
>   at java.io.File.checkAndCreate(File.java:1716)
>   at java.io.File.createTempFile(File.java:1804)
>   at org.apache.hadoop.util.RunJar.main(RunJar.java:115)

-- 
This message is automatically generated by JIRA.
-
For more information on JIRA, see: http://www.atlassian.com/software/jira




Re: Add Guava as a dependency?

2011-02-10 Thread Jakob Homan
+1

On Thu, Feb 10, 2011 at 10:04 PM, Konstantin Boudnik  wrote:
> Actually it seems that Pig uses this already, which perhaps means it
> is good enough for us as well ;)
> --
>   Take care,
> Konstantin (Cos) Boudnik
>
> On Thu, Feb 10, 2011 at 19:45, Todd Lipcon  wrote:
>> Anyone mind if I pull in the Guava library as a dependency for Common? It
>> has a bunch of very useful utilities - in this particular case the one I'm
>> itching to use is ThreadFactoryBuilder:
>>
>> http://guava-libraries.googlecode.com/svn/tags/release05/javadoc/com/google/common/util/concurrent/ThreadFactoryBuilder.html
>>
>> More info here:
>> http://code.google.com/p/guava-libraries/
>>
>> -Todd
>> --
>> Todd Lipcon
>> Software Engineer, Cloudera
>>
>


Re: Do Common developers use FindBugs?

2011-01-24 Thread Jakob Homan
Cesar-
   Theoretically, Findbugs is run as part of the test-patch process, a
correctness verification that all patches need to go through before
being committed, for all the projects.  However, test-patch has a
checkered history and has not always been reliable.  So yes, we do,
when we do.
-jg


2011/1/24 César Couto :
> Dear developers,
>
> I am a PhD student at UFMG, Brazil and as part of my research I am
> making a study  about the relevance of the warnings reported by the
> FindBugs bug finding tool.
>
> Since I am planning to use Common as a subject system in my research,
> I would like to know if Common's developers usually run FindBugs as
>
> part of the system  development process.
>
> Thanks in advance,
>
> Cesar Couto
>
> --
> http://www.decom.cefetmg.br/cesar
>


Re: Ready for review?

2011-01-07 Thread Jakob Homan
Niels-
   Thanks for the contribution.  For the moment your task is done.
Now it's up to a committer to review the patch and, either provide you
with feedback for its improvement, or commit it.  It's in the patch
available state, which is the flag for reviewers to know there's work
for them to do.  Since this is a volunteer effort, I'm afraid there's
no firm timeline for when this will get done.
-Jakob


On Fri, Jan 7, 2011 at 6:50 AM, Niels Basjes  wrote:
> Hi,
>
> I consider the patch I created to be ready for review by a code reviewer.
>
> It does what I want it to do and Hudson gives an overall +1.
> The http://wiki.apache.org/hadoop/HowToContribute was unclear to me on
> what to do next.
>
> So, what should I do?
> Simply wait / change the stated of the issue / ...
>
> https://issues.apache.org/jira/browse/HADOOP-7076
>
> --
> Met vriendelijke groeten,
>
> Niels Basjes
>


Re: Jira workflow problem.

2011-01-06 Thread Jakob Homan
Niels-
   For new patches, just cancel the current and submit will become
available again.  Also, there's no need to delete prior versions of
the patch; it's better to keep them around for posterity and
comparison.  Hudson will run against whatever is uploaded most
recently (this is important for backports, etc.)
-Jakob


On Thu, Jan 6, 2011 at 1:41 PM, Niels Basjes  wrote:
> I seem to have selected the wrong option in Jira to get the latest
> patch handled.
> For some reason the option to indicate a new patch has been made
> available is no longer present.
>
> https://issues.apache.org/jira/browse/HADOOP-7076
>
> What did I do wrong and what can I do to fix this?
>
> Thanks.
> --
> Met vriendelijke groeten,
>
> Niels Basjes
>


[jira] Created: (HADOOP-7069) Replace forrest with supported framework

2010-12-19 Thread Jakob Homan (JIRA)
Replace forrest with supported framework


 Key: HADOOP-7069
 URL: https://issues.apache.org/jira/browse/HADOOP-7069
 Project: Hadoop Common
  Issue Type: Improvement
  Components: documentation
Reporter: Jakob Homan
 Fix For: 0.23.0


It's time to burn down the forrest.  Apache forrest, which is used to generate 
the documentation for all three subprojects, has not had a release in several 
years (0.8, the version we use was released April 18, 2007), and requires JDK5, 
which was EOL'ed in November 2009.  Since it doesn't seem likely Forrest will 
be developed any more, and JDK5 is not shipped with recent OSX versions, or 
included by default in most linux distros, we should look to find a new 
documentation system and convert the current docs to it.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Created: (HADOOP-6989) TestSetFile is failing on trunk

2010-10-04 Thread Jakob Homan (JIRA)
TestSetFile is failing on trunk
---

 Key: HADOOP-6989
 URL: https://issues.apache.org/jira/browse/HADOOP-6989
 Project: Hadoop Common
  Issue Type: Bug
Affects Versions: 0.22.0
Reporter: Jakob Homan
 Fix For: 0.22.0


Testsuite: org.apache.hadoop.io.TestSetFile
Tests run: 1, Failures: 0, Errors: 1, Time elapsed: 1.015 sec
- Standard Output ---
2010-10-04 16:32:01,030 INFO  io.TestSetFile (TestSetFile.java:generate(56)) - 
generating 1 records in memory
2010-10-04 16:32:01,249 INFO  io.TestSetFile (TestSetFile.java:generate(63)) - 
sorting 1 records
2010-10-04 16:32:01,350 INFO  io.TestSetFile (TestSetFile.java:writeTest(72)) - 
creating with 1 records
-  ---

Testcase: testSetFile took 0.964 sec
Caused an ERROR
key class or comparator option must be set
java.lang.IllegalArgumentException: key class or comparator option must be set
at org.apache.hadoop.io.MapFile$Writer.(MapFile.java:247)
at org.apache.hadoop.io.SetFile$Writer.(SetFile.java:60)
at org.apache.hadoop.io.TestSetFile.writeTest(TestSetFile.java:73)
at org.apache.hadoop.io.TestSetFile.testSetFile(TestSetFile.java:45)

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Created: (HADOOP-6987) Use JUnit Rule to optionally fail test cases that run more than 10 seconds

2010-10-04 Thread Jakob Homan (JIRA)
Use JUnit Rule to optionally fail test cases that run more than 10 seconds
--

 Key: HADOOP-6987
 URL: https://issues.apache.org/jira/browse/HADOOP-6987
 Project: Hadoop Common
  Issue Type: Improvement
  Components: test
Affects Versions: 0.21.0
Reporter: Jakob Homan
Assignee: Jakob Homan
 Fix For: 0.22.0


Using JUnit Rules annotations we can fail tests cases that take longer than 10 
seconds (for instance) to run.  This provides a regression check against test 
cases taking longer than they had previously due to unintended code changes, as 
well as provides a membership criteria for unit tests versus integration tests 
in HDFS and MR.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Created: (HADOOP-6892) Common component of HDFS-1150 (Verify datanodes' identities to clients in secure clusters)

2010-07-30 Thread Jakob Homan (JIRA)
Common component of HDFS-1150 (Verify datanodes' identities to clients in 
secure clusters)
--

 Key: HADOOP-6892
 URL: https://issues.apache.org/jira/browse/HADOOP-6892
 Project: Hadoop Common
  Issue Type: New Feature
  Components: security
Affects Versions: 0.22.0
Reporter: Jakob Homan
Assignee: Jakob Homan
 Fix For: 0.22.0


HDFS-1150 will have changes to the start-up scripts and HttpServer.  These are 
handled here.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Created: (HADOOP-6853) Common component of HDFS-1045

2010-07-07 Thread Jakob Homan (JIRA)
Common component of HDFS-1045
-

 Key: HADOOP-6853
 URL: https://issues.apache.org/jira/browse/HADOOP-6853
 Project: Hadoop Common
  Issue Type: Bug
Reporter: Jakob Homan
Assignee: Jakob Homan
 Fix For: 0.22.0


HDFS-1045 modified UGI, which is in Common on trunk.  This JIRA is for that 
change.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Created: (HADOOP-6824) Improve NetUtils:normalizeHostName

2010-06-11 Thread Jakob Homan (JIRA)
Improve NetUtils:normalizeHostName
--

 Key: HADOOP-6824
 URL: https://issues.apache.org/jira/browse/HADOOP-6824
 Project: Hadoop Common
  Issue Type: Improvement
Reporter: Jakob Homan


HADOOP-6682 revealed a bug in NetUtils:normalizeHostName, which got fixed, but 
it may be worth replacing the current best-effort code with a more rigorous 
approach, as Hong suggested.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Created: (HADOOP-6823) Provide validation of security-related settings on startup

2010-06-11 Thread Jakob Homan (JIRA)
Provide validation of security-related settings on startup
--

 Key: HADOOP-6823
 URL: https://issues.apache.org/jira/browse/HADOOP-6823
 Project: Hadoop Common
  Issue Type: Improvement
Affects Versions: 0.22.0
Reporter: Jakob Homan
Priority: Minor


There are several new config settings for security, involving keytab files, 
kerberos principals, security filters, etc. that need to be set for a secure 
cluster to come up.  It'd be nice if on startup, we checked that the required 
config values have been set and have something vaguely reasonable for values.  
Displaying this information as well would provide a good summary for those 
trying to debug cluster security issues.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Created: (HADOOP-6822) Provide information as to whether or not security is enabled on web interface

2010-06-11 Thread Jakob Homan (JIRA)
Provide information as to whether or not security is enabled on web interface
-

 Key: HADOOP-6822
 URL: https://issues.apache.org/jira/browse/HADOOP-6822
 Project: Hadoop Common
  Issue Type: Sub-task
Reporter: Jakob Homan




-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Resolved: (HADOOP-6661) User document for UserGroupInformation.doAs

2010-06-03 Thread Jakob Homan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-6661?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jakob Homan resolved HADOOP-6661.
-

Resolution: Fixed

I've committed this.  Thanks, Jitendra!  Resolving as fixed.

> User document for UserGroupInformation.doAs
> ---
>
> Key: HADOOP-6661
> URL: https://issues.apache.org/jira/browse/HADOOP-6661
> Project: Hadoop Common
>  Issue Type: Improvement
>Reporter: Jitendra Nath Pandey
>Assignee: Jitendra Nath Pandey
> Attachments: HADOOP-6661-y20.1.patch, HADOOP-6661-y20.2.patch, 
> HADOOP-6661.1.patch, HADOOP-6661.2.patch, HADOOP-6661.3.patch, 
> HADOOP-6661.4.patch
>
>
> A user document should be added for secure impersonation feature. The 
> document should cover a code example about how  UserGroupInformation.doAs 
> should be used for secure impersonation.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Resolved: (HADOOP-6741) Refactor Hadoop Command line interface.

2010-05-03 Thread Jakob Homan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-6741?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jakob Homan resolved HADOOP-6741.
-

Resolution: Duplicate

I'll go ahead and say this duplicates HADOOP-6425, but that being said, 
HADOOP-6425 could probably be split into several subtasks that may be more 
manageable.

> Refactor Hadoop Command line interface.
> ---
>
> Key: HADOOP-6741
> URL: https://issues.apache.org/jira/browse/HADOOP-6741
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: fs
>Reporter: Ravi Phulari
> Fix For: 0.22.0
>
>
> Currently Hadoop fs commands are scattered between FsShell , FsPermissions  
> and DfsAdmin .  It will be great if refactor hadoop CLI for better debugging, 
> performance and code maintenance.
> Any thoughts? 

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Created: (HADOOP-6725) Evaluate HtmlUnit for unit and regression testing webpages

2010-04-25 Thread Jakob Homan (JIRA)
Evaluate HtmlUnit for unit and regression testing webpages
--

 Key: HADOOP-6725
 URL: https://issues.apache.org/jira/browse/HADOOP-6725
 Project: Hadoop Common
  Issue Type: Test
  Components: test
Reporter: Jakob Homan
Priority: Minor


HtmlUnit (http://htmlunit.sourceforge.net/) looks like it may be a good tool to 
help unit testing and evaluating our various webpages throughout the project.  
Currently this is done only occasionally in the code (usually falls to being a 
manual test during release cycles), and when it is done, usually the code to 
parse the webpage, etc. is re-written each time.  The framework is Apache 
licensed, so including it won't be an issue.  If it's found to be useful, new 
JIRAs for HDFS and MR should be opened.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Created: (HADOOP-6682) NetUtils:normalizeHostName does not process hostnames starting with [a-f] correctly

2010-04-05 Thread Jakob Homan (JIRA)
NetUtils:normalizeHostName does not process hostnames starting with [a-f] 
correctly
---

 Key: HADOOP-6682
 URL: https://issues.apache.org/jira/browse/HADOOP-6682
 Project: Hadoop Common
  Issue Type: Bug
  Components: io
Reporter: Jakob Homan


  public static String normalizeHostName(String name) {
if (Character.digit(name.charAt(0), 16) != -1) {
  return name;

This code is attempting to short-circuit the hostname->ip resolution on the 
assumption that if name starts with a digit, it's already an ip address.  This 
is of questionable value, but because it checks for a hex digit, it will fail 
on names starting with [a-f].  Such names will not be converted to an ip 
address, but be returned unchanged.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



Re: [VOTE] HADOOP-6671 - To use maven for hadoop common build

2010-04-01 Thread Jakob Homan
+1 for the the exploration.  *If* it works as advertised, it'll be a 
great time saver.

Giridharan Kesavan wrote:

I  would like call for a vote to created a development branch of common
trunk to work on mavenizing hadoop common.

-Giri





[jira] Resolved: (HADOOP-6611) Fix javac warnings introduced by HADOOP-6584

2010-03-03 Thread Jakob Homan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-6611?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jakob Homan resolved HADOOP-6611.
-

Resolution: Invalid

This patch should be attached to 6584, since the trunk patch won't have these 
issues...

> Fix javac warnings introduced by HADOOP-6584
> 
>
> Key: HADOOP-6611
> URL: https://issues.apache.org/jira/browse/HADOOP-6611
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: security
>    Reporter: Jakob Homan
>Assignee: Jakob Homan
>
> Some javac warnings slipped through in HADOOP-6584:
> {noformat}[exec]   [javadoc] 
> /Users/jhoman/work/git/hadoop-common/src/core/org/apache/hadoop/http/HttpServer.java:45
> 4: warning - @param argument "needClientAuth" is not a parameter name.
>  [exec]   [javadoc] 
> /Users/jhoman/work/git/hadoop-common/src/core/org/apache/hadoop/security/Krb5AndCertsSs
> lSocketConnector.java:183: warning - Tag @link: reference not found: 
> Krb5SslSocketConnector
>  [exec]   [javadoc] 
> /Users/jhoman/work/git/hadoop-common/src/core/org/apache/hadoop/security/Krb5AndCertsSs
> lSocketConnector.java:183: warning - Tag @link: reference not found: 
> Krb5SslSocketConnector
>  [exec]   [javadoc] 
> /Users/jhoman/work/git/hadoop-common/src/core/org/apache/hadoop/security/Krb5AndCertsSs
> lSocketConnector.java:183: warning - Tag @link: reference not found: 
> Krb5SslSocketConnector
>  [exec]   [javadoc] Building index for all the packages and classes...
>  [exec]   [javadoc] 
> /Users/jhoman/work/git/hadoop-common/src/core/org/apache/hadoop/security/Krb5AndCertsSs
> lSocketConnector.java:183: warning - Tag @link: reference not found: 
> Krb5SslSocketConnector
>  [exec]   [javadoc] Building index for all classes...
>  [exec]   [javadoc] Generating 
> /Users/jhoman/work/git/hadoop-common/build/docs/api/stylesheet.css...
>  [exec]   [javadoc] 5 warnings{noformat}

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Created: (HADOOP-6611) Fix javac warnings introduced by HADOOP-6584

2010-03-03 Thread Jakob Homan (JIRA)
Fix javac warnings introduced by HADOOP-6584


 Key: HADOOP-6611
 URL: https://issues.apache.org/jira/browse/HADOOP-6611
 Project: Hadoop Common
  Issue Type: Bug
  Components: security
Reporter: Jakob Homan
Assignee: Jakob Homan


Some javac warnings slipped through in HADOOP-6584:
{noformat}[exec]   [javadoc] 
/Users/jhoman/work/git/hadoop-common/src/core/org/apache/hadoop/http/HttpServer.java:45
4: warning - @param argument "needClientAuth" is not a parameter name.
 [exec]   [javadoc] 
/Users/jhoman/work/git/hadoop-common/src/core/org/apache/hadoop/security/Krb5AndCertsSs
lSocketConnector.java:183: warning - Tag @link: reference not found: 
Krb5SslSocketConnector
 [exec]   [javadoc] 
/Users/jhoman/work/git/hadoop-common/src/core/org/apache/hadoop/security/Krb5AndCertsSs
lSocketConnector.java:183: warning - Tag @link: reference not found: 
Krb5SslSocketConnector
 [exec]   [javadoc] 
/Users/jhoman/work/git/hadoop-common/src/core/org/apache/hadoop/security/Krb5AndCertsSs
lSocketConnector.java:183: warning - Tag @link: reference not found: 
Krb5SslSocketConnector
 [exec]   [javadoc] Building index for all the packages and classes...
 [exec]   [javadoc] 
/Users/jhoman/work/git/hadoop-common/src/core/org/apache/hadoop/security/Krb5AndCertsSs
lSocketConnector.java:183: warning - Tag @link: reference not found: 
Krb5SslSocketConnector
 [exec]   [javadoc] Building index for all classes...
 [exec]   [javadoc] Generating 
/Users/jhoman/work/git/hadoop-common/build/docs/api/stylesheet.css...
 [exec]   [javadoc] 5 warnings{noformat}


-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Created: (HADOOP-6603) Provide workaround for issue with Kerberos not resolving cross-realm principal

2010-03-01 Thread Jakob Homan (JIRA)
Provide workaround for issue with Kerberos not resolving cross-realm principal
--

 Key: HADOOP-6603
 URL: https://issues.apache.org/jira/browse/HADOOP-6603
 Project: Hadoop Common
  Issue Type: Bug
  Components: security
Reporter: Jakob Homan


Java's SSL-Kerberos implementation does not correctly obtain the principal for 
cross-realm principles when clients initiate connections to servers, resulting 
in the client being unable to authenticate the server.  We need a work-around 
until this bug gets fixed.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Created: (HADOOP-6594) Update hdfs script to provide fetchdt tool

2010-02-23 Thread Jakob Homan (JIRA)
Update hdfs script to provide fetchdt tool
--

 Key: HADOOP-6594
 URL: https://issues.apache.org/jira/browse/HADOOP-6594
 Project: Hadoop Common
  Issue Type: New Feature
Reporter: Jakob Homan
Assignee: Jakob Homan


Since the bin/hdfs script is still in common, this JIRA is needed to provide 
access to the new delegation token fetcher done in HDFS-994

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Created: (HADOOP-6584) Provide Kerberized SSL encryption for webservices

2010-02-21 Thread Jakob Homan (JIRA)
Provide Kerberized SSL encryption for webservices
-

 Key: HADOOP-6584
 URL: https://issues.apache.org/jira/browse/HADOOP-6584
 Project: Hadoop Common
  Issue Type: Improvement
Reporter: Jakob Homan
Assignee: Jakob Homan


Some web services should be authenticated via SSL backed by Kerberos, both to 
provide cross-cluster secure communication and to provide intra-cluster server 
authentication for services such as the 
{Name,SecondaryName,Backup,Checkpoint}node's image transfer and balancer.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Created: (HADOOP-6527) UserGroupInformation::createUserForTesting clobbers already defined group mappings

2010-01-30 Thread Jakob Homan (JIRA)
UserGroupInformation::createUserForTesting clobbers already defined group 
mappings
--

 Key: HADOOP-6527
 URL: https://issues.apache.org/jira/browse/HADOOP-6527
 Project: Hadoop Common
  Issue Type: Bug
  Components: security
Reporter: Jakob Homan


In UserGroupInformation::createUserForTesting the follow code creates a new 
groups instance, obliterating any groups that have been previously defined in 
the static groups field.
{code}if (!(groups instanceof TestingGroups)) {
  groups = new TestingGroups();
}
{code}
This becomes a problem in tests that start a Mini{DFS,MR}Cluster and then 
create a testing user.  The user that started the user (generally the real user 
running the test) immediately has their groups wiped out and is prevented from 
accessing files/folders/queues they should be able to.  Before the 
UserGroupInformation.createRemoteUserForTesting, calls to userA.getGroups may 
return {"a", "b", "c"} and immediately after the new fake user is created, the 
same call will return an empty array.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Created: (HADOOP-6521) FsPermission:SetUMask not updated to use new-style umask setting.

2010-01-29 Thread Jakob Homan (JIRA)
FsPermission:SetUMask not updated to use new-style umask setting.
-

 Key: HADOOP-6521
 URL: https://issues.apache.org/jira/browse/HADOOP-6521
 Project: Hadoop Common
  Issue Type: Bug
Reporter: Jakob Homan


FsPermission:
221   /** Set the user file creation mask (umask) */
222   public static void setUMask(Configuration conf, FsPermission umask) { 
   
223 conf.setInt(UMASK_LABEL, umask.toShort());
224   }
Needs to be updated to not use a decimal value. This is a bug introduced by 
HADOOP-6234.  

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Resolved: (HADOOP-6511) FsPermission.getDefault choice of 0777 is bad in all cases

2010-01-28 Thread Jakob Homan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-6511?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jakob Homan resolved HADOOP-6511.
-

Resolution: Invalid

Hearing no counter-arguments, I'm closing this as invalid...

> FsPermission.getDefault choice of 0777 is bad in all cases
> --
>
> Key: HADOOP-6511
> URL: https://issues.apache.org/jira/browse/HADOOP-6511
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs
>Affects Versions: 0.20.1
> Environment: hadoop 20.1, java 1.6.0_17, Fedora
>Reporter: robert Cook
>
>   /** Get the default permission. */
> 227 public static FsPermission getDefault() {
> 228   return new FsPermission((short)00777);
> 229 }
> setting any data file to execute-others is bad security under all 
> circumstances
> should be 664 or 644

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Resolved: (HADOOP-5436) job history directory grows without bound, locks up job tracker on new job submission

2010-01-27 Thread Jakob Homan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-5436?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jakob Homan resolved HADOOP-5436.
-

Resolution: Won't Fix

> job history directory grows without bound, locks up job tracker on new job 
> submission
> -
>
> Key: HADOOP-5436
> URL: https://issues.apache.org/jira/browse/HADOOP-5436
> Project: Hadoop Common
>  Issue Type: Bug
>Affects Versions: 0.19.0
>Reporter: Tim Williamson
> Attachments: HADOOP-5436.patch
>
>
> An unpleasant surprise upgrading to 0.19: requests to jobtracker.jsp would 
> take a long time or even time out whenever new jobs where submitted.  
> Investigation showed the call to JobInProgress.initTasks() was calling 
> JobHistory.JobInfo.logSubmitted() which in turn was calling 
> JobHistory.getJobHistoryFileName() which was pegging the CPU for a couple 
> minutes.  Further investigation showed the were 200,000+ files in the job 
> history folder -- and every submission was creating a FileStatus for them 
> all, then applying a regular expression to just the name.  All this just on 
> the off chance the job tracker had been restarted (see HADOOP-3245).  To make 
> matters worse, these files cannot be safely deleted while the job tracker is 
> running, as the disappearance of a history file at the wrong time causes a 
> FileNotFoundException.
> So to summarize the issues:
> - having Hadoop default to storing all the history files in a single 
> directory is a Bad Idea
> - doing expensive processing of every history file on every job submission is 
> a Worse Idea
> - doing expensive processing of every history file on every job submission 
> while holding a lock on the JobInProgress object and thereby blocking the 
> jobtracker.jsp from rendering is a Terrible Idea (note: haven't confirmed 
> this, but a cursory glance suggests that's what's going on)
> - not being able to clean up the mess without taking down the job tracker is 
> just Unfortunate

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Reopened: (HADOOP-5436) job history directory grows without bound, locks up job tracker on new job submission

2010-01-27 Thread Jakob Homan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-5436?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jakob Homan reopened HADOOP-5436:
-


> job history directory grows without bound, locks up job tracker on new job 
> submission
> -
>
> Key: HADOOP-5436
> URL: https://issues.apache.org/jira/browse/HADOOP-5436
> Project: Hadoop Common
>  Issue Type: Bug
>Affects Versions: 0.19.0
>Reporter: Tim Williamson
> Attachments: HADOOP-5436.patch
>
>
> An unpleasant surprise upgrading to 0.19: requests to jobtracker.jsp would 
> take a long time or even time out whenever new jobs where submitted.  
> Investigation showed the call to JobInProgress.initTasks() was calling 
> JobHistory.JobInfo.logSubmitted() which in turn was calling 
> JobHistory.getJobHistoryFileName() which was pegging the CPU for a couple 
> minutes.  Further investigation showed the were 200,000+ files in the job 
> history folder -- and every submission was creating a FileStatus for them 
> all, then applying a regular expression to just the name.  All this just on 
> the off chance the job tracker had been restarted (see HADOOP-3245).  To make 
> matters worse, these files cannot be safely deleted while the job tracker is 
> running, as the disappearance of a history file at the wrong time causes a 
> FileNotFoundException.
> So to summarize the issues:
> - having Hadoop default to storing all the history files in a single 
> directory is a Bad Idea
> - doing expensive processing of every history file on every job submission is 
> a Worse Idea
> - doing expensive processing of every history file on every job submission 
> while holding a lock on the JobInProgress object and thereby blocking the 
> jobtracker.jsp from rendering is a Terrible Idea (note: haven't confirmed 
> this, but a cursory glance suggests that's what's going on)
> - not being able to clean up the mess without taking down the job tracker is 
> just Unfortunate

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Resolved: (HADOOP-6500) Only the S3 test extends FileSystemContractBaseTest; the other FS tests do not

2010-01-26 Thread Jakob Homan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-6500?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jakob Homan resolved HADOOP-6500.
-

Resolution: Duplicate

This issue is currently being handled by HDFS-303.  Closing as duplicate.

> Only the S3 test extends FileSystemContractBaseTest; the other FS tests do not
> --
>
> Key: HADOOP-6500
> URL: https://issues.apache.org/jira/browse/HADOOP-6500
> Project: Hadoop Common
>  Issue Type: Test
>  Components: test
>Affects Versions: 0.20.1
> Environment: Fedora, java 1.6.0_17, hadoop 20.1
>Reporter: robert Cook
>
> TestChecksumFileSystem etc does not extend FileSystemContractBaseTest
> In fact, ChecksumFileSystem doesn't pass the ContractBaseTest

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Reopened: (HADOOP-6385) dfs does not support -rmdir (was HDFS-639)

2009-12-03 Thread Jakob Homan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-6385?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jakob Homan reopened HADOOP-6385:
-


Hey Scott, re-opening as resolved is used for issues that have been committed, 
which this hasn't.  Were you still wanting this feature?  I personally have 
some concerns about it, but haven't had a chance to look at the patch or 
comment.

> dfs does not support -rmdir (was HDFS-639)
> --
>
> Key: HADOOP-6385
> URL: https://issues.apache.org/jira/browse/HADOOP-6385
> Project: Hadoop Common
>  Issue Type: New Feature
>  Components: fs
>Reporter: Scott Phillips
>Priority: Minor
> Attachments: HADOOP-6385.patch
>
>
> From HDFS-639
> > Given we have a mkdir, we should have a rmdir. Using rmr is not
> > a reasonable substitute when you only want to delete empty
> > directories.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Created: (HADOOP-6396) Provide a description in the exception when an error is encountered parsing umask

2009-11-25 Thread Jakob Homan (JIRA)
Provide a description in the exception when an error is encountered parsing 
umask
-

 Key: HADOOP-6396
 URL: https://issues.apache.org/jira/browse/HADOOP-6396
 Project: Hadoop Common
  Issue Type: Bug
  Components: fs
Affects Versions: 0.21.0, 0.22.0
Reporter: Jakob Homan
Assignee: Jakob Homan
 Fix For: 0.21.0, 0.22.0


Currently when there is a problem parsing a umask, the exception text is just 
the offending umask with no other clue, which can be quite confusing as 
demonstrated in HDFS-763.  This message should include the nature of the 
problem and whether or not the umask parsing attempt was using old-style or 
new-style values.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



Re: HDFS-758 in Hadoop-21 , Updates to Namenode health page

2009-11-25 Thread Jakob Homan

+1. Backporting this does not in any way impact the release of 21.
-Jakob
Hairong Kuang wrote:

+1. Although this is a new feature, I'd like to have it committed to 0.21
since we have so many issues with delayed decomission recently.

Hairong 



On 11/24/09 6:06 PM, "Suresh Srinivas"  wrote:


+1. This will also help debug the issues when decommissioning takes a long
time to complete.


On 11/23/09 7:36 PM, "Jitendra Nath Pandey"  wrote:




Hi,

   We will be committing some changes to the Namenode Health page
(dfshealth.jsp) as part of the fix in HDFS-758. This will enable us to
monitor the progress of decommissioning of datanodes more effectively.
   Summary of changes :
   1. A new link on the page for Decommissioning nodes.
   2. This link will point to a new page with details about decommissioning
status for each node which include
a) Number of under-relplicated blocks in the node.
b) Number of blocks with only no live replica (i.e. All its replicas
are on decommissioning nodes).
c) Number of under-replicated blocks in open files.
   d) Time since decommissioning started.
   3. The main page will also contain total number of under-replicated
blocks
in the cluster.

Thanks
jitendra








[jira] Resolved: (HADOOP-6355) Make FileSystemContractBaseTest, LocalFileSystem, and FileSystem javadoc compliant

2009-11-03 Thread Jakob Homan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-6355?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jakob Homan resolved HADOOP-6355.
-

Resolution: Duplicate

I'm going to go ahead and say this does duplicate the other jira.  It would 
certainly be great to iron out those differences, but we ran into problems 
involving backwards compatibility when we worked on it previously.  If 
possible, stick to the FileSystemContract test and then rely on DFS behavior as 
guidance if there are differences.

> Make FileSystemContractBaseTest, LocalFileSystem, and FileSystem javadoc 
> compliant 
> ---
>
> Key: HADOOP-6355
> URL: https://issues.apache.org/jira/browse/HADOOP-6355
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs
>Reporter: Gregory Farnum
>Priority: Minor
>
> These do not all seem to play nicely together. ie, plugging the 
> LocalFileSystem into FileSystemContractBaseTest (via a subclass that just 
> overrides setUp to set fs=LocalFileSystem) results in 5 failures and 19 
> errors out of 28 tests.
> Additionally, the unit tests expect a few undocumented behaviors (ie, it 
> expects to get an IOException on mkdirs(subdir/of/a/file) rather than just 
> returning false) and some contradictory behaviors (it expects that 
> delete(directory, false) should succeed on an empty directory) compared to 
> FileSystem's javadoc.
> I suspect that most of the fixes should come from changing the unit tests to 
> be more sensible rather than modifying [Local]FileSystem, but however it's 
> done these should all agree on expected behavior!

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Created: (HADOOP-6345) Refactor Trash::moveToTrash() and its accompanying poor tests

2009-10-29 Thread Jakob Homan (JIRA)
Refactor Trash::moveToTrash() and its accompanying poor tests
-

 Key: HADOOP-6345
 URL: https://issues.apache.org/jira/browse/HADOOP-6345
 Project: Hadoop Common
  Issue Type: Improvement
  Components: fs
Reporter: Jakob Homan


We've had several issues relating to the Trash and its access via the shell, 
none of which were picked up by the unit tests.  The current moveToTrash method 
has 8 different ways it can terminate and sometimes uses a false value and 
sometimes uses an exception to indicate failure.  This method should be 
refactored to improve readability and testability, and new tests written to 
exercise all possible code paths.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Resolved: (HADOOP-3474) Need an option -rm and -rmr to *not* move to Trash, but actually delete

2009-10-28 Thread Jakob Homan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-3474?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jakob Homan resolved HADOOP-3474.
-

Resolution: Duplicate

This ended up being duplicated by HADOOP-6080.  Resolving.

> Need an option -rm and -rmr to *not* move to Trash, but actually delete
> ---
>
> Key: HADOOP-3474
> URL: https://issues.apache.org/jira/browse/HADOOP-3474
> Project: Hadoop Common
>  Issue Type: New Feature
>  Components: fs
>Affects Versions: 0.17.0
>Reporter: Milind Bhandarkar
>
> By default, when Trash is enabled, -rm and -rmr just move files/directories 
> to .Trash, which are actually deleted after some time set in the config.
> When DFS gets full, we need to quickly create space on it by deleting files, 
> so it will be good to have a --force option to both -rm and -rmr, so that 
> files are actually deleted and not moved.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Created: (HADOOP-6337) Update FilterInitializer class to be more visible and take a conf for further development

2009-10-27 Thread Jakob Homan (JIRA)
Update FilterInitializer class to be more visible and take a conf for further 
development
-

 Key: HADOOP-6337
 URL: https://issues.apache.org/jira/browse/HADOOP-6337
 Project: Hadoop Common
  Issue Type: New Feature
Reporter: Jakob Homan
Assignee: Jakob Homan
 Fix For: 0.22.0


Currently the FilterInitializer class, used to filter access to the Hadoop web 
interfaces, has its main method, initFilter, set to package private.  This 
means that any classes wishing to implement this type must be in the same 
package.  This should be public so FilterInitializers can reside in other 
packages.  

Also, currently all parameters to the FilterInitializer must be provided at 
compile time (or via a different configuration method).  It would be better if 
the FilterInitalizer::initFilter received a Configuration parameter so it can 
pull conf values out at run-time.  Alternatively, the class could implement 
Configured, but that seems heavier than is needed at this point.


-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



Re: Developing Hadoop and HDFS

2009-09-29 Thread Jakob Homan
Thanks for your interest, Geoff.  Yes, finding open JIRAS and 
contributing patches is very helpful.  We also maintain a wishlist of 
projects that one could work on: 
http://wiki.apache.org/hadoop/ProjectSuggestions.  In addition, please 
do consider documentation and example work as well, as this is very 
helpful both to new users and developers starting on the project.


Thanks,
Jakob
Hadoop at Yahoo!

Geoffrey Gallaway wrote:

Hello,

Yes, another person looking to contribute to and develop Hadoop. I'm looking
to start off small, fixing a few bugs before moving into larger stuff.

First, a bit of background:
Years ago I had the idea of creating a semi-decentralized distributed file
system. The idea came when I was working for a small/medium sized company
who was looking for a simple backup solution for their workstations. PC's
back then came with 100+ GB hard drives but, as simple workstations,
employees were using less than half that space. Why not have each
workstation backup to a few other workstations, duplicating files across
multiple machines for redundancy. RAID for the network. I started coming up
with design and architecture specs, protocol examples and even started
writing a bit of the system (in Java). I tried to find a few interested
developers but everyone seemed to think the task was much too large to be
accomplished as a side project (and I didn't think, given the IT industry of
the time, that anyone would fund it). Later, I realized such a distributed
system could be much more than a simple file backup solution.

It looks like Hadoop and HDFS are creating a lot of what I had wanted to
create, it's already surpassed what I had in mind in most ways.

So, where should I start? Just start fixing bugs listed in JIRA?

Geoff





[jira] Created: (HADOOP-6252) Provide method to determine if a deprecated key was set in the config file

2009-09-10 Thread Jakob Homan (JIRA)
Provide method to determine if a deprecated key was set in the config file
--

 Key: HADOOP-6252
 URL: https://issues.apache.org/jira/browse/HADOOP-6252
 Project: Hadoop Common
  Issue Type: Improvement
  Components: conf
Affects Versions: 0.21.0
Reporter: Jakob Homan
Assignee: Jakob Homan


HADOOP-6105 provided a method to deprecate config keys and transparently refer 
to the new key. However, it didn't provide a method to see if the deprecated 
key had been used in the config file.  This is useful when, if the deprecated 
key had been used, its value needs to be converted before use, for instance 
when we changed the umask format.  A method like "boolean 
wasDeprecatedKeySet()" would be great.  Patch shortly.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Created: (HADOOP-6246) Update umask code to use key deprecation facilities from HADOOP-6105

2009-09-08 Thread Jakob Homan (JIRA)
Update umask code to use key deprecation facilities from HADOOP-6105


 Key: HADOOP-6246
 URL: https://issues.apache.org/jira/browse/HADOOP-6246
 Project: Hadoop Common
  Issue Type: Improvement
  Components: fs
Reporter: Jakob Homan
Assignee: Jakob Homan


When the new octal/symbolic umask JIRA (HADOOP-6234) went through, the config 
key deprecation patch (HADOOP-6105) wasn't ready.  Now that both are committed, 
the home-brewed key deprecation system from 6234 should be updated to use the 
new system.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Created: (HADOOP-6203) Improve error message when moving to trash fails due to quota issue

2009-08-18 Thread Jakob Homan (JIRA)
Improve error message when moving to trash fails due to quota issue
---

 Key: HADOOP-6203
 URL: https://issues.apache.org/jira/browse/HADOOP-6203
 Project: Hadoop Common
  Issue Type: Improvement
  Components: fs
Reporter: Jakob Homan


HADOOP-6080 provided an option for deleting files even when overquota, but the 
error message that's returned in this situation is unhelpful and doesn't 
suggest skipTrash as a remediation:
{noformat}$ hdfs -rmr /foo/bar/bat/boo

rmr: Failed to move to trash:
hdfs://cluster/foo/bar/bat/boo{noformat}
In this situation, the error message should say there was a quote problem and 
suggest -skipTrash.


-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Created: (HADOOP-6201) FileSystem::ListStatus should throw FileNotFoundException

2009-08-18 Thread Jakob Homan (JIRA)
FileSystem::ListStatus should throw FileNotFoundException
-

 Key: HADOOP-6201
 URL: https://issues.apache.org/jira/browse/HADOOP-6201
 Project: Hadoop Common
  Issue Type: Improvement
  Components: fs
Reporter: Jakob Homan


As discussed in HDFS-538, it would be better for FileSystem and its 
implementations to throw FileNotFoundException when the file is not found, 
rather than returning null.  This will bring it in line with getFileStatus and, 
I expect, default user expectations.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Created: (HADOOP-6164) Test-patch's method of checking for new tests is not correct

2009-07-21 Thread Jakob Homan (JIRA)
Test-patch's method of checking for new tests is not correct


 Key: HADOOP-6164
 URL: https://issues.apache.org/jira/browse/HADOOP-6164
 Project: Hadoop Common
  Issue Type: Bug
  Components: test
Reporter: Jakob Homan


As happened in HDFS-458, test-patch/hudson saw the previous patch(HDFS-492) add 
a new test, bringing the total number of tests to 48.  Then it tested 458, 
which had no new tests as it was a change to the build file.  492 had not yet 
been committed yet, so there were still only 48 tests in trunk.  But test-patch 
failed the patch, expecting 49 tests.  test-patch should check the correct 
number of tests based on trunk, not on whatever number it last saw.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



Re: Developing cross-component patches post-split

2009-07-01 Thread Jakob Homan

Todd Lipcon wrote:
>> If someone could write up some kind of "post-split transition guide" 
>> on the Wiki I think that would be generally appreciated.
Here's something I wrote up for re-setting up Eclipse after the split in 
a way that gives relatively seamless access to the various projects' 
source code.  If it's works for other people, it could be part of the wiki.


Setting up Eclipse post-dev:
I believe my Eclipse dev environment is set up for near-seamless 
inter-project development following the Great Split of 2009. Here's how 
I did it, step-by-step, with no guarantee as to orthodoxy, correctness 
or appropriateness. Please let me know, or update, with corrections. 
Obligatory works-on-my-machine and YMMV.


With these instructions, you'll be able to work on all three projects at 
once. For instance, after the split, trying to look at the source of a 
class defined in Common, such as Text, will show you the bytecode in the 
included jar file. You can attach the source for review purposes, but 
this still leaves the problem of eclipse only running the code from the 
jar, rather than from the other project, in the event that you've 
modified both. These instructions fix this problem so that any changes, 
say in Common, are also applied to HDFS or MapReduce?.


These instructions assume svn and specifically svn from the command 
line. They will work fine with git once those repos are set up and also 
will work with the svn or git plugin from within eclipse.


1. Import each of the new repositories. Here're the directories that I 
picked.


svn checkout https://svn.apache.org/repos/asf/hadoop/common/trunk 
hadoop-common


svn checkout https://svn.apache.org/repos/asf/hadoop/hdfs/trunk hadoop-hdfs

svn checkout https://svn.apache.org/repos/asf/hadoop/mapreduce/trunk 
hadoop-mapreduce


2. Start Eclipse. For each of the new directories, create a new java 
project using that directory name, allowing Eclipse to do its standard 
work of importing the project. It was previously necessary to change 
Eclipse's default build directory (bin) to something else to keep it 
from wiping out Hadoop's bin directory. At the moment, only the common 
project has a bin directory. However, I still changed each of my eclipse 
build directories (set on the second screen of the new project wizard) 
to build/eclipse-files, just in case there is a bin added in the future 
and as it's tidier.


3. Ensure that ANT_HOME is defined in Eclipse's environment variables 
(Preferences -> Java -> Build Path -> Classpath Variables). Mine is set 
to the standard /usr/share/ant


4. From the Navigator window, right click on the build.xml and (#2 Run 
as ant build). Among the targets, specify compile, 
compile-{common,hdfs,mapred}-test, and eclipse-files. Let Eclipse do its 
work and after it's done, each of the projects should compile 
successfully and be working correctly independently.


5. To allow the projects to call directly into each other's code, rather 
than relying on the bundled libraries, connect each of the projects as 
dependencies. For each project set up the natural dependency (hdfs 
relies on common, mapred relies on common and hdfs). Right click on the 
each project and go build path -> configure build path -> projects tab. 
For HDFS, add common. For MapReduce, add Common and HDFS. Unfortunately, 
you can't just add everything to everything, as Eclipse detects this as 
a cycle and throws up errors.


6. Finally, in order to force Eclipse to look at the other projects' 
source code, rather than the included jar files, remove the jar files 
from the build path of each project, respectively. From HDFS, remove the 
common(core) jar from the build path. From MapReduce, remove the hdfs 
and common(core) jars from the build path. HDFS still has a dependency 
on MapReduce for tests, and I couldn't get Eclipse to allow me to remove 
the MapReduce jar from HDFS project; if anyone figures out how, please 
update this document or let me know.


7. All should be well. Now, for instance, if you control-click on a 
class defined in Common from hdfs (say Text), you are brought to its 
definition in the common project, as expected. Also, if you modify code 
in common and run a test from hdfs, it'll pull in the modified code from 
common. This doesn't solve the problem of needing to generated patches 
for each of the projects if your changes affect each of them, however.


[jira] Commented: (HADOOP-6080) Handling of Trash with quota

2009-06-30 Thread Jakob Homan (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-6080?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12725696#action_12725696
 ] 

Jakob Homan commented on HADOOP-6080:
-

bq. Should we consider excluding trash for files deleted from /tmp (ie make 
-skipTrash implicit when deleting from /tmp.)? 
I'm not a fan of special cases for certain directories, even for /tmp, and 
particularly when we're already straying away from the posix world with the 
trash feature.  Minimizing surprise seems a good goal, and I'd be very 
surprised if I were accustomed to explicitly skipping the trash when I want and 
discovering something I had expected to be trashed had been helpfully nuked by 
the system.

> Handling of  Trash with quota
> -
>
> Key: HADOOP-6080
> URL: https://issues.apache.org/jira/browse/HADOOP-6080
> Project: Hadoop Common
>  Issue Type: New Feature
>  Components: fs
>Reporter: Koji Noguchi
>Assignee: Jakob Homan
> Attachments: HADOOP-6080-v20.patch, HADOOP-6080.patch, 
> javac_warnings_diff.txt
>
>
> Currently with quota turned on, user cannot call '-rmr' on large directory 
> that causes over quota.
> {noformat}
> [knoguchi src]$ hadoop dfs -rmr /tmp/net2
> rmr: Failed to move to trash: hdfs://abc.def.com/tmp/net2
> [knoguchi src]$ hadoop dfs -mv /tmp/net2 /user/knoguchi/.Trash/Current
> mv: org.apache.hadoop.hdfs.protocol.QuotaExceededException: The quota of 
> /user/knoguchi is exceeded: namespace
> quota=37500 file count=37757, diskspace quota=-1 diskspace=1991250043353
> {noformat}
> Besides from error message being unfriendly, how should this be handled?

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HADOOP-6080) Handling of Trash with quota

2009-06-30 Thread Jakob Homan (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-6080?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12725659#action_12725659
 ] 

Jakob Homan commented on HADOOP-6080:
-

Updated patches are good to go on all commons unit tests for trunk and all 
tests for v20.  Test-patch is fine except the incorrect javac warnings, which 
are not related.
{noformat} [exec] -1 overall.  
 [exec] 
 [exec] +1 @author.  The patch does not contain any @author tags.
 [exec] 
 [exec] +1 tests included.  The patch appears to include 3 new or 
modified tests.
 [exec] 
 [exec] +1 javadoc.  The javadoc tool did not generate any warning 
messages.
 [exec] 
 [exec] -1 javac.  The applied patch generated 64 javac compiler 
warnings (more than the trunk's current 124 warnings).
 [exec] 
 [exec] +1 findbugs.  The patch does not introduce any new Findbugs 
warnings.
 [exec] 
 [exec] +1 release audit.  The applied patch does not increase the 
total number of release audit warnings.{noformat}

> Handling of  Trash with quota
> -
>
> Key: HADOOP-6080
> URL: https://issues.apache.org/jira/browse/HADOOP-6080
> Project: Hadoop Common
>  Issue Type: New Feature
>  Components: fs
>Reporter: Koji Noguchi
>Assignee: Jakob Homan
> Attachments: HADOOP-6080-v20.patch, HADOOP-6080.patch, 
> javac_warnings_diff.txt
>
>
> Currently with quota turned on, user cannot call '-rmr' on large directory 
> that causes over quota.
> {noformat}
> [knoguchi src]$ hadoop dfs -rmr /tmp/net2
> rmr: Failed to move to trash: hdfs://abc.def.com/tmp/net2
> [knoguchi src]$ hadoop dfs -mv /tmp/net2 /user/knoguchi/.Trash/Current
> mv: org.apache.hadoop.hdfs.protocol.QuotaExceededException: The quota of 
> /user/knoguchi is exceeded: namespace
> quota=37500 file count=37757, diskspace quota=-1 diskspace=1991250043353
> {noformat}
> Besides from error message being unfriendly, how should this be handled?

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HADOOP-6080) Handling of Trash with quota

2009-06-30 Thread Jakob Homan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-6080?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jakob Homan updated HADOOP-6080:


Attachment: HADOOP-6080.patch
HADOOP-6080-v20.patch

Ran into a problem that I didn't notice with the HDFS version of TestTrash.  I 
think there's an issue with the FileSystem.listStatus methods between 
LocalFileSystem and DistributedFileSystem, which I'll look into.  In the 
meantime, modified test so that it doesn't rely on that method and works on 
both local and distributed file systems.  
Will run full test suite tonight, report tomorrow morning.  Also, deleted old 
patches to avoid confusion.  New patches for both trunk and v20 should be good 
to go.

> Handling of  Trash with quota
> -
>
> Key: HADOOP-6080
> URL: https://issues.apache.org/jira/browse/HADOOP-6080
> Project: Hadoop Common
>  Issue Type: New Feature
>  Components: fs
>    Reporter: Koji Noguchi
>Assignee: Jakob Homan
> Attachments: HADOOP-6080-v20.patch, HADOOP-6080.patch, 
> javac_warnings_diff.txt
>
>
> Currently with quota turned on, user cannot call '-rmr' on large directory 
> that causes over quota.
> {noformat}
> [knoguchi src]$ hadoop dfs -rmr /tmp/net2
> rmr: Failed to move to trash: hdfs://abc.def.com/tmp/net2
> [knoguchi src]$ hadoop dfs -mv /tmp/net2 /user/knoguchi/.Trash/Current
> mv: org.apache.hadoop.hdfs.protocol.QuotaExceededException: The quota of 
> /user/knoguchi is exceeded: namespace
> quota=37500 file count=37757, diskspace quota=-1 diskspace=1991250043353
> {noformat}
> Besides from error message being unfriendly, how should this be handled?

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HADOOP-6080) Handling of Trash with quota

2009-06-30 Thread Jakob Homan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-6080?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jakob Homan updated HADOOP-6080:


Attachment: (was: HADOOP-6080-v20.patch)

> Handling of  Trash with quota
> -
>
> Key: HADOOP-6080
> URL: https://issues.apache.org/jira/browse/HADOOP-6080
> Project: Hadoop Common
>  Issue Type: New Feature
>  Components: fs
>Reporter: Koji Noguchi
>    Assignee: Jakob Homan
> Attachments: javac_warnings_diff.txt
>
>
> Currently with quota turned on, user cannot call '-rmr' on large directory 
> that causes over quota.
> {noformat}
> [knoguchi src]$ hadoop dfs -rmr /tmp/net2
> rmr: Failed to move to trash: hdfs://abc.def.com/tmp/net2
> [knoguchi src]$ hadoop dfs -mv /tmp/net2 /user/knoguchi/.Trash/Current
> mv: org.apache.hadoop.hdfs.protocol.QuotaExceededException: The quota of 
> /user/knoguchi is exceeded: namespace
> quota=37500 file count=37757, diskspace quota=-1 diskspace=1991250043353
> {noformat}
> Besides from error message being unfriendly, how should this be handled?

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HADOOP-6080) Handling of Trash with quota

2009-06-30 Thread Jakob Homan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-6080?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jakob Homan updated HADOOP-6080:


Attachment: (was: HADOOP-6080.patch)

> Handling of  Trash with quota
> -
>
> Key: HADOOP-6080
> URL: https://issues.apache.org/jira/browse/HADOOP-6080
> Project: Hadoop Common
>  Issue Type: New Feature
>  Components: fs
>Reporter: Koji Noguchi
>    Assignee: Jakob Homan
> Attachments: javac_warnings_diff.txt
>
>
> Currently with quota turned on, user cannot call '-rmr' on large directory 
> that causes over quota.
> {noformat}
> [knoguchi src]$ hadoop dfs -rmr /tmp/net2
> rmr: Failed to move to trash: hdfs://abc.def.com/tmp/net2
> [knoguchi src]$ hadoop dfs -mv /tmp/net2 /user/knoguchi/.Trash/Current
> mv: org.apache.hadoop.hdfs.protocol.QuotaExceededException: The quota of 
> /user/knoguchi is exceeded: namespace
> quota=37500 file count=37757, diskspace quota=-1 diskspace=1991250043353
> {noformat}
> Besides from error message being unfriendly, how should this be handled?

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HADOOP-6080) Handling of Trash with quota

2009-06-30 Thread Jakob Homan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-6080?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jakob Homan updated HADOOP-6080:


Attachment: (was: HADOOP-6080.patch)

> Handling of  Trash with quota
> -
>
> Key: HADOOP-6080
> URL: https://issues.apache.org/jira/browse/HADOOP-6080
> Project: Hadoop Common
>  Issue Type: New Feature
>  Components: fs
>Reporter: Koji Noguchi
>    Assignee: Jakob Homan
> Attachments: javac_warnings_diff.txt
>
>
> Currently with quota turned on, user cannot call '-rmr' on large directory 
> that causes over quota.
> {noformat}
> [knoguchi src]$ hadoop dfs -rmr /tmp/net2
> rmr: Failed to move to trash: hdfs://abc.def.com/tmp/net2
> [knoguchi src]$ hadoop dfs -mv /tmp/net2 /user/knoguchi/.Trash/Current
> mv: org.apache.hadoop.hdfs.protocol.QuotaExceededException: The quota of 
> /user/knoguchi is exceeded: namespace
> quota=37500 file count=37757, diskspace quota=-1 diskspace=1991250043353
> {noformat}
> Besides from error message being unfriendly, how should this be handled?

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HADOOP-6080) Handling of Trash with quota

2009-06-29 Thread Jakob Homan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-6080?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jakob Homan updated HADOOP-6080:


Status: Open  (was: Patch Available)

Canceling patch to double check something.  

> Handling of  Trash with quota
> -
>
> Key: HADOOP-6080
> URL: https://issues.apache.org/jira/browse/HADOOP-6080
> Project: Hadoop Common
>  Issue Type: New Feature
>  Components: fs
>Reporter: Koji Noguchi
>    Assignee: Jakob Homan
> Attachments: HADOOP-6080-v20.patch, HADOOP-6080.patch, 
> HADOOP-6080.patch, javac_warnings_diff.txt
>
>
> Currently with quota turned on, user cannot call '-rmr' on large directory 
> that causes over quota.
> {noformat}
> [knoguchi src]$ hadoop dfs -rmr /tmp/net2
> rmr: Failed to move to trash: hdfs://abc.def.com/tmp/net2
> [knoguchi src]$ hadoop dfs -mv /tmp/net2 /user/knoguchi/.Trash/Current
> mv: org.apache.hadoop.hdfs.protocol.QuotaExceededException: The quota of 
> /user/knoguchi is exceeded: namespace
> quota=37500 file count=37757, diskspace quota=-1 diskspace=1991250043353
> {noformat}
> Besides from error message being unfriendly, how should this be handled?

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HADOOP-6080) Handling of Trash with quota

2009-06-29 Thread Jakob Homan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-6080?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jakob Homan updated HADOOP-6080:


Attachment: HADOOP-6080-v20.patch
HADOOP-6080.patch

Attaching two new files:
  * Updated patch. Previous patch missed updating the help text for rmr to 
include -skipTrash option. No change to actual code.
  * Patch for Hadoop 20 off of the Hadoop-20 branch from svn.  Nothing had to 
be changed for patch, just file locations were different.  Code is still the 
same.  Passes unit tests.


> Handling of  Trash with quota
> -
>
> Key: HADOOP-6080
> URL: https://issues.apache.org/jira/browse/HADOOP-6080
> Project: Hadoop Common
>  Issue Type: New Feature
>  Components: fs
>Reporter: Koji Noguchi
>    Assignee: Jakob Homan
> Attachments: HADOOP-6080-v20.patch, HADOOP-6080.patch, 
> HADOOP-6080.patch, javac_warnings_diff.txt
>
>
> Currently with quota turned on, user cannot call '-rmr' on large directory 
> that causes over quota.
> {noformat}
> [knoguchi src]$ hadoop dfs -rmr /tmp/net2
> rmr: Failed to move to trash: hdfs://abc.def.com/tmp/net2
> [knoguchi src]$ hadoop dfs -mv /tmp/net2 /user/knoguchi/.Trash/Current
> mv: org.apache.hadoop.hdfs.protocol.QuotaExceededException: The quota of 
> /user/knoguchi is exceeded: namespace
> quota=37500 file count=37757, diskspace quota=-1 diskspace=1991250043353
> {noformat}
> Besides from error message being unfriendly, how should this be handled?

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HADOOP-6080) Handling of Trash with quota

2009-06-29 Thread Jakob Homan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-6080?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jakob Homan updated HADOOP-6080:


Attachment: javac_warnings_diff.txt

Test-patch:
{noformat}[exec] -1 overall.  
[exec] 
[exec] +1 @author.  The patch does not contain any @author tags.
[exec] 
[exec] +1 tests included.  The patch appears to include 3 new or modified 
tests.
[exec] 
[exec] +1 javadoc.  The javadoc tool did not generate any warning messages.
[exec] 
[exec] -1 javac.  The applied patch generated 64 javac compiler warnings 
(more than the trunk's current 124 warnings).
[exec] 
[exec] +1 findbugs.  The patch does not introduce any new Findbugs warnings.
[exec] 
[exec] +1 release audit.  The applied patch does not increase the total 
number of release audit warnings.
{noformat}
This is weird.  I've attached the javac warnings it says are new and they have 
nothing to do with this patch.  test-patch must be broken in this regard.  I 
believe the patch is ready to go.

> Handling of  Trash with quota
> -
>
> Key: HADOOP-6080
> URL: https://issues.apache.org/jira/browse/HADOOP-6080
> Project: Hadoop Common
>  Issue Type: New Feature
>  Components: fs
>Reporter: Koji Noguchi
>Assignee: Jakob Homan
> Attachments: HADOOP-6080.patch, javac_warnings_diff.txt
>
>
> Currently with quota turned on, user cannot call '-rmr' on large directory 
> that causes over quota.
> {noformat}
> [knoguchi src]$ hadoop dfs -rmr /tmp/net2
> rmr: Failed to move to trash: hdfs://abc.def.com/tmp/net2
> [knoguchi src]$ hadoop dfs -mv /tmp/net2 /user/knoguchi/.Trash/Current
> mv: org.apache.hadoop.hdfs.protocol.QuotaExceededException: The quota of 
> /user/knoguchi is exceeded: namespace
> quota=37500 file count=37757, diskspace quota=-1 diskspace=1991250043353
> {noformat}
> Besides from error message being unfriendly, how should this be handled?

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HADOOP-6080) Handling of Trash with quota

2009-06-29 Thread Jakob Homan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-6080?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jakob Homan updated HADOOP-6080:


Status: Patch Available  (was: Open)

submitting patch.

> Handling of  Trash with quota
> -
>
> Key: HADOOP-6080
> URL: https://issues.apache.org/jira/browse/HADOOP-6080
> Project: Hadoop Common
>  Issue Type: New Feature
>  Components: fs
>Reporter: Koji Noguchi
>    Assignee: Jakob Homan
> Attachments: HADOOP-6080.patch
>
>
> Currently with quota turned on, user cannot call '-rmr' on large directory 
> that causes over quota.
> {noformat}
> [knoguchi src]$ hadoop dfs -rmr /tmp/net2
> rmr: Failed to move to trash: hdfs://abc.def.com/tmp/net2
> [knoguchi src]$ hadoop dfs -mv /tmp/net2 /user/knoguchi/.Trash/Current
> mv: org.apache.hadoop.hdfs.protocol.QuotaExceededException: The quota of 
> /user/knoguchi is exceeded: namespace
> quota=37500 file count=37757, diskspace quota=-1 diskspace=1991250043353
> {noformat}
> Besides from error message being unfriendly, how should this be handled?

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HADOOP-6080) Handling of Trash with quota

2009-06-29 Thread Jakob Homan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-6080?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jakob Homan updated HADOOP-6080:


Attachment: HADOOP-6080.patch

Patch adds a new option to the fsshell rm and rmr commands: -skipTrash, which 
performs as expected. Adds to trash unit test to verify correct execution.  
Changes documentation to reflect new option.  Docs suggest this option as being 
a solution when a directory is over quota.

Passes all commons unit tests.  Running test patch now.  Will post those 
results when done.

> Handling of  Trash with quota
> -
>
> Key: HADOOP-6080
> URL: https://issues.apache.org/jira/browse/HADOOP-6080
> Project: Hadoop Common
>  Issue Type: New Feature
>  Components: fs
>Reporter: Koji Noguchi
>    Assignee: Jakob Homan
> Attachments: HADOOP-6080.patch
>
>
> Currently with quota turned on, user cannot call '-rmr' on large directory 
> that causes over quota.
> {noformat}
> [knoguchi src]$ hadoop dfs -rmr /tmp/net2
> rmr: Failed to move to trash: hdfs://abc.def.com/tmp/net2
> [knoguchi src]$ hadoop dfs -mv /tmp/net2 /user/knoguchi/.Trash/Current
> mv: org.apache.hadoop.hdfs.protocol.QuotaExceededException: The quota of 
> /user/knoguchi is exceeded: namespace
> quota=37500 file count=37757, diskspace quota=-1 diskspace=1991250043353
> {noformat}
> Besides from error message being unfriendly, how should this be handled?

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.