On Sun, Jan 30, 2011 at 23:19, Owen O'Malley <omal...@apache.org> wrote:
>
> On Jan 30, 2011, at 7:42 PM, Nigel Daley wrote:
>
>> Now that http://apache-extras.org is launched
>> (https://blogs.apache.org/foundation/entry/the_apache_software_foundation_launches)
>> I'd like to start a discussion on moving contrib components out of common,
>> mapreduce, and hdfs.
>
> The PMC can't "move" code to Apache extras. It can only choose to abandon
> code that it doesn't want to support any longer. As a separate action some
> group of developers may create projects in Apache Extras based on the code
> from Hadoop.
>
> Therefore the question is really what if any code Hadoop wants to abandon.
> That is a good question and one that we should ask ourselves occasionally.
>
> After a quick consideration, my personal list would look like:
>
> failmon
> fault injection

This is the best way to kill a project as tightly coupled with the
core code as fault injection.

So, if you really want to kill it - then move it.

> fuse-dfs
> hod
> kfs
>
> Also note that pushing code out of Hadoop has a high cost. There are at
> least 3 forks of the hadoop-gpl-compression code. That creates a lot of
> confusion for the users. A lot of users never go to the work to figure out
> which fork and branch of hadoop-gpl-compression work with the version of
> Hadoop they installed.
>
> -- Owen
>
>

Reply via email to