Yes, Sanjay. The reason is 'tight coupling'. In fact this was I who
was opposing it - not Nigel. I guess you misread the thread ;)

On Wed, Feb 16, 2011 at 20:25, Sanjay Radia <sra...@yahoo-inc.com> wrote:
>
> On Jan 31, 2011, at 10:19 PM, Konstantin Boudnik wrote:
>
>> On Sun, Jan 30, 2011 at 23:19, Owen O'Malley <omal...@apache.org> wrote:
>>>
>>> On Jan 30, 2011, at 7:42 PM, Nigel Daley wrote:
>>>
>>>> Now that http://apache-extras.org is launched
>>>>
>>>> (https://blogs.apache.org/foundation/entry/the_apache_software_foundation_launches)
>>>> I'd like to start a discussion on moving contrib components out of
>>>> common,
>>>> mapreduce, and hdfs.
>>>
>>> The PMC can't "move" code to Apache extras. It can only choose to abandon
>>> code that it doesn't want to support any longer. As a separate action
>>> some
>>> group of developers may create projects in Apache Extras based on the
>>> code
>>> from Hadoop.
>>>
>>> Therefore the question is really what if any code Hadoop wants to
>>> abandon.
>>> That is a good question and one that we should ask ourselves
>>> occasionally.
>>>
>>> After a quick consideration, my personal list would look like:
>>>
>>> failmon
>>> fault injection
>>
>> This is the best way to kill a project as tightly coupled with the
>> core code as fault injection.
>>
>> So, if you really want to kill it - then move it.
>
>
> Nigel/Owen did not say "kill it". Folks were simply listing potential
> projects to move out.
> If you feel that it should stay in then simply say so and give the reasons
> -- looks like your reason is "tight coupling".
>
>
> sanjay
>
>>
>>> fuse-dfs
>>> hod
>>> kfs
>>>
>>> Also note that pushing code out of Hadoop has a high cost. There are at
>>> least 3 forks of the hadoop-gpl-compression code. That creates a lot of
>>> confusion for the users. A lot of users never go to the work to figure
>>> out
>>> which fork and branch of hadoop-gpl-compression work with the version of
>>> Hadoop they installed.
>>>
>>> -- Owen
>>>
>>>
>
>

Reply via email to