I think you have a misunderstanding of the reserved parameter. As I
commented on hadoop-1463, remember that dfs.du.reserve is the space for
non-dfs usage, including the space for map/reduce, other application, fs
meta-data etc. In your case since /usr already takes 45GB, it far exceeds
the reserved limit 1G. You should set the reserved space to be 50G.

Hairong


On 3/10/08 4:54 PM, "Joydeep Sen Sarma" <[EMAIL PROTECTED]> wrote:

> Filed https://issues.apache.org/jira/browse/HADOOP-2991
> 
> -----Original Message-----
> From: Joydeep Sen Sarma [mailto:[EMAIL PROTECTED]
> Sent: Monday, March 10, 2008 12:56 PM
> To: core-user@hadoop.apache.org; core-user@hadoop.apache.org
> Cc: Pete Wyckoff
> Subject: RE: Does Hadoop Honor Reserved Space?
> 
> folks - Jimmy is right - as we have unfortunately hit it as well:
> 
> https://issues.apache.org/jira/browse/HADOOP-1463 caused a regression.
> we have left some comments on the bug - but can't reopen it.
> 
> this is going to be affecting all 0.15 and 0.16 deployments!
> 
> 
> -----Original Message-----
> From: Hairong Kuang [mailto:[EMAIL PROTECTED]
> Sent: Thu 3/6/2008 2:01 PM
> To: core-user@hadoop.apache.org
> Subject: Re: Does Hadoop Honor Reserved Space?
>  
> In addition to the version, could you please send us a copy of the
> datanode
> report by running the command bin/hadoop dfsadmin -report?
> 
> Thanks,
> Hairong
> 
> 
> On 3/6/08 11:56 AM, "Joydeep Sen Sarma" <[EMAIL PROTECTED]> wrote:
> 
>> but intermediate data is stored in a different directory from dfs/data
>> (something like mapred/local by default i think).
>> 
>> what version are u running?
>> 
>> 
>> -----Original Message-----
>> From: Ashwinder Ahluwalia on behalf of [EMAIL PROTECTED]
>> Sent: Thu 3/6/2008 10:14 AM
>> To: core-user@hadoop.apache.org
>> Subject: RE: Does Hadoop Honor Reserved Space?
>>  
>> I've run into a similar issue in the past. From what I understand,
> this
>> parameter only controls the HDFS space usage. However, the
> intermediate data
>> in
>> the map reduce job is stored on the local file system (not HDFS) and
> is not
>> subject to this configuration.
>> 
>> In the past I have used mapred.local.dir.minspacekill and
>> mapred.local.dir.minspacestart to control the amount of space that is
>> allowable
>> for use by this temporary data.
>> 
>> Not sure if that is the best approach though, so I'd love to hear what
> other
>> people have done. In your case, you have a map-red job that will
> consume too
>> much space (without setting a limit, you didn't have enough disk
> capacity for
>> the job), so looking at mapred.output.compress and
> mapred.compress.map.output
>> might be useful to decrease the job's disk requirements.
>> 
>> --Ash
>> 
>> -----Original Message-----
>> From: Jimmy Wan [mailto:[EMAIL PROTECTED]
>> Sent: Thursday, March 06, 2008 9:56 AM
>> To: core-user@hadoop.apache.org
>> Subject: Does Hadoop Honor Reserved Space?
>> 
>> I've got 2 datanodes setup with the following configuration parameter:
>> <property>
>>  <name>dfs.datanode.du.reserved</name>
>>  <value>429496729600</value>
>>  <description>Reserved space in bytes per volume. Always leave this
>> much  
>> space free for non dfs use.
>>  </description>
>> </property>
>> 
>> Both are housed on 800GB volumes, so I thought this would keep about
> half
>> the volume free for non-HDFS usage.
>> 
>> After some long running jobs last night, both disk volumes were
> completely
>> filled. The bulk of the data was in:
>> ${my.hadoop.tmp.dir}/hadoop-hadoop/dfs/data
>> 
>> This is running as the user hadoop.
>> 
>> Am I interpretting these parameters incorrectly?
>> 
>> I noticed this issue, but it is marked as closed:
>> http://issues.apache.org/jira/browse/HADOOP-2549
> 
> 
> 

Reply via email to