Yes you did. Thanks in advance. Is there any way to push this config to all the 
nodes from master or should I make to script to do this?

Regards 

Enviado desde mi iPhone

> El 12 feb. 2017, a las 02:30, Brahma Reddy Battula <bra...@apache.org> 
> escribió:
> 
> Hi Alexis Fidalgo
> 
> 1) I did not seen this query recently
> 
> 2) you need to configure this property in slaves ( DataNode).
> 
> dfs.datanode.du.reserved : The number of bytes will be left free on the 
> volumes used by the DataNodes. By Default,it's zero.
> 
> 
> For example if the disk capacity is 1TB and dfs.datanode.du.reserved 
> configured with 100GB.So DataNode will not use 100GB for block allocation,so 
> this data can be used by nodemanager intermittent files, log files .
> May be you can plan your MR jobs accordingly this. Hope I cleared your doubts.
> 
>> On Sat, Feb 11, 2017 at 7:26 PM, Alexis Fidalgo <alz...@gmail.com> wrote:
>> Hello, i’ve tried to search archives (and google) regarding this issue but 
>> had no luck. After some changes in our mapreduce code, it takes all the 
>> available disk space on datanodes, before this change we hade no problem at 
>> all, but since then, every few days, disks on datanodes (we have 4, all with 
>> same configuration regarding disk, memory, processor, OS) becomes full and 
>> we have no more mapreduce jobs completed. so i need to wipe datanodes and 
>> format namenode and start all over again.
>> 
>> Reading documentation i found this configuration for hdfs-site.xml
>> 
>> <property>
>> <name>dfs.datanode.du.reserved</name>
>> <value>32212254720</value>
>> </property>
>> </configuration>
>> 
>> Questions regarding this
>> 
>> 1. is there any thread already on this issue to read and not to ask again 
>> about it?
>> 2. if not 1, do i need to set up this property only on master or every slave 
>> too?
>> 3. will this fix the problem or just avoid the disk become full but the MR 
>> jobs will fail the same (no more space to work so we need to review our code)
>> 
>> 
>> thanks in advance, sorry if im asking about an already discussed issue, i 
>> just suscribed to the list.
>> 
>> regards
>> 
>> 
>> ---------------------------------------------------------------------
>> To unsubscribe, e-mail: user-unsubscr...@hadoop.apache.org
>> For additional commands, e-mail: user-h...@hadoop.apache.org
>> 
> 
> 
> 
> -- 
> 
> 
> 
> --Brahma Reddy Battula

Reply via email to