Re: wrong remaining space reported by Data Nodes

2016-07-26 Thread Ophir Etzion
Thank you very much for response! Do you know if its just reporting badly in that command or will hdfs not have that space available if it needs it? Thank you, Ophir On Fri, Jul 22, 2016 at 9:35 AM, Vinayakumar B wrote: > Hi > > > > You might be hitting

Re: Node Manager crashes with OutOfMemory error

2016-07-26 Thread Ravi Prakash
Hi Rahul! Which version of Hadoop are you using? What non-default values of configuration are you setting? You can set HeapDumpOnOutOfMemoryError on the command line while starting up your nodemanagers and see the resulting heap dump in Eclipse MAT / jvisualvm / yourkit to see where are the

Re: FileUtil.fullyDelete does ?

2016-07-26 Thread praveenesh kumar
https://hadoop.apache.org/docs/r2.7.1/api/org/apache/hadoop/fs/FileUtil.html#fullyDelete(java.io.File) On Tue, Jul 26, 2016 at 12:09 PM, Divya Gehlot wrote: > Resending to right list > -- Forwarded message -- > From: "Divya Gehlot"

Node Manager crashes with OutOfMemory error

2016-07-26 Thread Rahul Chhiber
Hi All, I am running a Hadoop cluster with following configuration :- Master (Resource Manager) - 16GB RAM + 8 vCPU Slave 1 (Node manager 1) - 8GB RAM + 4 vCPU Slave 2 (Node manager 2) - 8GB RAM + 4 vCPU Memory allocated for container use per slave i.e. yarn.nodemanager.resource.memory-mb is