I don’t know of any python process in the Hadoop path that gobbles up that much memory. Would it be possible that you have some kind of memory (–mx) flags in Namenode Options (Probably in HADOOP_OPTS), such that the min and max is set to the same value, say 32GB – so that namenode when it boots up is reserving that much memory?
Thanks Anu From: Dmitry Goldenberg <dgoldenb...@hexastax.com> Date: Tuesday, February 19, 2019 at 11:25 AM To: "user@hadoop.apache.org" <user@hadoop.apache.org> Subject: Memory error during hdfs dfs -format Hi, We've got a task in Ansible which returns a MemoryError during HDFS installation on a box with 64 GB memory total, 30 GB free at the moment. It appears that during the execution of the hdfs dfs -format command, a Python process is spawned which gobbles up ~32ish GB of memory and then the Ansible deploy fails. Any ideas as to how we could curtail / manage memory consumption better? Thanks