Check how much memory is allocated for the JVM running namenode.
In a file HADOOP_INSTALL/conf/hadoop-env.sh
you should change a line that starts with export HADOOP_HEAPSIZE=1000
It's set to 1GB by default.
On Fri, Jul 25, 2008 at 2:51 AM, Gert Pfeifer [EMAIL PROTECTED]
wrote:
Update on this one...
I put some more memory in the machine running the name node. Now fsck is
running. Unfortunately ls fails with a time-out.
I identified one directory that causes the trouble. I can run fsck on it
but not ls.
What could be the problem?
Gert
Gert Pfeifer schrieb:
Hi,
I am running a Hadoop DFS on a cluster of 5 data nodes with a name node
and one secondary name node.
I have 1788874 files and directories, 1465394 blocks = 3254268 total.
Heap Size max is 3.47 GB.
My problem is that I produce many small files. Therefore I have a cron
job which just runs daily across the new files and copies them into
bigger files and deletes the small files.
Apart from this program, even a fsck kills the cluster.
The problem is that, as soon as I start this program, the heap space of
the name node reaches 100 %.
What could be the problem? There are not many small files right now and
still it doesn't work. I guess we have this problem since the upgrade to
0.17.
Here is some additional data about the DFS:
Capacity : 2 TB
DFS Remaining : 1.19 TB
DFS Used: 719.35 GB
DFS Used% : 35.16 %
Thanks for hints,
Gert