Check how much memory is allocated for the JVM running namenode.

In a file HADOOP_INSTALL/conf/hadoop-env.sh
you should change a line that starts with "export HADOOP_HEAPSIZE=1000"

It's set to 1GB by default.


On Fri, Jul 25, 2008 at 2:51 AM, Gert Pfeifer <[EMAIL PROTECTED]>
wrote:

> Update on this one...
>
> I put some more memory in the machine running the name node. Now fsck is
> running. Unfortunately ls fails with a time-out.
>
> I identified one directory that causes the trouble. I can run fsck on it
> but not ls.
>
> What could be the problem?
>
> Gert
>
> Gert Pfeifer schrieb:
>
> Hi,
>> I am running a Hadoop DFS on a cluster of 5 data nodes with a name node
>> and one secondary name node.
>>
>> I have 1788874 files and directories, 1465394 blocks = 3254268 total.
>> Heap Size max is 3.47 GB.
>>
>> My problem is that I produce many small files. Therefore I have a cron
>> job which just runs daily across the new files and copies them into
>> bigger files and deletes the small files.
>>
>> Apart from this program, even a fsck kills the cluster.
>>
>> The problem is that, as soon as I start this program, the heap space of
>> the name node reaches 100 %.
>>
>> What could be the problem? There are not many small files right now and
>> still it doesn't work. I guess we have this problem since the upgrade to
>> 0.17.
>>
>> Here is some additional data about the DFS:
>> Capacity         :       2 TB
>> DFS Remaining   :       1.19 TB
>> DFS Used        :       719.35 GB
>> DFS Used%       :       35.16 %
>>
>> Thanks for hints,
>> Gert
>>
>
>

Reply via email to