Jesse Hires wrote:
I tried asking this over at the nutch-user alias, but I am seeing very
little traction, so I thought I'd ask the developers. I realize this is
most likely a configuration problem on my end, but I am very new to
using nutch, so I am having a difficult time understanding where I need
to look.
Does anyone have any insight into the following error I am seeing in the
hadoop logs? Is this something I should be concerned with, or is it
expected that this shows up in the logs from time to time? If it is not
expected, where can I look for more information on what is going on?
It's not expected at all - this usually indicates some config error, or
FS corruption, or it may be also caused by conflicting DNS (e.g. the
same name resolving to different addresses on different nodes), or a
problem with permissions (e.g. daemon started remotely uses
uid/permissions/env that doesn't allow it to create/delete files in data
dir). This may be also some weird corner case when processes run out of
file descriptors - you should check ulimit -n and set it to a value
higher than 4096.
Please also run fsck / and see what it says.
I can also provide config files if needed.
We need just the modifications in hadoop-site.xml, that's where the
problem may be located.
--
Best regards,
Andrzej Bialecki <><
___. ___ ___ ___ _ _ __________________________________
[__ || __|__/|__||\/| Information Retrieval, Semantic Web
___|||__|| \| || | Embedded Unix, System Integration
http://www.sigram.com Contact: info at sigram dot com