pl.java:39)
>> task_0017_m_00_1: at
>> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:27)
>> task_0017_m_00_1: at
>> java.lang.reflect.Constructor.newInstance(Constructor.java:513)
>> task_0017_m_00_1: at
>> org.apache.
I think you need to check the conf/crawl-urlfilter.txt file
On Thursday 20 December 2007 04:55, jibjoice wrote:
> please, help me to solve it
>
> jibjoice wrote:
> > where i should solve this? why it generated 0 records?
> >
> > pvvpr wrote:
> >> basically
basically your indexes are empty since no URLs were generated and fetched. See
this,
> > - Generator: 0 records selected for fetching, exiting ...
> > - Stopping at depth=0 - no more URLs to fetch.
> > - No URLs to fetch - check your seed list and URL filters.
> > - crawl finished: crawled
when
Dedup tries to acquire lock to the file system from IndexReader.java of
lucene. Latest versions of hadoop dont support this. I think this error is
because of that. What you can try is comment out acquire lock functionality
in lucene IndexReader, compile lucene and replace the old lucene jar with
Hi jibjoice,
You are using a GNU JVM which might be causing a problem. Type "which java"
or "echo $JAVA_HOME" to check which JVM you are using. Using Sun's JVM is a
good idea.
- Prasad Pingali.
On Monday 22 October 2007 05:19, jibjoice wrote:
> i can not solve it now, i don't know why?
>
> ji
Hi there,
I have upgraded from Hadoop 0.12.2 to 0.14.1 and am working on nutch. All
seems to be fine except for Dedup jobs fail with "java.io.IOException".
I found someone else get this error which is posted here,
http://www.nabble.com/Re:-Hadoop-and-Lucene-write-lock-p11827278.html
was just won