Re: Nutch crawl problem

2007-12-25 Thread pvvpr
pl.java:39) >> task_0017_m_00_1: at >> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:27) >> task_0017_m_00_1: at >> java.lang.reflect.Constructor.newInstance(Constructor.java:513) >> task_0017_m_00_1: at >> org.apache.

Re: Nutch crawl problem

2007-12-20 Thread pvvpr
I think you need to check the conf/crawl-urlfilter.txt file On Thursday 20 December 2007 04:55, jibjoice wrote: > please, help me to solve it > > jibjoice wrote: > > where i should solve this? why it generated 0 records? > > > > pvvpr wrote: > >> basically

Re: Nutch crawl problem

2007-12-18 Thread pvvpr
basically your indexes are empty since no URLs were generated and fetched. See this, > > - Generator: 0 records selected for fetching, exiting ... > > - Stopping at depth=0 - no more URLs to fetch. > > - No URLs to fetch - check your seed list and URL filters. > > - crawl finished: crawled when

Re: Error when i index by nutchwax-0.10.0

2007-11-23 Thread pvvpr
Dedup tries to acquire lock to the file system from IndexReader.java of lucene. Latest versions of hadoop dont support this. I think this error is because of that. What you can try is comment out acquire lock functionality in lucene IndexReader, compile lucene and replace the old lucene jar with

Re: Error Nutch Search

2007-10-22 Thread pvvpr
Hi jibjoice,   You are using a GNU JVM which might be causing a problem. Type "which java" or "echo $JAVA_HOME" to check which JVM you are using. Using Sun's JVM is a good idea. - Prasad Pingali. On Monday 22 October 2007 05:19, jibjoice wrote: > i can not solve it now, i don't know why? > > ji

Hadoop 0.14.1 Unable to obtain Lock error

2007-10-15 Thread pvvpr
Hi there, I have upgraded from Hadoop 0.12.2 to 0.14.1 and am working on nutch. All seems to be fine except for Dedup jobs fail with "java.io.IOException". I found someone else get this error which is posted here, http://www.nabble.com/Re:-Hadoop-and-Lucene-write-lock-p11827278.html was just won