I am not at all a Nutch expert, I am just experimenting a little bit,
but as far as I understood it
you can remove the indexes directory and re-index again the segments:
In may case ofter step 8 of the (see below) I have only one segment :
test/segments/20060522144050
after step 9 I will have a second segment
test/segments/20060522144050
Now what we can do is to remove the test/indexes directory and
re-index the two segments:
this what I did :
hadoop dfs -rm test/indexes
nutch index test/indexes test/crawldb linkdb
test/segments/20060522144050 test/segments/20060522144050
Hope it helps
-Corrqdo
Jacob Brunson wrote:
I looked at the referenced messaged at
http://www.mail-archive.com/nutch-user@lucene.apache.org/msg03990.html
but I am still having problems.
I am running the latest checkout from subversion.
These are the commands which I've run:
bin/nutch crawl myurls/ -dir crawl -threads 4 -depth 3 -topN 10000
bin/nutch generate crawl/crawldb crawl/segments -topN 500
lastsegment=`ls -d crawl/segments/2* | tail -1`
bin/nutch fetch $lastsegment
bin/nutch updatedb crawl/crawldb $lastsegment
bin/nutch index crawl/indexes crawl/crawldb crawl/linkdb $lastsegment
This last command fails with a java.io.IOException saying: "Output
directory /home/nutch/nutch/crawl/indexes already exists"
So I'm confused because it seems like I did exactly what was described
in the referenced email, but it didn't work for me. Can someone help
me figure out what I'm doing wrong or what I need to do instead?
Thanks,
Jacob
On 5/22/06, sudhendra seshachala <[EMAIL PROTECTED]> wrote:
Please do follow the link below..
http://www.mail-archive.com/nutch-user@lucene.apache.org/msg03990.html
I have been able to follow the threads as explained and merge
multiple crawl.. It works like a champ.
Thanks
Sudhi
zzcgiacomini <[EMAIL PROTECTED]> wrote:
I am currently using the last nightly nutch-0.8-dev build and
I am really confused about how to proceed after I have done two
different "whole web" incremental crawl
The tutorial to me is not clear on how to merge the results after the
two crawls in order to be able to
make a search operation.
Could some one please give me an Hints on what is the right procedure ?!
here is what I am doing:
1. create an initial urls file /tmp/dmoz/urls.txt
2. hadoop dfs -put /tmp/urls/ url
3. nutch inject test/crawldb dmoz
4. nutch generate test/crawldb test/segments
5. nutch fetch test/segments/20060522144050
6. nutch updatedb test/crawldb test/segments/20060522144050
7. nutch invertlinks linkdb test/segments/20060522144050
8. nutch index test/indexes test/crawldb linkdb
test/segments/20060522144050
..and now I am able to search...
Now I run
9. nutch generate test/crawldb test/segments -topN 1000
and I will end up to have a new segment : test/segments/20060522151957
10. nutch fetch test/segments/20060522151957
11. nutch updatedb test/crawldb test/segments/20060522151957
From this point on I cannot make any progresses much
A) I have tried to merge the two segments into a new one with the
idea to rerun an invertlinks and index on it but:
nutch mergesegs test/segments -dir test/segments
whatever I specify as outputdir or outputsegment I get errors
B) I have also tried to make invertlinks on all test/segments with
the goal to run nutch index command to produce a second
indexes directory, let say test/indexes1, an finally run the merge
index on index2
nutch invertlinks test/linkdb -dir test/segments
This as created a new linkdb directory *NOT* under test as specified
but as /linkdb-1108390519
nutch index test/indexes1 test/crawldb linkdb
test/segments/20060522144050
nutch merge index2 test/indexes test/indexes1
now I am not sure what to do; If I rename test/index2 to be
test/indexes after having removed test/indexes
I will not able to search anymore.
-Corrado
Sudhi Seshachala
http://sudhilogs.blogspot.com/
__________________________________________________
Do You Yahoo!?
Tired of spam? Yahoo! Mail has the best spam protection around
http://mail.yahoo.com