Hi Marcus, I do see the same URLs in the logs being fetched multiple times. Verified it by reading the sequence file in the content directory through the following piece of code.
String segment = "/home/IMPETUS/hpirosha/softwares/apache-nutch-1.11/runtime/local/bin/ftalk-db/segments/20160125164343"; File outDir = new File("/home/IMPETUS/hpirosha/softwares/apache-nutch-1.11/runtime/local/bin/crawled-content"); Path file = new Path(segment, Content.DIR_NAME + "/part-00000/data"); SequenceFile.Reader reader = new SequenceFile.Reader(fs, file, conf); Text key = new Text(); Content content = new Content(); while (reader.next(key, content)) { String filename = key.toString().replaceFirst("http://", "").replaceAll("/", "___").trim(); File f = new File(outDir.getCanonicalPath() + "/" + filename); FileOutputStream fos = new FileOutputStream(f); fos.write(content.getContent()); fos.close(); System.out.println("URL :"+key.toString()); System.out.println(f.getAbsolutePath()); } Through this code I can see that a web-page is present multiple times in the downloaded content. Please could you give me some pointers where should I look into nutch's code or configuration ? Thanks, Hussain ________________________________________ From: Markus Jelsma <markus.jel...@openindex.io> Sent: Monday, January 25, 2016 9:04 PM To: user@nutch.apache.org Subject: RE: Webpages are fetched multiple times Hi - do you see the same URL's written to stdout when fetching? I have see that too a few times, but in no case was the URL actually downloaded twice, nor do they appear multiple times in the segment or CrawlDB. Markus -----Original message----- > From:Hussain Pirosha <hussain.piro...@impetus.co.in> > Sent: Monday 25th January 2016 14:30 > To: user@nutch.apache.org > Subject: Webpages are fetched multiple times > > Hello, > > > I have been experimenting with Apache Nutch version 1.11 for few days. My use > case is to crawl a forum in local mode. Seed url text just contains one entry > : > > > http://www.flyertalk.com/forum/united-airlines-mileageplus/1736400-have-simple-question-about-united-airlines-mileageplus-ask-here-2016-a.html > > > Nutch config is pasted @ http://pasted.co/782e59ad > > > I issue the following commands :- > > 1. nutch generate ftalk-db/ ftalk-db/segments/ -depth 5 -topN 500 > > 2. nutch fetch ftalk-db/segments/20160125154244/ > > > I am struggling to find, why nutch keeps on fetching same page multiple > times. Instead of getting unique web-pages at the end of crawl, I get lot of > duplicates. > > > Please suggest what am I doing wrong ? > > > Thanks, > > Hussain > > ________________________________ > > > > > > > NOTE: This message may contain information that is confidential, proprietary, > privileged or otherwise protected by law. The message is intended solely for > the named addressee. If received in error, please destroy and notify the > sender. Any use of this email is prohibited when received in error. Impetus > does not represent, warrant and/or guarantee, that the integrity of this > communication has been maintained nor that the communication is free of > errors, virus, interception or interference. > ________________________________ NOTE: This message may contain information that is confidential, proprietary, privileged or otherwise protected by law. The message is intended solely for the named addressee. If received in error, please destroy and notify the sender. Any use of this email is prohibited when received in error. Impetus does not represent, warrant and/or guarantee, that the integrity of this communication has been maintained nor that the communication is free of errors, virus, interception or interference.