Hi Marcus,

I do see the same URLs in the logs being fetched multiple times. Verified it by 
reading the sequence file in the content directory through the following piece 
of code.

String segment = 
"/home/IMPETUS/hpirosha/softwares/apache-nutch-1.11/runtime/local/bin/ftalk-db/segments/20160125164343";
File outDir = new 
File("/home/IMPETUS/hpirosha/softwares/apache-nutch-1.11/runtime/local/bin/crawled-content");

        Path file = new Path(segment, Content.DIR_NAME + "/part-00000/data");
        SequenceFile.Reader reader = new SequenceFile.Reader(fs, file, conf);

        Text key = new Text();
        Content content = new Content();

        while (reader.next(key, content)) {
            String filename = key.toString().replaceFirst("http://";, 
"").replaceAll("/", "___").trim();

            File f = new File(outDir.getCanonicalPath() + "/" + filename);
            FileOutputStream fos = new FileOutputStream(f);
            fos.write(content.getContent());
            fos.close();
            System.out.println("URL :"+key.toString());
            System.out.println(f.getAbsolutePath());
        }

Through this code I can see that a web-page is present multiple times in the 
downloaded content.
Please could you give me some pointers where should I look into nutch's code or 
configuration ?

Thanks,
Hussain

________________________________________
From: Markus Jelsma <markus.jel...@openindex.io>
Sent: Monday, January 25, 2016 9:04 PM
To: user@nutch.apache.org
Subject: RE: Webpages are fetched multiple times

Hi - do you see the same URL's written to stdout when fetching? I have see that 
too a few times, but in no case was the URL actually downloaded twice, nor do 
they appear multiple times in the segment or CrawlDB.
Markus



-----Original message-----
> From:Hussain Pirosha <hussain.piro...@impetus.co.in>
> Sent: Monday 25th January 2016 14:30
> To: user@nutch.apache.org
> Subject: Webpages are fetched multiple times
>
> Hello,
>
>
> I have been experimenting with Apache Nutch version 1.11 for few days. My use 
> case is to crawl a forum in local mode. Seed url text just contains one entry 
> :
>
>
> http://www.flyertalk.com/forum/united-airlines-mileageplus/1736400-have-simple-question-about-united-airlines-mileageplus-ask-here-2016-a.html
>
>
> Nutch config is pasted @ http://pasted.co/782e59ad
>
>
> I issue the following commands :-
>
> 1. nutch generate ftalk-db/ ftalk-db/segments/ -depth 5 -topN 500
>
> 2. nutch fetch ftalk-db/segments/20160125154244/
>
>
> I am struggling to find, why nutch keeps on fetching same page multiple 
> times. Instead of getting unique web-pages at the end of crawl, I get lot of 
> duplicates.
>
>
> Please suggest what am I doing wrong ?
>
>
> Thanks,
>
> Hussain
>
> ________________________________
>
>
>
>
>
>
> NOTE: This message may contain information that is confidential, proprietary, 
> privileged or otherwise protected by law. The message is intended solely for 
> the named addressee. If received in error, please destroy and notify the 
> sender. Any use of this email is prohibited when received in error. Impetus 
> does not represent, warrant and/or guarantee, that the integrity of this 
> communication has been maintained nor that the communication is free of 
> errors, virus, interception or interference.
>

________________________________






NOTE: This message may contain information that is confidential, proprietary, 
privileged or otherwise protected by law. The message is intended solely for 
the named addressee. If received in error, please destroy and notify the 
sender. Any use of this email is prohibited when received in error. Impetus 
does not represent, warrant and/or guarantee, that the integrity of this 
communication has been maintained nor that the communication is free of errors, 
virus, interception or interference.

Reply via email to