[ 
http://issues.apache.org/jira/browse/NUTCH-289?page=comments#action_12413996 ] 

Andrzej Bialecki  commented on NUTCH-289:
-----------------------------------------

Re: lookup in ParseOutputFormat: I respectfully disagree. Consider the scenario 
when you run Fetcher in non-parsing mode. This means that you have to make two 
DNS lookups - once when fetching, and the second time when parsing. These 
lookups will be executed from different processes, so there is no benefit from 
caching inside Java resolver, i.e. the process will have to call the DNS server 
twice. The solution I proposed (record IP-s in Fetcher, but somewhere else than 
in ParseOutputFormat, e.g. crawl_fetch CrawlDatum) avoids this problem.

Another issue is virtual hosting, i.e. many sites resolving to a single IP (web 
hotels). It's true that in many cases these are spam sites, but often as not 
they are real, legitimate sites. If we generate/fetch by IP address we run the 
risk of dropping legitimate sites.

Regarding the timing: it's true that during the first run we won't have IP-s 
during generate (and subsequently for any newly injected URLs). In fact, since 
usually a significant part of crawlDB is unfetched we won't have this 
information for many URLs - unless we run this step in Generator to resolve ALL 
hosts, and then run an equivalent of updatedb to actually record them in 
crawldb.

And the last issue that needs to be discussed: should we use metadata, or add a 
dedicated field in CrawlDatum? If the core should rely on IP addresses, we 
should add it as a dedicated field. If it would be purely optional (e.g. for 
the use by optional plugins), then metadata seems a better place.

> CrawlDatum should store IP address
> ----------------------------------
>
>          Key: NUTCH-289
>          URL: http://issues.apache.org/jira/browse/NUTCH-289
>      Project: Nutch
>         Type: Bug

>   Components: fetcher
>     Versions: 0.8-dev
>     Reporter: Doug Cutting

>
> If the CrawlDatum stored the IP address of the host of it's URL, then one 
> could:
> - partition fetch lists on the basis of IP address, for better politeness;
> - truncate pages to fetch per IP address, rather than just hostname.  This 
> would be a good way to limit the impact of domain spammers.
> The IP addresses could be resolved when a CrawlDatum is first created for a 
> new outlink, or perhaps during CrawlDB update.

-- 
This message is automatically generated by JIRA.
-
If you think it was sent incorrectly contact one of the administrators:
   http://issues.apache.org/jira/secure/Administrators.jspa
-
For more information on JIRA, see:
   http://www.atlassian.com/software/jira

Reply via email to