[ https://issues.apache.org/jira/browse/NUTCH-1247?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13185908#comment-13185908 ]
Andrzej Bialecki commented on NUTCH-1247: ------------------------------------------ Originally the reason for a byte was compactness, but we can get the same effect using vint. Markus, something seems off in your setup if you get such high values of retries ... usually CrawlDbReducer will set STATUS_DB_GONE if the number of retries reaches db.fetch.retry.max, so the page will not be tried again until FetchSchedule.forceRefetch resets its status (and the number of retries). > CrawlDatum.retries should be int > -------------------------------- > > Key: NUTCH-1247 > URL: https://issues.apache.org/jira/browse/NUTCH-1247 > Project: Nutch > Issue Type: Bug > Affects Versions: 1.4 > Reporter: Markus Jelsma > Fix For: 1.5 > > > CrawlDatum.retries is a byte and goes bad with larger values. > 12/01/12 18:35:22 INFO crawl.CrawlDbReader: retry -127: 1 > 12/01/12 18:35:22 INFO crawl.CrawlDbReader: retry -128: 1 -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira