[ https://issues.apache.org/jira/browse/NUTCH-2011?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14547676#comment-14547676 ]
Sebastian Nagel commented on NUTCH-2011: ---------------------------------------- Yes, that's because of the nodeDB feature (the crawl has run before without any problems). Of course, a parsing fetcher needs more heap memory, esp., if there are huge files to parse. I traced the memory usage of the running fetch process: it started with 350 MB and reached 1 MB after 35.000 documents fetched. After that the process slowed down, I had to kill it after 39.000 fetched docs when the fetch speed went down to 0.1 docs / sec. while the process used more and more CPU to run the garbage collector. Roughly, 16kB are required for one FetchNode entry. Seems to be a lot at a first glance, but in fact: URL, title, outlinks + anchors take some space (including the overhead of Java objects). Also, a hash map <Integer, FetchNode> is not the ideal structure to hold a list of consecutively enumerated elements. > Endpoint to support realtime JSON output from the fetcher > --------------------------------------------------------- > > Key: NUTCH-2011 > URL: https://issues.apache.org/jira/browse/NUTCH-2011 > Project: Nutch > Issue Type: Sub-task > Components: fetcher, REST_api > Reporter: Sujen Shah > Assignee: Chris A. Mattmann > Labels: memex > Fix For: 1.11 > > > This fix will create an endpoint to query the Nutch REST service and get a > real-time JSON response of the current/past Fetched URLs. > This endpoint also includes pagination of the output to reduce data transfer > bw in large crawls. -- This message was sent by Atlassian JIRA (v6.3.4#6332)