[
https://issues.apache.org/jira/browse/NUTCH-392?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#action_12508812
]
Doğacan Güney commented on NUTCH-392:
-------------------------------------
OK, I have done a bit of testing on compression but I'm stuck. Here it is:
* I changed Content to be a regular Writable instead of a CompressedWritable
and turned on BLOCK compression. Results were pretty impressive. Content size
went down from ~1GB to ~500MB. Unfortunately, I haven't figured out how we can
change Content in a backward compatible way. Reading first byte as version
won't work (because first byte is not version, the first thing written is the
size of the compressed data as int).
* This is where it gets strange. I was trying to test the performance impact of
BLOCK compression (when generating summaries). I fetched a sample 250000 url
segment (a subset of dmoz). Then I made a small modification to
ParseOutputFormat so that it outputs parse_text in all three compression
formats ( http://www.ceng.metu.edu.tr/~e1345172/comp_parse.patch ). After
parsing, segment looks like this:
828M crawl/segments/20070626163143/content
35M crawl/segments/20070626163143/crawl_fetch
23M crawl/segments/20070626163143/crawl_generate
345M crawl/segments/20070626163143/crawl_parse
196M crawl/segments/20070626163143/parse_data
244M crawl/segments/20070626163143/parse_text # NONE
232M crawl/segments/20070626163143/parse_text_block # BLOCK
246M crawl/segments/20070626163143/parse_text_record # RECORD
Not only parse_text_record is larger than parse_text and parse_text_block is
only slightly smaller, but also crawl_parse is larger than any of them!
I probably messed up somewhere and I can't see it. Any help would be welcome.
> OutputFormat implementations should pass on Progressable
> --------------------------------------------------------
>
> Key: NUTCH-392
> URL: https://issues.apache.org/jira/browse/NUTCH-392
> Project: Nutch
> Issue Type: New Feature
> Components: fetcher
> Reporter: Doug Cutting
> Assignee: Andrzej Bialecki
> Fix For: 1.0.0
>
> Attachments: NUTCH-392.patch
>
>
> OutputFormat implementations should pass the Progressable they are passed to
> underlying SequenceFile implementations. This will keep reduce tasks from
> timing out when block writes are slow. This issue depends on
> http://issues.apache.org/jira/browse/HADOOP-636.
--
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.
-------------------------------------------------------------------------
This SF.net email is sponsored by DB2 Express
Download DB2 Express C - the FREE version of DB2 express and take
control of your XML. No limits. Just data. Click to get it now.
http://sourceforge.net/powerbar/db2/
_______________________________________________
Nutch-developers mailing list
[email protected]
https://lists.sourceforge.net/lists/listinfo/nutch-developers