Hickey, Larry uttered:
I have a blob structure which is primarily doubles. Is there anyone with some experience with doing data compression to make the blobs smaller?
No experience with compressing blobs...
Tests I have run so far indicate that compression is too slow on blobs of a few meg to be practical. I get now at least 20 to 40 inserts per second but if a single compression takes over a second, it's clearly not worth the trouble. Does anybody have experience with a compression scheme with blobs that consist of mostly arrays of doubles?
If you have lots of common values, such as 0.0 or 1.0, then you may get very good compression. However, if the doubles are the result of arbitrary calculations, that may hit the binary representation of your data, so that two values close to 1.0 may actually be 0.9999999999 and 0.999999999996.
The data may compress better if stored as a textual representation, and be more portable to boot. But if the format is out of your hands, that may not be an option.
Some schemes ( ibsen) offer lightening speed decompression so if the database was primarily used to read, this would be good choice but very expensive to do the compression required to make it.
You don't indicate what compression algorithms you've tried (other than ibsen).
You may find LZO an alternative, as it is designed for high performance (run time) compression. It is GPL, but commercial licenses are available. You might also want to try zlib, which may or may not perform better than ibsen, but has liberal licensing.
If your blobs are big, and your compression is fast, you may have higher performance due to the reduction in IO. But something like bzip2 will likely reduce your performance too much.
Christian -- /"\ \ / ASCII RIBBON CAMPAIGN - AGAINST HTML MAIL X - AGAINST MS ATTACHMENTS / \ ----------------------------------------------------------------------------- To unsubscribe, send email to [EMAIL PROTECTED] -----------------------------------------------------------------------------