Hello ajb,

Wednesday, July 11, 2007, 7:55:22 AM, you wrote:
> Not really.  LZW is basically PPM with a static (and flat!) frequency
> prediction model.  The contexts are build in a very similar way.

what you mean by "flat" and "static" applied to PPM? static PPM models
exist - they carry probabilities as separate table very like static
Huffman encoding. is "flat" the same as order-0?

>> the devil in details. just imagine size of huffman table with 64k
>> entries :)

> If you're allowed to pick your code values, you can do this extremely
> compactly.  Some compression schemes optimised for English separates
> words from the intermediate space and assigns a Huffman code to each
> word.  The tables aren't that big (though the space to hold the words
> might be).

can you give a link? i never heard about such algorithm


-- 
Best regards,
 Bulat                            mailto:[EMAIL PROTECTED]

_______________________________________________
Haskell-Cafe mailing list
Haskell-Cafe@haskell.org
http://www.haskell.org/mailman/listinfo/haskell-cafe

Reply via email to