Oh, yes, of course. That's in build_binary.
> On May 13, 2016, at 4:39 PM, kellen sunderland <kellen.sunderl...@gmail.com> > wrote: > > Could we also use quantization with the language model to reduce the size? > KenLM supports this right? > > On Fri, May 13, 2016 at 1:19 PM, Matt Post <p...@cs.jhu.edu> wrote: > >> Great idea, hadn't thought of that. >> >> I think we could also get some leverage out of: >> >> - Reducing the language model to a 4-gram one >> - Doing some filtering of the phrase table to reduce low-probability >> translation options >> >> These would be a bit lossier but I doubt it would matter much at all. >> >> matt >> >> >>> On May 13, 2016, at 4:02 PM, Tom Barber <t...@analytical-labs.com> wrote: >>> >>> Out of curiosity more than anything else I tested XZ compression on a >> model >>> instead of Gzip, it takes the Spain pack down from 1.9GB to 1.5GB, not >> the >>> most ever, but obviously does mean 400MB+ less in remote storage and data >>> going over the wire. >>> >>> Worth considering I guess. >>> >>> Tom >>> -------------- >>> >>> Director Meteorite.bi - Saiku Analytics Founder >>> Tel: +44(0)5603641316 >>> >>> (Thanks to the Saiku community we reached our Kickstart >>> < >> http://kickstarter.com/projects/2117053714/saiku-reporting-interactive-report-designer/ >>> >>> goal, but you can always help by sponsoring the project >>> <http://www.meteorite.bi/products/saiku/sponsorship>) >> >>