2015-03-11 19:21 GMT+00:00 Marcin Junczys-Dowmunt <junc...@amu.edu.pl>:
> Maybe someone will correct me, but if I am not wrong, the gziped version > already calculates the future score while loading (i.e. the phrase is being > scored by the language model). The compact phrase table cannot do this > during loading and doing this on-line. This will be the reason for the slow > speed. I suppose your phrase table has not been pruned? So, for instance > function words like "the" can have hundreds of thousands of counterparts > that need to be scored by the LM during collection. > That makes sense. You can limit your phrase table using Barry's prunePhraseTable tool. With > this you can limit it to, say, the 20 best phrases (corresponds to the > ttable limit) and only score this 20 phrases during collection. That should > be orders of magnitude faster. > OK. > Best, > Marcin > > W dniu 11.03.2015 o 20:12, Jesús González Rubio pisze: > > Thanks for the quick response, I will try as you suggest. > > Nevertheless, my main concern is the time spent collecting options. Is > it normal the difference observed respect to the gzip'ed tables? being the > tables cached, shouldn't they be closer? > > 2015-03-11 18:52 GMT+00:00 Marcin Junczys-Dowmunt <junc...@amu.edu.pl>: > >> Hi, >> Try measuring the differences again after a full system reboot (fresh >> reboot before each mesurement) or after purging OS read/write caches. Your >> phrase tables are most likely cached, which means they are in fact in >> memory. >> Best, >> Marcin >> >> W dniu 11.03.2015 o 19:31, Jesús González Rubio pisze: >> >> Hi, >> >> I'm obtaining some unintuitive timing results when using compact phrase >> tables. The average translation time per sentence is much higher for them >> in comparison to using gzip'ed phrase tables. Particularly important is the >> difference in time required to collect the options. This table summarizes >> the timings (in seconds): >> >> Compact Gzip'ed >> on-disk in-memory >> Init: 5.9 6.3 1882.8 >> Per-sentence: >> - Collect: 5.9 5.8 0.2 >> - Search: 1.6 1.6 3.3 >> >> Results in the table were computed using Moses v2.1 with one single >> thread (-th 1) but I've seen similar results using the pre-compiled binary >> for moses v3.0. The model comprises two phrase-tables (~2G and ~3M), two >> lexicalized reordering tables (~700M and ~1M) and two language models (~31G >> and ~38M). You can see the exact configuration in the attached moses.ini >> file. >> >> Interestingly, there is virtually no difference for the compact table >> between the the on-disk and in-memory options. Additionally, timings were >> higher for the initial sentences in both cases which I think should not be >> the case for the in-memory option. >> >> May be the case that the in-memory option of compact tables >> (-minpht-memory -minlexr-memory) is not working properly? >> >> Cheers. >> -- >> Jesús >> >> >> _______________________________________________ >> Moses-support mailing >> listMoses-support@mit.eduhttp://mailman.mit.edu/mailman/listinfo/moses-support >> >> >> >> _______________________________________________ >> Moses-support mailing list >> Moses-support@mit.edu >> http://mailman.mit.edu/mailman/listinfo/moses-support >> >> > > -- Jesús
_______________________________________________ Moses-support mailing list Moses-support@mit.edu http://mailman.mit.edu/mailman/listinfo/moses-support