Hi Mark,

I have checked that the weight of the laguage model is -0.000345 in
the tuned moses.ini,
which is almost the same as you anaylzed :-)

I have replaced the source langauge model in moses.ini file generated
by the training  step with the in-domain target langauge model, and
then went on to do the tuning and evaluation steps again. Hope this
can get its quality to be better.

I have read the Moses manual again, it seems the training steps
actually has no releationship with the langauge model, so it should be
OK if I replace the language model with the right one in the moses.ini
to proceed the tuning steps, right?

I also used the in-domain language model this time (the traning step
was performed by using both in-domain and out-domain training data),
because I saw one paper said that the in-domain LM will help to
improve the quality for our domain, hope this helps.

Thank you Mark ;-)


2010/8/11, Mark Fishel <fis...@ut.ee>:
> Hi,
>
> I would guess that your tuning (i.e. MERT) would push the weight of
> the language model towards 0, thus kind of translating without a
> language model. You can check the LM weight from the resulting
> moses.ini file.
>
> As for the quality of the engine, during translation the Spanish
> tokens would mostly be unknown to the language model, which means that
> for most n-grams the language model would give you the same weight
> (i.e. the weight of the p(UNK|UNK, ..., UNK) n-gram), which again means
> that this turns into translation without a language model.
>
> I'd suggest simply to train a correct language model and repeat the
> MERT and evaluation steps, unless you are interested in your setup as
> a researcher :)
>
> Best,
> Mark
>
> On Tue, Aug 10, 2010 at 5:04 PM, Wenlong Yang <welons...@gmail.com> wrote:
>> Hi,
>>
>> I have trained a Moses Engine for the language pair English to
>> Spanish, but for the training script, during below step:
>> "nohup nice
>> tools/moses-scripts/scripts-20100203-0505/training/train-factored-phrase-model.perl
>> -scripts-root-dir tools/moses-scripts/scripts-20100203-0505 -root-dir
>> Work-SP -corpus Work-SP/corpus/news-commentary.lowercased -f en -e sp
>> -alignment grow-diag-final-and -reordering msd-bidirectional-fe  -lm
>> 0:3:/home/user/Work-SP/lm/news-commentary.lm >& Work-SP/training.out
>> &"
>>
>> The "news-commentary.lm" in above command line should be the language
>> model of the target language, but by mistake, I was using the language
>> model for the source language, then I went on to finish the tuning and
>> evaluation steps for this engine.
>>
>> The weird thing is that this trained engine still can work for English
>> to Spanish, but I am doubting that its quality is getting lower
>> because of the wrong language model used.
>>
>> Can any of you help to analyze what will be influenced by the wrongly
>> used language model? Will it influence the engine's quality?
>>
>> Thanks.
>> _______________________________________________
>> Moses-support mailing list
>> Moses-support@mit.edu
>> http://mailman.mit.edu/mailman/listinfo/moses-support
>>
>
_______________________________________________
Moses-support mailing list
Moses-support@mit.edu
http://mailman.mit.edu/mailman/listinfo/moses-support

Reply via email to