On 17/06/2015 20:13, Kenneth Heafield wrote:
> I'll bite.
>
> The moses.ini files ship with bogus feature weights.  One is required to
> tune the system to discover good weights for their system.  You did not
> tune.  The results of an untuned system are meaningless.
>
> So for example if the feature weights are all zeros, then the scores are
> all zero.  The system will arbitrarily pick some awful translation from
> a large space of translations.
>
> The filter looks at one feature p(target | source).  So now you've
> constrained the awful untuned model to a slightly better region of the
> search space.
>
> In other words, all you've done is a poor approximation to manually
> setting the weight to 1.0 on p(target | source) and the rest to 0.
>
> The problem isn't that you are running without a language model (though
> we generally do not care what happens without one).  The problem is that
> you did not tune the feature weights.
>
> Moreover, as Marcin is pointing out, I wouldn't necessarily expect
> tuning to work without an LM.
Tuning does work without a LM. The results aren't half bad. fr-en 
europarl (pb):
   with LM: 22.84
   retuned without LM: 18.33
>
> On 06/17/15 11:56, Read, James C wrote:
>> Actually the approximation I expect to be:
>>
>> p(e|f)=p(f|e)
>>
>> Why would you expect this to give poor results if the TM is well trained? 
>> Surely the results of my filtering experiments provve otherwise.
>>
>> James
>>
>> ________________________________________
>> From: moses-support-boun...@mit.edu <moses-support-boun...@mit.edu> on 
>> behalf of Rico Sennrich <rico.sennr...@gmx.ch>
>> Sent: Wednesday, June 17, 2015 5:32 PM
>> To: moses-support@mit.edu
>> Subject: Re: [Moses-support] Major bug found in Moses
>>
>> Read, James C <jcread@...> writes:
>>
>>> I have been unable to find a logical explanation for this behaviour other
>> than to conclude that there must be some kind of bug in Moses which causes a
>> TM only run of Moses to perform poorly in finding the most likely
>> translations according to the TM when
>>>   there are less likely phrase pairs included in the race.
>> I may have overlooked something, but you seem to have removed the language
>> model from your config, and used default weights. your default model will
>> thus (roughly) implement the following model:
>>
>> p(e|f) = p(e|f)*p(f|e)
>>
>> which is obviously wrong, and will give you poor results. This is not a bug
>> in the code, but a poor choice of models and weights. Standard steps in SMT
>> (like tuning the model weights on a development set, and including a
>> language model) will give you the desired results.
>>
>> _______________________________________________
>> Moses-support mailing list
>> Moses-support@mit.edu
>> http://mailman.mit.edu/mailman/listinfo/moses-support
>>
>> _______________________________________________
>> Moses-support mailing list
>> Moses-support@mit.edu
>> http://mailman.mit.edu/mailman/listinfo/moses-support
> _______________________________________________
> Moses-support mailing list
> Moses-support@mit.edu
> http://mailman.mit.edu/mailman/listinfo/moses-support
>

-- 
Hieu Hoang
Researcher
New York University, Abu Dhabi
http://www.hoang.co.uk/hieu

_______________________________________________
Moses-support mailing list
Moses-support@mit.edu
http://mailman.mit.edu/mailman/listinfo/moses-support

Reply via email to