Re: [Moses-support] Tuning crashed

2016-06-08 Thread Barry Haddow

Hi Tomasz

The error message about missing the ini file is a consequence of the 
tuning crash, so just ignore this.


To find out why Moses is failing, run it again in the console like this:

/home/moses/src/mosesdecoder/bin/moses -threads 16 -v 0  -config 
/home/moses/working/experiments/NGRAM5/model/moses.bin.ini.2 -show-weights


If necessary, increase the verbosity of the debug (-v).

cheers - Barry

On 08/06/16 08:42, Tomasz Gawryl wrote:


Hi,

I have a problem with tuning crashed. It seems that moses.ini is 
missing in temporary folder but I have no idea why. I attached link to 
my config file.


Please help.

Regards

Thomas

moses@smtserver:~/working/experiments/NGRAM5/steps/2$ more 
TUNING_tune.2.STDERR


Using SCRIPTS_ROOTDIR: /home/moses/src/mosesdecoder/scripts

Asking moses for feature names and values from 
/home/moses/working/experiments/NGRAM5/model/moses.bin.ini.2


Executing: /home/moses/src/mosesdecoder/bin/moses -threads 16 -v 0 
-config /home/moses/working/experiments/NGRAM5/model/moses.bin.ini.2 
-show-weights


exec: /home/moses/src/mosesdecoder/bin/moses -threads 16 -v 0 -config 
/home/moses/working/experiments/NGRAM5/model/moses.bin.ini.2 -show-weights


Executing: /home/moses/src/mosesdecoder/bin/moses -threads 16 -v 0 
-config /home/moses/working/experiments/NGRAM5/model/moses.bin.ini.2 
-show-weights > ./features.list 2> /dev/null


Exit code: 1

ERROR: Failed to run '/home/moses/src/mosesdecoder/bin/moses -threads 
16 -v 0 -config 
/home/moses/working/experiments/NGRAM5/model/moses.bin.ini.2 
-show-weights'. at 
/home/moses/src/mosesdecoder/scripts/training/mert-moses.pl line 1748.


cp: cannot stat 
‘/home/moses/working/experiments/NGRAM5/tuning/tmp.2/moses.ini’: No 
such file or directory


https://docs.google.com/document/d/1gI7YVUx8VoktIfIQvvU54jKSm5Ta6UZjYcszBFtP-V8/edit?usp=sharing



___
Moses-support mailing list
Moses-support@mit.edu
http://mailman.mit.edu/mailman/listinfo/moses-support


The University of Edinburgh is a charitable body, registered in
Scotland, with registration number SC005336.
___
Moses-support mailing list
Moses-support@mit.edu
http://mailman.mit.edu/mailman/listinfo/moses-support


[Moses-support] Tuning crashed

2016-06-08 Thread Tomasz Gawryl
Hi, 

 

I have a problem with tuning crashed. It seems that moses.ini is missing in
temporary folder but I have no idea why. I attached link to my config file.

Please help.

 

Regards

Thomas

 

moses@smtserver:~/working/experiments/NGRAM5/steps/2$ more
TUNING_tune.2.STDERR

Using SCRIPTS_ROOTDIR: /home/moses/src/mosesdecoder/scripts

Asking moses for feature names and values from
/home/moses/working/experiments/NGRAM5/model/moses.bin.ini.2

Executing: /home/moses/src/mosesdecoder/bin/moses -threads 16 -v 0  -config
/home/moses/working/experiments/NGRAM5/model/moses.bin.ini.2 -show-weights

exec: /home/moses/src/mosesdecoder/bin/moses -threads 16 -v 0  -config
/home/moses/working/experiments/NGRAM5/model/moses.bin.ini.2 -show-weights

Executing: /home/moses/src/mosesdecoder/bin/moses -threads 16 -v 0  -config
/home/moses/working/experiments/NGRAM5/model/moses.bin.ini.2 -show-weights >
./features.list 2> /dev/null

Exit code: 1

ERROR: Failed to run '/home/moses/src/mosesdecoder/bin/moses -threads 16 -v
0  -config /home/moses/working/experiments/NGRAM5/model/moses.bin.ini.2
-show-weights'. at
/home/moses/src/mosesdecoder/scripts/training/mert-moses.pl line 1748.

cp: cannot stat
'/home/moses/working/experiments/NGRAM5/tuning/tmp.2/moses.ini': No such
file or directory

 

 

https://docs.google.com/document/d/1gI7YVUx8VoktIfIQvvU54jKSm5Ta6UZjYcszBFtP
-V8/edit?usp=sharing

 

 

 

___
Moses-support mailing list
Moses-support@mit.edu
http://mailman.mit.edu/mailman/listinfo/moses-support


Re: [Moses-support] Tuning crashed

2014-05-24 Thread Wei Qiu
Thanks, Barry,

Both batches-mira and pro tuning don't crash. But the result of
batch-miratuning is very unstable due to the size of dev set.

Best,
Wei



On Sat, May 24, 2014 at 11:34 AM, Wei Qiu  wrote:

> Thanks, Barry. I will try mira.
>
> Anyway, I share my model file here
>
> https://www.dropbox.com/s/jko7aitnk9mqa8a/model.tar.gz
>
> Best,
> Wei
>
>
> On Sat, May 24, 2014 at 10:13 AM, Barry Haddow  > wrote:
>
>>  Hi Wei
>>
>> This error (as far as I recall) tends to mean that one of your features
>> is flat (ie uninformative). I don't see that this is the case for your
>> features. The only thing I could think of is that your tuning set is quite
>> small (500 sentences) but that doesn't usually cause a problem.
>>
>> My suggestion would be to try tuning with batch mira or pro.
>>
>> cheers - Barry
>>
>>
>> On 24/05/14 08:31, Wei Qiu wrote:
>>
>> Hi,
>>
>>  I tried to train a phrase-based baseline system from German to English.
>>
>>  But during the tuning step, mert crashed in the first run. I tracked it
>> to this error,
>>
>>  Check abs(leftmost->first-gradient.rbegin()->first) < 0.0001 failed in
>> mert/Optimizer.cpp:170
>>
>>  I've already checked similar posts on the mailing list, but my run1. outand 
>> run1. Scores
>> .data look alright.
>>
>>  The tuning directory is shared here:
>> https://www.dropbox.com/s/j6tflbtft3uqvuu/tmp.1.tar.gz
>>
>>  The model directory will be shared as well(it's still uploading).
>>
>>  Thanks in advance.
>>
>>  Best,
>> Wei Qiu
>>
>>
>>
>>
>> ___
>> Moses-support mailing 
>> listMoses-support@mit.eduhttp://mailman.mit.edu/mailman/listinfo/moses-support
>>
>>
>>
>
___
Moses-support mailing list
Moses-support@mit.edu
http://mailman.mit.edu/mailman/listinfo/moses-support


Re: [Moses-support] Tuning crashed

2014-05-24 Thread Wei Qiu
Thanks, Barry. I will try mira.

Anyway, I share my model file here

https://www.dropbox.com/s/jko7aitnk9mqa8a/model.tar.gz

Best,
Wei


On Sat, May 24, 2014 at 10:13 AM, Barry Haddow
wrote:

>  Hi Wei
>
> This error (as far as I recall) tends to mean that one of your features is
> flat (ie uninformative). I don't see that this is the case for your
> features. The only thing I could think of is that your tuning set is quite
> small (500 sentences) but that doesn't usually cause a problem.
>
> My suggestion would be to try tuning with batch mira or pro.
>
> cheers - Barry
>
>
> On 24/05/14 08:31, Wei Qiu wrote:
>
> Hi,
>
>  I tried to train a phrase-based baseline system from German to English.
>
>  But during the tuning step, mert crashed in the first run. I tracked it
> to this error,
>
>  Check abs(leftmost->first-gradient.rbegin()->first) < 0.0001 failed in
> mert/Optimizer.cpp:170
>
>  I've already checked similar posts on the mailing list, but my run1. outand 
> run1. Scores
> .data look alright.
>
>  The tuning directory is shared here:
> https://www.dropbox.com/s/j6tflbtft3uqvuu/tmp.1.tar.gz
>
>  The model directory will be shared as well(it's still uploading).
>
>  Thanks in advance.
>
>  Best,
> Wei Qiu
>
>
>
>
> ___
> Moses-support mailing 
> listMoses-support@mit.eduhttp://mailman.mit.edu/mailman/listinfo/moses-support
>
>
>
___
Moses-support mailing list
Moses-support@mit.edu
http://mailman.mit.edu/mailman/listinfo/moses-support


Re: [Moses-support] Tuning crashed

2014-05-24 Thread Barry Haddow

Hi Wei

This error (as far as I recall) tends to mean that one of your features 
is flat (ie uninformative). I don't see that this is the case for your 
features. The only thing I could think of is that your tuning set is 
quite small (500 sentences) but that doesn't usually cause a problem.


My suggestion would be to try tuning with batch mira or pro.

cheers - Barry

On 24/05/14 08:31, Wei Qiu wrote:

Hi,

I tried to train a phrase-based baseline system from German to English.

But during the tuning step, mert crashed in the first run. I tracked 
it to this error,


Check abs(leftmost->first-gradient.rbegin()->first) < 0.0001 failed in 
mert/Optimizer.cpp:170


I've already checked similar posts on the mailing list, but my run1. 
out and run1. Scores.data look alright.


The tuning directory is shared here:
https://www.dropbox.com/s/j6tflbtft3uqvuu/tmp.1.tar.gz

The model directory will be shared as well(it's still uploading).

Thanks in advance.

Best,
Wei Qiu




___
Moses-support mailing list
Moses-support@mit.edu
http://mailman.mit.edu/mailman/listinfo/moses-support


___
Moses-support mailing list
Moses-support@mit.edu
http://mailman.mit.edu/mailman/listinfo/moses-support


[Moses-support] Tuning crashed

2014-05-24 Thread Wei Qiu
Hi,

I tried to train a phrase-based baseline system from German to English.

But during the tuning step, mert crashed in the first run. I tracked it to
this error,

Check abs(leftmost->first-gradient.rbegin()->first) < 0.0001 failed in mert
/Optimizer.cpp:170

I've already checked similar posts on the mailing list, but my run1.
outand run1. Scores
.data look alright.

The tuning directory is shared here:
https://www.dropbox.com/s/j6tflbtft3uqvuu/tmp.1.tar.gz

The model directory will be shared as well(it's still uploading).

Thanks in advance.

Best,
Wei Qiu
___
Moses-support mailing list
Moses-support@mit.edu
http://mailman.mit.edu/mailman/listinfo/moses-support


Re: [Moses-support] Tuning crashed, mert.out - no such file or directory. intial BLEU=0

2013-06-13 Thread jayendra rakesh
Thank you barry that was most helpful.

I was running the EMS for factored models(POS), there have been few
mistakes on my part while giving the inputs properly, which I have
corrected. The EMS now has run fine.

I gave a factorized reference file (target language) where as the output is
just surface form of words, because of which the BLEU score is initialized
to zero.

Similar thing with the language model input, instead of plain file I have
given a factorized file as input for language model leading to a lot of
OOVs.

Though the final BLEU score seems to be lower than the base run with just
surface forms.

Thanking you,
Y. Jayendra Rakesh,
B.Tech. Computer Science Dual Degree,
LTRC, IIIT Hyderabad.


On Sun, Jun 9, 2013 at 10:45 PM, Barry Haddow wrote:

> Hi Jayendra
>
> The fact that your bleu precisions are all zero (run1.scores.dat) suggests
> that something went badly wrong
> with the translation, or your reference set is incorrect. What do the
> translations (in run1.out) look like?
>
> The zeros in run1.features.dat are normal, but the very low lm scores (eg
> -1804.31 in first row) are not. This
> suggests that all the tokens in your input sentences are OOVs. Check your
> setup to make sure that you have not,
> for example, swapped the source and target language files accidentally at
> some point,
>
> cheers - Barry
>
>
> Quoting jayendra rakesh  on Sun, 9 Jun 2013
> 00:08:10 +0530:
>
>  Sorry for bumping the thread again.
>> here's what I found after running the command by hand as suggested.
>>
>> run1.init.opt:
>> * 0.30 0.30 0.30 0.30 0.30 0.30 0.30 0.50
>>
>> -1.00 0.20 0.20 0.20 0.20 0.20
>>  0 0 0 0 0 0 0 0 0 0 0 0 0 0
>>  1 1 1 1 1 1 1 1 1 1 1 1 1 1
>> *
>> Things noticed.
>> *run1.features.dat has columns 3,6 all  zeroes
>> *
>> *run1.scores.dat has columns 1,3,5,7 all zeroes
>> *
>>
>> run1.features.dat:
>> *FEATURES_TXT_BEGIN_0 0 100 14 d_0 d_1 d_2 d_3 d_4 d_5 d_6 lm_0 w_0 tm_0
>>
>> tm_1 tm_2 tm_3 tm_4
>> -8 -16.1621 0 -1.28775 -23.5316 0 -1.16898 -1804.31 -18 -79.6613 -80.0959
>> -48.9898 -27.8995 17.9981
>> 0 -20.6133 0 0 -27.4523 0 0 -1804.31 -18 -81.3889 -80.0948 -50.4425
>> -27.871
>> 17.9981
>> -8 -16.9628 0 -1.28775 -22.4881 0 -1.16898 -1804.31 -18 -77.4669 -80.9714
>> -49.7377 -28.9981 17.9981
>> 0 -21.414 0 0 -26.4089 0 0 -1804.31 -18 -79.1945 -80.9702 -51.1904
>> -28.9696
>> 17.9981
>> -8 -16.8088 0 -1.28775 -23.0796 0 -1.16898 -1804.31 -18 -78.6891 -79.6433
>> -50.0519 -28.2473 17.9981
>> -8 -16.1621 0 -1.28775 -23.5316 0 -1.16898 -1804.31 -18 -79.1464 -79.9766
>> -48.9898 -28.83 17.9981*
>>
>>
>> run1.scores.dat
>>
>> *
>> SCORES_TXT_BEGIN_0 0 100 9 BLEU
>> 0 18 0 17 0 16 0 15 31
>> 0 18 0 17 0 16 0 15 31
>> 0 18 0 17 0 16 0 15 31
>> 0 18 0 17 0 16 0 15 31
>> 0 18 0 17 0 16 0 15 31*
>>
>> *Issue: in the MIRA run BLEU score is initialized to zero
>>
>> *
>> Hoping the information would help in solving the issue.
>>
>> Thanking you,
>> Jayendra Rakesh.
>>
>>
>> On Wed, Jun 5, 2013 at 1:52 PM, jayendra rakesh
>> **wrote:
>>
>>  Hi Phi,
>>>
>>> Thanks for the reply, I made a hand run of the command as you have
>>> suggested and was able to repeat the crash. I have checked run1.init.opt
>>> file, it seems to be fine
>>>
>>> run1.init.opt:
>>> * 0.30 0.30 0.30 0.30 0.30 0.30 0.30 0.50
>>>
>>> -1.00 0.20 0.20 0.20 0.20 0.20
>>>  0 0 0 0 0 0 0 0 0 0 0 0 0 0
>>>  1 1 1 1 1 1 1 1 1 1 1 1 1 1
>>> *
>>> But I noticed a thing in both run1.scores.dat and run1.features.dat
>>> files,
>>> that many *features*(*columns 3,6 are all zero)* and *scores*(*columns
>>> 1,3,5,7 are all zero*) are having zero values, because of which i think
>>>
>>> BLEU score is initialized to zero in mert.log
>>>
>>> run1.features.dat:
>>> *FEATURES_TXT_BEGIN_0 0 100 14 d_0 d_1 d_2 d_3 d_4 d_5 d_6 lm_0 w_0 tm_0
>>>
>>> tm_1 tm_2 tm_3 tm_4
>>> -8 -16.1621 0 -1.28775 -23.5316 0 -1.16898 -1804.31 -18 -79.6613 -80.0959
>>> -48.9898 -27.8995 17.9981
>>> 0 -20.6133 0 0 -27.4523 0 0 -1804.31 -18 -81.3889 -80.0948 -50.4425
>>> -27.871 17.9981
>>> -8 -16.9628 0 -1.28775 -22.4881 0 -1.16898 -1804.31 -18 -77.4669 -80.9714
>>> -49.7377 -28.9981 17.9981
>>> 0 -21.414 0 0 -26.4089 0 0 -1804.31 -18 -79.1945 -80.9702 -51.1904
>>> -28.9696 17.9981
>>> -8 -16.8088 0 -1.28775 -23.0796 0 -1.16898 -1804.31 -18 -78.6891 -79.6433
>>> -50.0519 -28.2473 17.9981
>>> -8 -16.1621 0 -1.28775 -23.5316 0 -1.16898 -1804.31 -18 -79.1464 -79.9766
>>> -48.9898 -28.83 17.9981*
>>>
>>>
>>> run1.scores.dat
>>>
>>> *
>>> SCORES_TXT_BEGIN_0 0 100 9 BLEU
>>> 0 18 0 17 0 16 0 15 31
>>> 0 18 0 17 0 16 0 15 31
>>> 0 18 0 17 0 16 0 15 31
>>> 0 18 0 17 0 16 0 15 31
>>> 0 18 0 17 0 16 0 15 31*
>>>
>>>
>>>
>>> Hoping the information would help in pointing out the issue.
>>>
>>> Regards,
>>> Jayendra Rakesh
>>>
>>>
>>>
>>> On Sun, Jun 2, 2013 at 5:19 PM, Philipp Koehn 
>>> wrote:
>>>
>>>  Hi,

 I have been running kbest MIRA with factored 

Re: [Moses-support] Tuning crashed, mert.out - no such file or directory. intial BLEU=0

2013-06-09 Thread Barry Haddow
Hi Jayendra

The fact that your bleu precisions are all zero (run1.scores.dat)  
suggests that something went badly wrong
with the translation, or your reference set is incorrect. What do the  
translations (in run1.out) look like?

The zeros in run1.features.dat are normal, but the very low lm scores  
(eg -1804.31 in first row) are not. This
suggests that all the tokens in your input sentences are OOVs. Check  
your setup to make sure that you have not,
for example, swapped the source and target language files accidentally  
at some point,

cheers - Barry

Quoting jayendra rakesh  on Sun, 9 Jun 2013  
00:08:10 +0530:

> Sorry for bumping the thread again.
> here's what I found after running the command by hand as suggested.
>
> run1.init.opt:
> * 0.30 0.30 0.30 0.30 0.30 0.30 0.30 0.50
> -1.00 0.20 0.20 0.20 0.20 0.20
>  0 0 0 0 0 0 0 0 0 0 0 0 0 0
>  1 1 1 1 1 1 1 1 1 1 1 1 1 1
> *
> Things noticed.
> *run1.features.dat has columns 3,6 all  zeroes
> *
> *run1.scores.dat has columns 1,3,5,7 all zeroes
> *
>
> run1.features.dat:
> *FEATURES_TXT_BEGIN_0 0 100 14 d_0 d_1 d_2 d_3 d_4 d_5 d_6 lm_0 w_0 tm_0
> tm_1 tm_2 tm_3 tm_4
> -8 -16.1621 0 -1.28775 -23.5316 0 -1.16898 -1804.31 -18 -79.6613 -80.0959
> -48.9898 -27.8995 17.9981
> 0 -20.6133 0 0 -27.4523 0 0 -1804.31 -18 -81.3889 -80.0948 -50.4425 -27.871
> 17.9981
> -8 -16.9628 0 -1.28775 -22.4881 0 -1.16898 -1804.31 -18 -77.4669 -80.9714
> -49.7377 -28.9981 17.9981
> 0 -21.414 0 0 -26.4089 0 0 -1804.31 -18 -79.1945 -80.9702 -51.1904 -28.9696
> 17.9981
> -8 -16.8088 0 -1.28775 -23.0796 0 -1.16898 -1804.31 -18 -78.6891 -79.6433
> -50.0519 -28.2473 17.9981
> -8 -16.1621 0 -1.28775 -23.5316 0 -1.16898 -1804.31 -18 -79.1464 -79.9766
> -48.9898 -28.83 17.9981*
>
>
> run1.scores.dat
> *
> SCORES_TXT_BEGIN_0 0 100 9 BLEU
> 0 18 0 17 0 16 0 15 31
> 0 18 0 17 0 16 0 15 31
> 0 18 0 17 0 16 0 15 31
> 0 18 0 17 0 16 0 15 31
> 0 18 0 17 0 16 0 15 31*
>
> *Issue: in the MIRA run BLEU score is initialized to zero
> *
> Hoping the information would help in solving the issue.
>
> Thanking you,
> Jayendra Rakesh.
>
>
> On Wed, Jun 5, 2013 at 1:52 PM, jayendra rakesh
> wrote:
>
>> Hi Phi,
>>
>> Thanks for the reply, I made a hand run of the command as you have
>> suggested and was able to repeat the crash. I have checked run1.init.opt
>> file, it seems to be fine
>>
>> run1.init.opt:
>> * 0.30 0.30 0.30 0.30 0.30 0.30 0.30 0.50
>> -1.00 0.20 0.20 0.20 0.20 0.20
>>  0 0 0 0 0 0 0 0 0 0 0 0 0 0
>>  1 1 1 1 1 1 1 1 1 1 1 1 1 1
>> *
>> But I noticed a thing in both run1.scores.dat and run1.features.dat files,
>> that many *features*(*columns 3,6 are all zero)* and *scores*(*columns
>> 1,3,5,7 are all zero*) are having zero values, because of which i think
>> BLEU score is initialized to zero in mert.log
>>
>> run1.features.dat:
>> *FEATURES_TXT_BEGIN_0 0 100 14 d_0 d_1 d_2 d_3 d_4 d_5 d_6 lm_0 w_0 tm_0
>> tm_1 tm_2 tm_3 tm_4
>> -8 -16.1621 0 -1.28775 -23.5316 0 -1.16898 -1804.31 -18 -79.6613 -80.0959
>> -48.9898 -27.8995 17.9981
>> 0 -20.6133 0 0 -27.4523 0 0 -1804.31 -18 -81.3889 -80.0948 -50.4425
>> -27.871 17.9981
>> -8 -16.9628 0 -1.28775 -22.4881 0 -1.16898 -1804.31 -18 -77.4669 -80.9714
>> -49.7377 -28.9981 17.9981
>> 0 -21.414 0 0 -26.4089 0 0 -1804.31 -18 -79.1945 -80.9702 -51.1904
>> -28.9696 17.9981
>> -8 -16.8088 0 -1.28775 -23.0796 0 -1.16898 -1804.31 -18 -78.6891 -79.6433
>> -50.0519 -28.2473 17.9981
>> -8 -16.1621 0 -1.28775 -23.5316 0 -1.16898 -1804.31 -18 -79.1464 -79.9766
>> -48.9898 -28.83 17.9981*
>>
>>
>> run1.scores.dat
>> *
>> SCORES_TXT_BEGIN_0 0 100 9 BLEU
>> 0 18 0 17 0 16 0 15 31
>> 0 18 0 17 0 16 0 15 31
>> 0 18 0 17 0 16 0 15 31
>> 0 18 0 17 0 16 0 15 31
>> 0 18 0 17 0 16 0 15 31*
>>
>>
>> Hoping the information would help in pointing out the issue.
>>
>> Regards,
>> Jayendra Rakesh
>>
>>
>>
>> On Sun, Jun 2, 2013 at 5:19 PM, Philipp Koehn  wrote:
>>
>>> Hi,
>>>
>>> I have been running kbest MIRA with factored models many times, never
>>> with any problems, so "this should work".
>>>
>>> The error is in the step:  /tools/mosesdecoder-master_2/bin/kbmira -J
>>> 100 -C 0.001  --dense-init run1.init.opt  --ffile run1.features.data
>>> [...]
>>>
>>> so that's where to start.
>>>
>>> Check if the features file looks sane.
>>> Check the run1.init.opt file.
>>> Run the step by hand.
>>>
>>> If this does not work, send us the input files for this command (maybe
>>> even a smaller subset, if you can reproduce the error).
>>>
>>> -phi
>>>
>>> On Sun, Jun 2, 2013 at 10:56 AM, jayendra rakesh
>>>  wrote:
>>> > Hi,
>>> > My EMS setup (factored,MIRA) crashes at tuning stage after single run.
>>> > config.toy: (attaching only training and tuning sections)
>>> > # TRANSLATION MODEL TRAINING
>>> >
>>> > [TRAINING]
>>> >
>>> > ### training script to be used: either a legacy script or
>>> > # current moses training script (default)
>>> > #
>>> > script = $m

Re: [Moses-support] Tuning crashed, mert.out - no such file or directory. intial BLEU=0

2013-06-09 Thread Hieu Hoang
It's quite difficult to debug from this output. Can you make your model
files available for download via dropbox or google drive and i will try and
run it on my laptop.

Does it run ok with mert or pro?

Hieu
Sent while bumping into things

On 8 Jun 2013, at 07:39 PM, jayendra rakesh 
wrote:

Sorry for bumping the thread again.
here's what I found after running the command by hand as suggested.

run1.init.opt:
* 0.30 0.30 0.30 0.30 0.30 0.30 0.30 0.50
-1.00 0.20 0.20 0.20 0.20 0.20
 0 0 0 0 0 0 0 0 0 0 0 0 0 0
 1 1 1 1 1 1 1 1 1 1 1 1 1 1
*
Things noticed.
*run1.features.dat has columns 3,6 all  zeroes
*
*run1.scores.dat has columns 1,3,5,7 all zeroes
*

run1.features.dat:
*FEATURES_TXT_BEGIN_0 0 100 14 d_0 d_1 d_2 d_3 d_4 d_5 d_6 lm_0 w_0 tm_0
tm_1 tm_2 tm_3 tm_4
-8 -16.1621 0 -1.28775 -23.5316 0 -1.16898 -1804.31 -18 -79.6613 -80.0959
-48.9898 -27.8995 17.9981
0 -20.6133 0 0 -27.4523 0 0 -1804.31 -18 -81.3889 -80.0948 -50.4425 -27.871
17.9981
-8 -16.9628 0 -1.28775 -22.4881 0 -1.16898 -1804.31 -18 -77.4669 -80.9714
-49.7377 -28.9981 17.9981
0 -21.414 0 0 -26.4089 0 0 -1804.31 -18 -79.1945 -80.9702 -51.1904 -28.9696
17.9981
-8 -16.8088 0 -1.28775 -23.0796 0 -1.16898 -1804.31 -18 -78.6891 -79.6433
-50.0519 -28.2473 17.9981
-8 -16.1621 0 -1.28775 -23.5316 0 -1.16898 -1804.31 -18 -79.1464 -79.9766
-48.9898 -28.83 17.9981*


run1.scores.dat
*
SCORES_TXT_BEGIN_0 0 100 9 BLEU
0 18 0 17 0 16 0 15 31
0 18 0 17 0 16 0 15 31
0 18 0 17 0 16 0 15 31
0 18 0 17 0 16 0 15 31
0 18 0 17 0 16 0 15 31*

*Issue: in the MIRA run BLEU score is initialized to zero
*
Hoping the information would help in solving the issue.

Thanking you,
Jayendra Rakesh.


On Wed, Jun 5, 2013 at 1:52 PM, jayendra rakesh
wrote:

> Hi Phi,
>
> Thanks for the reply, I made a hand run of the command as you have
> suggested and was able to repeat the crash. I have checked run1.init.opt
> file, it seems to be fine
>
> run1.init.opt:
> * 0.30 0.30 0.30 0.30 0.30 0.30 0.30 0.50
> -1.00 0.20 0.20 0.20 0.20 0.20
>  0 0 0 0 0 0 0 0 0 0 0 0 0 0
>  1 1 1 1 1 1 1 1 1 1 1 1 1 1
> *
> But I noticed a thing in both run1.scores.dat and run1.features.dat files,
> that many *features*(*columns 3,6 are all zero)* and *scores*(*columns
> 1,3,5,7 are all zero*) are having zero values, because of which i think
> BLEU score is initialized to zero in mert.log
>
> run1.features.dat:
> *FEATURES_TXT_BEGIN_0 0 100 14 d_0 d_1 d_2 d_3 d_4 d_5 d_6 lm_0 w_0 tm_0
> tm_1 tm_2 tm_3 tm_4
> -8 -16.1621 0 -1.28775 -23.5316 0 -1.16898 -1804.31 -18 -79.6613 -80.0959
> -48.9898 -27.8995 17.9981
> 0 -20.6133 0 0 -27.4523 0 0 -1804.31 -18 -81.3889 -80.0948 -50.4425
> -27.871 17.9981
> -8 -16.9628 0 -1.28775 -22.4881 0 -1.16898 -1804.31 -18 -77.4669 -80.9714
> -49.7377 -28.9981 17.9981
> 0 -21.414 0 0 -26.4089 0 0 -1804.31 -18 -79.1945 -80.9702 -51.1904
> -28.9696 17.9981
> -8 -16.8088 0 -1.28775 -23.0796 0 -1.16898 -1804.31 -18 -78.6891 -79.6433
> -50.0519 -28.2473 17.9981
> -8 -16.1621 0 -1.28775 -23.5316 0 -1.16898 -1804.31 -18 -79.1464 -79.9766
> -48.9898 -28.83 17.9981*
>
>
> run1.scores.dat
> *
> SCORES_TXT_BEGIN_0 0 100 9 BLEU
> 0 18 0 17 0 16 0 15 31
> 0 18 0 17 0 16 0 15 31
> 0 18 0 17 0 16 0 15 31
> 0 18 0 17 0 16 0 15 31
> 0 18 0 17 0 16 0 15 31*
>
>
> Hoping the information would help in pointing out the issue.
>
> Regards,
> Jayendra Rakesh
>
>
>
> On Sun, Jun 2, 2013 at 5:19 PM, Philipp Koehn  wrote:
>
>> Hi,
>>
>> I have been running kbest MIRA with factored models many times, never
>> with any problems, so "this should work".
>>
>> The error is in the step:  /tools/mosesdecoder-master_2/bin/kbmira -J
>> 100 -C 0.001  --dense-init run1.init.opt  --ffile run1.features.data
>> [...]
>>
>> so that's where to start.
>>
>> Check if the features file looks sane.
>> Check the run1.init.opt file.
>> Run the step by hand.
>>
>> If this does not work, send us the input files for this command (maybe
>> even a smaller subset, if you can reproduce the error).
>>
>> -phi
>>
>> On Sun, Jun 2, 2013 at 10:56 AM, jayendra rakesh
>>  wrote:
>> > Hi,
>> > My EMS setup (factored,MIRA) crashes at tuning stage after single run.
>> > config.toy: (attaching only training and tuning sections)
>> > # TRANSLATION MODEL TRAINING
>> >
>> > [TRAINING]
>> >
>> > ### training script to be used: either a legacy script or
>> > # current moses training script (default)
>> > #
>> > script = $moses-script-dir/training/train-model.perl
>> >
>> > ### general options
>> > # these are options that are passed on to train-model.perl, for instance
>> > # * "-mgiza -mgiza-cpus 8" to use mgiza instead of giza
>> > # * "-sort-buffer-size 8G -sort-compress gzip" to reduce on-disk sorting
>> > # * "-sort-parallel 8 -cores 8" to speed up phrase table building
>> > #
>> > #training-options = ""
>> >
>> > ### factored training: specify here which factors used
>> > # if none specified, single factor training is ass

Re: [Moses-support] Tuning crashed, mert.out - no such file or directory. intial BLEU=0

2013-06-08 Thread jayendra rakesh
Sorry for bumping the thread again.
here's what I found after running the command by hand as suggested.

run1.init.opt:
* 0.30 0.30 0.30 0.30 0.30 0.30 0.30 0.50
-1.00 0.20 0.20 0.20 0.20 0.20
 0 0 0 0 0 0 0 0 0 0 0 0 0 0
 1 1 1 1 1 1 1 1 1 1 1 1 1 1
*
Things noticed.
*run1.features.dat has columns 3,6 all  zeroes
*
*run1.scores.dat has columns 1,3,5,7 all zeroes
*

run1.features.dat:
*FEATURES_TXT_BEGIN_0 0 100 14 d_0 d_1 d_2 d_3 d_4 d_5 d_6 lm_0 w_0 tm_0
tm_1 tm_2 tm_3 tm_4
-8 -16.1621 0 -1.28775 -23.5316 0 -1.16898 -1804.31 -18 -79.6613 -80.0959
-48.9898 -27.8995 17.9981
0 -20.6133 0 0 -27.4523 0 0 -1804.31 -18 -81.3889 -80.0948 -50.4425 -27.871
17.9981
-8 -16.9628 0 -1.28775 -22.4881 0 -1.16898 -1804.31 -18 -77.4669 -80.9714
-49.7377 -28.9981 17.9981
0 -21.414 0 0 -26.4089 0 0 -1804.31 -18 -79.1945 -80.9702 -51.1904 -28.9696
17.9981
-8 -16.8088 0 -1.28775 -23.0796 0 -1.16898 -1804.31 -18 -78.6891 -79.6433
-50.0519 -28.2473 17.9981
-8 -16.1621 0 -1.28775 -23.5316 0 -1.16898 -1804.31 -18 -79.1464 -79.9766
-48.9898 -28.83 17.9981*


run1.scores.dat
*
SCORES_TXT_BEGIN_0 0 100 9 BLEU
0 18 0 17 0 16 0 15 31
0 18 0 17 0 16 0 15 31
0 18 0 17 0 16 0 15 31
0 18 0 17 0 16 0 15 31
0 18 0 17 0 16 0 15 31*

*Issue: in the MIRA run BLEU score is initialized to zero
*
Hoping the information would help in solving the issue.

Thanking you,
Jayendra Rakesh.


On Wed, Jun 5, 2013 at 1:52 PM, jayendra rakesh
wrote:

> Hi Phi,
>
> Thanks for the reply, I made a hand run of the command as you have
> suggested and was able to repeat the crash. I have checked run1.init.opt
> file, it seems to be fine
>
> run1.init.opt:
> * 0.30 0.30 0.30 0.30 0.30 0.30 0.30 0.50
> -1.00 0.20 0.20 0.20 0.20 0.20
>  0 0 0 0 0 0 0 0 0 0 0 0 0 0
>  1 1 1 1 1 1 1 1 1 1 1 1 1 1
> *
> But I noticed a thing in both run1.scores.dat and run1.features.dat files,
> that many *features*(*columns 3,6 are all zero)* and *scores*(*columns
> 1,3,5,7 are all zero*) are having zero values, because of which i think
> BLEU score is initialized to zero in mert.log
>
> run1.features.dat:
> *FEATURES_TXT_BEGIN_0 0 100 14 d_0 d_1 d_2 d_3 d_4 d_5 d_6 lm_0 w_0 tm_0
> tm_1 tm_2 tm_3 tm_4
> -8 -16.1621 0 -1.28775 -23.5316 0 -1.16898 -1804.31 -18 -79.6613 -80.0959
> -48.9898 -27.8995 17.9981
> 0 -20.6133 0 0 -27.4523 0 0 -1804.31 -18 -81.3889 -80.0948 -50.4425
> -27.871 17.9981
> -8 -16.9628 0 -1.28775 -22.4881 0 -1.16898 -1804.31 -18 -77.4669 -80.9714
> -49.7377 -28.9981 17.9981
> 0 -21.414 0 0 -26.4089 0 0 -1804.31 -18 -79.1945 -80.9702 -51.1904
> -28.9696 17.9981
> -8 -16.8088 0 -1.28775 -23.0796 0 -1.16898 -1804.31 -18 -78.6891 -79.6433
> -50.0519 -28.2473 17.9981
> -8 -16.1621 0 -1.28775 -23.5316 0 -1.16898 -1804.31 -18 -79.1464 -79.9766
> -48.9898 -28.83 17.9981*
>
>
> run1.scores.dat
> *
> SCORES_TXT_BEGIN_0 0 100 9 BLEU
> 0 18 0 17 0 16 0 15 31
> 0 18 0 17 0 16 0 15 31
> 0 18 0 17 0 16 0 15 31
> 0 18 0 17 0 16 0 15 31
> 0 18 0 17 0 16 0 15 31*
>
>
> Hoping the information would help in pointing out the issue.
>
> Regards,
> Jayendra Rakesh
>
>
>
> On Sun, Jun 2, 2013 at 5:19 PM, Philipp Koehn  wrote:
>
>> Hi,
>>
>> I have been running kbest MIRA with factored models many times, never
>> with any problems, so "this should work".
>>
>> The error is in the step:  /tools/mosesdecoder-master_2/bin/kbmira -J
>> 100 -C 0.001  --dense-init run1.init.opt  --ffile run1.features.data
>> [...]
>>
>> so that's where to start.
>>
>> Check if the features file looks sane.
>> Check the run1.init.opt file.
>> Run the step by hand.
>>
>> If this does not work, send us the input files for this command (maybe
>> even a smaller subset, if you can reproduce the error).
>>
>> -phi
>>
>> On Sun, Jun 2, 2013 at 10:56 AM, jayendra rakesh
>>  wrote:
>> > Hi,
>> > My EMS setup (factored,MIRA) crashes at tuning stage after single run.
>> > config.toy: (attaching only training and tuning sections)
>> > # TRANSLATION MODEL TRAINING
>> >
>> > [TRAINING]
>> >
>> > ### training script to be used: either a legacy script or
>> > # current moses training script (default)
>> > #
>> > script = $moses-script-dir/training/train-model.perl
>> >
>> > ### general options
>> > # these are options that are passed on to train-model.perl, for instance
>> > # * "-mgiza -mgiza-cpus 8" to use mgiza instead of giza
>> > # * "-sort-buffer-size 8G -sort-compress gzip" to reduce on-disk sorting
>> > # * "-sort-parallel 8 -cores 8" to speed up phrase table building
>> > #
>> > #training-options = ""
>> >
>> > ### factored training: specify here which factors used
>> > # if none specified, single factor training is assumed
>> > # (one translation step, surface to surface)
>> > #
>> > input-factors = word pos
>> > output-factors = word pos
>> > alignment-factors = "word -> word"
>> > translation-factors = "word+pos -> word+pos"
>> > reordering-factors = "word -> word"
>> > #generation-factors = "pos -> word"

Re: [Moses-support] Tuning crashed, mert.out - no such file or directory. intial BLEU=0

2013-06-05 Thread jayendra rakesh
Hi Phi,

Thanks for the reply, I made a hand run of the command as you have
suggested and was able to repeat the crash. I have checked run1.init.opt
file, it seems to be fine

run1.init.opt:
* 0.30 0.30 0.30 0.30 0.30 0.30 0.30 0.50
-1.00 0.20 0.20 0.20 0.20 0.20
 0 0 0 0 0 0 0 0 0 0 0 0 0 0
 1 1 1 1 1 1 1 1 1 1 1 1 1 1
*
But I noticed a thing in both run1.scores.dat and run1.features.dat files,
that many *features*(*columns 3,6 are all zero)* and *scores*(*columns
1,3,5,7 are all zero*) are having zero values, because of which i think
BLEU score is initialized to zero in mert.log

run1.features.dat:
*FEATURES_TXT_BEGIN_0 0 100 14 d_0 d_1 d_2 d_3 d_4 d_5 d_6 lm_0 w_0 tm_0
tm_1 tm_2 tm_3 tm_4
-8 -16.1621 0 -1.28775 -23.5316 0 -1.16898 -1804.31 -18 -79.6613 -80.0959
-48.9898 -27.8995 17.9981
0 -20.6133 0 0 -27.4523 0 0 -1804.31 -18 -81.3889 -80.0948 -50.4425 -27.871
17.9981
-8 -16.9628 0 -1.28775 -22.4881 0 -1.16898 -1804.31 -18 -77.4669 -80.9714
-49.7377 -28.9981 17.9981
0 -21.414 0 0 -26.4089 0 0 -1804.31 -18 -79.1945 -80.9702 -51.1904 -28.9696
17.9981
-8 -16.8088 0 -1.28775 -23.0796 0 -1.16898 -1804.31 -18 -78.6891 -79.6433
-50.0519 -28.2473 17.9981
-8 -16.1621 0 -1.28775 -23.5316 0 -1.16898 -1804.31 -18 -79.1464 -79.9766
-48.9898 -28.83 17.9981*


run1.scores.dat
*
SCORES_TXT_BEGIN_0 0 100 9 BLEU
0 18 0 17 0 16 0 15 31
0 18 0 17 0 16 0 15 31
0 18 0 17 0 16 0 15 31
0 18 0 17 0 16 0 15 31
0 18 0 17 0 16 0 15 31*


Hoping the information would help in pointing out the issue.

Regards,
Jayendra Rakesh


On Sun, Jun 2, 2013 at 5:19 PM, Philipp Koehn  wrote:

> Hi,
>
> I have been running kbest MIRA with factored models many times, never
> with any problems, so "this should work".
>
> The error is in the step:  /tools/mosesdecoder-master_2/bin/kbmira -J
> 100 -C 0.001  --dense-init run1.init.opt  --ffile run1.features.data
> [...]
>
> so that's where to start.
>
> Check if the features file looks sane.
> Check the run1.init.opt file.
> Run the step by hand.
>
> If this does not work, send us the input files for this command (maybe
> even a smaller subset, if you can reproduce the error).
>
> -phi
>
> On Sun, Jun 2, 2013 at 10:56 AM, jayendra rakesh
>  wrote:
> > Hi,
> > My EMS setup (factored,MIRA) crashes at tuning stage after single run.
> > config.toy: (attaching only training and tuning sections)
> > # TRANSLATION MODEL TRAINING
> >
> > [TRAINING]
> >
> > ### training script to be used: either a legacy script or
> > # current moses training script (default)
> > #
> > script = $moses-script-dir/training/train-model.perl
> >
> > ### general options
> > # these are options that are passed on to train-model.perl, for instance
> > # * "-mgiza -mgiza-cpus 8" to use mgiza instead of giza
> > # * "-sort-buffer-size 8G -sort-compress gzip" to reduce on-disk sorting
> > # * "-sort-parallel 8 -cores 8" to speed up phrase table building
> > #
> > #training-options = ""
> >
> > ### factored training: specify here which factors used
> > # if none specified, single factor training is assumed
> > # (one translation step, surface to surface)
> > #
> > input-factors = word pos
> > output-factors = word pos
> > alignment-factors = "word -> word"
> > translation-factors = "word+pos -> word+pos"
> > reordering-factors = "word -> word"
> > #generation-factors = "pos -> word"
> > decoding-steps = "t0"
> >
> > ### parallelization of data preparation step
> > # the two directions of the data preparation can be run in parallel
> > # comment out if not needed
> > #
> > parallel = yes
> >
> > ### pre-computation for giza++
> > # giza++ has a more efficient data structure that needs to be
> > # initialized with snt2cooc. if run in parallel, this may reduces
> > # memory requirements. set here the number of parts
> > #
> > #run-giza-in-parts = 5
> >
> > ### symmetrization method to obtain word alignments from giza output
> > # (commonly used: grow-diag-final-and)
> > #
> > alignment-symmetrization-method = grow-diag-final-and
> >
> > ### use of berkeley aligner for word alignment
> > #
> > #use-berkeley = true
> > #alignment-symmetrization-method = berkeley
> > #berkeley-train = $moses-script-dir/ems/support/berkeley-train.sh
> > #berkeley-process =  $moses-script-dir/ems/support/berkeley-process.sh
> > #berkeley-jar = /your/path/to/berkeleyaligner-1.1/berkeleyaligner.jar
> > #berkeley-java-options = "-server -mx3m -ea"
> > #berkeley-training-options = "-Main.iters 5 5 -EMWordAligner.numThreads
> 8"
> > #berkeley-process-options = "-EMWordAligner.numThreads 8"
> > #berkeley-posterior = 0.5
> >
> > ### use of baseline alignment model (incremental training)
> > #
> > #baseline = 68
> > #baseline-alignment-model =
> > "$working-dir/training/prepared.$baseline/$input-extension.vcb \
> > #  $working-dir/training/prepared.$baseline/$output-extension.vcb \
> > #
> >
> $working-dir/training/giza.$baseline/${output-extension}-$input-extension.cooc
> > \
> > #
> >
> $working-dir/trainin

Re: [Moses-support] Tuning crashed, mert.out - no such file or directory. intial BLEU=0

2013-06-02 Thread Philipp Koehn
Hi,

I have been running kbest MIRA with factored models many times, never
with any problems, so "this should work".

The error is in the step:  /tools/mosesdecoder-master_2/bin/kbmira -J
100 -C 0.001  --dense-init run1.init.opt  --ffile run1.features.data
[...]

so that's where to start.

Check if the features file looks sane.
Check the run1.init.opt file.
Run the step by hand.

If this does not work, send us the input files for this command (maybe
even a smaller subset, if you can reproduce the error).

-phi

On Sun, Jun 2, 2013 at 10:56 AM, jayendra rakesh
 wrote:
> Hi,
> My EMS setup (factored,MIRA) crashes at tuning stage after single run.
> config.toy: (attaching only training and tuning sections)
> # TRANSLATION MODEL TRAINING
>
> [TRAINING]
>
> ### training script to be used: either a legacy script or
> # current moses training script (default)
> #
> script = $moses-script-dir/training/train-model.perl
>
> ### general options
> # these are options that are passed on to train-model.perl, for instance
> # * "-mgiza -mgiza-cpus 8" to use mgiza instead of giza
> # * "-sort-buffer-size 8G -sort-compress gzip" to reduce on-disk sorting
> # * "-sort-parallel 8 -cores 8" to speed up phrase table building
> #
> #training-options = ""
>
> ### factored training: specify here which factors used
> # if none specified, single factor training is assumed
> # (one translation step, surface to surface)
> #
> input-factors = word pos
> output-factors = word pos
> alignment-factors = "word -> word"
> translation-factors = "word+pos -> word+pos"
> reordering-factors = "word -> word"
> #generation-factors = "pos -> word"
> decoding-steps = "t0"
>
> ### parallelization of data preparation step
> # the two directions of the data preparation can be run in parallel
> # comment out if not needed
> #
> parallel = yes
>
> ### pre-computation for giza++
> # giza++ has a more efficient data structure that needs to be
> # initialized with snt2cooc. if run in parallel, this may reduces
> # memory requirements. set here the number of parts
> #
> #run-giza-in-parts = 5
>
> ### symmetrization method to obtain word alignments from giza output
> # (commonly used: grow-diag-final-and)
> #
> alignment-symmetrization-method = grow-diag-final-and
>
> ### use of berkeley aligner for word alignment
> #
> #use-berkeley = true
> #alignment-symmetrization-method = berkeley
> #berkeley-train = $moses-script-dir/ems/support/berkeley-train.sh
> #berkeley-process =  $moses-script-dir/ems/support/berkeley-process.sh
> #berkeley-jar = /your/path/to/berkeleyaligner-1.1/berkeleyaligner.jar
> #berkeley-java-options = "-server -mx3m -ea"
> #berkeley-training-options = "-Main.iters 5 5 -EMWordAligner.numThreads 8"
> #berkeley-process-options = "-EMWordAligner.numThreads 8"
> #berkeley-posterior = 0.5
>
> ### use of baseline alignment model (incremental training)
> #
> #baseline = 68
> #baseline-alignment-model =
> "$working-dir/training/prepared.$baseline/$input-extension.vcb \
> #  $working-dir/training/prepared.$baseline/$output-extension.vcb \
> #
> $working-dir/training/giza.$baseline/${output-extension}-$input-extension.cooc
> \
> #
> $working-dir/training/giza-inverse.$baseline/${input-extension}-$output-extension.cooc
> \
> #
> $working-dir/training/giza.$baseline/${output-extension}-$input-extension.thmm.5
> \
> #
> $working-dir/training/giza.$baseline/${output-extension}-$input-extension.hhmm.5
> \
> #
> $working-dir/training/giza-inverse.$baseline/${input-extension}-$output-extension.thmm.5
> \
> #
> $working-dir/training/giza-inverse.$baseline/${input-extension}-$output-extension.hhmm.5"
>
> ### if word alignment should be skipped,
> # point to word alignment files
> #
> #word-alignment = $working-dir/model/aligned.1
>
> ### filtering some corpora with modified Moore-Lewis
> # specify corpora to be filtered and ratio to be kept, either before or
> after word alignment
> #mml-filter-corpora = toy
> #mml-before-wa = "-proportion 0.9"
> #mml-after-wa = "-proportion 0.9"
>
> ### create a bilingual concordancer for the model
> #
> #biconcor = $moses-script-dir/ems/biconcor/biconcor
>
> ### lexicalized reordering: specify orientation type
> # (default: only distance-based reordering model)
> #
> lexicalized-reordering = msd-bidirectional-fe
>
> ### hierarchical rule set
> #
> #hierarchical-rule-set = true
>
> ### settings for rule extraction
> #
> #extract-settings = ""
> max-phrase-length = 5
>
> ### add extracted phrases from baseline model
> #
> #baseline-extract = $working-dir/model/extract.$baseline
> #
> # requires aligned parallel corpus for re-estimating lexical translation
> probabilities
> #baseline-corpus = $working-dir/training/corpus.$baseline
> #baseline-alignment =
> $working-dir/model/aligned.$baseline.$alignment-symmetrization-method
>
> ### unknown word labels (target syntax only)
> # enables use of unknown word labels during decoding
> # label file is generated during rule extraction
> #
> #use-unknown-word-labels = true
>
> ##

[Moses-support] Tuning crashed, mert.out - no such file or directory. intial BLEU=0

2013-06-02 Thread jayendra rakesh
Hi,
My EMS setup (factored,MIRA) crashes at tuning stage after single run.
config.toy: (attaching only training and tuning sections)
*# TRANSLATION MODEL TRAINING

[TRAINING]

### training script to be used: either a legacy script or
# current moses training script (default)
#
script = $moses-script-dir/training/train-model.perl

### general options
# these are options that are passed on to train-model.perl, for instance
# * "-mgiza -mgiza-cpus 8" to use mgiza instead of giza
# * "-sort-buffer-size 8G -sort-compress gzip" to reduce on-disk sorting
# * "-sort-parallel 8 -cores 8" to speed up phrase table building
#
#training-options = ""

### factored training: specify here which factors used
# if none specified, single factor training is assumed
# (one translation step, surface to surface)
#
input-factors = word pos
output-factors = word pos
alignment-factors = "word -> word"
translation-factors = "word+pos -> word+pos"
reordering-factors = "word -> word"
#generation-factors = "pos -> word"
decoding-steps = "t0"

### parallelization of data preparation step
# the two directions of the data preparation can be run in parallel
# comment out if not needed
#
parallel = yes

### pre-computation for giza++
# giza++ has a more efficient data structure that needs to be
# initialized with snt2cooc. if run in parallel, this may reduces
# memory requirements. set here the number of parts
#
#run-giza-in-parts = 5

### symmetrization method to obtain word alignments from giza output
# (commonly used: grow-diag-final-and)
#
alignment-symmetrization-method = grow-diag-final-and

### use of berkeley aligner for word alignment
#
#use-berkeley = true
#alignment-symmetrization-method = berkeley
#berkeley-train = $moses-script-dir/ems/support/berkeley-train.sh
#berkeley-process =  $moses-script-dir/ems/support/berkeley-process.sh
#berkeley-jar = /your/path/to/berkeleyaligner-1.1/berkeleyaligner.jar
#berkeley-java-options = "-server -mx3m -ea"
#berkeley-training-options = "-Main.iters 5 5 -EMWordAligner.numThreads 8"
#berkeley-process-options = "-EMWordAligner.numThreads 8"
#berkeley-posterior = 0.5

### use of baseline alignment model (incremental training)
#
#baseline = 68
#baseline-alignment-model =
"$working-dir/training/prepared.$baseline/$input-extension.vcb \
#  $working-dir/training/prepared.$baseline/$output-extension.vcb \
#
$working-dir/training/giza.$baseline/${output-extension}-$input-extension.cooc
\
#
$working-dir/training/giza-inverse.$baseline/${input-extension}-$output-extension.cooc
\
#
$working-dir/training/giza.$baseline/${output-extension}-$input-extension.thmm.5
\
#
$working-dir/training/giza.$baseline/${output-extension}-$input-extension.hhmm.5
\
#
$working-dir/training/giza-inverse.$baseline/${input-extension}-$output-extension.thmm.5
\
#
$working-dir/training/giza-inverse.$baseline/${input-extension}-$output-extension.hhmm.5"

### if word alignment should be skipped,
# point to word alignment files
#
#word-alignment = $working-dir/model/aligned.1

### filtering some corpora with modified Moore-Lewis
# specify corpora to be filtered and ratio to be kept, either before or
after word alignment
#mml-filter-corpora = toy
#mml-before-wa = "-proportion 0.9"
#mml-after-wa = "-proportion 0.9"

### create a bilingual concordancer for the model
#
#biconcor = $moses-script-dir/ems/biconcor/biconcor

### lexicalized reordering: specify orientation type
# (default: only distance-based reordering model)
#
lexicalized-reordering = msd-bidirectional-fe

### hierarchical rule set
#
#hierarchical-rule-set = true

### settings for rule extraction
#
#extract-settings = ""
max-phrase-length = 5

### add extracted phrases from baseline model
#
#baseline-extract = $working-dir/model/extract.$baseline
#
# requires aligned parallel corpus for re-estimating lexical translation
probabilities
#baseline-corpus = $working-dir/training/corpus.$baseline
#baseline-alignment =
$working-dir/model/aligned.$baseline.$alignment-symmetrization-method

### unknown word labels (target syntax only)
# enables use of unknown word labels during decoding
# label file is generated during rule extraction
#
#use-unknown-word-labels = true

### if phrase extraction should be skipped,
# point to stem for extract files
#
# extracted-phrases =

### settings for rule scoring
#
score-settings = "--GoodTuring"

### include word alignment in phrase table
#
include-word-alignment-in-rules = yes

### sparse lexical features
#
#sparse-lexical-features = "target-word-insertion top 50,
source-word-deletion top 50, word-translation top 50 50, phrase-length"

### domain adaptation settings
# options: sparse, any of: indicator, subset, ratio
#domain-features = "subset"

### if phrase table training should be skipped,
# point to phrase translation table
#
# phrase-translation-table =

### if reordering table training should be skipped,
# point to reordering table
#
# reordering-table =

### filtering the phrase table based on significance tests
# Johnson, Martin,