[Moses-support] Moses 3.0 cannot start with configuration file and models of moses 2.0

2016-07-05 Thread Eleftherios Avramidis
Hi,

I have a freshly installed Moses 3.0. I am trying to load some existing 
compact models from moses 2.

The only error i get is "Exception: bad lexical cast: source type value 
could not be interpreted as target", right after I start moses.

Could you please instruct me how to further locate the issue?

best
Lefteris
___
Moses-support mailing list
Moses-support@mit.edu
http://mailman.mit.edu/mailman/listinfo/moses-support


[Moses-support] boost::thread_resource_error

2016-07-05 Thread arefeh kazemi
Hi all

I have implemented a feature function for Hiero moses, which has 8 scores.
I'm trying to tune the system using mira, I use a node with 192G memory for
tuning. After the first run of the tuning process, the program exits with
boost::thread_resource_error.

Any help would be appreciated.

-- 
Arefeh
___
Moses-support mailing list
Moses-support@mit.edu
http://mailman.mit.edu/mailman/listinfo/moses-support


[Moses-support] boost::thread_resource_error

2016-07-05 Thread Arefeh Kazemi
Hi all

I have implemented a feature function for Hiero moses, which has 8 scores. I'm 
trying to tune the system using mira, I use a node with 192G memory for tuning. 
After the first run of the tuning process, the program exits with 
boost::thread_resource_error. 

Any help would be appreciate.

___
Moses-support mailing list
Moses-support@mit.edu
http://mailman.mit.edu/mailman/listinfo/moses-support


[Moses-support] Parameters in server mode

2016-07-05 Thread Mathias Müller
Hi all

I have two questions about using Moses in server mode.

1) behaviour of certain parameters in server mode

It seems that parameters related to word alignment do not work in server
mode. For instance, "*--print-alignment-info*" does not have any effect if
I start the server as follows:

*$ mosesserver -f model/moses.ini --print-alignment-info*

If a translation is requested, the server does not output alignment
information after translation. In non-server mode, the word alignment is
displayed.

I then hypothesized that perhaps all options that relate to direct output
are disabled in server mode, because output is usually sent as a server
response. But other parameters that have to do with output apparently do
work in server mode, e.g. "*--n-best-list*".

So, my question is:

*Why do certain parameters not have any effect in server mode and is there
any documentation about this?*

2) Parameters that are accepted in a request

The examples on Github and elsewhere suggest that the only two parameters
to the translate() function are "text" and "align". But apparently,
"word-align" is also possible. Now I am wondering:

*Does anyone have a complete list of parameters of the translate() function
that can be sent to a Moses server in an XML-RPC request? Again, is there
documentation about this?*

XML-RPC introspection functions are not really helpful at the moment, since

*system.methodHelp("translate")*

returns:

*'Does translation'*

and

*system.methodSignature("translate")*

returns

*[['struct', 'struct']]*

Are there plans to make this more informative?

Thanks a lot for your help.
Mathias

—

Mathias Müller
AND-2-20
Institute of Computational Linguistics
University of Zurich
+41 44 635 75 81
mathias.muel...@uzh.ch
___
Moses-support mailing list
Moses-support@mit.edu
http://mailman.mit.edu/mailman/listinfo/moses-support


Re: [Moses-support] Moses Installation

2016-07-05 Thread Tom Hoar

Irene,

It looks like you need a primer course in how to work with the terminal, 
especially on Linux. Many of the utilites (including tokenizer.perl) 
require you to pipe input and output. This requires use of the < and > 
characters in your command line. In your case, the command line should be:


:~/mosesdecoder/scripts/tokenizer$ ./tokenizer.perl -l en < 
~/mosesdecoder/corpus/training/news-commentary-v8.fr-en.en > 
~/mosesdecoder/corpus/news-commentary-v8.fr-en.tok.en


If you don't pipe the input/output, the command will sit there waiting 
for you to type something into the terminal.


Tom



On 7/5/2016 5:07 PM, moses-support-requ...@mit.edu wrote:

Date: Tue, 5 Jul 2016 07:28:14 +0300
From: Irene Nandutu
Subject: Re: [Moses-support] Fwd: Moses Installation
To: Sa?o Kuntaric
Cc: moses-support
Message-ID:

Content-Type: text/plain; charset="utf-8"

Thanks.
Am preparing the data for the translation system under tokenization it
gives this ourput and can go 24+ minus stopping yet the data is only 20kb

irene@irene-HP-650-Notebook-PC:~/mosesdecoder/scripts/tokenizer$
./tokenizer.perl -l en
~/mosesdecoder/corpus/training/news-commentary-v8.fr-en.en
~/mosesdecoder/corpus/news-commentary-v8.fr-en.tok.en
Tokenizer Version 1.1
Language: en
Number of threads: 1

Any insights please


Irene Nandutu
+256781595675
Twitter

Google+

On Mon, Jul 4, 2016 at 11:40 PM, Sa?o Kuntaric
wrote:


>Hi Irene,
>
>The command you posted is for training a translation model ... the error
>states that you didn't tell Moses which corpus you wanna use for the model
>and where it is.
>
>Best regards,
>
>Sa?o
>
>2016-07-04 20:43 GMT+02:00 Irene Nandutu:
>

>>Hi all, am seeking support
>>Am installing moses and am using thes link
>>http://www.statmt.org/moses/?n=moses.baseline.
>>All the steps were fine but when i came to run the training, i wanted to
>>to tell the training script where GIZA++ was installed using the
>>-external-bin-dir argument.
>>
>>Here is the output in the commandline:
>>
>>irene@irene-HP-650-Notebook-PC:~/mosesdecoder/scripts/training$
>>./train-model.perl -external-bin-dir $HOME/mosesdecoder/tools
>>Using SCRIPTS_ROOTDIR: /home/irene/mosesdecoder/scripts
>>Using single-thread GIZA
>>using gzip
>>ERROR: use --corpus to specify corpus at ./train-model.perl line 499.
>>irene@irene-HP-650-Notebook-PC:~/mosesdecoder/scripts/training$
>>
>>I will be grateful for any help and advice.
>>Thanks
>>
>>Irene Nandutu
>>+256781595675
>>Twitter
>>
>>Google+
>>
>>___
>>Moses-support mailing list
>>Moses-support@mit.edu
>>http://mailman.mit.edu/mailman/listinfo/moses-support
>>
>>

>
>
>--
>lp,
>
>Sa?o
>
>
>
>--
>lp,
>
>Sa?o
>
>___
>Moses-support mailing list
>Moses-support@mit.edu
>http://mailman.mit.edu/mailman/listinfo/moses-support
>
>


___
Moses-support mailing list
Moses-support@mit.edu
http://mailman.mit.edu/mailman/listinfo/moses-support


[Moses-support] snt2cooc option fails the training

2016-07-05 Thread Gmehlin Floran
Hi,

It seems that whenever I use the option "-snt2cooc snt2cooc.pl" the training 
fails (see below for error log). When I try on the same corpora (rather small 
because of memory limitation) it works.

Does anyone have a clue about this ?

Reading more sentence pairs into memory ...
[sent:340]
 Train total # sentence pairs (weighted): 6.74606e+06
Size of source portion of the training corpus: 2.05275e+08 tokens
Size of the target portion of the training corpus: 2.40869e+08 tokens
In source portion of the training corpus, only 4208312 unique tokens appeared
In target portion of the training corpus, only 1428380 unique tokens appeared
lambda for PP calculation in IBM-1,IBM-2,HMM:= 
2.40869e+08/(2.12021e+08-6.74606e+06)== 1.1734
Dictionary Loading complete
Inputfile in /local/para_corpora/pattr/claims/giza.en-de/en-de.cooc
ERROR: Execution of: /nas/fgmehlin/bin/mgizapp/mgiza  -CoocurrenceFile 
/local/para_corpora/pattr/claims/giza.en-de/en-de.cooc -c 
/local/para_corpora/pattr/claims/corpus/en-de-int-train.snt -m1 5 -m2 0 -m3 3 
-m4 3 -model1dumpfrequency 1 -model4smoothfactor 0.4 -ncpus 4 -nodumps 1 
-nsmooth 4 -o /local/para_corpora/pattr/claims/giza.en-de/en-de -onlyaldumps 1 
-p0 0.999 -s /local/para_corpora/pattr/claims/corpus/de.vcb -t 
/local/para_corpora/pattr/claims/corpus/en.vcb
  died with signal 11, without coredump
...
Reading more sentence pairs into memory ...
Reading more sentence pairs into memory ...
...
[sent:350]Compacted Vocabulary, eliminated 1 entries 1428381 remains
Compacted Vocabulary, eliminated 1 entries 4208312 remains
 Train total # sentence pairs (weighted): 6.74606e+06
Size of source portion of the training corpus: 2.40869e+08 tokens
Size of the target portion of the training corpus: 2.05275e+08 tokens
In source portion of the training corpus, only 1428381 unique tokens appeared
In target portion of the training corpus, only 4208311 unique tokens appeared
lambda for PP calculation in IBM-1,IBM-2,HMM:= 
2.05275e+08/(2.47615e+08-6.74606e+06)== 0.852227
Dictionary Loading complete
Inputfile in /local/para_corpora/pattr/claims/giza.de-en/de-en.cooc
ERROR: Execution of: /nas/fgmehlin/bin/mgizapp/mgiza  -CoocurrenceFile 
/local/para_corpora/pattr/claims/giza.de-en/de-en.cooc -c 
/local/para_corpora/pattr/claims/corpus/de-en-int-train.snt -m1 5 -m2 0 -m3 3 
-m4 3 -model1dumpfrequency 1 -model4smoothfactor 0.4 -ncpus 4 -nodumps 1 
-nsmooth 4 -o /local/para_corpora/pattr/claims/giza.de-en/de-en -onlyaldumps 1 
-p0 0.999 -s /local/para_corpora/pattr/claims/corpus/en.vcb -t 
/local/para_corpora/pattr/claims/corpus/de.vcb
  died with signal 11, without coredump

The command I used to start the training is the following :

/local/moses/mosesdecoder/scripts/training/train-model.perl --root-dir . 
--corpus clean_lc_utf8 --f de --e en -cores 4 --parallel -external-bin-dir 
/nas/fgmehlin/bin/mgizapp -mgiza -mgiza-cpus 4 -snt2cooc snt2cooc.pl -lm 
0:5:/local/para_corpora/pattr/claims/lm.arpa

Thank you in advance,

Floran
___
Moses-support mailing list
Moses-support@mit.edu
http://mailman.mit.edu/mailman/listinfo/moses-support


Re: [Moses-support] Tuning giving zero weights in Factored Based SMT

2016-07-05 Thread Saumitra Yadav
Hi Sašo,
Thanks will try that.
Regards,
Saumitra


Regards,
Saumitra Yadav
MS CSE
IIIT-Hyderabad

On Tue, Jul 5, 2016 at 2:09 AM, Sašo Kuntaric 
wrote:

> Hi Saumitra,
>
> Had the same issue a few days ago. What worked in my case was that I used
> a tagged source side of the corpus and an untagged translated side of the
> tuning corpus. The other thing I changed was the language model I used, so
> you might want to experiment with those two things and see if you get a
> resolution.
>
> Best regards,
>
> Sašo
>
> 2016-07-04 2:03 GMT+02:00 Saumitra Yadav :
>
>> Hi,
>> I'm trying Factored Based SMT (punjabi to english, with pos tag as factor
>> and translation factor 0,1-0) and while tuning i get BEST at 2: 0 0 0 0 0 0
>> 0 0 ==> 0 as weights. Is there a reason for this?
>> Regards,
>> Saumitra Yadav
>> MS CSE
>> IIIT-Hyderabad
>>
>> ___
>> Moses-support mailing list
>> Moses-support@mit.edu
>> http://mailman.mit.edu/mailman/listinfo/moses-support
>>
>>
>
>
> --
> lp,
>
> Sašo
>
___
Moses-support mailing list
Moses-support@mit.edu
http://mailman.mit.edu/mailman/listinfo/moses-support