Hi Everybody, I am also facing problem with Mosesdecoder.v211 moseserver. I have compiled the moseserver with out any error and the moseserver is listenening to the port also. But when a translation request is going from the interface, there is no responds from the mosesserver.
I am getting the folowing exception.. org.apache.xmlrpc.XmlRpcException: Failed to read server's response: Connection refused Is there any difference in the moseswerver connection of Mosesdecoder Release 1.0 and Mosesdecoder.V211. ? could anybody please clarify these doubts and how can I establish the moseserver connection.. Regards Lakshya Message: 1 Date: Mon, 7 Apr 2014 18:25:35 +0100 From: kamel nebhi <k.ne...@sheffield.ac.uk> Subject: [Moses-support] moses server segmentation fault (core dumped) To: moses-support <moses-support@mit.edu> Message-ID: <CAG66Y3c2UFq5+ 2w4eV00RrwVrMWTtVxpQ7=jgxfwng+afnr...@mail.gmail.com> Content-Type: text/plain; charset="utf-8" Hi, I try to install mosesserver on localhost. I have installed xml-rpc and rebuild moses with no problem. Next i use this command to run the server : *~/mosesdecoder/bin/mosesserver -f working/model/moses.ini --server-port 8999* But it failed with this message : Defined parameters (per moses.ini or switch): config: /home/kamelnebhi/recaser/training/moses.ini distortion-limit: 6 feature: UnknownWordPenalty WordPenalty PhrasePenalty PhraseDictionaryMemory name=TranslationModel0 table-limit=20 num-features=4 path=/home/kamelnebhi/recaser/training/phrase-table.gz input-factor=0 output-factor=0 Distortion KENLM lazyken=0 name=LM0 factor=0 path=/home/kamelnebhi/recaser/training//cased.srilm.gz order=3 input-factors: 0 mapping: 0 T 0 weight: UnknownWordPenalty0= 1 WordPenalty0= -1 PhrasePenalty0= 0.2 TranslationModel0= 0.2 0.2 0.2 0.2 Distortion0= 0.3 LM0= 0.5 /home/kamelnebhi/mosesdecoder/bin line=UnknownWordPenalty FeatureFunction: UnknownWordPenalty0 start: 0 end: 0 line=WordPenalty FeatureFunction: WordPenalty0 start: 1 end: 1 line=PhrasePenalty FeatureFunction: PhrasePenalty0 start: 2 end: 2 line=PhraseDictionaryMemory name=TranslationModel0 table-limit=20 num-features=4 path=/home/kamelnebhi/recaser/training/phrase-table.gz input-factor=0 output-factor=0 FeatureFunction: TranslationModel0 start: 3 end: 6 line=Distortion FeatureFunction: Distortion0 start: 7 end: 7 line=KENLM lazyken=0 name=LM0 factor=0 path=/home/kamelnebhi/recaser/training//cased.srilm.gz order=3 FeatureFunction: LM0 start: 8 end: 8 Loading the LM will be faster if you build a binary file. Reading /home/kamelnebhi/recaser/training//cased.srilm.gz ----5---10---15---20---25---30---35---40---45---50---55---60---65---70---75---80---85---90---95--100 *The ARPA file is missing <unk>. Substituting log10 probability -100. *************************************************************************************************** Loading UnknownWordPenalty0 Loading WordPenalty0 Loading PhrasePenalty0 Loading Distortion0 Loading LM0 Loading TranslationModel0 Start loading text SCFG phrase table. Moses format : [3.69361] seconds Reading /home/kamelnebhi/recaser/training/phrase-table.gz ----5---10---15---20---25---30---35---40---45---50---55---60---65---70---75---80---85---90---95--100 **************************************************************************************************** Erreur de segmentation (core dumped) root@kamelnebhi-MacBookPro:/home/kamelnebhi# /home/kamelnebhi/mosesdecoder/bin/mosesserver -f /home/kamelnebhi/recaser/training/moses.ini --server-port 80 Defined parameters (per moses.ini or switch): config: /home/kamelnebhi/recaser/training/moses.ini distortion-limit: 6 feature: UnknownWordPenalty WordPenalty PhrasePenalty PhraseDictionaryMemory name=TranslationModel0 table-limit=20 num-features=4 path=/home/kamelnebhi/recaser/training/phrase-table.gz input-factor=0 output-factor=0 Distortion KENLM lazyken=0 name=LM0 factor=0 path=/home/kamelnebhi/recaser/training//cased.srilm.gz order=3 input-factors: 0 mapping: 0 T 0 weight: UnknownWordPenalty0= 1 WordPenalty0= -1 PhrasePenalty0= 0.2 TranslationModel0= 0.2 0.2 0.2 0.2 Distortion0= 0.3 LM0= 0.5 /home/kamelnebhi/mosesdecoder/bin line=UnknownWordPenalty FeatureFunction: UnknownWordPenalty0 start: 0 end: 0 line=WordPenalty FeatureFunction: WordPenalty0 start: 1 end: 1 line=PhrasePenalty FeatureFunction: PhrasePenalty0 start: 2 end: 2 line=PhraseDictionaryMemory name=TranslationModel0 table-limit=20 num-features=4 path=/home/kamelnebhi/recaser/training/phrase-table.gz input-factor=0 output-factor=0 FeatureFunction: TranslationModel0 start: 3 end: 6 line=Distortion FeatureFunction: Distortion0 start: 7 end: 7 line=KENLM lazyken=0 name=LM0 factor=0 path=/home/kamelnebhi/recaser/training//cased.srilm.gz order=3 FeatureFunction: LM0 start: 8 end: 8 Loading the LM will be faster if you build a binary file. Reading /home/kamelnebhi/recaser/training//cased.srilm.gz ----5---10---15---20---25---30---35---40---45---50---55---60---65---70---75---80---85---90---95--100 *The ARPA file is missing <unk>. Substituting log10 probability -100. *************************************************************************************************** Loading UnknownWordPenalty0 Loading WordPenalty0 Loading PhrasePenalty0 Loading Distortion0 Loading LM0 Loading TranslationModel0 Start loading text SCFG phrase table. Moses format : [3.69152] seconds Reading /home/kamelnebhi/recaser/training/phrase-table.gz ----5---10---15---20---25---30---35---40---45---50---55---60---65---70---75---80---85---90---95--100 **************************************************************************************************** Segmentation fault (core dumped) *Thanks for your help* -------------- next part -------------- An HTML attachment was scrubbed... URL: http://mailman.mit.edu/mailman/private/moses-support/attachments/20140407/82302fab/attachment-0001.htm ------------------------------ ---------- Forwarded message ---------- From: <moses-support-requ...@mit.edu> Date: Tue, Apr 8, 2014 at 3:57 AM Subject: Moses-support Digest, Vol 90, Issue 19 To: moses-support@mit.edu Send Moses-support mailing list submissions to moses-support@mit.edu To subscribe or unsubscribe via the World Wide Web, visit http://mailman.mit.edu/mailman/listinfo/moses-support or, via email, send a message with subject or body 'help' to moses-support-requ...@mit.edu You can reach the person managing the list at moses-support-ow...@mit.edu When replying, please edit your Subject line so it is more specific than "Re: Contents of Moses-support digest..." Today's Topics: 1. moses server segmentation fault (core dumped) (kamel nebhi) 2. Re: Monolingual Word alignment (Philipp Koehn) 3. Call for Participation: Automatic and Manual Metrics for Operational Translation Evaluation (Lucia Specia) ---------------------------------------------------------------------- Message: 1 Date: Mon, 7 Apr 2014 18:25:35 +0100 From: kamel nebhi <k.ne...@sheffield.ac.uk> Subject: [Moses-support] moses server segmentation fault (core dumped) To: moses-support <moses-support@mit.edu> Message-ID: <CAG66Y3c2UFq5+2w4eV00RrwVrMWTtVxpQ7=jgxfwng+afnr...@mail.gmail.com> Content-Type: text/plain; charset="utf-8" Hi, I try to install mosesserver on localhost. I have installed xml-rpc and rebuild moses with no problem. Next i use this command to run the server : *~/mosesdecoder/bin/mosesserver -f working/model/moses.ini --server-port 8999* But it failed with this message : Defined parameters (per moses.ini or switch): config: /home/kamelnebhi/recaser/training/moses.ini distortion-limit: 6 feature: UnknownWordPenalty WordPenalty PhrasePenalty PhraseDictionaryMemory name=TranslationModel0 table-limit=20 num-features=4 path=/home/kamelnebhi/recaser/training/phrase-table.gz input-factor=0 output-factor=0 Distortion KENLM lazyken=0 name=LM0 factor=0 path=/home/kamelnebhi/recaser/training//cased.srilm.gz order=3 input-factors: 0 mapping: 0 T 0 weight: UnknownWordPenalty0= 1 WordPenalty0= -1 PhrasePenalty0= 0.2 TranslationModel0= 0.2 0.2 0.2 0.2 Distortion0= 0.3 LM0= 0.5 /home/kamelnebhi/mosesdecoder/bin line=UnknownWordPenalty FeatureFunction: UnknownWordPenalty0 start: 0 end: 0 line=WordPenalty FeatureFunction: WordPenalty0 start: 1 end: 1 line=PhrasePenalty FeatureFunction: PhrasePenalty0 start: 2 end: 2 line=PhraseDictionaryMemory name=TranslationModel0 table-limit=20 num-features=4 path=/home/kamelnebhi/recaser/training/phrase-table.gz input-factor=0 output-factor=0 FeatureFunction: TranslationModel0 start: 3 end: 6 line=Distortion FeatureFunction: Distortion0 start: 7 end: 7 line=KENLM lazyken=0 name=LM0 factor=0 path=/home/kamelnebhi/recaser/training//cased.srilm.gz order=3 FeatureFunction: LM0 start: 8 end: 8 Loading the LM will be faster if you build a binary file. Reading /home/kamelnebhi/recaser/training//cased.srilm.gz ----5---10---15---20---25---30---35---40---45---50---55---60---65---70---75---80---85---90---95--100 *The ARPA file is missing <unk>. Substituting log10 probability -100. *************************************************************************************************** Loading UnknownWordPenalty0 Loading WordPenalty0 Loading PhrasePenalty0 Loading Distortion0 Loading LM0 Loading TranslationModel0 Start loading text SCFG phrase table. Moses format : [3.69361] seconds Reading /home/kamelnebhi/recaser/training/phrase-table.gz ----5---10---15---20---25---30---35---40---45---50---55---60---65---70---75---80---85---90---95--100 **************************************************************************************************** Erreur de segmentation (core dumped) root@kamelnebhi-MacBookPro:/home/kamelnebhi# /home/kamelnebhi/mosesdecoder/bin/mosesserver -f /home/kamelnebhi/recaser/training/moses.ini --server-port 80 Defined parameters (per moses.ini or switch): config: /home/kamelnebhi/recaser/training/moses.ini distortion-limit: 6 feature: UnknownWordPenalty WordPenalty PhrasePenalty PhraseDictionaryMemory name=TranslationModel0 table-limit=20 num-features=4 path=/home/kamelnebhi/recaser/training/phrase-table.gz input-factor=0 output-factor=0 Distortion KENLM lazyken=0 name=LM0 factor=0 path=/home/kamelnebhi/recaser/training//cased.srilm.gz order=3 input-factors: 0 mapping: 0 T 0 weight: UnknownWordPenalty0= 1 WordPenalty0= -1 PhrasePenalty0= 0.2 TranslationModel0= 0.2 0.2 0.2 0.2 Distortion0= 0.3 LM0= 0.5 /home/kamelnebhi/mosesdecoder/bin line=UnknownWordPenalty FeatureFunction: UnknownWordPenalty0 start: 0 end: 0 line=WordPenalty FeatureFunction: WordPenalty0 start: 1 end: 1 line=PhrasePenalty FeatureFunction: PhrasePenalty0 start: 2 end: 2 line=PhraseDictionaryMemory name=TranslationModel0 table-limit=20 num-features=4 path=/home/kamelnebhi/recaser/training/phrase-table.gz input-factor=0 output-factor=0 FeatureFunction: TranslationModel0 start: 3 end: 6 line=Distortion FeatureFunction: Distortion0 start: 7 end: 7 line=KENLM lazyken=0 name=LM0 factor=0 path=/home/kamelnebhi/recaser/training//cased.srilm.gz order=3 FeatureFunction: LM0 start: 8 end: 8 Loading the LM will be faster if you build a binary file. Reading /home/kamelnebhi/recaser/training//cased.srilm.gz ----5---10---15---20---25---30---35---40---45---50---55---60---65---70---75---80---85---90---95--100 *The ARPA file is missing <unk>. Substituting log10 probability -100. *************************************************************************************************** Loading UnknownWordPenalty0 Loading WordPenalty0 Loading PhrasePenalty0 Loading Distortion0 Loading LM0 Loading TranslationModel0 Start loading text SCFG phrase table. Moses format : [3.69152] seconds Reading /home/kamelnebhi/recaser/training/phrase-table.gz ----5---10---15---20---25---30---35---40---45---50---55---60---65---70---75---80---85---90---95--100 **************************************************************************************************** Segmentation fault (core dumped) *Thanks for your help* -------------- next part -------------- An HTML attachment was scrubbed... URL: http://mailman.mit.edu/mailman/private/moses-support/attachments/20140407/82302fab/attachment-0001.htm ------------------------------ Message: 2 Date: Mon, 7 Apr 2014 16:00:47 -0400 From: Philipp Koehn <pko...@inf.ed.ac.uk> Subject: Re: [Moses-support] Monolingual Word alignment To: Mostafa Dehghani <dehghani.most...@gmail.com> Cc: "moses-support@mit.edu" <moses-support@mit.edu> Message-ID: <CAAFADDBXoZv7=u5rqjay3brfwob92ywkrvfekdz6yrx29m7...@mail.gmail.com> Content-Type: text/plain; charset=ISO-8859-1 Hi, this outcome is not that surprising to me. If you align identical sentences, then translating each word to itself is a pretty good model. Since your goal is paraphrasing words into synonyms, you should rather use methods such as the one proposed by Bannard and Callison-Burch: http://acl.ldc.upenn.edu/P/P05/P05-1074.pdf -phi On Sun, Apr 6, 2014 at 11:11 AM, Mostafa Dehghani <dehghani.most...@gmail.com> wrote: > Dear all, > > I am working on a method for Multilingual Information Retrieval. In my > method I expand the text of each document by probabilistically translating > its words to other languages' words (interlingual expansion). However, to > pass some axioms, I need to expand text of each document in its own language > (intralingual expansion). So, beside bilingual word alignments, I need > monolingual word alignments table (that probably contains the alignment of > each word to the words those are related/concurred with that word). To do > so, I used one side of each language sentences and their copy as parallel > corpus. Then I used the following command: > > > train-model.perl -root-dir train -corpus corpus/fr-fr -f fr1 -e fr2 > -alignment grow-diag-final-and -reordering msd-bidirectional-fe > -external-bin-dir externalbin -last-step 4 > > > such that fr-fr.fr1 fr-fr.fr2 are the same files containing French > sentences. > However, I got f2e and e2f files that are only contain alignments of each > word to itself with probability of 1. > I am wondering is there any parameter that I should set to achieve words > alignments (e2f/f2e) those are proper for intralingual expansion? > > Regards, > > -- > Mostafa > , > > http://khorshid.ut.ac.ir/~m.dehghani > > _______________________________________________ > Moses-support mailing list > Moses-support@mit.edu > http://mailman.mit.edu/mailman/listinfo/moses-support > ------------------------------ Message: 3 Date: Mon, 7 Apr 2014 23:27:07 +0100 From: Lucia Specia <lspe...@gmail.com> Subject: [Moses-support] Call for Participation: Automatic and Manual Metrics for Operational Translation Evaluation To: moses-support@mit.edu, wmt-ta...@googlegroups.com Message-ID: <caaleuxzvnsv0xp-uvxok0z8qwtkd-jt11ftoporho0ras9u...@mail.gmail.com> Content-Type: text/plain; charset="iso-8859-1" Dear all, This workshop may be relevant for those of you interested in MT evaluation metrics. ---- Automatic and Manual Metrics for Operational Translation Evaluation http://mte2014.github.io/ 26 May 2014 Workshop at Language Resources and Evaluation Conference (LREC) 2014 http://lrec2014.lrec-conf.org In brief: We invite you to join us for an interesting day of work (and play!) as we discuss metrics for machine translation quality assessment and participate in some hands-on task-based translation evaluation. This workshop on Automatic and Manual Metrics for Operational Translation Evaluation (MTE 2014) will be a full-day LREC workshop to be held on Monday, May 26, 2014 in Reykjavik, Iceland. The format of MTE 2014 will be interactive and energizing: a half-day of short presentations and discussion of recent work on machine translation quality assessment, followed by a half-day of hands-on collaborative work with MT metrics that show promise for the prediction of task suitability of MT output. The afternoon hands-on work will follow from the morning's presentations, with some of the hands-on exercises developed directly from the submissions to the workshop. Details: While a significant body of work has been done by the machine translation (MT) research community towards the development and meta-evaluation of automatic metrics to assess overall MT quality, less attention has been dedicated to more operational evaluation metrics aimed at testing whether translations are adequate within a specific context: purpose, end-user, task, etc., and why the MT system fails in some cases. Both of these can benefit from some form of manual analysis. Most work in this area is limited to productivity tests (e.g. contrasting time for human translation and MT post-editing). A few initiatives consider more detailed metrics for the problem, which can also be used to understand and diagnose errors in MT systems. These include the Multidimensional Quality Metrics (MQM) recently proposed by the EU F7 project QTLaunchPad, the TAUS Dynamic Quality Framework, and past projects such as FEMTI, EAGLES and ISLE. Some of these metrics are also applicable to human translation evaluation. A number of task-based metrics have also been proposed for applications such as topic ID / triage and reading comprehension. The purpose of this workshop is to bring together representatives from academia, industry and government institutions to discuss and assess metrics for manual and automatic quality evaluation, with an eye toward how they might be leveraged or further developed into task-based metrics for more objective "fitness for purpose" assessment. We will also consider comparisons to well-established metrics for automatic evaluation such as BLEU, METEOR and others, including reference-less metrics for quality prediction. The workshop will benefit from datasets already collected and manually annotated for translation errors by the QTLaunchPad project (http://www.qt21.eu/launchpad/) and will cover concepts from many the metrics proposed by participants in the half-day of hands-on tasks. Up-to-the-minute information and (most importantly) Registration: Additional details and schedule will be posted at the workshop website http://mte2014.github.io/ as they become available. Register to attend via the LREC registration site at http://lrec2014.lrec-conf.org/en/registration/ . We look forward to seeing you there! The MTE 2014 Organizing Committee Keith J. Miller (MITRE) Lucia Specia (University of Sheffield) Kim Harris (GALA and text & form) Stacey Bailey (MITRE) -------------- next part -------------- An HTML attachment was scrubbed... URL: http://mailman.mit.edu/mailman/private/moses-support/attachments/20140407/a8dbf5d0/attachment.htm ------------------------------ _______________________________________________ Moses-support mailing list Moses-support@mit.edu http://mailman.mit.edu/mailman/listinfo/moses-support End of Moses-support Digest, Vol 90, Issue 19 *********************************************
_______________________________________________ Moses-support mailing list Moses-support@mit.edu http://mailman.mit.edu/mailman/listinfo/moses-support