[please excuse cross-posting]

Hi Everyone,


We have exciting news to share with you as part of our effort to support 
research on low resource machine translation: 1) we are releasing Flores 101, a 
large evaluation benchmark for multilingual machine translation in over 100 
languages, 2) we are organizing a multilingual machine translation task at WMT 
and 3) we are supporting WMT participants to the large-scale multilingual task 
with compute credits that we encourage people to apply for.

If you want to learn more, please read below for more details and feel free to 
respond to this email if you have any questions.


We hope you’ll participate at this year’s WMT MMT task and leverage the 
upcoming Flores 101 in your future research, and that together as a community 
we can make collective progress on improving translation quality on low 
resource languages.


Best regards,

Paco Guzmán

on behalf of the FLORES 101 Team

flo...@fb.com



Context

Translation is a key technology to connect people and ideas together across 
language barriers. However, current translation technology works very well 
mostly in a few languages and it covers only a few domains. Many people around 
the world still lack access, partly, due to the lack of compute and data 
resources to create translation models.


A prerequisite for developing new modeling techniques is having reliable 
evaluation. As a baby step in this direction back in 2019, we started 
FLORES<https://arxiv.org/abs/1902.01382>, which came with two evaluation 
datasets for Nepali-English and Sinhala-English, that we later 
expanded<https://github.com/facebookresearch/flores> to include Pashto and 
Khmer.


Today, we announce the FLORES101 evaluation benchmark:  a full Many-to-Many 
evaluation dataset across over 100 languages, most of which are low-resource. 
True to the original multi-domain spirit of FLORES, this dataset consists of 
3000 English sentences across several domains (news, books, and travel) all 
taken from Wikipedia, maintaining document-level context as well as document 
metadata, such as topics, hyperlinks, etc. These sentences are then 
professionally translated though several rounds of thorough evaluation. You can 
see the full list of languages at the bottom of the page 
[here<http://www.statmt.org/wmt21/large-scale-multilingual-translation-task.html>].


We are making the entire dev and devtest splits of FLORES101 available to the 
research community (2000 sentences total, aligned Many to Many), in June 2021, 
along with a tech report describing the dataset in detail.


To ensure robust and fair evaluation, we’ll keep the test split blind and not 
publicly accessible. Instead, we’ll host an evaluation server based on 
open-source code, which will enable us to track the progress of the community 
in low-resource languages. Importantly, such a setup will enable comparison of 
models on several axes besides translation quality, such as compute and memory 
efficiency. The evaluation server will also be available starting June 2021.


Finally, we want to continue encouraging the research community to work on 
low-resource translation. As part of this, we are launching a WMT multilingual 
machine translation track and encourage people to apply for compute grants so 
that GPU compute is less of a barrier for translation research. You can see 
more detailed information and apply for the compute grant  
[here<http://www.statmt.org/wmt21/flores-compute-grants.html>]. We propose two 
small tracks --- one for low-resource European languages and another one for

low-resource Southeast Asian languages --- along with the full track of 100+ 
languages.

​
_______________________________________________
Moses-support mailing list
Moses-support@mit.edu
http://mailman.mit.edu/mailman/listinfo/moses-support

Reply via email to