[please excuse cross-posting]

Hi Everyone,

We have exciting news to share with you as part of our effort to support
research on low resource machine translation: 1) we are releasing Flores
101, a large evaluation benchmark for multilingual machine translation in
over 100 languages, 2) we are organizing a multilingual machine translation
task at WMT and 3) we are supporting WMT participants to the large-scale
multilingual task with compute credits that we encourage people to apply
for.

If you want to learn more, please read below for more details and feel free
to respond to this email if you have any questions.

We hope you’ll participate at this year’s WMT MMT task and leverage the
upcoming Flores 101 in your future research, and that together as a
community we can make collective progress on improving translation quality
on low resource languages.

Best regards,

Paco Guzmán on behalf of the FLORES 101 Team

flo...@fb.com


Context

Translation is a key technology to connect people and ideas together across
language barriers. However, current translation technology works very well
mostly in a few languages and it covers only few domains. Many people
around the world still lack access, partly, due to the lack of compute and
data resources to create translation models.

A prerequisite for developing new modeling techniques is having reliable
evaluation. As a baby step in this direction back in 2019, we started FLORES
<https://urldefense.com/v3/__https://arxiv.org/abs/1902.01382__;!!Bt8RZUm9aw!pF4KnNrHjHNSFek_lDpt6hqt8cAKqShptahb0OdNBUrc0LHv1zjtelB1_3IF$>,
which came with two evaluation datasets for Nepali-English and
Sinhala-English, that we later expanded
<https://github.com/facebookresearch/flores> to include Pashto and Khmer.

Today, we announce the FLORES101 evaluation benchmark:  a full Many-to-Many
evaluation dataset across over 100 languages, most of which are
low-resource. True to the original multi-domain spirit of FLORES, this
dataset consists of 3000 English sentences across several domains (news,
books, and travel) all taken from Wikipedia, maintaining document-level
context as well as document metadata, such as topics, hyperlinks, etc.
These sentences are then professionally translated though several rounds of
thorough evaluation. You can see the full list of languages at the bottom
of the page [here
<https://urldefense.com/v3/__http://www.statmt.org/wmt21/large-scale-multilingual-translation-task.html__;!!Bt8RZUm9aw!pF4KnNrHjHNSFek_lDpt6hqt8cAKqShptahb0OdNBUrc0LHv1zjtegIOCWT4$>
].

We are making the entire dev and devtest splits of FLORES101 available to
the research community (2000 sentences total, aligned Many to Many), in
June 2021, along with a tech report describing the dataset in detail.

To ensure robust and fair evaluation, we’ll keep the test split blind and
not publicly accessible. Instead, we’ll host an evaluation server based on
open-source code, which will enable us to track the progress of the
community in low-resource languages. Importantly, such a setup will enable
comparison of models on several axes besides translation quality, such as
compute and memory efficiency. The evaluation server will also be available
starting June 2021.

Finally, we want to continue encouraging the research community to work on
low-resource translation. As part of this, we are launching a WMT
multilingual machine translation track and encourage people to apply for
compute grants so that GPU compute is less of a barrier for translation
research. You can see more detailed information and apply for the compute
grant  [here
<https://urldefense.com/v3/__http://www.statmt.org/wmt21/flores-compute-grants.html__;!!Bt8RZUm9aw!pF4KnNrHjHNSFek_lDpt6hqt8cAKqShptahb0OdNBUrc0LHv1zjtehir9L9K$>].
We propose two small tracks --- one low-resource European languages and
another low-resource Southeast Asian languages --- along with the full
track of 100+ languages.
_______________________________________________
Mt-list site list
Mt-list@eamt.org
http://lists.eamt.org/mailman/listinfo/mt-list

Reply via email to