Machine Translation

Covid-19 MLIA Eval

Task Description

The goal of the Machine Translation (MT) task is to evaluate systems focused on the Covid-19 related text. The first round of the Covid-19 MT task addresses the following language pairs:

  • English-German
  • English-French
  • English-Spanish
  • English-Italian
  • English-Modern Greek
  • English-Swedish
All languages pairs only in the direction translating from English to the other language. The main challenge is that the text to be translated is specialized on the new and high-relevant topic of Covid-19. The task is open for beginners and established research groups from any area of interest in the scientific community, the public administration and the industry. At the end of each round, participants will write/update an incremental report explaining their system. The report will highlight which methods data have been used.

To participate in the Machine Translation task, groups need to register at the following link:


Important Dates - Round 1:

Round starts: October 23, 2020

Release of training data: October 23, 2020

Release of test data: November 20, 2020

Translations submission deadline: November 27, 2020 -> extended to December 2, 2020

Translations scored: December 4, 2020

Rolling report submission deadline (peliminary version): December 23, 2020

Rolling report submission deadline (camera ready): January 8, 2021

Slot for a virtual meeting to discuss the results: January 12-14, 2021

Round ends: January 15, 2021

Participation Guidelines

Organizers will provide training data for all language pairs. Participants must submit at least a system trained only with the provided data (constrained) for each of the language pairs they would like to participate. Additionally, participants can use additional training data (not provided by the organisers) or existing translation systems specifying a flag that the system uses additional data (unconstrained). System submissions that used the provided training data (constrained) will be distinguished from submissions that used additional data resources (unconstrained). Note that basic linguistic tools such as taggers, parsers, or morphological analyzers or multilingual systems are allowed in the constrained condition.

Participants will use their systems to translate a test set of unseen sentences in the source language. The translation quality is measured by various automatic evaluation metrics (BLEU will be the main evaluation metric for the 1st round). You may participate in any or all of the language pairs. Organizers will provide a framework to show the results to be compared.

Participant Repository:

Participants are provided with a single repository for all the tasks they take part in. The repository contains the runs, resources, code, and report of each participant.

The repository is organised as follows:

Covid-19 MLIA Eval consists of three tasks run in three rounds. Therefore, the submission and score folders are organized into sub-folders for each task and round as follows:

Participants which do not take part in a given task or round can simply delete the corresponding sub-folders.

The goal of Covid-19 MLIA Eval is to speed up the creation of multilingual information access systems and (language) resources for Covid-19 as well as openly share these systems and resources as much as possible. Therefore, participants are more than encouraged to share their code and any additional (language) resources they have used or created.

All the contents of these repositories are released under the Creative Commons Attribution-ShareAlike 4.0 International License.

Rolling Technical Report:

The rolling technical report should be formatted according to the Springer LNCS format, using either the LaTeX template or the Word template. LaTeX is the preferred format.



Automatic Evaluation:

A ranking with the results of the automatic evaluation is available at this website. This ranking will be updated periodically until the translations submission deadline is passed. New: Final results have been published in a very rough version of the findings rolling report.

Submission Guidelines

Participating teams should satisfy the following guidelines:

Submission Upload:

Runs should be uploaded in the repository provided by the organizers. Following the repository structure discussed above, for example, a run submitted for the first round of the Machine Translation task should be included in submission/task3/round1.

Runs should be uploaded with the following name convention: <teamname>_task3_<round>_<languagedirection>_<constrainedfield>_<descriptionfield>.sgm where:

For example, a complete run identifier may look like pangeanic_task3_round1_en2de_constrained_bt.sgm where:

Performance scores for the submitted runs will be returned by the organizers in the score folder, which follows the same structure as the submission folder.

The rolling technical report has to be uploaded and kept update in the report folder.

Here, you can find a sample participant repository to get a better idea of its layout.


The quality of the submitted systems will be evaluated with the following measures:


Francisco Casacuberta, Universitat Politècnica de València, Spain

Miguel Domingo, Universitat Politècnica de València, Spain

Mercedes García-Martínez, Pangeanic, Spain

Manuel Herranz, Pangeanic, Spain