D 2025

Efficient Architectures For Low-Resource Machine Translation

SIGNORONI, Edoardo; Pavel RYCHLÝ and Ruggero SIGNORONI

Basic information

Original name

Efficient Architectures For Low-Resource Machine Translation

Authors

SIGNORONI, Edoardo; Pavel RYCHLÝ and Ruggero SIGNORONI

Edition

Shoumen, BULGARIA, Proceedings of the First Workshop on Advancing NLP for Low-Resource Languages associated with the International Conference RANLP 2025, p. 39-64, 26 pp. 2025

Publisher

INCOMA Ltd.

Other information

Language

English

Type of outcome

Proceedings paper

Country of publisher

Bulgaria

Confidentiality degree

is not subject to a state or trade secret

Publication form

electronic version available online

References:

Marked to be transferred to RIV

No

Organization

Fakulta informatiky – Repository – Repository

ISBN

978-954-452-100-4

Keywords in English

Machine Translation; Low-Resource Languages

Links

LM2023062, research and development project.
Changed: 18/3/2026 00:50, RNDr. Daniel Jakubík

Abstract

In the original language

Low-resource Neural Machine Translation is highly sensitive to hyperparameters and needs careful tuning to achieve the best results with small amounts of training data. We focus on exploring the impact of changes in the Transformer architecture on downstream translation quality, and propose a metric to score the computational efficiency of such changes. By experimenting on English-Akkadian, German-Lower Sorbian, English-Italian, and English-Manipuri, we confirm previous finding in low-resource machine translation optimization, and show that smaller and more parameter-efficient models can achieve the same translation quality of larger and unwieldy ones at a fraction of the computational cost. Optimized models have around 95% less parameters, while dropping only up to 14.8% ChrF. We compile a list of optimal ranges for each hyperparameter.

Files attached