Rethinking Low-Resource MT: The Surprising Effectiveness of Fine-Tuned Multilingual Models in the LLM Age

Barbara Scalvini, Iben Nyholm Debess, Annika Simonsen, Hafsteinn Einarsson

Research output: Contribution to conferencePaperpeer-review

1 Downloads (Pure)

Abstract

This study challenges the current paradigm shift in machine translation, where large language models (LLMs) are gaining prominence over traditional neural machine translation models, with a focus on English-to-Faroese translation. We compare the performance of various models, including fine-tuned multilingual models, LLMs (GPT-SW3, Llama 3.1), and closed-source models (Claude 3.5, GPT-4). Our findings show that a fine-tuned NLLB model outperforms most LLMs, including some larger models, in both automatic and human evaluations. We also demonstrate the effectiveness of using LLM-generated synthetic data for fine-tuning. While closed-source models like Claude 3.5 perform best overall, the competitive performance of smaller, fine-tuned models suggests a more nuanced approach to low-resource machine translation. Our results highlight the potential of specialized multilingual models and the importance of language-specific knowledge. We discuss implications for resource allocation in low-resource settings and suggest future directions for improving low-resource machine translation, including targeted data creation and more comprehensive evaluation methodologies.
Original languageEnglish
Pages609-621
Number of pages13
Publication statusPublished - 2025
EventNodalida/Baltic-HLT 2025: The Joint 25th Nordic Conference on Computational Linguistics and 11th Baltic Conference on Human Language Technologies - Tallinn, Estonia
Duration: 2 Mar 20255 Mar 2025
Conference number: 25
https://www.nodalida-bhlt2025.eu/conference

Conference

ConferenceNodalida/Baltic-HLT 2025
Country/TerritoryEstonia
CityTallinn
Period2/03/255/03/25
Internet address

Keywords

  • LLM
  • Language models
  • multilingual models

Fingerprint

Dive into the research topics of 'Rethinking Low-Resource MT: The Surprising Effectiveness of Fine-Tuned Multilingual Models in the LLM Age'. Together they form a unique fingerprint.

Cite this