Rethinking Low-Resource MT: The Surprising Effectiveness of Fine-Tuned Multilingual Models in the LLM Age
| dc.contributor.author | Scalvini, Barbara | |
| dc.contributor.author | Debess, Iben Nyholm | |
| dc.contributor.author | Simonsen, Annika | |
| dc.contributor.author | Einarsson, Hafsteinn | |
| dc.contributor.editor | Johansson, Richard | |
| dc.contributor.editor | Stymne, Sara | |
| dc.coverage.spatial | Tallinn, Estonia | |
| dc.date.accessioned | 2025-02-19T08:20:16Z | |
| dc.date.available | 2025-02-19T08:20:16Z | |
| dc.date.issued | 2025-03 | |
| dc.description.abstract | This study challenges the current paradigm shift in machine translation, where large language models (LLMs) are gaining prominence over traditional neural machine translation models, with a focus on English-to-Faroese translation. We compare the performance of various models, including fine-tuned multilingual models, LLMs (GPT-SW3, Llama 3.1), and closed-source models (Claude 3.5, GPT-4). Our findings show that a fine-tuned NLLB model outperforms most LLMs, including some larger models, in both automatic and human evaluations. We also demonstrate the effectiveness of using LLM-generated synthetic data for fine-tuning. While closed-source models like Claude 3.5 perform best overall, the competitive performance of smaller, fine-tuned models suggests a more nuanced approach to low-resource machine translation. Our results highlight the potential of specialized multilingual models and the importance of language-specific knowledge. We discuss implications for resource allocation in low-resource settings and suggest future directions for improving low-resource machine translation, including targeted data creation and more comprehensive evaluation methodologies. | |
| dc.identifier.uri | https://hdl.handle.net/10062/107255 | |
| dc.language.iso | en | |
| dc.publisher | University of Tartu Library | |
| dc.relation.ispartofseries | NEALT Proceedings Series, No. 57 | |
| dc.rights | Attribution-NonCommercial-NoDerivatives 4.0 International | |
| dc.rights.uri | https://creativecommons.org/licenses/by-nc-nd/4.0/ | |
| dc.title | Rethinking Low-Resource MT: The Surprising Effectiveness of Fine-Tuned Multilingual Models in the LLM Age | |
| dc.type | Article |
Failid
Originaal pakett
1 - 1 1