Using Reinforcement Learning LoRA For Interactive Machine Translation Systems

ACL ARR 2024 December Submission1235 Authors

16 Dec 2024 (modified: 05 Feb 2025)ACL ARR 2024 December SubmissionEveryoneRevisionsBibTeXCC BY 4.0
Abstract: The use of large language models (LLMs) is growing due to their impressive performance on a wide range of tasks. As new versions of these models appear to achieve better results, their size often increases, making it more challenging to maintain different versions specialized in specific domains. However, by employing the Low-Rank Adaptation (LoRA) method, we can bypass this space limitation, as the fine-tuning changes of the model are stored in a file of just a few megabytes. In the Machine Translation (MT) field, it is common to have models specialized for particular domains or language pairs. In our case, we apply these models within Interactive Machine Translation (IMT), where it is crucial that the model generates high-quality translations and adapts to user modifications. We have incorporated Reinforcement Learning (RL) techniques to optimize the model using various metrics to enhance this adaptability further. Our results demonstrate that these methods effectively improve the quality of translations generated by the models, though in some cases, this comes at the cost of a slight reduction in generalization capability.
Paper Type: Long
Research Area: Machine Translation
Research Area Keywords: fine-tuning, interactive MT, reinforcement learning
Contribution Types: Model analysis & interpretability
Languages Studied: english, spanish, french, basque, swahili, lingala, yoruba
Submission Number: 1235
Loading

OpenReview is a long-term project to advance science through improved peer review with legal nonprofit status. We gratefully acknowledge the support of the OpenReview Sponsors. © 2025 OpenReview