Machine Translation with Large Language Models: Prompting, Few-shot Learning, and Fine-tuning with QLoRA

14Citations
Citations of this article
23Readers
Mendeley users who have this article in their library.

Abstract

While large language models have made remarkable advancements in natural language generation, their potential in machine translation, especially when fine-tuned, remains under-explored. In our study, we conduct comprehensive experiments, evaluating 15 publicly available language models on machine translation tasks. We compare the performance across three methodologies: zero-shot prompting, few-shot learning, and fine-tuning. Central to our approach is the use of QLoRA, an efficient finetuning method. On French-English, QLoRA fine-tuning outperforms both few-shot learning and models trained from scratch. This superiority is highlighted in both sentence-level and document-level translations, with a significant BLEU score improvement of 28.93 over the prompting method. Impressively, with QLoRA, the enhanced performance is achieved by finetuning a mere 0.77% of the model's parameters.

Cite

CITATION STYLE

APA

Zhang, X., Rajabi, N., Duh, K., & Koehn, P. (2023). Machine Translation with Large Language Models: Prompting, Few-shot Learning, and Fine-tuning with QLoRA. In Conference on Machine Translation - Proceedings (pp. 466–479). Association for Computational Linguistics. https://doi.org/10.18653/v1/2023.wmt-1.43

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free