Embed_Llama: using LLM embeddings for the Metrics Shared Task

9Citations
Citations of this article
12Readers
Mendeley users who have this article in their library.

Abstract

Embed_llama is an assessment metric for language translation that hinges upon the utilization of the recently introduced Llama 2 Large Language Model (LLM), specifically focusing on its embedding layer, to transform sentences into a vector space that establishes connections between geometric and semantic proximities. Investigations utilizing previous WMT datasets have revealed that within the Llama 2 architecture, relying solely on the initial embedding layer does not result in the highest degree of correlation when assessing machine translations. The incorporation of additional layers, however, holds the potential to augment the contextual understanding of sentences. As a contribution to the WMT23 challenge, this study delves into the advantages derived from employing a pre-trained LLM that has not undergone fine-tuning specifically for translation evaluation tasks, to provide a metric conducive to operation on readily accessible consumer-grade hardware. This research digs into the observation that deeper layers within the model do not result in a linear increase in the spatial proximity between sentences within the vector space.

Cite

CITATION STYLE

APA

Dréano, S., Molloy, D., & Murphy, N. (2023). Embed_Llama: using LLM embeddings for the Metrics Shared Task. In Conference on Machine Translation - Proceedings (pp. 736–743). Association for Computational Linguistics. https://doi.org/10.18653/v1/2023.wmt-1.60

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free