MIPE: A Metric Independent Pipeline for Effective Code-Mixed NLG Evaluation

2Citations
Citations of this article
18Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Code-mixing is a phenomenon of mixing words and phrases from two or more languages in a single utterance of speech and text. Due to the high linguistic diversity, code-mixing presents several challenges in evaluating standard natural language generation (NLG) tasks. Various widely popular metrics perform poorly with the code-mixed NLG tasks. To address this challenge, we present a metric independent evaluation pipeline MIPE that significantly improves the correlation between evaluation metrics and human judgments on the generated code-mixed text. As a use case, we demonstrate the performance of MIPE on the machine-generated Hinglish (code-mixing of Hindi and English languages) sentences from the HinGE corpus. We can extend the proposed evaluation strategy to other code-mixed language pairs, NLG tasks, and evaluation metrics with minimal to no effort.

Cite

CITATION STYLE

APA

Garg, A., Kagi, S. S., Srivastava, V., & Singh, M. (2021). MIPE: A Metric Independent Pipeline for Effective Code-Mixed NLG Evaluation. In Eval4NLP 2021 - Evaluation and Comparison of NLP Systems, Proceedings of the 2nd Workshop (pp. 123–132). Association for Computational Linguistics (ACL). https://doi.org/10.26615/978-954-452-056-4_013

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free