Towards Interpretable and Efficient Automatic Reference-Based Summarization Evaluation

11Citations
Citations of this article
16Readers
Mendeley users who have this article in their library.

Abstract

Interpretability and efficiency are two important considerations for the adoption of neural automatic metrics. In this work, we develop strong-performing automatic metrics for reference-based summarization evaluation, based on a two-stage evaluation pipeline that first extracts basic information units from one text sequence and then checks the extracted units in another sequence. The metrics we developed include two-stage metrics that can provide high interpretability at both the fine-grained unit level and summary level, and one-stage metrics that achieve a balance between efficiency and interpretability. We make the developed tools publicly available at https://github.com/Yale-LILY/AutoACU.

Cite

CITATION STYLE

APA

Liu, Y., Fabbri, A. R., Zhao, Y., Liu, P., Joty, S., Wu, C. S., … Radev, D. (2023). Towards Interpretable and Efficient Automatic Reference-Based Summarization Evaluation. In EMNLP 2023 - 2023 Conference on Empirical Methods in Natural Language Processing, Proceedings (pp. 16360–16368). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2023.emnlp-main.1018

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free