Interpretability and efficiency are two important considerations for the adoption of neural automatic metrics. In this work, we develop strong-performing automatic metrics for reference-based summarization evaluation, based on a two-stage evaluation pipeline that first extracts basic information units from one text sequence and then checks the extracted units in another sequence. The metrics we developed include two-stage metrics that can provide high interpretability at both the fine-grained unit level and summary level, and one-stage metrics that achieve a balance between efficiency and interpretability. We make the developed tools publicly available at https://github.com/Yale-LILY/AutoACU.
CITATION STYLE
Liu, Y., Fabbri, A. R., Zhao, Y., Liu, P., Joty, S., Wu, C. S., … Radev, D. (2023). Towards Interpretable and Efficient Automatic Reference-Based Summarization Evaluation. In EMNLP 2023 - 2023 Conference on Empirical Methods in Natural Language Processing, Proceedings (pp. 16360–16368). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2023.emnlp-main.1018
Mendeley helps you to discover research relevant for your work.