An Exploration of Encoder-Decoder Approaches to Multi-Label Classification for Legal and Biomedical Text

14Citations
Citations of this article
32Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Standard methods for multi-label text classification largely rely on encoder-only pre-trained language models, whereas encoder-decoder models have proven more effective in other classification tasks. In this study, we compare four methods for multi-label classification, two based on an encoder only, and two based on an encoder-decoder. We carry out experiments on four datasets-two in the legal domain and two in the biomedical domain, each with two levels of label granularity- and always depart from the same pre-trained model, T5. Our results show that encoder-decoder methods outperform encoder-only methods, with a growing advantage on more complex datasets and labeling schemes of finer granularity. Using encoder-decoder models in a non-autoregressive fashion, in particular, yields the best performance overall, so we further study this approach through ablations to better understand its strengths.

Cite

CITATION STYLE

APA

Kementchedjhieva, Y., & Chalkidis, I. (2023). An Exploration of Encoder-Decoder Approaches to Multi-Label Classification for Legal and Biomedical Text. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (pp. 5828–5843). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2023.findings-acl.360

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free