Abstract
To understand a legal document for real-world applications, SemEval-2023 Task 6 proposes a shared Subtask A, rhetorical roles (RRs) prediction, which requires a system to automatically assign a RR label for each semantical segment in a legal text. In this paper, we propose a LEGAL-BERT based hierarchical BiLSTM model with conditional random field (CRF) for RR prediction, which primarily consists of two parts: word-level and sentence-level encoders. The word-level encoder first adopts a legal-domain pre-trained language model, LEGAL-BERT, initially word-embedding words in each sentence in a document and a word-level BiLSTM further encoding such sentence representation. The sentence-level encoder then uses an attentive pooling method for sentence embedding and a sentence-level BiLSTM for document modeling. Finally, a CRF is utilized to predict RRs for each sentence. The officially released results show that our method outperformed the baseline systems. Our team won 7th rank out of 27 participants in Subtask A.
Cite
CITATION STYLE
Chen, Y., Zhang, Y., Wang, J., & Zhang, X. (2023). YNU-HPCC at SemEval-2023 Task 6: LEGAL-BERT based Hierarchical BiLSTM with CRF for Rhetorical Roles Prediction. In 17th International Workshop on Semantic Evaluation, SemEval 2023 - Proceedings of the Workshop (pp. 2075–2081). Association for Computational Linguistics. https://doi.org/10.18653/v1/2023.semeval-1.286
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.