Harnessing GPT-3.5-Turbo for Rhetorical Role Prediction in Legal Cases

6Citations
Citations of this article
13Readers
Mendeley users who have this article in their library.

Abstract

We propose a comprehensive study of one-stage elicitation techniques for querying a large pre-trained generative transformer (GPT-3.5-turbo) in the rhetorical role prediction task of legal cases. This task is known as requiring textual context to be addressed. Our study explores strategies such as zero-few shots, task specification with definitions and clarification of annotation ambiguities, textual context and reasoning with general prompts and specific questions. We show that the number of examples, the definition of labels, the presentation of the (labelled) textual context and specific questions about this context have a positive influence on the performance of the model. Given non-equivalent test set configurations, we observed that prompting with a few labelled examples from direct context can lead the model to a better performance than a supervised fined-tuned multi-class classifier based on the BERT encoder (weighted F1 score of ≈72%). But there is still a gap to reach the performance of the best systems ≈86%) in the LegalEval 2023 task which, on the other hand, require dedicated resources, architectures and training.

Cite

CITATION STYLE

APA

Belfathi, A., Hernandez, N., & Monceaux, L. (2023). Harnessing GPT-3.5-Turbo for Rhetorical Role Prediction in Legal Cases. In Frontiers in Artificial Intelligence and Applications (Vol. 379, pp. 187–196). IOS Press BV. https://doi.org/10.3233/FAIA230964

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free