Fine-tuning Transformers with Additional Context to Classify Discursive Moves in Mathematics Classrooms

31Citations
Citations of this article
49Readers
Mendeley users who have this article in their library.
Get full text

Abstract

“Talk moves” are specific discursive strategies used by teachers and students to facilitate conversations in which students share their thinking, and actively consider the ideas of others, and engage in rich discussions. Experts in instructional practices often rely on cues to identify and document these strategies, for example by annotating classroom transcripts. Prior efforts to develop automated systems to classify teacher talk moves using transformers achieved a performance of 76.32% F1. In this paper, we investigate the feasibility of using enriched contextual cues to improve model performance. We applied state-of-the-art deep learning approaches for Natural Language Processing (NLP), including Robustly optimized bidirectional encoder representations from transformers (Roberta) with a special input representation that supports previous and subsequent utterances as context for talk moves classification. We worked with the publically available TalkMoves dataset, which contains utterances sourced from real-world classroom sessions (human- transcribed and annotated). Through a series of experimentations, we found that a combination of previous and subsequent utterances improved the transformers’ ability to differentiate talk moves (by 2.6% F1). These results constitute a new state of the art over previously published results and provide actionable insights to those in the broader NLP community who are working to develop similar transformer-based classification models.

Cite

CITATION STYLE

APA

Suresh, A., Jacobs, J., Perkoff, M., Martin, J. H., & Sumner, T. (2022). Fine-tuning Transformers with Additional Context to Classify Discursive Moves in Mathematics Classrooms. In BEA 2022 - 17th Workshop on Innovative Use of NLP for Building Educational Applications, Proceedings (pp. 71–81). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2022.bea-1.11

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free