Strategies for deploying unreliable AI graders in high-transparency high-stakes exams

27Citations
Citations of this article
39Readers
Mendeley users who have this article in their library.

This article is free to access.

Abstract

We describe the deployment of an imperfect NLP-based automatic short answer grading system on an exam in a large-enrollment introductory college course. We characterize this deployment as both high stakes (the questions were on an mid-term exam worth 10% of students’ final grade) and high transparency (the question was graded interactively during the computer-based exam and correct solutions were shown to students that could be compared to their answer). We study two techniques designed to mitigate the potential student dissatisfaction resulting from students incorrectly not granted credit by the imperfect AI grader. We find (1) that providing multiple attempts can eliminate first-attempt false negatives at the cost of additional false positives, and (2) that students not granted credit from the algorithm cannot reliably determine if their answer was mis-scored.

Cite

CITATION STYLE

APA

Azad, S., Chen, B., Fowler, M., West, M., & Zilles, C. (2020). Strategies for deploying unreliable AI graders in high-transparency high-stakes exams. In Lecture Notes in Computer Science (Vol. 12163 LNAI, pp. 16–28). Springer Science and Business Media Deutschland GmbH. https://doi.org/10.1007/978-3-030-52237-7_2

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free