Animated videos in assessment: comparing validity evidence from and test-takers’ reactions to an animated and a text-based situational judgment test

14Citations
Citations of this article
43Readers
Mendeley users who have this article in their library.

This article is free to access.

Abstract

The linguistic complexity of many text-based tests can be a source of construct-irrelevant variance, as test-takers’ performance may be affected by factors that are beyond the focus of the assessment itself, such as reading comprehension skills. This experimental study examined the extent to which the use of animated videos, as opposed to written text, could (i) reduce construct-irrelevant variance attributed to language and reading skills and (ii) impact test-takers’ reactions to a situational judgment test. The results indicated that the variance attributed to construct-irrelevant factors was lower by 9.5% in the animated version of the test. In addition, those who took the animated test perceived it to be more valid, fair, and enjoyable, than those who took the text-based test. They also rated the language used as less difficult to understand. The implications of these findings are discussed.

Cite

CITATION STYLE

APA

Karakolidis, A., O’Leary, M., & Scully, D. (2021). Animated videos in assessment: comparing validity evidence from and test-takers’ reactions to an animated and a text-based situational judgment test. International Journal of Testing, 21(2), 57–79. https://doi.org/10.1080/15305058.2021.1916505

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free