Wrong Answers for Wrong Reasons: The Risks of Ad Hoc Instruments

2Citations
Citations of this article
6Readers
Mendeley users who have this article in their library.
Get full text

Abstract

To evaluate novel pedagogies, approaches, and tools, Computer Science Education researchers often conduct experiments to look for differences among groups treated with different interventions. The methodological rigor of such experiments affects the soundness of the conclusions the researchers can draw. In this paper we focus on a central aspect of such experimental research: the instruments used to assess participants' knowledge. Specifically, we study the use of ad hoc instruments and the risks due to their insufficient validation. We present a literature survey that highlights how, even though standardized instruments exist, the majority of published experiments in the last five years at major Computer Science Education conferences carries out pre/post-tests using ad hoc instruments, often with multiple-choice as question type. We demonstrate the risks of such commonly used but insufficiently validated multiple-choice instruments. We propose a richer way to analyze and assess the correctness of answers to multiple-choice questions, requiring participants to add brief explanation texts as a justification of each answer. We run an experiment and analyze the collected answers using the two approaches, with and without explanations, to show that the risk of drawing opposite conclusions from the statistical analysis is real.

Cite

CITATION STYLE

APA

Chiodini, L., & Hauswirth, M. (2021). Wrong Answers for Wrong Reasons: The Risks of Ad Hoc Instruments. In ACM International Conference Proceeding Series. Association for Computing Machinery. https://doi.org/10.1145/3488042.3488045

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free