In this paper, we examine a challenge that arises in the application of peer-based tutoring: coping with inappropriate advice from peers. We examine an environment where students are presented with those learning objects predicted to improve their learning (on the basis of the success of previous, like-minded students) but where peers can additionally inject annotations. To avoid presenting annotations that would detract from student learning (e.g. those found confusing by other students) we integrate trust modeling, to detect over time the reputation of the annotation (as voted by previous students) and the reputability of the annotator. We empirically demonstrate, through simulation, that even when the environment is populated with a large number of poor annotations, our algorithm for directing the learning of the students is effective, confirming the value of our proposed approach for student modeling. In addition, the research introduces a valuable integration of trust modeling into educational applications. © 2011 Springer-Verlag.
CITATION STYLE
Champaign, J., Zhang, J., & Cohen, R. (2011). Coping with poor advice from peers in peer-based intelligent tutoring: The case of avoiding bad annotations of learning objects. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 6787 LNCS, pp. 38–49). https://doi.org/10.1007/978-3-642-22362-4_4
Mendeley helps you to discover research relevant for your work.