Abstract
Individual taste plays an important role in assessing music performances. Aggregating ratings from multiple evaluators is commonly applied to overcome individual biases and to obtain near-objective judgments. However, high degrees of subjectivity in the evaluation of music performances inevitability lowers the agreement between music evaluators and the reliability of aggregated ratings. The present project aims at contributing to the research on music performance assessment by designing and implementing a novel procedure, the Multi-Perspective Assessment Protocol (MPAP) in music, for reliably measuring performance quality. The approach is inspired by the Noise Framework (Kahneman et al., 2021) and Generalizability Theory (G Theory; Brennan, 2001) and results show how interrater agreement can be conceptualized and operationalized efficiently in terms of rater-related variability in numerical judgments. In particular, the Generalizability and Dependability coefficients from G Theory seem well suited for providing useful information about the necessary sample size to ensure reliable performance score estimates. Moreover, the MPAP allowed to collect and compare data fromdifferent cohorts of evaluators, assessing performance quality from complementary viewpoints: aggregate judgments from expert evaluators were strongly correlated to those from a general sample of raters without expert experience, r =.786, p
Author supplied keywords
Cite
CITATION STYLE
Passarotto, E., Altenmüller, E., & Müllensiefen, D. (2023). Music Performance Assessment: Noise in Judgments and Reliability of Measurements. Psychology of Aesthetics, Creativity, and the Arts. https://doi.org/10.1037/aca0000574
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.