Principled Multi-Aspect Evaluation Measures of Rankings

1Citations
Citations of this article
13Readers
Mendeley users who have this article in their library.

Abstract

Information Retrieval evaluation has traditionally focused on defining principled ways of assessing the relevance of a ranked list of documents with respect to a query. Several methods extend this type of evaluation beyond relevance, making it possible to evaluate different aspects of a document ranking (e.g., relevance, usefulness, or credibility) using a single measure (multi-aspect evaluation). However, these methods either are (i) tailor-made for specific aspects and do not extend to other types or numbers of aspects, or (ii) have theoretical anomalies, e.g. assign maximum score to a ranking where all documents are labelled with the lowest grade with respect to all aspects (e.g., not relevant, not credible, etc.). We present a theoretically principled multi-aspect evaluation method that can be used for any number, and any type, of aspects. A thorough empirical evaluation using up to 5 aspects and a total of 425 runs officially submitted to 10 TREC tracks shows that our method is more discriminative than the state-of-the-art and overcomes theoretical limitations of the state-of-the-art.

Cite

CITATION STYLE

APA

Maistro, M., Lima, L. C., Simonsen, J. G., & Lioma, C. (2021). Principled Multi-Aspect Evaluation Measures of Rankings. In International Conference on Information and Knowledge Management, Proceedings (pp. 1232–1242). Association for Computing Machinery. https://doi.org/10.1145/3459637.3482287

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free