Abstract
In this paper we report preliminary results of a study to develop, and subsequently to automate, new metrics for assessment of information quality in text documents, particularly in news. Through focus group studies, quality judgment experiments, and textual feature extraction and analysis, we were able to generate nine quality aspects and apply them in human assessments. Experts and students participated quality experiments, during which 1000 TREC documents were evaluated by participants from two sites Albany and Rutgers. Data showed good interjudge agreement between judges from both sites. Principal component analysis revealed that the nine aspects form clusters of "content" and "presentation." Automatic quality prediction has been derived based on statistical analysis on the association between textual features and human quality judgments.
Cite
CITATION STYLE
Tang, R., Ng, K. B., Strzalkowski, T., & Kantor, P. B. (2003). Toward machine understanding of information quality. In Proceedings of the ASIST Annual Meeting (Vol. 40, pp. 213–220). John Wiley and Sons Inc. https://doi.org/10.1002/meet.1450400127
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.