Abstract
Industry datasets used for text classification are rarely created for that purpose. In most cases, the data and target predictions are a byproduct of accumulated historical data, typically fraught with noise, present in both the text-based document, as well as in the targeted labels. In this work, we address the question of how well performance metrics computed on noisy, historical data reflect the performance on the intended future machine learning model input. The results demonstrate the utility of dirty training datasets used to build prediction models for cleaner (and different) prediction inputs.
Cite
CITATION STYLE
Apostolova, E., & Andrew Kreek, R. (2018). Training and Prediction Data Discrepancies: Challenges of Text Classification with Noisy, Historical Data. In 4th Workshop on Noisy User-Generated Text, W-NUT 2018 - Proceedings of the Workshop (pp. 104–109). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/w18-6114
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.