Abstract
In many applications, training data is provided in the form of related datasets obtained from several sources, which typically affects the sample distribution. The learned classification models, which are expected to perform well on similar data coming from new sources, often suffer due to bias introduced by what we call 'spurious' samples - those due to source characteristics and not representative of any other part of the data. As standard outlier detection and robust classification usually fall short of determining groups of spurious samples, we propose a procedure which identifies the common structure across datasets by minimizing a multi-dataset divergence metric, increasing accuracy for new datasets.
Cite
CITATION STYLE
Barnes, M., Gisolfi, N., Fiterau, M., & Dubrawski, A. (2015). Leveraging common structure to improve prediction across related datasets. In Proceedings of the National Conference on Artificial Intelligence (Vol. 6, pp. 4144–4145). AI Access Foundation. https://doi.org/10.1609/aaai.v29i1.9746
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.