This paper introduces MixtBoost, a variant of AdaBoost dedicated to solve problems in which both labeled and unlabeled data are available. We propose several definitions of loss for unlabeled data, from which margins are defined. The resulting boosting schemes implement mixture models as base classifiers. Preliminary experiments are analyzed and the relevance of loss choices is discussed. MixtBoost improves on both mixture models and AdaBoost provided classes are structured, and is otherwise similar to AdaBoost.
CITATION STYLE
Grandvalet, Y., D’alché-Buc, F., & Ambroise, C. (2001). Boosting mixture models for semi-supervised learning. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 2130, pp. 41–48). Springer Verlag. https://doi.org/10.1007/3-540-44668-0_7
Mendeley helps you to discover research relevant for your work.