On Semi-Supervised Learning of Gaussian Mixture Models for Phonetic Classification∗

7Citations
Citations of this article
92Readers
Mendeley users who have this article in their library.
Get full text

Abstract

This paper investigates semi-supervised learning of Gaussian mixture models using an unified objective function taking both labeled and unlabeled data into account. Two methods are compared in this work – the hybrid discriminative/generative method and the purely generative method. They differ in the criterion type on labeled data; the hybrid method uses the class posterior probabilities and the purely generative method uses the data likelihood. We conducted experiments on the TIMIT database and a standard synthetic data set from UCI Machine Learning repository. The results show that the two methods behave similarly in various conditions. For both methods, unlabeled data improve training on models of higher complexity in which the supervised method performs poorly. In addition, there is a trend that more unlabeled data results in more improvement in classification accuracy over the supervised model. We also provided experimental observations on the relative weights of labeled and unlabeled parts of the training objective and suggested a critical value which could be useful for selecting a good weighing factor.

Cite

CITATION STYLE

APA

Huang, J. T., & Hasegawa-Johnson, M. (2009). On Semi-Supervised Learning of Gaussian Mixture Models for Phonetic Classification∗. In NAACL HLT 2009 - Semi-Supervised Learning for Natural Language Processing, Proceedings of the Workshop (pp. 75–83). Association for Computational Linguistics (ACL). https://doi.org/10.3115/1621829.1621839

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free