From topic models to semi-supervised learning: Biasing mixed-membership models to exploit topic-indicative features in entity clustering

8Citations
Citations of this article
26Readers
Mendeley users who have this article in their library.

This article is free to access.

Abstract

We present methods to introduce different forms of supervision into mixed-membership latent variable models. Firstly, we introduce a technique to bias the models to exploit topic-indicative features, i.e. features which are apriori known to be good indicators of the latent topics that generated them. Next, we present methods to modify the Gibbs sampler used for approximate inference in such models to permit injection of stronger forms of supervision in the form of labels for features and documents, along with a description of the corresponding change in the underlying generative process. This ability allows us to span the range from unsupervised topic models to semi-supervised learning in the same mixed membership model. Experimental results from an entity-clustering task demonstrate that the biasing technique and the introduction of feature and document labels provide a significant increase in clustering performance over baseline mixed-membership methods. © 2013 Springer-Verlag.

Cite

CITATION STYLE

APA

Balasubramanyan, R., Dalvi, B., & Cohen, W. W. (2013). From topic models to semi-supervised learning: Biasing mixed-membership models to exploit topic-indicative features in entity clustering. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 8189 LNAI, pp. 628–642). https://doi.org/10.1007/978-3-642-40991-2_40

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free