Multimodal poisson gamma belief network

10Citations
Citations of this article
27Readers
Mendeley users who have this article in their library.

Abstract

To learn a deep generative model of multimodal data, we propose a multimodal Poisson gamma belief network (mPGBN) that tightly couple the data of different modalities at multiple hidden layers. The mPGBN unsupervisedly extracts a nonnegative latent representation using an upward-downward Gibbs sampler. It imposes sparse connections between different layers, making it simple to visualize the generative process and the relationships between the latent features of different modalities. Our experimental results on bi-modal data consisting of images and tags show that the mPGBN can easily impute a missing modality and hence is useful for both image annotation and retrieval. We further demonstrate that the mPGBN achieves state-of-the-art results on unsupervisedly extracting latent features from multimodal data.

Cite

CITATION STYLE

APA

Wang, C., Chen, B., & Zhou, M. (2018). Multimodal poisson gamma belief network. In 32nd AAAI Conference on Artificial Intelligence, AAAI 2018 (pp. 2492–2499). AAAI press. https://doi.org/10.1609/aaai.v32i1.11846

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free