Sum-product autoencoding: Encoding and decoding representations using sum-product networks

14Citations
Citations of this article
31Readers
Mendeley users who have this article in their library.

Abstract

Sum-Product Networks (SPNs) are a deep probabilistic architecture that up to now has been successfully employed for tractable inference. Here, we extend their scope towards unsupervised representation learning: we encode samples into continuous and categorical embeddings and show that they can also be decoded back into the original input space by leveraging MPE inference. We characterize when this Sum-Product Autoencoding (SPAE) leads to equivalent reconstructions and extend it towards dealing with missing embedding information. Our experimental results on several multi-label classification problems demonstrate that SPAE is competitive with state-of-the-art autoencoder architectures, even if the SPNs were never trained to reconstruct their inputs.

Cite

CITATION STYLE

APA

Vergari, A., Molina, A., Peharz, R., Kersting, K., Mauro, N. D., & Esposito, F. (2018). Sum-product autoencoding: Encoding and decoding representations using sum-product networks. In 32nd AAAI Conference on Artificial Intelligence, AAAI 2018 (pp. 4163–4170). AAAI press. https://doi.org/10.1609/aaai.v32i1.11734

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free