From Monte Carlo to Las Vegas: Improving restricted Boltzmann machine training through stopping sets

2Citations
Citations of this article
22Readers
Mendeley users who have this article in their library.

Abstract

We propose a Las Vegas transformation of Markov Chain Monte Carlo (MCMC) estimators of Restricted Boltzmann Machines (RBMs). We denote our approach Markov Chain Las Vegas (MCLV). MCLV gives statistical guarantees in exchange for random running times. MCLV uses a stopping set built from the training data and has maximum number of Markov chain steps K (referred as MCLV-K). We present a MCLV-K gradient estimator (LVS-K) for RBMs and explore the correspondence and differences between LVS-K and Contrastive Divergence (CD-K), with LVS-K significantly outperforming CD-K training RBMs over the MNIST dataset, indicating MCLV to be a promising direction in learning generative models.

Cite

CITATION STYLE

APA

Savarese, P. H. P., Kakodkar, M., & Ribeiro, B. (2018). From Monte Carlo to Las Vegas: Improving restricted Boltzmann machine training through stopping sets. In 32nd AAAI Conference on Artificial Intelligence, AAAI 2018 (pp. 4016–4025). AAAI press. https://doi.org/10.1609/aaai.v32i1.11821

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free