Abstract
Message-passing algorithms based on the belief propagation (BP) equations constitute a well-known distributed computational scheme. They yield exact marginals on tree-like graphical models and have also proven to be effective in many problems defined on loopy graphs, from inference to optimization, from signal processing to clustering. The BP-based schemes are fundamentally different from stochastic gradient descent (SGD), on which the current success of deep networks is based. In this paper, we present and adapt to mini-batch training on GPUs a family of BP-based message-passing algorithms with a reinforcement term that biases distributions towards locally entropic solutions. These algorithms are capable of training multi-layer neural networks with performance comparable to SGD heuristics in a diverse set of experiments on natural datasets including multi-class image classification and continual learning, while being capable of yielding improved performances on sparse networks. Furthermore, they allow to make approximate Bayesian predictions that have higher accuracy than point-wise ones.
Author supplied keywords
Cite
CITATION STYLE
Lucibello, C., Pittorino, F., Perugini, G., & Zecchina, R. (2022). Deep learning via message passing algorithms based on belief propagation. Machine Learning: Science and Technology, 3(3). https://doi.org/10.1088/2632-2153/ac7d3b
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.