STEIN VARIATIONAL GRADIENT DESCENT: MANY-PARTICLE AND LONG-TIME ASYMPTOTICS

6Citations
Citations of this article
16Readers
Mendeley users who have this article in their library.

Abstract

Stein variational gradient descent (SVGD) refers to a class of methods for Bayesian inference based on interacting particle systems. In this paper, we consider the originally proposed deterministic dynamics as well as a stochastic variant, each of which represent one of the two main paradigms in Bayesian computational statistics: variational inference and Markov chain Monte Carlo. As it turns out, these are tightly linked through a correspondence between gradient flow structures and large-deviation principles rooted in statistical physics. To expose this relationship, we develop the cotangent space construction for the Stein geometry, prove its basic properties, and determine the large-deviation functional governing the many-particle limit for the empirical measure. Moreover, we identify the Stein-Fisher information (or kernelised Stein discrepancy) as its leading order contribution in the longtime and many-particle regime in the sense of Γ-convergence, shedding some light on the finite-particle properties of SVGD. Finally, we establish a comparison principle between the Stein-Fisher information and RKHS-norms that might be of independent interest.

Cite

CITATION STYLE

APA

Nüsken, N., & Renger, D. R. M. (2023). STEIN VARIATIONAL GRADIENT DESCENT: MANY-PARTICLE AND LONG-TIME ASYMPTOTICS. Foundations of Data Science, 5(3), 286–320. https://doi.org/10.3934/fods.2022023

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free