Abstract
Stein variational gradient descent (SVGD) refers to a class of methods for Bayesian inference based on interacting particle systems. In this paper, we consider the originally proposed deterministic dynamics as well as a stochastic variant, each of which represent one of the two main paradigms in Bayesian computational statistics: variational inference and Markov chain Monte Carlo. As it turns out, these are tightly linked through a correspondence between gradient flow structures and large-deviation principles rooted in statistical physics. To expose this relationship, we develop the cotangent space construction for the Stein geometry, prove its basic properties, and determine the large-deviation functional governing the many-particle limit for the empirical measure. Moreover, we identify the Stein-Fisher information (or kernelised Stein discrepancy) as its leading order contribution in the longtime and many-particle regime in the sense of Γ-convergence, shedding some light on the finite-particle properties of SVGD. Finally, we establish a comparison principle between the Stein-Fisher information and RKHS-norms that might be of independent interest.
Author supplied keywords
Cite
CITATION STYLE
Nüsken, N., & Renger, D. R. M. (2023). STEIN VARIATIONAL GRADIENT DESCENT: MANY-PARTICLE AND LONG-TIME ASYMPTOTICS. Foundations of Data Science, 5(3), 286–320. https://doi.org/10.3934/fods.2022023
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.