{"title":"Stein transport for Bayesian inference","authors":"Nikolas Nüsken","doi":"arxiv-2409.01464","DOIUrl":null,"url":null,"abstract":"We introduce $\\textit{Stein transport}$, a novel methodology for Bayesian\ninference designed to efficiently push an ensemble of particles along a\npredefined curve of tempered probability distributions. The driving vector\nfield is chosen from a reproducing kernel Hilbert space and can be derived\neither through a suitable kernel ridge regression formulation or as an\ninfinitesimal optimal transport map in the Stein geometry. The update equations\nof Stein transport resemble those of Stein variational gradient descent (SVGD),\nbut introduce a time-varying score function as well as specific weights\nattached to the particles. While SVGD relies on convergence in the long-time\nlimit, Stein transport reaches its posterior approximation at finite time\n$t=1$. Studying the mean-field limit, we discuss the errors incurred by\nregularisation and finite-particle effects, and we connect Stein transport to\nbirth-death dynamics and Fisher-Rao gradient flows. In a series of experiments,\nwe show that in comparison to SVGD, Stein transport not only often reaches more\naccurate posterior approximations with a significantly reduced computational\nbudget, but that it also effectively mitigates the variance collapse phenomenon\ncommonly observed in SVGD.","PeriodicalId":501379,"journal":{"name":"arXiv - STAT - Statistics Theory","volume":"144 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-09-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - STAT - Statistics Theory","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.01464","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
We introduce $\textit{Stein transport}$, a novel methodology for Bayesian
inference designed to efficiently push an ensemble of particles along a
predefined curve of tempered probability distributions. The driving vector
field is chosen from a reproducing kernel Hilbert space and can be derived
either through a suitable kernel ridge regression formulation or as an
infinitesimal optimal transport map in the Stein geometry. The update equations
of Stein transport resemble those of Stein variational gradient descent (SVGD),
but introduce a time-varying score function as well as specific weights
attached to the particles. While SVGD relies on convergence in the long-time
limit, Stein transport reaches its posterior approximation at finite time
$t=1$. Studying the mean-field limit, we discuss the errors incurred by
regularisation and finite-particle effects, and we connect Stein transport to
birth-death dynamics and Fisher-Rao gradient flows. In a series of experiments,
we show that in comparison to SVGD, Stein transport not only often reaches more
accurate posterior approximations with a significantly reduced computational
budget, but that it also effectively mitigates the variance collapse phenomenon
commonly observed in SVGD.