Skip to main content
CenXiv.org
This website is in trial operation, support us!
We gratefully acknowledge support from all contributors.
Contribute
Donate
cenxiv logo > stat > arXiv:2506.05905

Help | Advanced Search

Statistics > Methodology

arXiv:2506.05905 (stat)
[Submitted on 6 Jun 2025 ]

Title: Sequential Monte Carlo approximations of Wasserstein--Fisher--Rao gradient flows

Title: Wasserstein-Fisher-Rao梯度流的顺序蒙特卡罗近似

Authors:Francesca R. Crucinio, Sahani Pathiraja
Abstract: We consider the problem of sampling from a probability distribution $\pi$. It is well known that this can be written as an optimisation problem over the space of probability distribution in which we aim to minimise the Kullback--Leibler divergence from $\pi$. We consider several partial differential equations (PDEs) whose solution is a minimiser of the Kullback--Leibler divergence from $\pi$ and connect them to well-known Monte Carlo algorithms. We focus in particular on PDEs obtained by considering the Wasserstein--Fisher--Rao geometry over the space of probabilities and show that these lead to a natural implementation using importance sampling and sequential Monte Carlo. We propose a novel algorithm to approximate the Wasserstein--Fisher--Rao flow of the Kullback--Leibler divergence which empirically outperforms the current state-of-the-art. We study tempered versions of these PDEs obtained by replacing the target distribution with a geometric mixture of initial and target distribution and show that these do not lead to a convergence speed up.
Abstract: 我们研究了从概率分布$\pi$抽样的问题。众所周知,这可以写成概率空间上的一个优化问题,其中我们旨在最小化 Kullback-Leibler 散度与$\pi$的距离。我们考虑了几种偏微分方程(PDEs),其解是最小化 Kullback-Leibler 散度与$\pi$距离的极小值点,并将它们与著名的蒙特卡罗算法联系起来。我们特别关注通过在概率空间上考虑 Wasserstein-Fisher-Rao 几何得到的 PDEs,并表明这些可以自然地通过重要性采样和序列蒙特卡罗方法实现。我们提出了一种新的算法来近似 Kullback-Leibler 散度的 Wasserstein-Fisher-Rao 流,该算法在实验中优于当前最先进的方法。我们研究了这些 PDEs 的温化版本,通过用初始分布和目标分布的几何混合替换目标分布得到,并证明这些并不会加速收敛速度。
Subjects: Methodology (stat.ME) ; Numerical Analysis (math.NA); Computation (stat.CO); Machine Learning (stat.ML)
MSC classes: 65C05, 62F15
Cite as: arXiv:2506.05905 [stat.ME]
  (or arXiv:2506.05905v1 [stat.ME] for this version)
  https://doi.org/10.48550/arXiv.2506.05905
arXiv-issued DOI via DataCite

Submission history

From: Francesca Romana Crucinio [view email]
[v1] Fri, 6 Jun 2025 09:24:46 UTC (994 KB)
Full-text links:

Access Paper:

    View a PDF of the paper titled
  • View Chinese PDF
  • View PDF
  • HTML (experimental)
  • TeX Source
license icon view license
Current browse context:
stat.ME
< prev   |   next >
new | recent | 2025-06
Change to browse by:
cs
cs.NA
math
math.NA
stat
stat.CO
stat.ML

References & Citations

  • NASA ADS
  • Google Scholar
  • Semantic Scholar
a export BibTeX citation Loading...

BibTeX formatted citation

×
Data provided by:

Bookmark

BibSonomy logo Reddit logo

Bibliographic and Citation Tools

Bibliographic Explorer (What is the Explorer?)
Connected Papers (What is Connected Papers?)
Litmaps (What is Litmaps?)
scite Smart Citations (What are Smart Citations?)

Code, Data and Media Associated with this Article

alphaXiv (What is alphaXiv?)
CatalyzeX Code Finder for Papers (What is CatalyzeX?)
DagsHub (What is DagsHub?)
Gotit.pub (What is GotitPub?)
Hugging Face (What is Huggingface?)
Papers with Code (What is Papers with Code?)
ScienceCast (What is ScienceCast?)

Demos

Replicate (What is Replicate?)
Hugging Face Spaces (What is Spaces?)
TXYZ.AI (What is TXYZ.AI?)

Recommenders and Search Tools

Influence Flower (What are Influence Flowers?)
CORE Recommender (What is CORE?)
IArxiv Recommender (What is IArxiv?)
  • Author
  • Venue
  • Institution
  • Topic

arXivLabs: experimental projects with community collaborators

arXivLabs is a framework that allows collaborators to develop and share new arXiv features directly on our website.

Both individuals and organizations that work with arXivLabs have embraced and accepted our values of openness, community, excellence, and user data privacy. arXiv is committed to these values and only works with partners that adhere to them.

Have an idea for a project that will add value for arXiv's community? Learn more about arXivLabs.

Which authors of this paper are endorsers? | Disable MathJax (What is MathJax?)
  • About
  • Help
  • contact arXivClick here to contact arXiv Contact
  • subscribe to arXiv mailingsClick here to subscribe Subscribe
  • Copyright
  • Privacy Policy
  • Web Accessibility Assistance
  • arXiv Operational Status
    Get status notifications via email or slack

京ICP备2025123034号