Skip to main content
CenXiv.org
This website is in trial operation, support us!
We gratefully acknowledge support from all contributors.
Contribute
Donate
cenxiv logo > cs > arXiv:2509.14052v1

Help | Advanced Search

Computer Science > Sound

arXiv:2509.14052v1 (cs)
[Submitted on 17 Sep 2025 ]

Title: AnyAccomp: Generalizable Accompaniment Generation via Quantized Melodic Bottleneck

Title: 任何伴奏:通过量化旋律瓶颈的可推广伴奏生成

Authors:Junan Zhang, Yunjia Zhang, Xueyao Zhang, Zhizheng Wu
Abstract: Singing Accompaniment Generation (SAG) is the process of generating instrumental music for a given clean vocal input. However, existing SAG techniques use source-separated vocals as input and overfit to separation artifacts. This creates a critical train-test mismatch, leading to failure on clean, real-world vocal inputs. We introduce AnyAccomp, a framework that resolves this by decoupling accompaniment generation from source-dependent artifacts. AnyAccomp first employs a quantized melodic bottleneck, using a chromagram and a VQ-VAE to extract a discrete and timbre-invariant representation of the core melody. A subsequent flow-matching model then generates the accompaniment conditioned on these robust codes. Experiments show AnyAccomp achieves competitive performance on separated-vocal benchmarks while significantly outperforming baselines on generalization test sets of clean studio vocals and, notably, solo instrumental tracks. This demonstrates a qualitative leap in generalization, enabling robust accompaniment for instruments - a task where existing models completely fail - and paving the way for more versatile music co-creation tools. Demo audio and code: https://anyaccomp.github.io
Abstract: 伴奏生成(SAG)是为给定的干净人声输入生成乐器音乐的过程。 然而,现有的SAG技术使用源分离后的人声作为输入,并过度拟合分离伪影。 这导致了训练和测试之间的严重不匹配,使得在干净的现实世界人声输入上失效。 我们引入了AnyAccomp,一种通过将伴奏生成与依赖源的伪影解耦来解决这一问题的框架。 AnyAccomp首先采用量化旋律瓶颈,使用音高图和VQ-VAE提取核心旋律的离散且与音色无关的表示。 随后的流匹配模型则根据这些稳健的代码生成伴奏。 实验表明 AnyAccomp在分离人声基准上表现出色,并在一般化测试集的干净录音室人声以及值得注意的独奏乐器轨道上显著优于基线模型。 这展示了泛化能力的质的飞跃,使乐器的稳健伴奏成为可能——这是一个现有模型完全失败的任务——并为更通用的音乐协同创作工具铺平了道路。 演示音频和代码:https://anyaccomp.github.io
Comments: Demo audio and code: https://anyaccomp.github.io
Subjects: Sound (cs.SD) ; Signal Processing (eess.SP)
Cite as: arXiv:2509.14052 [cs.SD]
  (or arXiv:2509.14052v1 [cs.SD] for this version)
  https://doi.org/10.48550/arXiv.2509.14052
arXiv-issued DOI via DataCite

Submission history

From: Junan Zhang [view email]
[v1] Wed, 17 Sep 2025 14:55:21 UTC (3,462 KB)
Full-text links:

Access Paper:

    View a PDF of the paper titled
  • View Chinese PDF
  • View PDF
  • HTML (experimental)
  • TeX Source
license icon view license
Current browse context:
cs.SD
< prev   |   next >
new | recent | 2025-09
Change to browse by:
cs
eess
eess.SP

References & Citations

  • NASA ADS
  • Google Scholar
  • Semantic Scholar
a export BibTeX citation Loading...

BibTeX formatted citation

×
Data provided by:

Bookmark

BibSonomy logo Reddit logo

Bibliographic and Citation Tools

Bibliographic Explorer (What is the Explorer?)
Connected Papers (What is Connected Papers?)
Litmaps (What is Litmaps?)
scite Smart Citations (What are Smart Citations?)

Code, Data and Media Associated with this Article

alphaXiv (What is alphaXiv?)
CatalyzeX Code Finder for Papers (What is CatalyzeX?)
DagsHub (What is DagsHub?)
Gotit.pub (What is GotitPub?)
Hugging Face (What is Huggingface?)
Papers with Code (What is Papers with Code?)
ScienceCast (What is ScienceCast?)

Demos

Replicate (What is Replicate?)
Hugging Face Spaces (What is Spaces?)
TXYZ.AI (What is TXYZ.AI?)

Recommenders and Search Tools

Influence Flower (What are Influence Flowers?)
CORE Recommender (What is CORE?)
IArxiv Recommender (What is IArxiv?)
  • Author
  • Venue
  • Institution
  • Topic

arXivLabs: experimental projects with community collaborators

arXivLabs is a framework that allows collaborators to develop and share new arXiv features directly on our website.

Both individuals and organizations that work with arXivLabs have embraced and accepted our values of openness, community, excellence, and user data privacy. arXiv is committed to these values and only works with partners that adhere to them.

Have an idea for a project that will add value for arXiv's community? Learn more about arXivLabs.

Which authors of this paper are endorsers? | Disable MathJax (What is MathJax?)
  • About
  • Help
  • contact arXivClick here to contact arXiv Contact
  • subscribe to arXiv mailingsClick here to subscribe Subscribe
  • Copyright
  • Privacy Policy
  • Web Accessibility Assistance
  • arXiv Operational Status
    Get status notifications via email or slack

京ICP备2025123034号