Skip to main content
CenXiv.org
This website is in trial operation, support us!
We gratefully acknowledge support from all contributors.
Contribute
Donate
cenxiv logo > cs > arXiv:2507.02606

Help | Advanced Search

Computer Science > Sound

arXiv:2507.02606 (cs)
[Submitted on 3 Jul 2025 ]

Title: De-AntiFake: Rethinking the Protective Perturbations Against Voice Cloning Attacks

Title: 去防伪:重新思考针对语音克隆攻击的保护性扰动

Authors:Wei Fan, Kejiang Chen, Chang Liu, Weiming Zhang, Nenghai Yu
Abstract: The rapid advancement of speech generation models has heightened privacy and security concerns related to voice cloning (VC). Recent studies have investigated disrupting unauthorized voice cloning by introducing adversarial perturbations. However, determined attackers can mitigate these protective perturbations and successfully execute VC. In this study, we conduct the first systematic evaluation of these protective perturbations against VC under realistic threat models that include perturbation purification. Our findings reveal that while existing purification methods can neutralize a considerable portion of the protective perturbations, they still lead to distortions in the feature space of VC models, which degrades the performance of VC. From this perspective, we propose a novel two-stage purification method: (1) Purify the perturbed speech; (2) Refine it using phoneme guidance to align it with the clean speech distribution. Experimental results demonstrate that our method outperforms state-of-the-art purification methods in disrupting VC defenses. Our study reveals the limitations of adversarial perturbation-based VC defenses and underscores the urgent need for more robust solutions to mitigate the security and privacy risks posed by VC. The code and audio samples are available at https://de-antifake.github.io.
Abstract: 语音生成模型的快速发展引发了与语音克隆(VC)相关的隐私和安全问题。 最近的研究通过引入对抗性扰动来干扰未经授权的语音克隆。 然而,有决心的攻击者可以减轻这些保护性扰动,并成功执行VC。 在本研究中,我们在包含扰动净化的现实威胁模型下,首次对这些保护性扰动进行了系统评估。 我们的研究结果表明,尽管现有的净化方法可以中和相当一部分保护性扰动,但它们仍然会导致VC模型特征空间中的失真,从而降低VC的性能。 从这个角度来看,我们提出了一种新的两阶段净化方法:(1)净化受扰的语音;(2)使用音素指导对其进行优化,使其与干净语音分布对齐。 实验结果表明,我们的方法在破坏VC防御方面优于最先进的净化方法。 我们的研究揭示了基于对抗性扰动的VC防御的局限性,并强调了迫切需要更稳健的解决方案来缓解VC带来的安全和隐私风险。 代码和音频样本可在 https://de-antifake.github.io 获取。
Comments: Accepted by ICML 2025
Subjects: Sound (cs.SD) ; Artificial Intelligence (cs.AI); Cryptography and Security (cs.CR); Machine Learning (cs.LG); Audio and Speech Processing (eess.AS)
Cite as: arXiv:2507.02606 [cs.SD]
  (or arXiv:2507.02606v1 [cs.SD] for this version)
  https://doi.org/10.48550/arXiv.2507.02606
arXiv-issued DOI via DataCite

Submission history

From: Wei Fan [view email]
[v1] Thu, 3 Jul 2025 13:30:58 UTC (4,105 KB)
Full-text links:

Access Paper:

    View a PDF of the paper titled
  • View Chinese PDF
  • View PDF
  • HTML (experimental)
  • TeX Source
  • Other Formats
view license
Current browse context:
cs.SD
< prev   |   next >
new | recent | 2025-07
Change to browse by:
cs
cs.AI
cs.CR
cs.LG
eess
eess.AS

References & Citations

  • NASA ADS
  • Google Scholar
  • Semantic Scholar
a export BibTeX citation Loading...

BibTeX formatted citation

×
Data provided by:

Bookmark

BibSonomy logo Reddit logo

Bibliographic and Citation Tools

Bibliographic Explorer (What is the Explorer?)
Connected Papers (What is Connected Papers?)
Litmaps (What is Litmaps?)
scite Smart Citations (What are Smart Citations?)

Code, Data and Media Associated with this Article

alphaXiv (What is alphaXiv?)
CatalyzeX Code Finder for Papers (What is CatalyzeX?)
DagsHub (What is DagsHub?)
Gotit.pub (What is GotitPub?)
Hugging Face (What is Huggingface?)
Papers with Code (What is Papers with Code?)
ScienceCast (What is ScienceCast?)

Demos

Replicate (What is Replicate?)
Hugging Face Spaces (What is Spaces?)
TXYZ.AI (What is TXYZ.AI?)

Recommenders and Search Tools

Influence Flower (What are Influence Flowers?)
CORE Recommender (What is CORE?)
IArxiv Recommender (What is IArxiv?)
  • Author
  • Venue
  • Institution
  • Topic

arXivLabs: experimental projects with community collaborators

arXivLabs is a framework that allows collaborators to develop and share new arXiv features directly on our website.

Both individuals and organizations that work with arXivLabs have embraced and accepted our values of openness, community, excellence, and user data privacy. arXiv is committed to these values and only works with partners that adhere to them.

Have an idea for a project that will add value for arXiv's community? Learn more about arXivLabs.

Which authors of this paper are endorsers? | Disable MathJax (What is MathJax?)
  • About
  • Help
  • contact arXivClick here to contact arXiv Contact
  • subscribe to arXiv mailingsClick here to subscribe Subscribe
  • Copyright
  • Privacy Policy
  • Web Accessibility Assistance
  • arXiv Operational Status
    Get status notifications via email or slack

京ICP备2025123034号