Skip to main content
CenXiv.org
This website is in trial operation, support us!
We gratefully acknowledge support from all contributors.
Contribute
Donate
cenxiv logo > cs > arXiv:2506.00562

Help | Advanced Search

Computer Science > Computer Vision and Pattern Recognition

arXiv:2506.00562 (cs)
[Submitted on 31 May 2025 ]

Title: SEED: A Benchmark Dataset for Sequential Facial Attribute Editing with Diffusion Models

Title: SEED:一种用于基于扩散模型的顺序面部属性编辑的基准数据集

Authors:Yule Zhu, Ping Liu, Zhedong Zheng, Wei Liu
Abstract: Diffusion models have recently enabled precise and photorealistic facial editing across a wide range of semantic attributes. Beyond single-step modifications, a growing class of applications now demands the ability to analyze and track sequences of progressive edits, such as stepwise changes to hair, makeup, or accessories. However, sequential editing introduces significant challenges in edit attribution and detection robustness, further complicated by the lack of large-scale, finely annotated benchmarks tailored explicitly for this task. We introduce SEED, a large-scale Sequentially Edited facE Dataset constructed via state-of-the-art diffusion models. SEED contains over 90,000 facial images with one to four sequential attribute modifications, generated using diverse diffusion-based editing pipelines (LEdits, SDXL, SD3). Each image is annotated with detailed edit sequences, attribute masks, and prompts, facilitating research on sequential edit tracking, visual provenance analysis, and manipulation robustness assessment. To benchmark this task, we propose FAITH, a frequency-aware transformer-based model that incorporates high-frequency cues to enhance sensitivity to subtle sequential changes. Comprehensive experiments, including systematic comparisons of multiple frequency-domain methods, demonstrate the effectiveness of FAITH and the unique challenges posed by SEED. SEED offers a challenging and flexible resource for studying progressive diffusion-based edits at scale. Dataset and code will be publicly released at: https://github.com/Zeus1037/SEED.
Abstract: 扩散模型最近使得在广泛的语义属性上实现精确且照片级真实的面部编辑成为可能。除了单步修改之外,现在有一类不断增长的应用需求能够分析和跟踪逐步编辑序列的能力,例如头发、化妆或配饰的分步变化。然而,顺序编辑在编辑归因和检测鲁棒性方面引入了重大挑战,进一步由于缺乏专门为此任务设计的大规模、精细标注的数据集而变得更加复杂。我们介绍了 SEED(Sequentially Edited facE Dataset,顺序编辑人脸数据集),这是一个通过最先进的扩散模型构建的大规模顺序编辑人脸数据集。SEED 包含超过 90,000 张具有一个到四个连续属性修改的面部图像,这些图像是使用多样化的基于扩散的编辑管道(LEdits、SDXL、SD3)生成的。每张图像都带有详细的编辑序列、属性掩码和提示,以促进关于顺序编辑跟踪、视觉来源分析和操作鲁棒性评估的研究。为了对该任务进行基准测试,我们提出了 FAITH(frequency-aware transformer-based model,基于频率感知变换器的模型),该模型结合高频线索以增强对细微顺序变化的敏感性。全面的实验,包括多种频域方法的系统比较,证明了 FAITH 的有效性以及 SEED 所带来的独特挑战。SEED 提供了一个具有挑战性和灵活性的资源,用于大规模研究基于扩散的渐进式编辑。数据集和代码将在以下地址公开发布:https://github.com/Zeus1037/SEED。
Subjects: Computer Vision and Pattern Recognition (cs.CV) ; Multimedia (cs.MM)
Cite as: arXiv:2506.00562 [cs.CV]
  (or arXiv:2506.00562v1 [cs.CV] for this version)
  https://doi.org/10.48550/arXiv.2506.00562
arXiv-issued DOI via DataCite

Submission history

From: YuL Zhu [view email]
[v1] Sat, 31 May 2025 13:39:45 UTC (16,255 KB)
Full-text links:

Access Paper:

    View a PDF of the paper titled
  • View Chinese PDF
  • View PDF
  • HTML (experimental)
  • TeX Source
  • Other Formats
view license
Current browse context:
cs.CV
< prev   |   next >
new | recent | 2025-06
Change to browse by:
cs
cs.MM

References & Citations

  • NASA ADS
  • Google Scholar
  • Semantic Scholar
a export BibTeX citation Loading...

BibTeX formatted citation

×
Data provided by:

Bookmark

BibSonomy logo Reddit logo

Bibliographic and Citation Tools

Bibliographic Explorer (What is the Explorer?)
Connected Papers (What is Connected Papers?)
Litmaps (What is Litmaps?)
scite Smart Citations (What are Smart Citations?)

Code, Data and Media Associated with this Article

alphaXiv (What is alphaXiv?)
CatalyzeX Code Finder for Papers (What is CatalyzeX?)
DagsHub (What is DagsHub?)
Gotit.pub (What is GotitPub?)
Hugging Face (What is Huggingface?)
Papers with Code (What is Papers with Code?)
ScienceCast (What is ScienceCast?)

Demos

Replicate (What is Replicate?)
Hugging Face Spaces (What is Spaces?)
TXYZ.AI (What is TXYZ.AI?)

Recommenders and Search Tools

Influence Flower (What are Influence Flowers?)
CORE Recommender (What is CORE?)
IArxiv Recommender (What is IArxiv?)
  • Author
  • Venue
  • Institution
  • Topic

arXivLabs: experimental projects with community collaborators

arXivLabs is a framework that allows collaborators to develop and share new arXiv features directly on our website.

Both individuals and organizations that work with arXivLabs have embraced and accepted our values of openness, community, excellence, and user data privacy. arXiv is committed to these values and only works with partners that adhere to them.

Have an idea for a project that will add value for arXiv's community? Learn more about arXivLabs.

Which authors of this paper are endorsers? | Disable MathJax (What is MathJax?)
  • About
  • Help
  • contact arXivClick here to contact arXiv Contact
  • subscribe to arXiv mailingsClick here to subscribe Subscribe
  • Copyright
  • Privacy Policy
  • Web Accessibility Assistance
  • arXiv Operational Status
    Get status notifications via email or slack

京ICP备2025123034号