Skip to main content
CenXiv.org
This website is in trial operation, support us!
We gratefully acknowledge support from all contributors.
Contribute
Donate
cenxiv logo > cs > arXiv:2411.00311

Help | Advanced Search

Computer Science > Machine Learning

arXiv:2411.00311 (cs)
[Submitted on 1 Nov 2024 ]

Title: C2A: Client-Customized Adaptation for Parameter-Efficient Federated Learning

Title: C2A:面向参数高效联邦学习的客户端定制适应

Authors:Yeachan Kim, Junho Kim, Wing-Lam Mok, Jun-Hyung Park, SangKeun Lee
Abstract: Despite the versatility of pre-trained language models (PLMs) across domains, their large memory footprints pose significant challenges in federated learning (FL), where the training model has to be distributed between a server and clients. One potential solution to bypass such constraints might be the use of parameter-efficient fine-tuning (PEFT) in the context of FL. However, we have observed that typical PEFT tends to severely suffer from heterogeneity among clients in FL scenarios, resulting in unstable and slow convergence. In this paper, we propose Client-Customized Adaptation (C2A), a novel hypernetwork-based FL framework that generates client-specific adapters by conditioning the client information. With the effectiveness of the hypernetworks in generating customized weights through learning to adopt the different characteristics of inputs, C2A can maximize the utility of shared model parameters while minimizing the divergence caused by client heterogeneity. To verify the efficacy of C2A, we perform extensive evaluations on FL scenarios involving heterogeneity in label and language distributions. Comprehensive evaluation results clearly support the superiority of C2A in terms of both efficiency and effectiveness in FL scenarios.
Abstract: 尽管预训练语言模型(PLMs)在不同领域具有多功能性,但其较大的内存占用在联邦学习(FL)中带来了重大挑战,因为在FL中,训练模型必须在服务器和客户端之间进行分发。 在FL背景下,使用参数高效的微调(PEFT)可能是绕过这些限制的一种潜在解决方案。 然而,我们观察到典型的PEFT在FL场景中容易受到客户端异质性的严重影响,导致不稳定和缓慢的收敛。 在本文中,我们提出了客户定制适应(C2A),这是一种基于超网络的新型FL框架,通过条件化客户端信息生成特定于客户端的适配器。 由于超网络在通过学习采用不同输入特征来生成定制权重方面的有效性,C2A可以在最大化共享模型参数效用的同时,最小化由客户端异质性引起的差异。 为了验证C2A的有效性,我们在涉及标签和语言分布异质性的FL场景中进行了广泛的评估。 全面的评估结果明确支持C2A在FL场景中的效率和效果方面的优越性。
Comments: Published at Findings of ACL 2023
Subjects: Machine Learning (cs.LG) ; Artificial Intelligence (cs.AI); Cryptography and Security (cs.CR)
Cite as: arXiv:2411.00311 [cs.LG]
  (or arXiv:2411.00311v1 [cs.LG] for this version)
  https://doi.org/10.48550/arXiv.2411.00311
arXiv-issued DOI via DataCite

Submission history

From: Junho Kim [view email]
[v1] Fri, 1 Nov 2024 02:07:38 UTC (1,146 KB)
Full-text links:

Access Paper:

    View a PDF of the paper titled
  • View Chinese PDF
  • View PDF
  • HTML (experimental)
  • TeX Source
  • Other Formats
license icon view license
Current browse context:
cs.LG
< prev   |   next >
new | recent | 2024-11
Change to browse by:
cs
cs.AI
cs.CR

References & Citations

  • NASA ADS
  • Google Scholar
  • Semantic Scholar
a export BibTeX citation Loading...

BibTeX formatted citation

×
Data provided by:

Bookmark

BibSonomy logo Reddit logo

Bibliographic and Citation Tools

Bibliographic Explorer (What is the Explorer?)
Connected Papers (What is Connected Papers?)
Litmaps (What is Litmaps?)
scite Smart Citations (What are Smart Citations?)

Code, Data and Media Associated with this Article

alphaXiv (What is alphaXiv?)
CatalyzeX Code Finder for Papers (What is CatalyzeX?)
DagsHub (What is DagsHub?)
Gotit.pub (What is GotitPub?)
Hugging Face (What is Huggingface?)
Papers with Code (What is Papers with Code?)
ScienceCast (What is ScienceCast?)

Demos

Replicate (What is Replicate?)
Hugging Face Spaces (What is Spaces?)
TXYZ.AI (What is TXYZ.AI?)

Recommenders and Search Tools

Influence Flower (What are Influence Flowers?)
CORE Recommender (What is CORE?)
IArxiv Recommender (What is IArxiv?)
  • Author
  • Venue
  • Institution
  • Topic

arXivLabs: experimental projects with community collaborators

arXivLabs is a framework that allows collaborators to develop and share new arXiv features directly on our website.

Both individuals and organizations that work with arXivLabs have embraced and accepted our values of openness, community, excellence, and user data privacy. arXiv is committed to these values and only works with partners that adhere to them.

Have an idea for a project that will add value for arXiv's community? Learn more about arXivLabs.

Which authors of this paper are endorsers? | Disable MathJax (What is MathJax?)
  • About
  • Help
  • contact arXivClick here to contact arXiv Contact
  • subscribe to arXiv mailingsClick here to subscribe Subscribe
  • Copyright
  • Privacy Policy
  • Web Accessibility Assistance
  • arXiv Operational Status
    Get status notifications via email or slack

京ICP备2025123034号