Skip to main content
CenXiv.org
This website is in trial operation, support us!
We gratefully acknowledge support from all contributors.
Contribute
Donate
cenxiv logo > eess > arXiv:2501.06510

Help | Advanced Search

Electrical Engineering and Systems Science > Systems and Control

arXiv:2501.06510 (eess)
[Submitted on 11 Jan 2025 ]

Title: Cooperative Optimal Output Tracking for Discrete-Time Multiagent Systems: Stabilizing Policy Iteration Frameworks and Analysis

Title: 离散时间多智能体系统的合作最优输出跟踪:稳定策略迭代框架与分析

Authors:Dongdong Li, Jiuxiang Dong
Abstract: In this paper, two model-free optimal output tracking frameworks based on policy iteration for discrete-time multi-agent systems are proposed. First, we establish a framework of stabilizing policy iteration that can start from any initial feedback control policy, relaxing the dependence of traditional policy iteration on the initial stabilizing control policy. Then, another efficient and equivalent $Q$-learning policy iteration framework is developed, which is shown to require only less system data to get the same results as the stabilizing policy iteration. Both frameworks obtain stabilizing control policy by iterating the stabilizing virtual closed-loop system step-by-step to the actual closed-loop system. Multiple explicit schemes for the iteration step-size/coefficient are designed and their stability during the above iterations is analyzed. By using the generated closed-loop stabilizing control policy and two frameworks, the optimal feedback control gain is obtained. The approximate solution of the regulator equations is found by model-free iteration, which leads to the optimal feedforward gain. Finally, the cooperative optimal output tracking is realized by a distributed feedforward-feedback controller. The proposed algorithms are validated by simulation.
Abstract: 本文提出了两种基于策略迭代的离散时间多智能体系统的无模型最优输出跟踪框架。 首先,我们建立了一个可以从任何初始反馈控制策略开始的稳定策略迭代框架,放松了传统策略迭代对初始稳定控制策略的依赖。 然后,开发了另一种高效且等效的$Q$-学习策略迭代框架,该框架被证明只需要更少的系统数据就能得到与稳定策略迭代相同的结果。 两个框架通过逐步迭代稳定虚拟闭环系统到实际闭环系统来获得稳定控制策略。 设计了多种迭代步长/系数的显式方案,并分析了上述迭代过程中的稳定性。 通过使用生成的闭环稳定控制策略和两个框架,得到了最优反馈控制增益。 通过无模型迭代找到了调节器方程的近似解,从而得到最优前馈增益。 最后,通过分布式前馈-反馈控制器实现了合作最优输出跟踪。 所提出的算法通过仿真进行了验证。
Subjects: Systems and Control (eess.SY)
Cite as: arXiv:2501.06510 [eess.SY]
  (or arXiv:2501.06510v1 [eess.SY] for this version)
  https://doi.org/10.48550/arXiv.2501.06510
arXiv-issued DOI via DataCite

Submission history

From: Dongdong Li [view email]
[v1] Sat, 11 Jan 2025 11:11:26 UTC (2,965 KB)
Full-text links:

Access Paper:

    View a PDF of the paper titled
  • View Chinese PDF
  • View PDF
  • HTML (experimental)
  • TeX Source
view license
Current browse context:
eess.SY
< prev   |   next >
new | recent | 2025-01
Change to browse by:
cs
cs.SY
eess

References & Citations

  • NASA ADS
  • Google Scholar
  • Semantic Scholar
a export BibTeX citation Loading...

BibTeX formatted citation

×
Data provided by:

Bookmark

BibSonomy logo Reddit logo

Bibliographic and Citation Tools

Bibliographic Explorer (What is the Explorer?)
Connected Papers (What is Connected Papers?)
Litmaps (What is Litmaps?)
scite Smart Citations (What are Smart Citations?)

Code, Data and Media Associated with this Article

alphaXiv (What is alphaXiv?)
CatalyzeX Code Finder for Papers (What is CatalyzeX?)
DagsHub (What is DagsHub?)
Gotit.pub (What is GotitPub?)
Hugging Face (What is Huggingface?)
Papers with Code (What is Papers with Code?)
ScienceCast (What is ScienceCast?)

Demos

Replicate (What is Replicate?)
Hugging Face Spaces (What is Spaces?)
TXYZ.AI (What is TXYZ.AI?)

Recommenders and Search Tools

Influence Flower (What are Influence Flowers?)
CORE Recommender (What is CORE?)
IArxiv Recommender (What is IArxiv?)
  • Author
  • Venue
  • Institution
  • Topic

arXivLabs: experimental projects with community collaborators

arXivLabs is a framework that allows collaborators to develop and share new arXiv features directly on our website.

Both individuals and organizations that work with arXivLabs have embraced and accepted our values of openness, community, excellence, and user data privacy. arXiv is committed to these values and only works with partners that adhere to them.

Have an idea for a project that will add value for arXiv's community? Learn more about arXivLabs.

Which authors of this paper are endorsers? | Disable MathJax (What is MathJax?)
  • About
  • Help
  • contact arXivClick here to contact arXiv Contact
  • subscribe to arXiv mailingsClick here to subscribe Subscribe
  • Copyright
  • Privacy Policy
  • Web Accessibility Assistance
  • arXiv Operational Status
    Get status notifications via email or slack

京ICP备2025123034号