Skip to main content
CenXiv.org
This website is in trial operation, support us!
We gratefully acknowledge support from all contributors.
Contribute
Donate
cenxiv logo > eess > arXiv:2212.02715

Help | Advanced Search

Electrical Engineering and Systems Science > Systems and Control

arXiv:2212.02715 (eess)
[Submitted on 6 Dec 2022 ]

Title: Efficient Learning of Voltage Control Strategies via Model-based Deep Reinforcement Learning

Title: 基于模型的深度强化学习的电压控制策略高效学习

Authors:Ramij R. Hossain, Tianzhixi Yin, Yan Du, Renke Huang, Jie Tan, Wenhao Yu, Yuan Liu, Qiuhua Huang
Abstract: This article proposes a model-based deep reinforcement learning (DRL) method to design emergency control strategies for short-term voltage stability problems in power systems. Recent advances show promising results in model-free DRL-based methods for power systems, but model-free methods suffer from poor sample efficiency and training time, both critical for making state-of-the-art DRL algorithms practically applicable. DRL-agent learns an optimal policy via a trial-and-error method while interacting with the real-world environment. And it is desirable to minimize the direct interaction of the DRL agent with the real-world power grid due to its safety-critical nature. Additionally, state-of-the-art DRL-based policies are mostly trained using a physics-based grid simulator where dynamic simulation is computationally intensive, lowering the training efficiency. We propose a novel model-based-DRL framework where a deep neural network (DNN)-based dynamic surrogate model, instead of a real-world power-grid or physics-based simulation, is utilized with the policy learning framework, making the process faster and sample efficient. However, stabilizing model-based DRL is challenging because of the complex system dynamics of large-scale power systems. We solved these issues by incorporating imitation learning to have a warm start in policy learning, reward-shaping, and multi-step surrogate loss. Finally, we achieved 97.5% sample efficiency and 87.7% training efficiency for an application to the IEEE 300-bus test system.
Abstract: 本文提出了一种基于模型的深度强化学习(DRL)方法,用于设计电力系统短期电压稳定问题的紧急控制策略。 近期研究表明,无模型的基于DRL的方法在电力系统中取得了有前景的结果,但无模型方法存在样本效率低和训练时间长的问题,这两点对于使最先进的DRL算法具有实际应用价值至关重要。 DRL代理通过与真实环境交互的试错方法学习最优策略。由于其安全性要求高,希望尽量减少DRL代理直接与真实电网的交互。此外,最先进的基于DRL的策略大多使用基于物理的电网模拟器进行训练,动态仿真计算量大,降低了训练效率。 我们提出了一种新颖的基于模型的DRL框架,在该框架中,利用基于深度神经网络(DNN)的动态代理模型代替真实电网或基于物理的仿真,与策略学习框架结合,使得过程更快且样本高效。然而,由于大规模电力系统的复杂动力学特性,稳定基于模型的DRL具有挑战性。我们通过引入模仿学习来实现策略学习的热启动、奖励塑造以及多步代理损失来解决这些问题。最终,我们在IEEE 300节点测试系统中的应用达到了97.5%的样本效率和87.7%的训练效率。
Subjects: Systems and Control (eess.SY) ; Artificial Intelligence (cs.AI); Machine Learning (cs.LG); Optimization and Control (math.OC)
Cite as: arXiv:2212.02715 [eess.SY]
  (or arXiv:2212.02715v1 [eess.SY] for this version)
  https://doi.org/10.48550/arXiv.2212.02715
arXiv-issued DOI via DataCite

Submission history

From: Ramij Raja Hossain [view email]
[v1] Tue, 6 Dec 2022 02:50:53 UTC (1,144 KB)
Full-text links:

Access Paper:

    View a PDF of the paper titled
  • View Chinese PDF
  • View PDF
  • TeX Source
view license
Current browse context:
math
< prev   |   next >
new | recent | 2022-12
Change to browse by:
cs
cs.AI
cs.LG
cs.SY
eess
eess.SY
math.OC

References & Citations

  • NASA ADS
  • Google Scholar
  • Semantic Scholar
a export BibTeX citation Loading...

BibTeX formatted citation

×
Data provided by:

Bookmark

BibSonomy logo Reddit logo

Bibliographic and Citation Tools

Bibliographic Explorer (What is the Explorer?)
Connected Papers (What is Connected Papers?)
Litmaps (What is Litmaps?)
scite Smart Citations (What are Smart Citations?)

Code, Data and Media Associated with this Article

alphaXiv (What is alphaXiv?)
CatalyzeX Code Finder for Papers (What is CatalyzeX?)
DagsHub (What is DagsHub?)
Gotit.pub (What is GotitPub?)
Hugging Face (What is Huggingface?)
Papers with Code (What is Papers with Code?)
ScienceCast (What is ScienceCast?)

Demos

Replicate (What is Replicate?)
Hugging Face Spaces (What is Spaces?)
TXYZ.AI (What is TXYZ.AI?)

Recommenders and Search Tools

Influence Flower (What are Influence Flowers?)
CORE Recommender (What is CORE?)
IArxiv Recommender (What is IArxiv?)
  • Author
  • Venue
  • Institution
  • Topic

arXivLabs: experimental projects with community collaborators

arXivLabs is a framework that allows collaborators to develop and share new arXiv features directly on our website.

Both individuals and organizations that work with arXivLabs have embraced and accepted our values of openness, community, excellence, and user data privacy. arXiv is committed to these values and only works with partners that adhere to them.

Have an idea for a project that will add value for arXiv's community? Learn more about arXivLabs.

Which authors of this paper are endorsers? | Disable MathJax (What is MathJax?)
  • About
  • Help
  • contact arXivClick here to contact arXiv Contact
  • subscribe to arXiv mailingsClick here to subscribe Subscribe
  • Copyright
  • Privacy Policy
  • Web Accessibility Assistance
  • arXiv Operational Status
    Get status notifications via email or slack

京ICP备2025123034号