强化学习
钢筋
计算机科学
事件(粒子物理)
人工智能
心理学
物理
社会心理学
量子力学
作者
Shan Xue,Weidong Zhang,Biao Luo,Derong Liu
标识
DOI:10.1109/tcyb.2025.3533139
摘要
In this article, an integral reinforcement learning (IRL) method is developed for dynamic event-triggered nonzero-sum (NZS) games to achieve the Nash equilibrium of unmanned surface vehicles (USVs) with state and input constraints. Initially, a mapping function is designed to map the state and control of the USV into a safe environment. Subsequently, IRL-based coupled Hamilton-Jacobi equations, which avoid dependence on system dynamics, are derived to solve the Nash equilibrium. To conserve computational resources and reduce network transmission burdens, a static event-triggered control is initially designed, followed by the development of a more flexible dynamic form. Finally, a critic neural network is designed for each player to approximate its value function and control policy. Rigorous proofs are provided for the uniform ultimate boundedness of the state and the weight estimation errors. The effectiveness of the present method is demonstrated through simulation experiments.
科研通智能强力驱动
Strongly Powered by AbleSci AI