控制理论(社会学)
强化学习
执行机构
非线性系统
计算机科学
稳健性(进化)
初始化
机身
控制器(灌溉)
工程类
控制工程
人工智能
控制(管理)
航空航天工程
农学
生物化学
化学
物理
量子力学
生物
基因
程序设计语言
作者
Xiaoqi Qiu,Changsheng Gao,Kefan Wang,Wuxing Jing
标识
DOI:10.1061/(asce)as.1943-5525.0001381
摘要
A moving mass–actuated unmanned aerial vehicle (MAUAV) is controlled by mass sliders installed inside the airframe and has the advantages of high aerodynamic efficiency and good stealth performance. However, designing a controller for it faces severe challenges due to the strong nonlinearity and coupling of its dynamics. To this end, we proposed an attitude controller based on deep reinforcement learning for the MAUAV. It directly maps the states to the needed deflection of the actuators and is an end-to-end controller. For the sparse reward problem, the reward function required for training is reasonably designed through reward shaping to hasten the algorithm's training speed. In training, random initialization and parameter perturbation are used to strengthen the final policy's robustness further. The simulation results tentatively demonstrate that the proposed controller is not only robust but suboptimal. Compared with an active disturbance rejection controller (ADRC) optimized by the particle swarm algorithm, our controller still guarantees a 100% success rate in multiple unlearned scenarios, meaning it has good generalization ability.
科研通智能强力驱动
Strongly Powered by AbleSci AI