期刊:Journal of Aerospace Engineering [American Society of Civil Engineers] 日期:2021-12-02卷期号:35 (2)被引量:16
标识
DOI:10.1061/(asce)as.1943-5525.0001381
摘要
A moving mass–actuated unmanned aerial vehicle (MAUAV) is controlled by mass sliders installed inside the airframe and has the advantages of high aerodynamic efficiency and good stealth performance. However, designing a controller for it faces severe challenges due to the strong nonlinearity and coupling of its dynamics. To this end, we proposed an attitude controller based on deep reinforcement learning for the MAUAV. It directly maps the states to the needed deflection of the actuators and is an end-to-end controller. For the sparse reward problem, the reward function required for training is reasonably designed through reward shaping to hasten the algorithm's training speed. In training, random initialization and parameter perturbation are used to strengthen the final policy's robustness further. The simulation results tentatively demonstrate that the proposed controller is not only robust but suboptimal. Compared with an active disturbance rejection controller (ADRC) optimized by the particle swarm algorithm, our controller still guarantees a 100% success rate in multiple unlearned scenarios, meaning it has good generalization ability.