强化学习
维数之咒
计算机科学
马尔可夫决策过程
水准点(测量)
组分(热力学)
过程(计算)
比例(比率)
人工智能
机器学习
马尔可夫过程
数学
统计
物理
大地测量学
量子力学
热力学
地理
操作系统
作者
Yifan Zhou,Bangcheng Li,Tian Ran Lin
标识
DOI:10.1016/j.ress.2021.108078
摘要
The Markov decision process (MDP) is a widely used method to optimise the maintenance of multicomponent systems, which can provide a system-level maintenance action at each decision point to address various dependences among components. However, MDP suffers from the “curse of dimensionality” and can only process small-scale systems. This paper develops a hierarchical coordinated reinforcement learning (HCRL) algorithm to optimise the maintenance of large-scale multicomponent systems. Both parameters of agents and the coordination relationship among agents are designed based on system characteristics. Furthermore, the hierarchical structure of agents is established according to the structural importance measures of components. The effectiveness of the proposed HCRL algorithm is validated using two maintenance optimisation problems, one on a natural gas plant system and the other using a 12-component series system under dependant competing risks. Results show that the proposed HCRL outperforms methods in two recently published papers and other benchmark approaches including the new emerging deep reinforcement learning.
科研通智能强力驱动
Strongly Powered by AbleSci AI