计算机科学
强化学习
动态优先级调度
两级调度
公平份额计划
单调速率调度
分布式计算
调度(生产过程)
循环调度
作业车间调度
固定优先级先发制人调度
抽奖日程安排
人工智能
数学优化
计算机网络
数学
服务质量
布线(电子设计自动化)
作者
Xuemei Gan,Ying Zuo,Ansi Zhang,Shaobo Li,Fei Tao
标识
DOI:10.1007/s11431-022-2413-5
摘要
The modern complicated manufacturing industry and smart manufacturing tendency have imposed new requirements on the scheduling method, such as self-regulation and self-learning capabilities. While traditional scheduling methods cannot meet these needs due to their rigidity. Self-learning is an inherent ability of reinforcement learning (RL) algorithm inhered from its continuous learning and trial-and-error characteristics. Self-regulation of scheduling could be enabled by the emerging digital twin (DT) technology because of its virtual-real mapping and mutual control characteristics. This paper proposed a DT-enabled adaptive scheduling based on the improved proximal policy optimization RL algorithm, which was called explicit exploration and asynchronous update proximal policy optimization algorithm (E2APPO). Firstly, the DT-enabled scheduling system framework was designed to enhance the interaction between the virtual and the physical job shops, strengthening the self-regulation of the scheduling model. Secondly, an innovative action selection strategy and an asynchronous update mechanism were proposed to improve the optimization algorithm to strengthen the self-learning ability of the scheduling model. Lastly, the proposed scheduling model was extensively tested in comparison with heuristic and meta-heuristic algorithms, such as well-known scheduling rules and genetic algorithms, as well as other existing scheduling methods based on reinforcement learning. The comparisons have proved both the effectiveness and advancement of the proposed DT-enabled adaptive scheduling strategy.
科研通智能强力驱动
Strongly Powered by AbleSci AI