强化学习
计算机科学
马尔可夫决策过程
作业车间调度
调度(生产过程)
动态规划
数学优化
图形
人工智能
动态决策
工作车间
马尔可夫过程
机器学习
流水车间调度
算法
理论计算机科学
地铁列车时刻表
数学
操作系统
统计
作者
Chupeng Su,Cong Zhang,Dan Xia,Bao An Han,Chuang Wang,Gang Chen,Longhan Xie
标识
DOI:10.1016/j.asoc.2023.110596
摘要
The job shop scheduling problem (JSSP) with dynamic events and uncertainty is a strongly NP-hard combinatorial optimization problem (COP) with extensive applications in the manufacturing system. Recently, growing interest has been aroused in utilizing machine learning techniques to solve the JSSP. However, most prior arts cannot handle dynamic events and barely consider uncertainties. To close this gap, this paper proposes a framework to solve a dynamic JSSP (DJSP) with machine breakdown and stochastic processing time based on Graph Neural Network (GNN) and deep reinforcement learning (DRL). To this end, we first formulate the DJSP as a Markov Decision Process (MDP), where disjunctive graph represent the states. Secondly, we propose a GNN-based model to effectively extract the embeddings of the state by considering the features of the dynamic events and the stochasticity of the problem, e.g., the machine breakdown and stochastic processing time. Then, the model constructs solutions by dispatching optimal operations to machines based on the learned embeddings. Notably, we propose to use the evolution strategies (ES) to find optimal policies that are more stable and robust than conventional DRL algorithms. The extensive experiments show that our method substantially outperforms existing reinforcement learning-based and traditional methods on multiple classic benchmarks.
科研通智能强力驱动
Strongly Powered by AbleSci AI