TY - JOUR
T1 - A deep reinforcement learning based approach for dynamic distributed blocking flowshop scheduling with job insertions
AU - Sun, Xueyan
AU - Vogel-Heuser, Birgit
AU - Bi, Fandi
AU - Shen, Weiming
N1 - Publisher Copyright:
© 2022 The Authors. IET Collaborative Intelligent Manufacturing published by John Wiley & Sons Ltd on behalf of The Institution of Engineering and Technology.
PY - 2022/9
Y1 - 2022/9
N2 - The distributed blocking flowshop scheduling problem (DBFSP) with new job insertions is studied. Rescheduling all remaining jobs after a dynamic event like a new job insertion is unreasonable to an actual distributed blocking flowshop production process. A deep reinforcement learning (DRL) algorithm is proposed to optimise the job selection model, and local modifications are made on the basis of the original scheduling plan when new jobs arrive. The objective is to minimise the total completion time deviation of all products so that all jobs can be finished on time to reduce the cost of storage. First, according to the definitions of the dynamic DBFSP problem, a DRL framework based on multi-agent deep deterministic policy gradient (MADDPG) is proposed. In this framework, a full schedule is generated by the variable neighbourhood descent algorithm before a dynamic event occurs. Meanwhile, all newly added jobs are reordered before the agents make decisions to select the one that needs to be scheduled most urgently. This study defines the observations, actions and reward calculation methods and applies centralised training and distributed execution in MADDPG. Finally, a comprehensive computational experiment is carried out to compare the proposed method with the closely related and well-performing methods. The results indicate that the proposed method can solve the dynamic DBFSP effectively and efficiently.
AB - The distributed blocking flowshop scheduling problem (DBFSP) with new job insertions is studied. Rescheduling all remaining jobs after a dynamic event like a new job insertion is unreasonable to an actual distributed blocking flowshop production process. A deep reinforcement learning (DRL) algorithm is proposed to optimise the job selection model, and local modifications are made on the basis of the original scheduling plan when new jobs arrive. The objective is to minimise the total completion time deviation of all products so that all jobs can be finished on time to reduce the cost of storage. First, according to the definitions of the dynamic DBFSP problem, a DRL framework based on multi-agent deep deterministic policy gradient (MADDPG) is proposed. In this framework, a full schedule is generated by the variable neighbourhood descent algorithm before a dynamic event occurs. Meanwhile, all newly added jobs are reordered before the agents make decisions to select the one that needs to be scheduled most urgently. This study defines the observations, actions and reward calculation methods and applies centralised training and distributed execution in MADDPG. Finally, a comprehensive computational experiment is carried out to compare the proposed method with the closely related and well-performing methods. The results indicate that the proposed method can solve the dynamic DBFSP effectively and efficiently.
KW - deep reinforcement learning
KW - distributed blocking flowshop scheduling problem
KW - dynamic scheduling
KW - job insertions
KW - multi-agent deep deterministic policy gradient
UR - http://www.scopus.com/inward/record.url?scp=85137467442&partnerID=8YFLogxK
U2 - 10.1049/cim2.12060
DO - 10.1049/cim2.12060
M3 - Article
AN - SCOPUS:85137467442
SN - 2516-8398
VL - 4
SP - 166
EP - 180
JO - IET Collaborative Intelligent Manufacturing
JF - IET Collaborative Intelligent Manufacturing
IS - 3
ER -