{"title":"Robust-stable scheduling in dynamic flow shops based on deep reinforcement learning","date_created":"2023-01-10T13:08:37Z","publication":"Journal of Intelligent Manufacturing","citation":{"alphadin":"Grumbach, Felix ; Müller, Anna ; Reusch, Pascal ; Trojahn, Sebastian: Robust-stable scheduling in dynamic flow shops based on deep reinforcement learning. In: Journal of Intelligent Manufacturing, Springer Science and Business Media LLC (2022)","short":"F. Grumbach, A. Müller, P. Reusch, S. Trojahn, Journal of Intelligent Manufacturing (2022).","ieee":"F. Grumbach, A. Müller, P. Reusch, and S. Trojahn, “Robust-stable scheduling in dynamic flow shops based on deep reinforcement learning,” Journal of Intelligent Manufacturing, 2022.","apa":"Grumbach, F., Müller, A., Reusch, P., & Trojahn, S. (2022). Robust-stable scheduling in dynamic flow shops based on deep reinforcement learning. Journal of Intelligent Manufacturing. https://doi.org/10.1007/s10845-022-02069-x","chicago":"Grumbach, Felix, Anna Müller, Pascal Reusch, and Sebastian Trojahn. “Robust-Stable Scheduling in Dynamic Flow Shops Based on Deep Reinforcement Learning.” Journal of Intelligent Manufacturing, 2022. https://doi.org/10.1007/s10845-022-02069-x.","ama":"Grumbach F, Müller A, Reusch P, Trojahn S. Robust-stable scheduling in dynamic flow shops based on deep reinforcement learning. Journal of Intelligent Manufacturing. 2022. doi:10.1007/s10845-022-02069-x","mla":"Grumbach, Felix, et al. “Robust-Stable Scheduling in Dynamic Flow Shops Based on Deep Reinforcement Learning.” Journal of Intelligent Manufacturing, Springer Science and Business Media LLC, 2022, doi:10.1007/s10845-022-02069-x.","bibtex":"@article{Grumbach_Müller_Reusch_Trojahn_2022, title={Robust-stable scheduling in dynamic flow shops based on deep reinforcement learning}, DOI={10.1007/s10845-022-02069-x}, journal={Journal of Intelligent Manufacturing}, publisher={Springer Science and Business Media LLC}, author={Grumbach, Felix and Müller, Anna and Reusch, Pascal and Trojahn, Sebastian}, year={2022} }"},"user_id":"216459","article_type":"original","_id":"2295","status":"public","publication_status":"published","year":"2022","tmp":{"short":"CC BY (4.0)","image":"/images/cc_by.png","name":"Creative Commons Attribution 4.0 International Public License (CC-BY 4.0)","legal_code_url":"https://creativecommons.org/licenses/by/4.0/legalcode"},"type":"journal_article","main_file_link":[{"open_access":"1","url":"https://link.springer.com/article/10.1007/s10845-022-02069-x"}],"abstract":[{"text":"This proof-of-concept study provides a novel method for robust-stable scheduling in dynamic flow shops based on deep reinforcement learning (DRL) implemented with OpenAI frameworks. In realistic manufacturing environments, dynamic events endanger baseline schedules, which can require a cost intensive re-scheduling. Extensive research has been done on methods for generating proactive baseline schedules to absorb uncertainties in advance and in balancing the competing metrics of robustness and stability. Recent studies presented exact methods and heuristics based on Monte Carlo experiments (MCE), both of which are very computationally intensive. Furthermore, approaches based on surrogate measures were proposed, which do not explicitly consider uncertainties and robustness metrics. Surprisingly, DRL has not yet been scientifically investigated for generating robust-stable schedules in the proactive stage of production planning. The contribution of this article is a proposal on how DRL can be applied to manipulate operation slack times by stretching or compressing plan durations. The method is demonstrated using different flow shop instances with uncertain processing times, stochastic machine failures and uncertain repair times. Through a computational study, we found that DRL agents achieve about 98% result quality but only take about 2% of the time compared to traditional metaheuristics. This is a promising advantage for the use in real-time environments and supports the idea of improving proactive scheduling methods with machine learning based techniques.\r\n ","lang":"eng"}],"publication_identifier":{"issn":["0956-5515"],"eissn":["1572-8145"]},"date_updated":"2023-06-17T09:49:50Z","language":[{"iso":"eng"}],"jel":["C6"],"publisher":"Springer Science and Business Media LLC","oa":"1","author":[{"id":"243801","first_name":"Felix","orcid":"0000-0001-6348-7897","last_name":"Grumbach","full_name":"Grumbach, Felix"},{"last_name":"Müller","full_name":"Müller, Anna","first_name":"Anna"},{"first_name":"Pascal","last_name":"Reusch","full_name":"Reusch, Pascal"},{"last_name":"Trojahn","full_name":"Trojahn, Sebastian","first_name":"Sebastian"}],"keyword":["Dynamic flow shop","Predictive scheduling","Proactive scheduling","Robust scheduling","Reinforcement learning","Simheuristics"],"doi":"10.1007/s10845-022-02069-x","quality_controlled":"1"}