Policies for the dynamic traveling maintainer problem with alerts

European Journal of Operational Research - Tập 305 - Trang 1141-1152 - 2023
Paulo da Costa1,2, Peter Verleijsdonk3,2, Simon Voorberg1,2, Alp Akcay1,2, Stella Kapodistria3,2, Willem van Jaarsveld1,2, Yingqian Zhang1,2
1Department of Industrial Engineering and Innovation Sciences, the Netherlands
2Eindhoven University of Technology, PO Box 513, Eindhoven 5600 MB, the Netherlands
3Department of Mathematics and Computer Science, the Netherlands

Tài liệu tham khảo

Afrati, 1986, The complexity of the travelling repairman problem, RAIRO - Theoretical Informatics and Applications - Informatique Théorique et Applications, 20, 79, 10.1051/ita/1986200100791 Akcay, 2022, An alert-assisted inspection policy for a production process with imperfect condition signals, European Journal of Operational Research, 298, 510, 10.1016/j.ejor.2021.05.051 Andriotis, 2021, Deep reinforcement learning driven inspection and maintenance planning under incomplete information and constraints, Reliability Engineering & System Safety, 212, 107551, 10.1016/j.ress.2021.107551 Badia, 2020, Agent57: Outperforming the Atari human benchmark, 507 Bellemare, 2017, A distributional perspective on reinforcement learning, vol. 70, 449 Bertsimas, D., Van Ryzin, G. et al. (1989). The dynamic traveling repairman problem. Bhattacharya, 2020, Reinforcement learning for POMDP: Partitioned rollout and policy iteration with application to autonomous sequential repair problems, IEEE Robotics and Automation Letters, 5, 3967, 10.1109/LRA.2020.2978451 Breton, 2009, Status, plans and technologies for offshore wind turbines in Europe and North America, Renewable Energy, 34, 646, 10.1016/j.renene.2008.05.040 Camci, 2014, The travelling maintainer problem: Integration of condition-based maintenance with the travelling salesman problem, Journal of the Operational Research Society, 65, 1423, 10.1057/jors.2013.88 Camci, 2015, Maintenance scheduling of geographically distributed assets with prognostics information, European Journal of Operational Research, 245, 506, 10.1016/j.ejor.2015.03.023 Cartesius (accessed: 08.05.2021). Cartesius supercomputer. https://www.surf.nl/en/dutch-national-supercomputer-cartesius. Compare, 2018, Reinforcement learning-based flow management of gas turbine parts under stochastic failures, The International Journal of Advanced Manufacturing Technology, 99, 2981, 10.1007/s00170-018-2690-6 Dabney, 2018, Distributional reinforcement learning with quantile regression, vol. 32, 2892 De Asis, 2018, Multi-step reinforcement learning: A unifying algorithm, vol. 32, 2902 De Jonge, 2016, Reducing costs by clustering maintenance activities for multiple critical units, Reliability Engineering & System Safety, 145, 93, 10.1016/j.ress.2015.09.003 De Jonge, 2020, A review on maintenance optimization, European Journal of Operational Research, 285, 805, 10.1016/j.ejor.2019.09.047 Derman, 1963, On optimal replacement rules when changes of state are Markovian, Mathematical Optimization Techniques, 396, 201, 10.1525/9780520319875-011 Drent, 2020, Dynamic dispatching and repositioning policies for fast-response service networks, European Journal of Operational Research, 285, 583, 10.1016/j.ejor.2020.02.014 Havinga, 2020, Condition-based maintenance in the cyclic patrolling repairman problem, International Journal of Production Economics, 222, 107497, 10.1016/j.ijpe.2019.09.018 Hernandez-Garcia, J. F., & Sutton, R. S. (2019). Understanding multi-step deep reinforcement learning: a systematicstudy of the dqn target. arXiv preprint arXiv:1901.07510. Hessel, 2018, Rainbow: Combining improvements in deep reinforcement learning, vol. 32, 3215 Huber, 1964, Robust estimation of a location parameter, The Annals of Mathematical Statistics, 35, 73, 10.1214/aoms/1177703732 Jaakkola, 1993, Convergence of stochastic iterative dynamic programming algorithms, Advances in Neural Information Processing Systems, 6, 703 Keizer, 2017, Condition-based maintenance policies for systems with multiple dependent components: A review, European Journal of Operational Research, 261, 405, 10.1016/j.ejor.2017.02.044 Kenbeek, 2019, Data-driven online monitoring of wind turbines, 143150 Kingma, D. P., & Ba, J. (2014). Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980. Kuhnle, 2019, Reinforcement learning for opportunistic maintenance optimization, Production Engineering, 13, 33, 10.1007/s11740-018-0855-7 Liu, 2020, Dynamic selective maintenance optimization for multi-state systems over a finite horizon: A deep reinforcement learning approach, European Journal of Operational Research, 283, 166, 10.1016/j.ejor.2019.10.049 Mnih, 2015, Human-level control through deep reinforcement learning, Nature, 518, 529, 10.1038/nature14236 Puterman, 1990, Markov decision processes, Handbooks in Operations Research and Management Science, 2, 331, 10.1016/S0927-0507(05)80172-0 Sutton, 2018 Topan, 2020, A review of operational spare parts service logistics in service control towers, European Journal of Operational Research, 282, 401, 10.1016/j.ejor.2019.03.026 Tulabandhula, 2011, The machine learning and traveling repairman problem, 262 Van Hasselt, 2016, Deep reinforcement learning with double q-learning, vol. 30, 2094 Van Staden, 2021, The effect of multi-sensor data on condition-based maintenance policies, European Journal of Operational Research, 290, 585, 10.1016/j.ejor.2020.08.035 Wang, 2012, An overview of the recent advances in delay-time-based maintenance modelling, Reliability Engineering & System Safety, 106, 165, 10.1016/j.ress.2012.04.004