A deep inverse reinforcement learning approach to route choice modeling with context-dependent rewards

Zhan Zhao1,2, Yuebing Liang1
1Department of Urban Planning and Design, The University of Hong Kong, Hong Kong SAR, China
2Musketeers Foundation Institute of Data Science, The University of Hong Kong, Hong Kong SAR, China

Tài liệu tham khảo

Abbeel, 2004, Apprenticeship learning via inverse reinforcement learning, 1 Alwosheel, 2021, Why did you predict that? Towards explainable artificial neural networks for travel demand analysis, Transp. Res. C, 128, 10.1016/j.trc.2021.103143 Azevedo, 1993, An algorithm for the ranking of shortest paths, European J. Oper. Res., 69, 97, 10.1016/0377-2217(93)90095-5 Ben-Akiva, 1999, Discrete choice methods and their applications to short term travel decisions, 5 Cantarella, 2005, Multilayer feedforward networks for transportation mode choice analysis: An analysis and a comparison with random utility models, Transp. Res. C, 13, 121, 10.1016/j.trc.2005.04.002 Choi, 2021, TrajGAIL: Generating urban vehicle trajectories using generative adversarial imitation learning, Transp. Res. C, 128, 10.1016/j.trc.2021.103091 Doshi-Velez, 2017 Finn, 2016 Finn, 2016, Guided cost learning: deep inverse optimal control via policy optimization, 49 Fosgerau, 2013, A link based network route choice model with unrestricted choice set, Transp. Res. B, 56, 70, 10.1016/j.trb.2013.07.012 Frejinger, 2009, Sampling of alternatives for route choice modeling, Transp. Res. B, 43, 984, 10.1016/j.trb.2009.03.001 Fu, 2018 Goodfellow, 2014, Generative adversarial nets He, 2020, What is the human mobility in a new city: Transfer mobility knowledge across cities, 1355 Ho, 2016, Generative adversarial imitation learning Jan, 2000, Using global positioning system data to understand variations in path choice, Transp. Res. Rec., 1725, 37, 10.3141/1725-06 Koch, 2020, A review of methods to model route choice behavior of bicyclists: inverse reinforcement learning in spatial context and recursive logit, 30 Liang, 2022, NetTraj: A network-based vehicle trajectory prediction model with directional representation and spatiotemporal attention mechanisms, IEEE Trans. Intell. Transp. Syst., 23, 14470, 10.1109/TITS.2021.3129588 Liang, 2022, Modeling taxi cruising time based on multi-source data: a case study in Shanghai, Transportation, 10.1007/s11116-022-10348-y Lima, 2016, Understanding individual routing behaviour, J. R. Soc. Interface, 13, 10.1098/rsif.2016.0021 Liu, 2022, Personalized route recommendation for ride-hailing with deep inverse reinforcement learning and real-time traffic conditions, Transp. Res. E, 164, 10.1016/j.tre.2022.102780 Liu, 2020, Integrating Dijkstra’s algorithm into deep inverse reinforcement learning for food delivery route planning, Transp. Res. E, 142, 10.1016/j.tre.2020.102070 Lou, 2009, Map-matching for low-sampling-rate GPS trajectories, 352 Lundberg, 2017, A unified approach to interpreting model predictions, 4768 Mai, 2018, A decomposition method for estimating recursive logit based route choice models, EURO J. Transp. Logist., 7, 253, 10.1007/s13676-016-0102-3 Mai, 2015, A nested recursive logit model for route choice analysis, Transp. Res. B, 75, 100, 10.1016/j.trb.2015.03.015 Marra, 2021 Mnih, 2015, Human-level control through deep reinforcement learning, Nature, 518, 529, 10.1038/nature14236 Ng, 2000, Algorithms for inverse reinforcement learning, 663 Oyama, 2017, A discounted recursive logit model for dynamic gridlock network analysis, Transp. Res. C, 85, 509, 10.1016/j.trc.2017.10.001 Prato, 2009, Route choice modeling: past, present and future research directions, J. Choice Model., 2, 65, 10.1016/S1755-5345(13)70005-8 Prato, 2007, Modeling route choice behavior: How relevant is the composition of choice set?, Transp. Res. Rec., 2003, 64, 10.3141/2003-09 Rust, 1987, Optimal replacement of GMC bus engines: An empirical model of Harold Zurcher, Econometrica, 55, 999, 10.2307/1911259 Schulman, 2016, High-dimensional continuous control using generalized advantage estimation Schulman, 2017 Sifringer, 2020, Enhancing discrete choice models with representation learning, Transp. Res. B, 140, 236, 10.1016/j.trb.2020.08.006 Simini, 2021, A Deep Gravity model for mobility flows generation, Nature Commun., 12, 6576, 10.1038/s41467-021-26752-4 Sutton, 2018, xxii, 526 Wang, 2020, Deep neural networks for choice analysis: Extracting complete economic information for interpretation, Transp. Res. C, 118, 10.1016/j.trc.2020.102701 Wulfmeier, 2016 Yang, 2018, Fast map matching, an algorithm integrating hidden Markov model with precomputation, Int. J. Geogr. Inf. Sci., 10.1080/13658816.2017.1400548 Zhang, 2020, cGAIL: Conditional generative adversarial imitation learning—An application in taxi drivers’ strategy learning, IEEE Trans. Big Data, 1 Ziebart, 2008, Maximum entropy inverse reinforcement learning, 1433 Zimmermann, 2020, A tutorial on recursive models for analyzing and predicting path choice behavior, EURO J. Transp. Logist., 9, 10.1016/j.ejtl.2020.100004 Zimmermann, 2017, Bike route choice modeling using GPS data without choice sets of paths, Transp. Res. C, 75, 183, 10.1016/j.trc.2016.12.009