A deep inverse reinforcement learning approach to route choice modeling with context-dependent rewards
Tài liệu tham khảo
Abbeel, 2004, Apprenticeship learning via inverse reinforcement learning, 1
Alwosheel, 2021, Why did you predict that? Towards explainable artificial neural networks for travel demand analysis, Transp. Res. C, 128, 10.1016/j.trc.2021.103143
Azevedo, 1993, An algorithm for the ranking of shortest paths, European J. Oper. Res., 69, 97, 10.1016/0377-2217(93)90095-5
Ben-Akiva, 1999, Discrete choice methods and their applications to short term travel decisions, 5
Cantarella, 2005, Multilayer feedforward networks for transportation mode choice analysis: An analysis and a comparison with random utility models, Transp. Res. C, 13, 121, 10.1016/j.trc.2005.04.002
Choi, 2021, TrajGAIL: Generating urban vehicle trajectories using generative adversarial imitation learning, Transp. Res. C, 128, 10.1016/j.trc.2021.103091
Doshi-Velez, 2017
Finn, 2016
Finn, 2016, Guided cost learning: deep inverse optimal control via policy optimization, 49
Fosgerau, 2013, A link based network route choice model with unrestricted choice set, Transp. Res. B, 56, 70, 10.1016/j.trb.2013.07.012
Frejinger, 2009, Sampling of alternatives for route choice modeling, Transp. Res. B, 43, 984, 10.1016/j.trb.2009.03.001
Fu, 2018
Goodfellow, 2014, Generative adversarial nets
He, 2020, What is the human mobility in a new city: Transfer mobility knowledge across cities, 1355
Ho, 2016, Generative adversarial imitation learning
Jan, 2000, Using global positioning system data to understand variations in path choice, Transp. Res. Rec., 1725, 37, 10.3141/1725-06
Koch, 2020, A review of methods to model route choice behavior of bicyclists: inverse reinforcement learning in spatial context and recursive logit, 30
Liang, 2022, NetTraj: A network-based vehicle trajectory prediction model with directional representation and spatiotemporal attention mechanisms, IEEE Trans. Intell. Transp. Syst., 23, 14470, 10.1109/TITS.2021.3129588
Liang, 2022, Modeling taxi cruising time based on multi-source data: a case study in Shanghai, Transportation, 10.1007/s11116-022-10348-y
Lima, 2016, Understanding individual routing behaviour, J. R. Soc. Interface, 13, 10.1098/rsif.2016.0021
Liu, 2022, Personalized route recommendation for ride-hailing with deep inverse reinforcement learning and real-time traffic conditions, Transp. Res. E, 164, 10.1016/j.tre.2022.102780
Liu, 2020, Integrating Dijkstra’s algorithm into deep inverse reinforcement learning for food delivery route planning, Transp. Res. E, 142, 10.1016/j.tre.2020.102070
Lou, 2009, Map-matching for low-sampling-rate GPS trajectories, 352
Lundberg, 2017, A unified approach to interpreting model predictions, 4768
Mai, 2018, A decomposition method for estimating recursive logit based route choice models, EURO J. Transp. Logist., 7, 253, 10.1007/s13676-016-0102-3
Mai, 2015, A nested recursive logit model for route choice analysis, Transp. Res. B, 75, 100, 10.1016/j.trb.2015.03.015
Marra, 2021
Mnih, 2015, Human-level control through deep reinforcement learning, Nature, 518, 529, 10.1038/nature14236
Ng, 2000, Algorithms for inverse reinforcement learning, 663
Oyama, 2017, A discounted recursive logit model for dynamic gridlock network analysis, Transp. Res. C, 85, 509, 10.1016/j.trc.2017.10.001
Prato, 2009, Route choice modeling: past, present and future research directions, J. Choice Model., 2, 65, 10.1016/S1755-5345(13)70005-8
Prato, 2007, Modeling route choice behavior: How relevant is the composition of choice set?, Transp. Res. Rec., 2003, 64, 10.3141/2003-09
Rust, 1987, Optimal replacement of GMC bus engines: An empirical model of Harold Zurcher, Econometrica, 55, 999, 10.2307/1911259
Schulman, 2016, High-dimensional continuous control using generalized advantage estimation
Schulman, 2017
Sifringer, 2020, Enhancing discrete choice models with representation learning, Transp. Res. B, 140, 236, 10.1016/j.trb.2020.08.006
Simini, 2021, A Deep Gravity model for mobility flows generation, Nature Commun., 12, 6576, 10.1038/s41467-021-26752-4
Sutton, 2018, xxii, 526
Wang, 2020, Deep neural networks for choice analysis: Extracting complete economic information for interpretation, Transp. Res. C, 118, 10.1016/j.trc.2020.102701
Wulfmeier, 2016
Yang, 2018, Fast map matching, an algorithm integrating hidden Markov model with precomputation, Int. J. Geogr. Inf. Sci., 10.1080/13658816.2017.1400548
Zhang, 2020, cGAIL: Conditional generative adversarial imitation learning—An application in taxi drivers’ strategy learning, IEEE Trans. Big Data, 1
Ziebart, 2008, Maximum entropy inverse reinforcement learning, 1433
Zimmermann, 2020, A tutorial on recursive models for analyzing and predicting path choice behavior, EURO J. Transp. Logist., 9, 10.1016/j.ejtl.2020.100004
Zimmermann, 2017, Bike route choice modeling using GPS data without choice sets of paths, Transp. Res. C, 75, 183, 10.1016/j.trc.2016.12.009