Actively learning dynamical systems using Bayesian neural networks

Springer Science and Business Media LLC - Tập 53 Số 23 - Trang 29338-29362 - 2023
Shengbing Tang1, Kenji Fujimoto2, Ichiro Maruta2
1National Engineering Research Center for E-Learning, Central China Normal University, Wuhan, China
2Department of Aeronautics and Astronautics, Kyoto University, Kyoto-shi, Japan

Tóm tắt

Từ khóa


Tài liệu tham khảo

Moerland TM, Broekens J, Plaat A, Jonker CM (2023) Model-based reinforcement learning: a survey. Found Trends$${\circledR} $$ Mach Learn 16(1):1–118

Chatzilygeroudis K, Vassiliades V, Stulp F, Calinon S, Mouret JB (2019) A survey on policy search algorithms for learning robot controllers in a handful of trials. IEEE Trans Robot 36(2):328–347

Levine S, Abbeel P (2014) Learning neural network policies with guided policy search under unknown dynamics. Adv Neural Inf Process Syst 27

Nagabandi A, Kahn G, Fearing RS, Levine S (2018) Neural network dynamics for model-based deep reinforcement learning with model-free fine-tuning. In: 2018 IEEE international conference on robotics and automation, pp 7559–7566

Deisenroth MP, Fox D, Rasmussen CE (2013) Gaussian processes for data-efficient learning in robotics and control. IEEE Trans Pattern Anal Mach Intell 37(2):408–423

Kurutach T, Clavera I, Duan Y, Tamar A, Abbeel P (2018) Model-ensemble trust-region policy optimization. In: International conference on learning representations

Chua K, Calandra R, McAllister R, Levine S (2018) Deep reinforcement learning in a handful of trials using probabilistic dynamics models. Adv Neural Inf Process Syst 31

Tang S, Fujimoto K, Maruta I (2021) Learning dynamic systems using Gaussian process regression with analytic ordinary differential equations as prior information. IEICE Trans Inf Syst 104(9):1440–1449

Tang S, Fujimoto K, Maruta I (2022) Actively learning Gaussian process dynamical systems through global and local explorations. IEEE Access

Higuera JCG, Meger D, Dudek G (2018) Synthesizing neural network controllers with probabilistic model-based reinforcement learning. In: 2018 IEEE/RSJ international conference on intelligent robots and systems, pp 2538–2544

Wang T, Ba J (2019) Exploring model-based planning with policy networks

Zarrop MB (1979) Optimal experiment design for dynamic system identification. Springer, Berlin Heidelberg

Settles B (2009) Active learning literature survey. Technical report, University of Wisconsin-Madison Department of Computer Sciences

Schultheis M, Belousov B, Abdulsamad H, Peters J (2020) Receding horizon curiosity. Proceedings of the conference on robot learning, pp 278–1288

Buisson-Fenet M, Solowjow F, Trimpe S (2020) Actively learning gaussian process dynamics. Proceedings of the 2nd conference on learning for dynamics and control, pp 5–15

Gal Y, Ghahramani Z (2016) Dropout as a bayesian approximation: representing model uncertainty in deep learning. International conference on machine learning, pp 1050–1059

Haarnoja T, Zhou A, Abbeel P, Levine S (2018) Soft actor-critic: off-policy maximum entropy deep reinforcement learning with a stochastic actor. International conference on machine learning, pp 1861–1870

Botev ZI, Kroese DP, Rubinstein RY, L’Ecuyer P (2013) The cross-entropy method for optimization. In: Handbook of statistics. Elsevier, vol 31, pp 35–59

Camacho EF, Alba CB (2013) Model predictive control. Springer Science & Business Media

Rao AV (2009) A survey of numerical methods for optimal control. Adv Astronaut Sci

Jospin LV, Laga H, Boussaid F, Buntine W, Bennamoun M (2022) Hands-on Bayesian neural networks-A tutorial for deep learning users. IEEE Comput Intell Mag 17(2):29–48

Blundell C, Cornebise J, Kavukcuoglu K, Wierstra D (2015) Weight uncertainty in neural network. International conference on machine learning, pp 1613–1622

Damianou A, Lawrence N (2013) Deep gaussian processes. Proceedings of the sixteenth international conference on artificial intelligence and statistics, vol 31, pp 207–215

Schulman J, Levine S, Moritz P, Jordan M, Abbeel P (2015) Trust region policy optimization. Int Conf Mach Learn 37:1889–1897

Clavera I, Rothfuss J, Schulman J, Fujita Y, Asfour T, Abbeel P (2018) Model-based reinforcement learning via meta-policy optimization. In: Conference on robot learning, pp 617–629

Kobilarov M (2012) Cross-entropy motion planning. Int J Robot Res 31(7):855–871

Todorov E, Erez T, Tassa Y (2012) Mujoco: a physics engine for model-based control. 2012 IEEE/RSJ international conference on intelligent robots and systems, pp 5026–5033

Daxberger E, Kristiadi A, Immer A, Eschenhagen R, Bauer M, Hennig P (2021) Laplace redux-effortless bayesian deep learning. Adv Neural Inf Process Syst 34:20089–20103

Gal Y, Islam R, Ghahramani Z (2017) Deep bayesian active learning with image data. International conference on machine learning, pp 1183–1192

Lindley DV (1956) On a measure of the information provided by an experiment. Ann Math Stat 27(4):986–1005