Research on task decomposition and state abstraction in reinforcement learning

Artificial Intelligence Review - Tập 38 - Trang 119-127 - 2011
Yu Lasheng1, Jiang Zhongbin1, Liu Kang1
1School of Information Science and Engineering, Central South University, Hunan, China

Tóm tắt

Task decomposition and State abstraction are crucial parts in reinforcement learning. It allows an agent to ignore aspects of its current states that are irrelevant to its current decision, and therefore speeds up dynamic programming and learning. This paper presents the SVI algorithm that uses a dynamic Bayesian network model to construct an influence graph that indicates relationships between state variables. SVI performs state abstraction for each subtask by ignoring irrelevant state variables and lower level subtasks. Experiment results show that the decomposition of tasks introduced by SVI can significantly accelerate constructing a near-optimal policy. This general framework can be applied to a broad spectrum of complex real world problems such as robotics, industrial manufacturing, games and others.

Tài liệu tham khảo

Barto A, Mahadevan S (2003) Recent advances in hierarchical reinforcement learning. Discrete Event Syst (special issue on reinforcement learning) 13: 41–77 Bertsekas DP, Tsitsiklis JN (1996) Neuro-dynamic programming. Athena Scientific, Belmont Boutilier C, Dearden R, Goldszmidt M (1995) Exploiting structure in policy construction. IJCAI 14: 1104–1113 Dean T, Kanazawa K (1989) A model for reasoning about persistence and causation. Comput Intell 5(3): 142–150 Dietterich T (2000) Hierarchical reinforcement learning with the MAXQ value function decoposition. J Artif Intell Res 13: 227–303 Hengst B (2002) Discovering hierarchy in reinforcement learning with HEXQ. ICML 19: 243–250 Jonsson A, Barto A (2005) A causal approach to hierarchical decomposition of factored MDPs. In: Proceedings of the 22nd international conference on machine learning, pp 401–408 Makar R, Mahadevan S, Ghavamzadeh M (2001) Hierarchical multi-agent reinforcement learning. In: Proceedings of the 5th international conference on autonomous agents Parr R, Russell S (1998) Reinforcement learning with hierarchies of machines. Advances in neural information processing systems. MIT Press, Oxford, pp 1043–1049 Sutton R, Barto A (1998) Reinforcement learning. MIT Press, Oxford Sutton R, Precup D, Singh S (1999) Between MDPs and semi-MDPs: a framework for temporal abstraction in reinforcement learning. Artif Intell 112(1-2): 181–211