One objective of artificial intelligence is to model the behavior of an intelligent agent interacting with its environment. The environment's transformations could be modeled as a Markov chain, whose state is partially observable to the agent and affected by its actions; such processes are known as partially observable Markov decision processes (POMDPs). While the environment's dynamics are assumed to obey certain rules, the agent does not them and must learn.
In this dissertation we focus on the agent's adaptation as captured by the reinforcement learning framework. Reinforcement learning means learning a policy—a mapping of observations into actions-based on feedback from the environment. The learning can be viewed as browsing a set of policies while evaluating them by trial through interaction with the environment.
The set of policies being searched is constrained by the architecture of the agent's controller. POMDPs require a controller to have a memory. We investigate various architectures for controllers with memory, including controllers with external memory, finite state controllers and distributed controllers for multi-agent system. For these various controllers we work out the details of the algorithms which learn by ascending the gradient of expected cumulative reinforcement.
Building on statistical learning theory and experiment design theory, a policy evaluation algorithm is developed for the case of experience re-use. We address the question of sufficient experience for uniform convergence of policy evaluation and obtain sample complexity bounds for various estimators. Finally, we demonstrate the performance of the proposed algorithms on several domains, the most complex of which is simulated adaptive packet routing in a telecommunication network.
Cited By
- Yu Y, Hou P, Da Q and Qian Y Boosting Nonparametric Policies Proceedings of the 2016 International Conference on Autonomous Agents & Multiagent Systems, (477-484)
- Boularias A and Chaib-draa B Predictive representations for policy gradient in POMDPs Proceedings of the 26th Annual International Conference on Machine Learning, (65-72)
- Oliehoek F, Spaan M and Vlassis N (2008). Optimal and approximate Q-value functions for decentralized POMDPs, Journal of Artificial Intelligence Research, 32:1, (289-353), Online publication date: 1-May-2008.
- Varshavskaya P, Kaelbling L and Rus D (2008). Automated Design of Adaptive Controllers for Modular Robots using Reinforcement Learning, International Journal of Robotics Research, 27:3-4, (505-526), Online publication date: 1-Mar-2008.
- Khoussainov R and Kushmerick N Automated index management for distributed web search Proceedings of the twelfth international conference on Information and knowledge management, (386-393)
Index Terms
- Reinforcement learning by policy search
Recommendations
Continuous-action reinforcement learning with fast policy search and adaptive basis function selection
Special issue on Recent advances on machine learning and CyberneticsAs an important approach to solving complex sequential decision problems, reinforcement learning (RL) has been widely studied in the community of artificial intelligence and machine learning. However, the generalization ability of RL is still an open ...
Policy Synthesis and Reinforcement Learning for Discounted LTL
Computer Aided VerificationAbstractThe difficulty of manually specifying reward functions has led to an interest in using linear temporal logic (LTL) to express objectives for reinforcement learning (RL). However, LTL has the downside that it is sensitive to small perturbations in ...