Riassunto
This work addresses research in an area that is gaining popularity in the artificial intelligence and neural network communities. Reinforcement learning has become a primary paradigm of machine learning. It applies to problems in which an agent (such as a robot, a process controller, or an information-retrieval engine) has to learn how to behave given only information about the success of its current actions. This book is a collection of papers that address topics including the theoretical foundations of dynamic programming approaches, the role of prior knowledge, and methods for improving performance of reinforcement-learning techniques. These papers build on previous work and form a resource for students and researchers in the area.
Contenuti
Editorial; T.G. Dietterich. Introduction; L.P. Kaelbling. Efficient Reinforcement Learning Through Symbiotic Evolution; D.E. Moriarty, R. Mikkulainen. Linear Least-Squares Algorithms for Temporal Difference Learning; S.J. Bradtke, A.G. Barto. Feature-Based Methods for Large Scale Dynamic Programming; J.N. Tsitsiklis, B. Van Roy. On the Worst-Case Analysis of Temporal-Difference Learning Algorithms; R.E. Schapire, M.K. Warmuth. Reinforcement Learning with Replacing Eligibility Traces; S.P. Singh, R.S. Sutton. Average Reward Reinforcement Learning: Foundations, Algorithms, and Empirical Results; S. Mahadevan. The Loss from Imperfect Value Functions in Expectation-Based and Minimax-Based Tasks; M. Heger. The Effect of Representation and Knowledge on Goal-Directed Exploration with Reinforcement-Learning Algorithms; S. Koenig, R.G. Simmons. Creating Advice-Taking Reinforcement Learners; R. Maclin, J.W. Shavlik. Technical Note: Incremental Multi-Step Q-Learning; J. Peng, R.J. Williams.
Le informazioni nella sezione "Su questo libro" possono far riferimento a edizioni diverse di questo titolo.