Reinforcement Learning and Dynamic Programming Using Function Approximators

Reinforcement Learning and Dynamic Programming Using Function Approximators

Author: Lucian Busoniu

Publisher: CRC Press

Published: 2017-07-28

Total Pages: 280

ISBN-13: 1439821097

DOWNLOAD EBOOK

From household appliances to applications in robotics, engineered systems involving complex dynamics can only be as effective as the algorithms that control them. While Dynamic Programming (DP) has provided researchers with a way to optimally solve decision and control problems involving complex dynamic systems, its practical value was limited by algorithms that lacked the capacity to scale up to realistic problems. However, in recent years, dramatic developments in Reinforcement Learning (RL), the model-free counterpart of DP, changed our understanding of what is possible. Those developments led to the creation of reliable methods that can be applied even when a mathematical model of the system is unavailable, allowing researchers to solve challenging control problems in engineering, as well as in a variety of other disciplines, including economics, medicine, and artificial intelligence. Reinforcement Learning and Dynamic Programming Using Function Approximators provides a comprehensive and unparalleled exploration of the field of RL and DP. With a focus on continuous-variable problems, this seminal text details essential developments that have substantially altered the field over the past decade. In its pages, pioneering experts provide a concise introduction to classical RL and DP, followed by an extensive presentation of the state-of-the-art and novel methods in RL and DP with approximation. Combining algorithm development with theoretical guarantees, they elaborate on their work with illustrative examples and insightful comparisons. Three individual chapters are dedicated to representative algorithms from each of the major classes of techniques: value iteration, policy iteration, and policy search. The features and performance of these algorithms are highlighted in extensive experimental studies on a range of control applications. The recent development of applications involving complex systems has led to a surge of interest in RL and DP methods and the subsequent need for a quality resource on the subject. For graduate students and others new to the field, this book offers a thorough introduction to both the basics and emerging methods. And for those researchers and practitioners working in the fields of optimal and adaptive control, machine learning, artificial intelligence, and operations research, this resource offers a combination of practical algorithms, theoretical analysis, and comprehensive examples that they will be able to adapt and apply to their own work. Access the authors' website at www.dcsc.tudelft.nl/rlbook/ for additional material, including computer code used in the studies and information concerning new developments.


Algorithms for Reinforcement Learning

Algorithms for Reinforcement Learning

Author: Csaba Grossi

Publisher: Springer Nature

Published: 2022-05-31

Total Pages: 89

ISBN-13: 3031015517

DOWNLOAD EBOOK

Reinforcement learning is a learning paradigm concerned with learning to control a system so as to maximize a numerical performance measure that expresses a long-term objective. What distinguishes reinforcement learning from supervised learning is that only partial feedback is given to the learner about the learner's predictions. Further, the predictions may have long term effects through influencing the future state of the controlled system. Thus, time plays a special role. The goal in reinforcement learning is to develop efficient learning algorithms, as well as to understand the algorithms' merits and limitations. Reinforcement learning is of great interest because of the large number of practical applications that it can be used to address, ranging from problems in artificial intelligence to operations research or control engineering. In this book, we focus on those algorithms of reinforcement learning that build on the powerful theory of dynamic programming. We give a fairly comprehensive catalog of learning problems, describe the core ideas, note a large number of state of the art algorithms, followed by the discussion of their theoretical properties and limitations. Table of Contents: Markov Decision Processes / Value Prediction Problems / Control / For Further Exploration


Applied Dynamic Programming

Applied Dynamic Programming

Author: Richard E. Bellman

Publisher: Princeton University Press

Published: 2015-12-08

Total Pages: 389

ISBN-13: 1400874653

DOWNLOAD EBOOK

This comprehensive study of dynamic programming applied to numerical solution of optimization problems. It will interest aerodynamic, control, and industrial engineers, numerical analysts, and computer specialists, applied mathematicians, economists, and operations and systems analysts. Originally published in 1962. The Princeton Legacy Library uses the latest print-on-demand technology to again make available previously out-of-print books from the distinguished backlist of Princeton University Press. These editions preserve the original texts of these important books while presenting them in durable paperback and hardcover editions. The goal of the Princeton Legacy Library is to vastly increase access to the rich scholarly heritage found in the thousands of books published by Princeton University Press since its founding in 1905.


Optimal Stochastic Control, Stochastic Target Problems, and Backward SDE

Optimal Stochastic Control, Stochastic Target Problems, and Backward SDE

Author: Nizar Touzi

Publisher: Springer Science & Business Media

Published: 2012-09-25

Total Pages: 219

ISBN-13: 1461442869

DOWNLOAD EBOOK

This book collects some recent developments in stochastic control theory with applications to financial mathematics. We first address standard stochastic control problems from the viewpoint of the recently developed weak dynamic programming principle. A special emphasis is put on the regularity issues and, in particular, on the behavior of the value function near the boundary. We then provide a quick review of the main tools from viscosity solutions which allow to overcome all regularity problems. We next address the class of stochastic target problems which extends in a nontrivial way the standard stochastic control problems. Here the theory of viscosity solutions plays a crucial role in the derivation of the dynamic programming equation as the infinitesimal counterpart of the corresponding geometric dynamic programming equation. The various developments of this theory have been stimulated by applications in finance and by relevant connections with geometric flows. Namely, the second order extension was motivated by illiquidity modeling, and the controlled loss version was introduced following the problem of quantile hedging. The third part specializes to an overview of Backward stochastic differential equations, and their extensions to the quadratic case.​


Reinforcement Learning, second edition

Reinforcement Learning, second edition

Author: Richard S. Sutton

Publisher: MIT Press

Published: 2018-11-13

Total Pages: 549

ISBN-13: 0262352702

DOWNLOAD EBOOK

The significantly expanded and updated new edition of a widely used text on reinforcement learning, one of the most active research areas in artificial intelligence. Reinforcement learning, one of the most active research areas in artificial intelligence, is a computational approach to learning whereby an agent tries to maximize the total amount of reward it receives while interacting with a complex, uncertain environment. In Reinforcement Learning, Richard Sutton and Andrew Barto provide a clear and simple account of the field's key ideas and algorithms. This second edition has been significantly expanded and updated, presenting new topics and updating coverage of other topics. Like the first edition, this second edition focuses on core online learning algorithms, with the more mathematical material set off in shaded boxes. Part I covers as much of reinforcement learning as possible without going beyond the tabular case for which exact solutions can be found. Many algorithms presented in this part are new to the second edition, including UCB, Expected Sarsa, and Double Learning. Part II extends these ideas to function approximation, with new sections on such topics as artificial neural networks and the Fourier basis, and offers expanded treatment of off-policy learning and policy-gradient methods. Part III has new chapters on reinforcement learning's relationships to psychology and neuroscience, as well as an updated case-studies chapter including AlphaGo and AlphaGo Zero, Atari game playing, and IBM Watson's wagering strategy. The final chapter discusses the future societal impacts of reinforcement learning.


A Multigrid Tutorial

A Multigrid Tutorial

Author: William L. Briggs

Publisher: SIAM

Published: 2000-07-01

Total Pages: 318

ISBN-13: 9780898714623

DOWNLOAD EBOOK

Mathematics of Computing -- Numerical Analysis.


Algorithms for Optimization

Algorithms for Optimization

Author: Mykel J. Kochenderfer

Publisher: MIT Press

Published: 2019-03-12

Total Pages: 521

ISBN-13: 0262039427

DOWNLOAD EBOOK

A comprehensive introduction to optimization with a focus on practical algorithms for the design of engineering systems. This book offers a comprehensive introduction to optimization with a focus on practical algorithms. The book approaches optimization from an engineering perspective, where the objective is to design a system that optimizes a set of metrics subject to constraints. Readers will learn about computational approaches for a range of challenges, including searching high-dimensional spaces, handling problems where there are multiple competing objectives, and accommodating uncertainty in the metrics. Figures, examples, and exercises convey the intuition behind the mathematical approaches. The text provides concrete implementations in the Julia programming language. Topics covered include derivatives and their generalization to multiple dimensions; local descent and first- and second-order methods that inform local descent; stochastic methods, which introduce randomness into the optimization process; linear constrained optimization, when both the objective function and the constraints are linear; surrogate models, probabilistic surrogate models, and using probabilistic surrogate models to guide optimization; optimization under uncertainty; uncertainty propagation; expression optimization; and multidisciplinary design optimization. Appendixes offer an introduction to the Julia language, test functions for evaluating algorithm performance, and mathematical concepts used in the derivation and analysis of the optimization methods discussed in the text. The book can be used by advanced undergraduates and graduate students in mathematics, statistics, computer science, any engineering field, (including electrical engineering and aerospace engineering), and operations research, and as a reference for professionals.


Dynamic Economics

Dynamic Economics

Author: Jerome Adda

Publisher: MIT Press

Published: 2023-05-09

Total Pages: 297

ISBN-13: 0262547880

DOWNLOAD EBOOK

An integrated approach to the empirical application of dynamic optimization programming models, for students and researchers. This book is an effective, concise text for students and researchers that combines the tools of dynamic programming with numerical techniques and simulation-based econometric methods. Doing so, it bridges the traditional gap between theoretical and empirical research and offers an integrated framework for studying applied problems in macroeconomics and microeconomics. In part I the authors first review the formal theory of dynamic optimization; they then present the numerical tools and econometric techniques necessary to evaluate the theoretical models. In language accessible to a reader with a limited background in econometrics, they explain most of the methods used in applied dynamic research today, from the estimation of probability in a coin flip to a complicated nonlinear stochastic structural model. These econometric techniques provide the final link between the dynamic programming problem and data. Part II is devoted to the application of dynamic programming to specific areas of applied economics, including the study of business cycles, consumption, and investment behavior. In each instance the authors present the specific optimization problem as a dynamic programming problem, characterize the optimal policy functions, estimate the parameters, and use models for policy evaluation. The original contribution of Dynamic Economics: Quantitative Methods and Applications lies in the integrated approach to the empirical application of dynamic optimization programming models. This integration shows that empirical applications actually complement the underlying theory of optimization, while dynamic programming problems provide needed structure for estimation and policy evaluation.