Handbook of Reinforcement Learning and Control

Handbook of Reinforcement Learning and Control

Author: Kyriakos G. Vamvoudakis

Publisher: Springer Nature

Published: 2021-06-23

Total Pages: 833

ISBN-13: 3030609901

DOWNLOAD EBOOK

This handbook presents state-of-the-art research in reinforcement learning, focusing on its applications in the control and game theory of dynamic systems and future directions for related research and technology. The contributions gathered in this book deal with challenges faced when using learning and adaptation methods to solve academic and industrial problems, such as optimization in dynamic environments with single and multiple agents, convergence and performance analysis, and online implementation. They explore means by which these difficulties can be solved, and cover a wide range of related topics including: deep learning; artificial intelligence; applications of game theory; mixed modality learning; and multi-agent reinforcement learning. Practicing engineers and scholars in the field of machine learning, game theory, and autonomous control will find the Handbook of Reinforcement Learning and Control to be thought-provoking, instructive and informative.


Intelligent Optimal Adaptive Control for Mechatronic Systems

Intelligent Optimal Adaptive Control for Mechatronic Systems

Author: Marcin Szuster

Publisher: Springer

Published: 2017-12-28

Total Pages: 387

ISBN-13: 331968826X

DOWNLOAD EBOOK

The book deals with intelligent control of mobile robots, presenting the state-of-the-art in the field, and introducing new control algorithms developed and tested by the authors. It also discusses the use of artificial intelligent methods like neural networks and neuraldynamic programming, including globalised dual-heuristic dynamic programming, for controlling wheeled robots and robotic manipulators,and compares them to classical control methods.


Reinforcement Learning and Optimal Control

Reinforcement Learning and Optimal Control

Author: Dimitri Bertsekas

Publisher: Athena Scientific

Published: 2019-07-01

Total Pages: 388

ISBN-13: 1886529396

DOWNLOAD EBOOK

This book considers large and challenging multistage decision problems, which can be solved in principle by dynamic programming (DP), but their exact solution is computationally intractable. We discuss solution methods that rely on approximations to produce suboptimal policies with adequate performance. These methods are collectively known by several essentially equivalent names: reinforcement learning, approximate dynamic programming, neuro-dynamic programming. They have been at the forefront of research for the last 25 years, and they underlie, among others, the recent impressive successes of self-learning in the context of games such as chess and Go. Our subject has benefited greatly from the interplay of ideas from optimal control and from artificial intelligence, as it relates to reinforcement learning and simulation-based neural network methods. One of the aims of the book is to explore the common boundary between these two fields and to form a bridge that is accessible by workers with background in either field. Another aim is to organize coherently the broad mosaic of methods that have proved successful in practice while having a solid theoretical and/or logical foundation. This may help researchers and practitioners to find their way through the maze of competing ideas that constitute the current state of the art. This book relates to several of our other books: Neuro-Dynamic Programming (Athena Scientific, 1996), Dynamic Programming and Optimal Control (4th edition, Athena Scientific, 2017), Abstract Dynamic Programming (2nd edition, Athena Scientific, 2018), and Nonlinear Programming (Athena Scientific, 2016). However, the mathematical style of this book is somewhat different. While we provide a rigorous, albeit short, mathematical account of the theory of finite and infinite horizon dynamic programming, and some fundamental approximation methods, we rely more on intuitive explanations and less on proof-based insights. Moreover, our mathematical requirements are quite modest: calculus, a minimal use of matrix-vector algebra, and elementary probability (mathematically complicated arguments involving laws of large numbers and stochastic convergence are bypassed in favor of intuitive explanations). The book illustrates the methodology with many examples and illustrations, and uses a gradual expository approach, which proceeds along four directions: (a) From exact DP to approximate DP: We first discuss exact DP algorithms, explain why they may be difficult to implement, and then use them as the basis for approximations. (b) From finite horizon to infinite horizon problems: We first discuss finite horizon exact and approximate DP methodologies, which are intuitive and mathematically simple, and then progress to infinite horizon problems. (c) From deterministic to stochastic models: We often discuss separately deterministic and stochastic problems, since deterministic problems are simpler and offer special advantages for some of our methods. (d) From model-based to model-free implementations: We first discuss model-based implementations, and then we identify schemes that can be appropriately modified to work with a simulator. The book is related and supplemented by the companion research monograph Rollout, Policy Iteration, and Distributed Reinforcement Learning (Athena Scientific, 2020), which focuses more closely on several topics related to rollout, approximate policy iteration, multiagent problems, discrete and Bayesian optimization, and distributed computation, which are either discussed in less detail or not covered at all in the present book. The author's website contains class notes, and a series of videolectures and slides from a 2021 course at ASU, which address a selection of topics from both books.


Reinforcement Learning, second edition

Reinforcement Learning, second edition

Author: Richard S. Sutton

Publisher: MIT Press

Published: 2018-11-13

Total Pages: 549

ISBN-13: 0262352702

DOWNLOAD EBOOK

The significantly expanded and updated new edition of a widely used text on reinforcement learning, one of the most active research areas in artificial intelligence. Reinforcement learning, one of the most active research areas in artificial intelligence, is a computational approach to learning whereby an agent tries to maximize the total amount of reward it receives while interacting with a complex, uncertain environment. In Reinforcement Learning, Richard Sutton and Andrew Barto provide a clear and simple account of the field's key ideas and algorithms. This second edition has been significantly expanded and updated, presenting new topics and updating coverage of other topics. Like the first edition, this second edition focuses on core online learning algorithms, with the more mathematical material set off in shaded boxes. Part I covers as much of reinforcement learning as possible without going beyond the tabular case for which exact solutions can be found. Many algorithms presented in this part are new to the second edition, including UCB, Expected Sarsa, and Double Learning. Part II extends these ideas to function approximation, with new sections on such topics as artificial neural networks and the Fourier basis, and offers expanded treatment of off-policy learning and policy-gradient methods. Part III has new chapters on reinforcement learning's relationships to psychology and neuroscience, as well as an updated case-studies chapter including AlphaGo and AlphaGo Zero, Atari game playing, and IBM Watson's wagering strategy. The final chapter discusses the future societal impacts of reinforcement learning.


Learning-Based Adaptive Control

Learning-Based Adaptive Control

Author: Mouhacine Benosman

Publisher: Butterworth-Heinemann

Published: 2016-08-02

Total Pages: 284

ISBN-13: 0128031514

DOWNLOAD EBOOK

Adaptive control has been one of the main problems studied in control theory. The subject is well understood, yet it has a very active research frontier. This book focuses on a specific subclass of adaptive control, namely, learning-based adaptive control. As systems evolve during time or are exposed to unstructured environments, it is expected that some of their characteristics may change. This book offers a new perspective about how to deal with these variations. By merging together Model-Free and Model-Based learning algorithms, the author demonstrates, using a number of mechatronic examples, how the learning process can be shortened and optimal control performance can be reached and maintained. - Includes a good number of Mechatronics Examples of the techniques. - Compares and blends Model-free and Model-based learning algorithms. - Covers fundamental concepts, state-of-the-art research, necessary tools for modeling, and control.


Adaptive Dynamic Programming with Applications in Optimal Control

Adaptive Dynamic Programming with Applications in Optimal Control

Author: Derong Liu

Publisher: Springer

Published: 2017-01-04

Total Pages: 609

ISBN-13: 3319508156

DOWNLOAD EBOOK

This book covers the most recent developments in adaptive dynamic programming (ADP). The text begins with a thorough background review of ADP making sure that readers are sufficiently familiar with the fundamentals. In the core of the book, the authors address first discrete- and then continuous-time systems. Coverage of discrete-time systems starts with a more general form of value iteration to demonstrate its convergence, optimality, and stability with complete and thorough theoretical analysis. A more realistic form of value iteration is studied where value function approximations are assumed to have finite errors. Adaptive Dynamic Programming also details another avenue of the ADP approach: policy iteration. Both basic and generalized forms of policy-iteration-based ADP are studied with complete and thorough theoretical analysis in terms of convergence, optimality, stability, and error bounds. Among continuous-time systems, the control of affine and nonaffine nonlinear systems is studied using the ADP approach which is then extended to other branches of control theory including decentralized control, robust and guaranteed cost control, and game theory. In the last part of the book the real-world significance of ADP theory is presented, focusing on three application examples developed from the authors’ work: • renewable energy scheduling for smart power grids;• coal gasification processes; and• water–gas shift reactions. Researchers studying intelligent control methods and practitioners looking to apply them in the chemical-process and power-supply industries will find much to interest them in this thorough treatment of an advanced approach to control.


Handbook On Computational Intelligence (In 2 Volumes)

Handbook On Computational Intelligence (In 2 Volumes)

Author: Plamen Parvanov Angelov

Publisher: World Scientific

Published: 2016-03-18

Total Pages: 964

ISBN-13: 9814675024

DOWNLOAD EBOOK

With the Internet, the proliferation of Big Data, and autonomous systems, mankind has entered into an era of 'digital obesity'. In this century, computational intelligence, such as thinking machines, have been brought forth to process complex human problems in a wide scope of areas — from social sciences, economics and biology, medicine and social networks, to cyber security.The Handbook of Computational Intelligence (in two volumes) prompts readers to look at these problems from a non-traditional angle. It takes a step by step approach, supported by case studies, to explore the issues that have arisen in the process. The Handbook covers many classic paradigms, as well as recent achievements and future promising developments to solve some of these very complex problems. Volume one explores the subjects of fuzzy logic and systems, artificial neural networks, and learning systems. Volume two delves into evolutionary computation, hybrid systems, as well as the applications of computational intelligence in decision making, the process industry, robotics, and autonomous systems.This work is a 'one-stop-shop' for beginners, as well as an inspirational source for more advanced researchers. It is a useful resource for lecturers and learners alike.


Advanced Optimal Control and Applications Involving Critic Intelligence

Advanced Optimal Control and Applications Involving Critic Intelligence

Author: Ding Wang

Publisher: Springer Nature

Published: 2023-01-21

Total Pages: 283

ISBN-13: 9811972915

DOWNLOAD EBOOK

This book intends to report new optimal control results with critic intelligence for complex discrete-time systems, which covers the novel control theory, advanced control methods, and typical applications for wastewater treatment systems. Therein, combining with artificial intelligence techniques, such as neural networks and reinforcement learning, the novel intelligent critic control theory as well as a series of advanced optimal regulation and trajectory tracking strategies are established for discrete-time nonlinear systems, followed by application verifications to complex wastewater treatment processes. Consequently, developing such kind of critic intelligence approaches is of great significance for nonlinear optimization and wastewater recycling. The book is likely to be of interest to researchers and practitioners as well as graduate students in automation, computer science, and process industry who wish to learn core principles, methods, algorithms, and applications in the field of intelligent optimal control. It is beneficial to promote the development of intelligent optimal control approaches and the construction of high-level intelligent systems.


Adaptive Dynamic Programming: Single and Multiple Controllers

Adaptive Dynamic Programming: Single and Multiple Controllers

Author: Ruizhuo Song

Publisher: Springer

Published: 2018-12-28

Total Pages: 278

ISBN-13: 9811317127

DOWNLOAD EBOOK

This book presents a class of novel optimal control methods and games schemes based on adaptive dynamic programming techniques. For systems with one control input, the ADP-based optimal control is designed for different objectives, while for systems with multi-players, the optimal control inputs are proposed based on games. In order to verify the effectiveness of the proposed methods, the book analyzes the properties of the adaptive dynamic programming methods, including convergence of the iterative value functions and the stability of the system under the iterative control laws. Further, to substantiate the mathematical analysis, it presents various application examples, which provide reference to real-world practices.