Dynamic Programming and Its Applications

Dynamic Programming and Its Applications

Author: Martin L. Puterman

Publisher: Academic Press

Published: 2014-05-10

Total Pages: 427

ISBN-13: 1483258947

DOWNLOAD EBOOK

Dynamic Programming and Its Applications provides information pertinent to the theory and application of dynamic programming. This book presents the development and future directions for dynamic programming. Organized into four parts encompassing 23 chapters, this book begins with an overview of recurrence conditions for countable state Markov decision problems, which ensure that the optimal average reward exists and satisfies the functional equation of dynamic programming. This text then provides an extensive analysis of the theory of successive approximation for Markov decision problems. Other chapters consider the computational methods for deterministic, finite horizon problems, and present a unified and insightful presentation of several foundational questions. This book discusses as well the relationship between policy iteration and Newton's method. The final chapter deals with the main factors severely limiting the application of dynamic programming in practice. This book is a valuable resource for growth theorists, economists, biologists, mathematicians, and applied management scientists.


Approximate Dynamic Programming

Approximate Dynamic Programming

Author: Warren B. Powell

Publisher: John Wiley & Sons

Published: 2007-10-05

Total Pages: 487

ISBN-13: 0470182954

DOWNLOAD EBOOK

A complete and accessible introduction to the real-world applications of approximate dynamic programming With the growing levels of sophistication in modern-day operations, it is vital for practitioners to understand how to approach, model, and solve complex industrial problems. Approximate Dynamic Programming is a result of the author's decades of experience working in large industrial settings to develop practical and high-quality solutions to problems that involve making decisions in the presence of uncertainty. This groundbreaking book uniquely integrates four distinct disciplines—Markov design processes, mathematical programming, simulation, and statistics—to demonstrate how to successfully model and solve a wide range of real-life problems using the techniques of approximate dynamic programming (ADP). The reader is introduced to the three curses of dimensionality that impact complex problems and is also shown how the post-decision state variable allows for the use of classical algorithmic strategies from operations research to treat complex stochastic optimization problems. Designed as an introduction and assuming no prior training in dynamic programming of any form, Approximate Dynamic Programming contains dozens of algorithms that are intended to serve as a starting point in the design of practical solutions for real problems. The book provides detailed coverage of implementation challenges including: modeling complex sequential decision processes under uncertainty, identifying robust policies, designing and estimating value function approximations, choosing effective stepsize rules, and resolving convergence issues. With a focus on modeling and algorithms in conjunction with the language of mainstream operations research, artificial intelligence, and control theory, Approximate Dynamic Programming: Models complex, high-dimensional problems in a natural and practical way, which draws on years of industrial projects Introduces and emphasizes the power of estimating a value function around the post-decision state, allowing solution algorithms to be broken down into three fundamental steps: classical simulation, classical optimization, and classical statistics Presents a thorough discussion of recursive estimation, including fundamental theory and a number of issues that arise in the development of practical algorithms Offers a variety of methods for approximating dynamic programs that have appeared in previous literature, but that have never been presented in the coherent format of a book Motivated by examples from modern-day operations research, Approximate Dynamic Programming is an accessible introduction to dynamic modeling and is also a valuable guide for the development of high-quality solutions to problems that exist in operations research and engineering. The clear and precise presentation of the material makes this an appropriate text for advanced undergraduate and beginning graduate courses, while also serving as a reference for researchers and practitioners. A companion Web site is available for readers, which includes additional exercises, solutions to exercises, and data sets to reinforce the book's main concepts.


Dynamic Programming

Dynamic Programming

Author: Moshe Sniedovich

Publisher: CRC Press

Published: 1991-10-31

Total Pages: 438

ISBN-13: 9780824782450

DOWNLOAD EBOOK

Portrays dynamic programming as a methodology, identifying its constituent components, and explaining how it approaches problems and tackles them. Does not consider it as a practical tool, nor how it might address any actual situations in the real world. Assumes calculus, set theory, and some optimi


Neural Approximations for Optimal Control and Decision

Neural Approximations for Optimal Control and Decision

Author: Riccardo Zoppoli

Publisher: Springer Nature

Published: 2019-12-17

Total Pages: 532

ISBN-13: 3030296938

DOWNLOAD EBOOK

Neural Approximations for Optimal Control and Decision provides a comprehensive methodology for the approximate solution of functional optimization problems using neural networks and other nonlinear approximators where the use of traditional optimal control tools is prohibited by complicating factors like non-Gaussian noise, strong nonlinearities, large dimension of state and control vectors, etc. Features of the text include: • a general functional optimization framework; • thorough illustration of recent theoretical insights into the approximate solutions of complex functional optimization problems; • comparison of classical and neural-network based methods of approximate solution; • bounds to the errors of approximate solutions; • solution algorithms for optimal control and decision in deterministic or stochastic environments with perfect or imperfect state measurements over a finite or infinite time horizon and with one decision maker or several; • applications of current interest: routing in communications networks, traffic control, water resource management, etc.; and • numerous, numerically detailed examples. The authors’ diverse backgrounds in systems and control theory, approximation theory, machine learning, and operations research lend the book a range of expertise and subject matter appealing to academics and graduate students in any of those disciplines together with computer science and other areas of engineering.


Stochastic Digital Control System Techniques

Stochastic Digital Control System Techniques

Author:

Publisher: Academic Press

Published: 1996-05-16

Total Pages: 441

ISBN-13: 0080529925

DOWNLOAD EBOOK

Praise for the Series:"This book will be a useful reference to control engineers and researchers. The papers contained cover well the recent advances in the field of modern control theory."-IEEE Group Correspondence"This book will help all those researchers who valiantly try to keep abreast of what is new in the theory and practice of optimal control."--Control


Mathematical Aspects of Scheduling and Applications

Mathematical Aspects of Scheduling and Applications

Author: R. Bellman

Publisher: Elsevier

Published: 2014-05-20

Total Pages: 345

ISBN-13: 1483137449

DOWNLOAD EBOOK

Mathematical Aspects of Scheduling and Applications addresses the perennial problem of optimal utilization of finite resources in the accomplishment of an assortment of tasks or objectives. The book provides ways to uncover the core of these problems, presents them in mathematical terms, and devises mathematical solutions for them. The book consists of 12 chapters. Chapter 1 deals with network problems, the shortest path problem, and applications to control theory. Chapter 2 stresses the role and use of computers based on the decision-making problems outlined in the preceding chapter. Chapter 3 classifies scheduling problems and their solution approaches. Chapters 4 to 6 discuss machine sequencing problems and techniques. Chapter 5 tackles capacity expansion problems and introduces the technique of embedded state space dynamic programming for reducing dimensionality so that larger problems can be solved. Chapter 6 then examines an important class of network problems with non-serial phase structures and exploits dimensionality reduction techniques, such as the pseudo-stage concept, branch compression, and optimal order elimination methods to solve large-scale, nonlinear network scheduling problems. Chapters 7 to 11 consider the flow-shop scheduling problem under different objectives and constraints. Chapter 12 discusses the job-shop-scheduling problem. The book will be useful to economists, planners, and graduate students in the fields of mathematics, operations research, management science, computer science, and engineering.