This book presents a thorough development of the modern theory of stochastic approximation or recursive stochastic algorithms for both constrained and unconstrained problems. This second edition is a thorough revision, although the main features and structure remain unchanged. It contains many additional applications and results as well as more detailed discussion.
The book deals with a powerful and convenient approach to a great variety of types of problems of the recursive monte-carlo or stochastic approximation type. Such recu- sive algorithms occur frequently in stochastic and adaptive control and optimization theory and in statistical esti- tion theory. Typically, a sequence {X } of estimates of a n parameter is obtained by means of some recursive statistical th st procedure. The n estimate is some function of the n_l estimate and of some new observational data, and the aim is to study the convergence, rate of convergence, and the pa- metric dependence and other qualitative properties of the - gorithms. In this sense, the theory is a statistical version of recursive numerical analysis. The approach taken involves the use of relatively simple compactness methods. Most standard results for Kiefer-Wolfowitz and Robbins-Monro like methods are extended considerably. Constrained and unconstrained problems are treated, as is the rate of convergence problem. While the basic method is rather simple, it can be elaborated to allow a broad and deep coverage of stochastic approximation like problems. The approach, relating algorithm behavior to qualitative properties of deterministic or stochastic differ ential equations, has advantages in algorithm conceptualiza tion and design. It is often possible to obtain an intuitive understanding of algorithm behavior or qualitative dependence upon parameters, etc., without getting involved in a great deal of deta~l.
Estimating unknown parameters based on observation data conta- ing information about the parameters is ubiquitous in diverse areas of both theory and application. For example, in system identification the unknown system coefficients are estimated on the basis of input-output data of the control system; in adaptive control systems the adaptive control gain should be defined based on observation data in such a way that the gain asymptotically tends to the optimal one; in blind ch- nel identification the channel coefficients are estimated using the output data obtained at the receiver; in signal processing the optimal weighting matrix is estimated on the basis of observations; in pattern classifi- tion the parameters specifying the partition hyperplane are searched by learning, and more examples may be added to this list. All these parameter estimation problems can be transformed to a root-seeking problem for an unknown function. To see this, let - note the observation at time i. e. , the information available about the unknown parameters at time It can be assumed that the parameter under estimation denoted by is a root of some unknown function This is not a restriction, because, for example, may serve as such a function.
This book is devoted to sequential methods of solving a class of problems to which belongs, for example, the problem of finding a maximum point of a function if each measured value of this function contains a random error. Some basic procedures of stochastic approximation are investigated from a single point of view, namely the theory of Markov processes and martingales. Examples are considered of applications of the theorems to some problems of estimation theory, educational theory and control theory, and also to some problems of information transmission in the presence of inverse feedback.
Adaptive systems are widely encountered in many applications ranging through adaptive filtering and more generally adaptive signal processing, systems identification and adaptive control, to pattern recognition and machine intelligence: adaptation is now recognised as keystone of "intelligence" within computerised systems. These diverse areas echo the classes of models which conveniently describe each corresponding system. Thus although there can hardly be a "general theory of adaptive systems" encompassing both the modelling task and the design of the adaptation procedure, nevertheless, these diverse issues have a major common component: namely the use of adaptive algorithms, also known as stochastic approximations in the mathematical statistics literature, that is to say the adaptation procedure (once all modelling problems have been resolved). The juxtaposition of these two expressions in the title reflects the ambition of the authors to produce a reference work, both for engineers who use these adaptive algorithms and for probabilists or statisticians who would like to study stochastic approximations in terms of problems arising from real applications. Hence the book is organised in two parts, the first one user-oriented, and the second providing the mathematical foundations to support the practice described in the first part. The book covers the topcis of convergence, convergence rate, permanent adaptation and tracking, change detection, and is illustrated by various realistic applications originating from these areas of applications.
* Unique in its survey of the range of topics. * Contains a strong, interdisciplinary format that will appeal to both students and researchers. * Features exercises and web links to software and data sets.
In this book we analyze the error caused by numerical schemes for the approximation of semilinear stochastic evolution equations (SEEq) in a Hilbert space-valued setting. The numerical schemes considered combine Galerkin finite element methods with Euler-type temporal approximations. Starting from a precise analysis of the spatio-temporal regularity of the mild solution to the SEEq, we derive and prove optimal error estimates of the strong error of convergence in the first part of the book. The second part deals with a new approach to the so-called weak error of convergence, which measures the distance between the law of the numerical solution and the law of the exact solution. This approach is based on Bismut’s integration by parts formula and the Malliavin calculus for infinite dimensional stochastic processes. These techniques are developed and explained in a separate chapter, before the weak convergence is proven for linear SEEq.