Optimal Solution of Nonlinear Equations is a text/monograph designed to provide an overview of optimal computational methods for the solution of nonlinear equations, fixed points of contractive and noncontractive mapping, and for the computation of the topological degree. It is of interest to any reader working in the area of Information-Based Complexity. The worst-case settings are analyzed here. Several classes of functions are studied with special emphasis on tight complexity bounds and methods which are close to or achieve these bounds. Each chapter ends with exercises, including companies and open-ended research based exercises.
This book is the first on the topic and explains the most cutting-edge methods needed for precise calculations and explores the development of powerful algorithms to solve research problems. Multipoint methods have an extensive range of practical applications significant in research areas such as signal processing, analysis of convergence rate, fluid mechanics, solid state physics, and many others. The book takes an introductory approach in making qualitative comparisons of different multipoint methods from various viewpoints to help the reader understand applications of more complex methods. Evaluations are made to determine and predict efficiency and accuracy of presented models useful to wide a range of research areas along with many numerical examples for a deep understanding of the usefulness of each method. This book will make it possible for the researchers to tackle difficult problems and deepen their understanding of problem solving using numerical methods. Multipoint methods are of great practical importance, as they determine sequences of successive approximations for evaluative purposes. This is especially helpful in achieving the highest computational efficiency. The rapid development of digital computers and advanced computer arithmetic have provided a need for new methods useful to solving practical problems in a multitude of disciplines such as applied mathematics, computer science, engineering, physics, financial mathematics, and biology. - Provides a succinct way of implementing a wide range of useful and important numerical algorithms for solving research problems - Illustrates how numerical methods can be used to study problems which have applications in engineering and sciences, including signal processing, and control theory, and financial computation - Facilitates a deeper insight into the development of methods, numerical analysis of convergence rate, and very detailed analysis of computational efficiency - Provides a powerful means of learning by systematic experimentation with some of the many fascinating problems in science - Includes highly efficient algorithms convenient for the implementation into the most common computer algebra systems such as Mathematica, MatLab, and Maple
Solving nonlinear equations in Banach spaces (real or complex nonlinear equations, nonlinear systems, and nonlinear matrix equations, among others), is a non-trivial task that involves many areas of science and technology. Usually the solution is not directly affordable and require an approach using iterative algorithms. This Special Issue focuses mainly on the design, analysis of convergence, and stability of new schemes for solving nonlinear problems and their application to practical problems. Included papers study the following topics: Methods for finding simple or multiple roots either with or without derivatives, iterative methods for approximating different generalized inverses, real or complex dynamics associated to the rational functions resulting from the application of an iterative method on a polynomial. Additionally, the analysis of the convergence has been carried out by means of different sufficient conditions assuring the local, semilocal, or global convergence. This Special issue has allowed us to present the latest research results in the area of iterative processes for solving nonlinear equations as well as systems and matrix equations. In addition to the theoretical papers, several manuscripts on signal processing, nonlinear integral equations, or partial differential equations, reveal the connection between iterative methods and other branches of science and engineering.
Linear and nonlinear systems of equations are the basis for many, if not most, of the models of phenomena in science and engineering, and their efficient numerical solution is critical to progress in these areas. This is the first book to be published on nonlinear equations since the mid-1980s. Although it stresses recent developments in this area, such as Newton-Krylov methods, considerable material on linear equations has been incorporated. This book focuses on a small number of methods and treats them in depth. The author provides a complete analysis of the conjugate gradient and generalized minimum residual iterations as well as recent advances including Newton-Krylov methods, incorporation of inexactness and noise into the analysis, new proofs and implementations of Broyden's method, and globalization of inexact Newton methods. Examples, methods, and algorithmic choices are based on applications to infinite dimensional problems such as partial differential equations and integral equations. The analysis and proof techniques are constructed with the infinite dimensional setting in mind and the computational examples and exercises are based on the MATLAB environment.
This textbook on Linear and Nonlinear Optimization is intended for graduate and advanced undergraduate students in operations research and related fields. It is both literate and mathematically strong, yet requires no prior course in optimization. As suggested by its title, the book is divided into two parts covering in their individual chapters LP Models and Applications; Linear Equations and Inequalities; The Simplex Algorithm; Simplex Algorithm Continued; Duality and the Dual Simplex Algorithm; Postoptimality Analyses; Computational Considerations; Nonlinear (NLP) Models and Applications; Unconstrained Optimization; Descent Methods; Optimality Conditions; Problems with Linear Constraints; Problems with Nonlinear Constraints; Interior-Point Methods; and an Appendix covering Mathematical Concepts. Each chapter ends with a set of exercises. The book is based on lecture notes the authors have used in numerous optimization courses the authors have taught at Stanford University. It emphasizes modeling and numerical algorithms for optimization with continuous (not integer) variables. The discussion presents the underlying theory without always focusing on formal mathematical proofs (which can be found in cited references). Another feature of this book is its inclusion of cultural and historical matters, most often appearing among the footnotes. "This book is a real gem. The authors do a masterful job of rigorously presenting all of the relevant theory clearly and concisely while managing to avoid unnecessary tedious mathematical details. This is an ideal book for teaching a one or two semester masters-level course in optimization – it broadly covers linear and nonlinear programming effectively balancing modeling, algorithmic theory, computation, implementation, illuminating historical facts, and numerous interesting examples and exercises. Due to the clarity of the exposition, this book also serves as a valuable reference for self-study." Professor Ilan Adler, IEOR Department, UC Berkeley "A carefully crafted introduction to the main elements and applications of mathematical optimization. This volume presents the essential concepts of linear and nonlinear programming in an accessible format filled with anecdotes, examples, and exercises that bring the topic to life. The authors plumb their decades of experience in optimization to provide an enriching layer of historical context. Suitable for advanced undergraduates and masters students in management science, operations research, and related fields." Michael P. Friedlander, IBM Professor of Computer Science, Professor of Mathematics, University of British Columbia
From the Preface (1964): ``This book presents a general theory of iteration algorithms for the numerical solution of equations and systems of equations. The relationship between the quantity and the quality of information used by an algorithm and the efficiency of the algorithm is investigated. Iteration functions are divided into four classes depending on whether they use new information at one or at several points and whether or not they reuse old information. Known iteration functions are systematized and new classes of computationally effective iteration functions are introduced. Our interest in the efficient use of information is influenced by the widespread use of computing machines ... The mathematical foundations of our subject are treated with rigor, but rigor in itself is not the main object. Some of the material is of wider application ... Most of the material is new and unpublished. Every attempt has been made to keep the subject in proper historical perspective ... ''
Optimization is an important tool used in decision science and for the analysis of physical systems used in engineering. One can trace its roots to the Calculus of Variations and the work of Euler and Lagrange. This natural and reasonable approach to mathematical programming covers numerical methods for finite-dimensional optimization problems. It begins with very simple ideas progressing through more complicated concepts, concentrating on methods for both unconstrained and constrained optimization.
Here is a collection of nonlinear optimization applications from the real world, expressed in the General Algebraic Modeling System (GAMS). The concepts are presented so that the reader can quickly modify and update them to represent real-world situations.
Many engineering, operations, and scientific applications include a mixture of discrete and continuous decision variables and nonlinear relationships involving the decision variables that have a pronounced effect on the set of feasible and optimal solutions. Mixed-integer nonlinear programming (MINLP) problems combine the numerical difficulties of handling nonlinear functions with the challenge of optimizing in the context of nonconvex functions and discrete variables. MINLP is one of the most flexible modeling paradigms available for optimization; but because its scope is so broad, in the most general cases it is hopelessly intractable. Nonetheless, an expanding body of researchers and practitioners — including chemical engineers, operations researchers, industrial engineers, mechanical engineers, economists, statisticians, computer scientists, operations managers, and mathematical programmers — are interested in solving large-scale MINLP instances.