This book focuses on Augmented Lagrangian techniques for solving practical constrained optimization problems. The authors: rigorously delineate mathematical convergence theory based on sequential optimality conditions and novel constraint qualifications; orient the book to practitioners by giving priority to results that provide insight on the practical behavior of algorithms and by providing geometrical and algorithmic interpretations of every mathematical result; and fully describe a freely available computational package for constrained optimization and illustrate its usefulness with applications.
This book focuses on Augmented Lagrangian techniques for solving practical constrained optimization problems. The authors rigorously delineate mathematical convergence theory based on sequential optimality conditions and novel constraint qualifications. They also orient the book to practitioners by giving priority to results that provide insight on the practical behavior of algorithms and by providing geometrical and algorithmic interpretations of every mathematical result, and they fully describe a freely available computational package for constrained optimization and illustrate its usefulness with applications.
Computer Science and Applied Mathematics: Constrained Optimization and Lagrange Multiplier Methods focuses on the advancements in the applications of the Lagrange multiplier methods for constrained minimization. The publication first offers information on the method of multipliers for equality constrained problems and the method of multipliers for inequality constrained and nondifferentiable optimization problems. Discussions focus on approximation procedures for nondifferentiable and ill-conditioned optimization problems; asymptotically exact minimization in the methods of multipliers; duality framework for the method of multipliers; and the quadratic penalty function method. The text then examines exact penalty methods, including nondifferentiable exact penalty functions; linearization algorithms based on nondifferentiable exact penalty functions; differentiable exact penalty functions; and local and global convergence of Lagrangian methods. The book ponders on the nonquadratic penalty functions of convex programming. Topics include large scale separable integer programming problems and the exponential method of multipliers; classes of penalty functions and corresponding methods of multipliers; and convergence analysis of multiplier methods. The text is a valuable reference for mathematicians and researchers interested in the Lagrange multiplier methods.
In the intervening years since this book was published in 1981, the field of optimization has been exceptionally lively. This fertility has involved not only progress in theory, but also faster numerical algorithms and extensions into unexpected or previously unknown areas such as semidefinite programming. Despite these changes, many of the important principles and much of the intuition can be found in this Classics version of Practical Optimization. This book provides model algorithms and pseudocode, useful tools for users who prefer to write their own code as well as for those who want to understand externally provided code. It presents algorithms in a step-by-step format, revealing the overall structure of the underlying procedures and thereby allowing a high-level perspective on the fundamental differences. And it contains a wealth of techniques and strategies that are well suited for optimization in the twenty-first century, and particularly in the now-flourishing fields of data science, big data, and machine learning. Practical Optimization is appropriate for advanced undergraduates, graduate students, and researchers interested in methods for solving optimization problems.
Many of the most challenging problems in the applied sciences involve non-differentiable structures as well as partial differential operators, thus leading to non-smooth distributed parameter systems. This edited volume aims to establish a theoretical and numerical foundation and develop new algorithmic paradigms for the treatment of non-smooth phenomena and associated parameter influences. Other goals include the realization and further advancement of these concepts in the context of robust and hierarchical optimization, partial differential games, and nonlinear partial differential complementarity problems, as well as their validation in the context of complex applications. Areas for which applications are considered include optimal control of multiphase fluids and of superconductors, image processing, thermoforming, and the formation of rivers and networks. Chapters are written by leading researchers and present results obtained in the first funding phase of the DFG Special Priority Program on Nonsmooth and Complementarity Based Distributed Parameter Systems: Simulation and Hierarchical Optimization that ran from 2016 to 2019.
In the late forties, Mathematical Programming became a scientific discipline in its own right. Since then it has experienced a tremendous growth. Beginning with economic and military applications, it is now among the most important fields of applied mathematics with extensive use in engineering, natural sciences, economics, and biological sciences. The lively activity in this area is demonstrated by the fact that as early as 1949 the first "Symposium on Mathe matical Programming" took place in Chicago. Since then mathematical programmers from all over the world have gath ered at the intfrnational symposia of the Mathematical Programming Society roughly every three years to present their recent research, to exchange ideas with their colleagues and to learn about the latest developments in their own and related fields. In 1982, the XI. International Symposium on Mathematical Programming was held at the University of Bonn, W. Germany, from August 23 to 27. It was organized by the Institut fUr Okonometrie und Operations Re search of the University of Bonn in collaboration with the Sonderforschungs bereich 21 of the Deutsche Forschungsgemeinschaft. This volume constitutes part of the outgrowth of this symposium and docu ments its scientific activities. Part I of the book contains information about the symposium, welcoming addresses, lists of committees and sponsors and a brief review about the Ful kerson Prize and the Dantzig Prize which were awarded during the opening ceremony.
Surveys the theory and history of the alternating direction method of multipliers, and discusses its applications to a wide variety of statistical and machine learning problems of recent interest, including the lasso, sparse logistic regression, basis pursuit, covariance selection, support vector machines, and many others.