The contributions by leading experts in this book focus on a variety of topics of current interest related to information-based complexity, ranging from function approximation, numerical integration, numerical methods for the sphere, and algorithms with random information, to Bayesian probabilistic numerical methods and numerical methods for stochastic differential equations.
The contributions by leading experts in this book focus on a variety of topics of current interest related to information-based complexity, ranging from function approximation, numerical integration, numerical methods for the sphere, and algorithms with random information, to Bayesian probabilistic numerical methods and numerical methods for stochastic differential equations.
This book contains five essays on the complexity of continuous problems, written for a wider audience. The first four essays are based on talks presented in 2008 when Henryk Wozniakowski received an honorary doctoral degree from the Friedrich Schiller University of Jena. The focus is on the introduction and history of the complexity of continuous problems, as well as on recent progress concerning the complexity of high-dimensional numerical problems. The last essay provides a brief and informal introduction to the basic notions and concepts of information-based complexity addressed to a general readership.
Parameterized complexity is currently a thriving field in complexity theory and algorithm design. A significant part of the success of the field can be attributed to Michael R. Fellows. This Festschrift has been published in honor of Mike Fellows on the occasion of his 60th birthday. It contains 20 papers that showcase the important scientific contributions of this remarkable man, describes the history of the field of parameterized complexity, and also reflects on other parts of Mike Fellows’s unique and broad range of interests, including his work on the popularization of discrete mathematics for young children. The volume contains several surveys that introduce the reader to the field of parameterized complexity and discuss important notions, results, and developments in this field.
New and classical results in computational complexity, including interactive proofs, PCP, derandomization, and quantum computation. Ideal for graduate students.
This book provides a comprehensive treatment of information-based complexity, the branch of computational complexity that deals with the intrinsic difficulty of the approximate solution of problems for which the information is partial, noisy, and priced. Such problems arise in many areas including economics, physics, human and robotic vision, scientific and engineering computation, geophysics, decision theory, signal processing and control theory.
Multivariate problems occur in many applications. These problems are defined on spaces of $d$-variate functions and $d$ can be huge--in the hundreds or even in the thousands. Some high-dimensional problems can be solved efficiently to within $\varepsilon$, i.e., the cost increases polynomially in $\varepsilon^{-1}$ and $d$. However, there are many multivariate problems for which even the minimal cost increases exponentially in $d$. This exponential dependence on $d$ is called intractability or the curse of dimensionality. This is the first volume of a three-volume set comprising a comprehensive study of the tractability of multivariate problems. It is devoted to tractability in the case of algorithms using linear information and develops the theory for multivariate problems in various settings: worst case, average case, randomized and probabilistic. A problem is tractable if its minimal cost is not exponential in $\varepsilon^{-1}$ and $d$. There are various notions of tractability, depending on how we measure the lack of exponential dependence. For example, a problem is polynomially tractable if its minimal cost is polynomial in $\varepsilon^{-1}$ and $d$. The study of tractability was initiated about 15 years ago. This is the first and only research monograph on this subject. Many multivariate problems suffer from the curse of dimensionality when they are defined over classical (unweighted) spaces. In this case, all variables and groups of variables play the same role, which causes the minimal cost to be exponential in $d$. But many practically important problems are solved today for huge $d$ in a reasonable time. One of the most intriguing challenges of the theory is to understand why this is possible. Multivariate problems may become weakly tractable, polynomially tractable or even strongly polynomially tractable if they are defined over weighted spaces with properly decaying weights. One of the main purposes of this book is to study weighted spaces and obtain necessary and sufficient conditions on weights for various notions of tractability. The book is of interest for researchers working in computational mathematics, especially in approximation of high-dimensional problems. It may be also suitable for graduate courses and seminars. The text concludes with a list of thirty open problems that can be good candidates for future tractability research.
This comprehensive and self-contained textbook presents an accessible overview of the state of the art of multivariate algorithmics and complexity. Increasingly, multivariate algorithmics is having significant practical impact in many application domains, with even more developments on the horizon. The text describes how the multivariate framework allows an extended dialog with a problem, enabling the reader who masters the complexity issues under discussion to use the positive and negative toolkits in their own research. Features: describes many of the standard algorithmic techniques available for establishing parametric tractability; reviews the classical hardness classes; explores the various limitations and relaxations of the methods; showcases the powerful new lower bound techniques; examines various different algorithmic solutions to the same problems, highlighting the insights to be gained from each approach; demonstrates how complexity methods and ideas have evolved over the past 25 years.
The purpose of this monograph is to create a general framework for the study of optimal algorithms for problems that are solved approximately. For generality the setting is abstract, but we present many applications to practical problems and provide examples to illustrate concepts and major theorems. The work presented here is motivated by research in many fields. Influential have been questions, concepts, and results from complexity theory, algorithmic analysis, applied mathematics and numerical analysis, the mathematical theory of approximation (particularly the work on n-widths in the sense of Gelfand and Kolmogorov), applied approximation theory (particularly the theory of splines), as well as earlier work on optimal algorithms. But many of the questions we ask (see Overview) are new. We present a different view of algorithms and complexity and must request the reader's