Introduction to Empirical Processes and Semiparametric Inference

Introduction to Empirical Processes and Semiparametric Inference

Author: Michael R. Kosorok

Publisher: Springer Science & Business Media

Published: 2007-12-29

Total Pages: 482

ISBN-13: 0387749780

DOWNLOAD EBOOK

Kosorok’s brilliant text provides a self-contained introduction to empirical processes and semiparametric inference. These powerful research techniques are surprisingly useful for developing methods of statistical inference for complex models and in understanding the properties of such methods. This is an authoritative text that covers all the bases, and also a friendly and gradual introduction to the area. The book can be used as research reference and textbook.


Benchmark Estimation for Markov Chain Monte Carlo Samplers

Benchmark Estimation for Markov Chain Monte Carlo Samplers

Author: Subharup Guha

Publisher:

Published: 2004

Total Pages:

ISBN-13:

DOWNLOAD EBOOK

Abstract: While studying various features of the posterior distribution of a vector-valued parameter using an MCMC sample, systematically subsampling of the MCMC output can only lead to poorer estimation. Nevertheless, a 1-in-k subsample is often all that is retained in investigations where intensive computations are involved or where speed is essential. The goal of benchmark estimation is to produce a number of estimates based on the best available information, i.e. the entire MCMC sample, and to use these to improve other estimates made on the basis of the subsample. We take a simple approach by creating a weighted subsample where the weights are quickly obtained as a solution to a system of linear equations. We provide a theoretical basis for the method and illustrate the technique using examples from the literature. For a subsampling rate of 1-in-10, the observed reductions in MSE often exceed 50% for a number of posterior features. Much larger gains are expected for certain complex estimation methods and for the commonly used thinner subsampling rates. Benchmark estimation can be used wherever other fast or efficient estimation strategies like importance sampling already exist. We show how the two strategies can be used in conjunction with each other. We discuss some asymptotic properties of benchmark estimators that provide insight into the gains associated with the technique. The observed gains are found to closely match the theoretical values predicted by the asymptotic, even for k as small as 10.


Statistical Advances in the Biomedical Sciences

Statistical Advances in the Biomedical Sciences

Author: Atanu Biswas

Publisher: John Wiley & Sons

Published: 2007-12-14

Total Pages: 623

ISBN-13: 0470181192

DOWNLOAD EBOOK

The Most Comprehensive and Cutting-Edge Guide to Statistical Applications in Biomedical Research With the increasing use of biotechnology in medical research and the sophisticated advances in computing, it has become essential for practitioners in the biomedical sciences to be fully educated on the role statistics plays in ensuring the accurate analysis of research findings. Statistical Advances in the Biomedical Sciences explores the growing value of statistical knowledge in the management and comprehension of medical research and, more specifically, provides an accessible introduction to the contemporary methodologies used to understand complex problems in the four major areas of modern-day biomedical science: clinical trials, epidemiology, survival analysis, and bioinformatics. Composed of contributions from eminent researchers in the field, this volume discusses the application of statistical techniques to various aspects of modern medical research and illustrates how these methods ultimately prove to be an indispensable part of proper data collection and analysis. A structural uniformity is maintained across all chapters, each beginning with an introduction that discusses general concepts and the biomedical problem under focus and is followed by specific details on the associated methods, algorithms, and applications. In addition, each chapter provides a summary of the main ideas and offers a concluding remarks section that presents novel ideas, approaches, and challenges for future research. Complete with detailed references and insight on the future directions of biomedical research, Statistical Advances in the Biomedical Sciences provides vital statistical guidance to practitioners in the biomedical sciences while also introducing statisticians to new, multidisciplinary frontiers of application. This text is an excellent reference for graduate- and PhD-level courses in various areas of biostatistics and the medical sciences and also serves as a valuable tool for medical researchers, statisticians, public health professionals, and biostatisticians.


Evaluating the Efficiency of Markov Chain Monte Carlo Algorithms

Evaluating the Efficiency of Markov Chain Monte Carlo Algorithms

Author: Thuy Scanlon

Publisher:

Published: 2021

Total Pages: 76

ISBN-13:

DOWNLOAD EBOOK

Markov chain Monte Carlo (MCMC) is a simulation technique that produces a Markov chain designed to converge to a stationary distribution. In Bayesian statistics, MCMC is used to obtain samples from a posterior distribution for inference. To ensure the accuracy of estimates using MCMC samples, the convergence to the stationary distribution of an MCMC algorithm has to be checked. As computation time is a resource, optimizing the efficiency of an MCMC algorithm in terms of effective sample size (ESS) per time unit is an important goal for statisticians. In this paper, we use simulation studies to demonstrate how the Gibbs sampler and the Metropolis-Hasting algorithm works and how MCMC diagnostic tests are used to check for MCMC convergence. We investigated and compared the efficiency of different MCMC algorithms fit to a linear and a spatial model. Our results showed that the Gibbs sampler and the Metropolis-Hasting algorithm give estimates similar to the maximum likelihood estimates, validating the accuracy of MCMC. The results also imply that the efficiency of an MCMC algorithm can be affected by different factors. In particular, a model with more parameters could still be more efficient in terms of ESS per time unit. For fitting large datasets, algorithms whose computation involves dividing a large matrix into smaller matrices can be more efficient than algorithms that use the entire large matrix.


Hamiltonian Markov Chain Monte Carlo Schemes for Rare Event Estimation

Hamiltonian Markov Chain Monte Carlo Schemes for Rare Event Estimation

Author: Hamed Nikbakht

Publisher:

Published: 2020

Total Pages:

ISBN-13:

DOWNLOAD EBOOK

Estimating rare event probabilities is a commonly encountered important problem in several engineering and scientific applications, most often observed in the form of probability of failure (PF) estimation or, alternatively and better sounding for the public, reliability estimation. In many practical applications, such as for structures, airplanes, mechanical equipment, and many more, failure probabilities are fortunately very low, from 10-4 to even 10-9 and less. Such estimations are of utmost importance for design choices, emergency preparedness, safety regulations, maintenance suggestions and more. Calculating such small numbers with accuracy however presents many numerical and mathematical challenges. To make matters worse, these estimations in realistic applications are usually based on high dimensional random spaces with numerous random variables and processes involved. A single simulation of such a model, or else a single model call, may also require several minutes to hours of computing time. As such, reducing the number of model calls is of great importance in these problems and one of the critical parameters that limits or prohibits use of several available techniques in the literature. This research is motivated by efficiently and precisely quantifying these probabilities, often encountered in reliability analysis of complex engineering systems, based on a developed framework termed Approximate Sampling Target with Postprocessing Adjustment (ASTPA), which herein is integrated with and supported by gradient-based Hamiltonian Markov Chain Monte Carlo (HMCMC) methods. Hamiltonian Markov Chain Monte Carlo sampling is characterized by much better scalability, faster mixing rates, is capable of generating samples with much weaker auto-correlation, even in complex high-dimensional parameter spaces, and has enjoyed broad-spectrum successes in most general settings. HMCMC adopts physical system dynamics, rather than a proposal probability distribution, and can be used to produce distant proposal samples for the integrated Metropolis step, thereby avoiding the slow exploration of the state space that results from the diffusive behavior of simple random-walk proposals. In this work, we aim to advance knowledge on Hamiltonian Markov Chain Monte Carlo methods, in general, with particular emphasis on its efficient utilization for rare event probability estimation in both Gaussian and Non-Gaussian spaces. This research also seeks to offer significant advancements in probabilistic inference and reliability predictions. Thus, in this context, we develop various Quasi-Newton based HMCMC schemes, which can sample very adeptly, particularly in difficult cases of high curvature, high-dimensionality and very small failure probabilities. The methodology is formally introduced, and the key theoretical aspects, and the underlying assumptions are discussed. Performance of the proposed methodology is then compared against state-of-the-art Subset Simulation in a series of challenging static and dynamic (time-dependent reliability) low- and high-dimensional benchmark problems. In the last phase of this work, with an aim to avoid using analytical gradients, within the proposed HMCMC-based framework, we investigate application of the Automatic Differentiation (AD) technique. In addition, to avoid use of gradients altogether and to improve the performance of the original SuS algorithm, we study the application of Quasi-Newton based HMCMC within the Subset Simulation framework. Various numerical examples are then presented to showcase the performance of the aforementioned approaches.


Methods in Markov Chain Monte Carlo and Spatiotemporal Data Analysis

Methods in Markov Chain Monte Carlo and Spatiotemporal Data Analysis

Author: Jianan Hui

Publisher:

Published: 2017

Total Pages: 138

ISBN-13: 9780355471779

DOWNLOAD EBOOK

Markov chain Monte Carlo (MCMC) methods are highly desirable when the sampling distribution is intractable. Among all MCMC methods, the fundamental one is the Metropolis-Hastings algorithm. Despite its extensive application in approximating any distribution, the Markov chain often suffers from slow mixing, which then causes insufficient estimation. We address this issue by proposing modifications to the Metropolis-Hastings algorithm that, under specified conditions, induces substantial improvements in jump distances and statistical efficiency while preserving the overall quality of convergence. This dissertation starts with an introduction of the MCMC methods and continues by proposing the Efficient Conditional Metropolis-Hastings algorithm (ECMH) and a variation of ECMH under a uniform setting (ECMHu). We further investigate their properties through a series of models, including a Bivariate normal model, a Bayesian random effects model, and a Bayesian dynamic spatiotemporal model. Simulation results are compared across all algorithms.


Modern Monte Carlo Methods and Their Application in Semiparametric Regression

Modern Monte Carlo Methods and Their Application in Semiparametric Regression

Author: Samuel Joseph Thomas

Publisher:

Published: 2021

Total Pages: 208

ISBN-13:

DOWNLOAD EBOOK

The essence of Bayesian data analysis is to ascertain posterior distributions. Posteriors generally do not have closed-form expressions for direct computation in practical applications. Analysts, therefore, resort to Markov Chain Monte Carlo (MCMC) methods for the generation of sample observations that approximate the desired posterior distribution. Standard MCMC methods simulate sample values from the desired posterior distribution via random proposals. As a result, the mechanism used to generate the proposals inevitably determines the efficiency of the algorithm. One of the modern MCMC techniques designed to explore the high-dimensional space more efficiently is Hamiltonian Monte Carlo (HMC), based on the Hamiltonian differential equations. Inspired by classical mechanics, these equations incorporate a latent variable to generate MCMC proposals that are likely to be accepted. This dissertation discusses how such a powerful computational approach can be used for implementing statistical models. Along this line, I created a unified computational procedure for using HMC to fit various types of statistical models. The procedure that I proposed can be applied to a broad class of models, including linear models, generalized linear models, mixed-effects models, and various types of semiparametric regression models. To facilitate the fitting of a diverse set of models, I incorporated new parameterization and decomposition schemes to ensure the numerical performance of Bayesian model fitting without sacrificing the procedure's general applicability. As a concrete application, I demonstrate how to use the proposed procedure to fit a multivariate generalized additive model (GAM), a nonstandard statistical model with a complex covariance structure and numerous parameters. Byproducts of the research include two software packages that all practical data analysts to use the proposed computational method to fit their own models. The research's main methodological contribution is the unified computational approach that it presents for Bayesian model fitting that can be used for standard and nonstandard statistical models. Availability of such a procedure has greatly enhanced statistical modelers' toolbox for implementing new and nonstandard statistical models.