The Generalized Method of Moments for Mixture and Mixed Models

The Generalized Method of Moments for Mixture and Mixed Models

Author: Zhiyue Huang

Publisher:

Published: 2015

Total Pages:

ISBN-13:

DOWNLOAD EBOOK

Mixture models can be found in a wide variety of statistical applications. However, undertaking statistical inference in mixture models, especially non-parametric mixture models, can be challenging. A general, or nonparametric, mixture model has effectively an infinite dimensional parameter space. In frequentist statistics, the maximum likelihood estimator with an infinite dimensional parameter may not be consistent or efficient in the sense that the Cramer-Rao bound is not attained even asymptotically. In Bayesian statistics, a prior on an infinite dimensional space is not well defined and can be highly informative even with large amounts of data. In this thesis, we mainly consider mixture and mixed-effects models, when the mixing distribution is non-parametric. Following the dimensionality reduction idea in [Marriott, 2002], we propose a reparameterization-approximation framework with a complete orthonormal basis in a Hilbert space. The parameters in the reparameterized models are interpreted as the generalized moments of a mixing distribution. We consider different orthonormal bases, including the families of orthogonal polynomials and the eigenfunctions of positive self-adjoint integral operators. We also study the approximation errors of the truncation approximations of the reparameterized models in some special cases. The generalized moments in the truncated approximations of the reparameterized models have a natural parameter space, called the generalized moment space. We study the geometric properties of the generalized moment space and obtain two important geometric properties: the positive representation and the gradient characterization. The positive representation reveals the identifiability of the mixing distribution by its generalized moments and provides an upper bound of the number of the support points of the mixing distribution. On the other hand, the gradient characterization provides the foundation of the class of gradient-based algorithms when the feasible set is the generalized moment space. Next, we aim to fit a non-parametric mixture model by a set of generalized moment conditions, which are from the proposed reparameterization-approximation procedure. We propose a new estimation method, called the generalized method of moments for mixture models. The proposed estimation method involves minimizing a quadratic objective function over the generalized moment space. The proposed estimators can be easily computed through the gradient-based algorithms. We show the convergence rate of the mean squared error of the proposed estimators, as the sample size goes to infinity. Moreover, we design the quadratic objective function to ensure that the proposed estimators are robust to the outliers. Compared to the other existing estimation methods for mixture models, the GMM for mixture models is more computationally friendly and robust to outliers. Lastly, we consider the hypothesis testing problem on the regression parameter in a mixed-effects model with univariate random effects. Through our new procedures, we obtain a series of estimating equations parameterized in the regression parameter and the generalized moments of the random-effects distribution. These parameters are estimated under the framework of the generalized method of moments. In the case that the number of the generalized moments diverges with the sample size and the dimension of the regression parameter is fixed, we compute the convergence rate of the generalized method of moments estimators for the mixed-effects models with univariate random effects. Since the regularity conditions in [Wilks, 1938] fail under our context, it is challenging to construct an asymptotically $\chi^2$ test statistic. We propose using ensemble inference, in which an asymptotically $\chi^2$ test statistic is constructed from a series of the estimators obtained from the generalized estimating equations with different working correlation matrices.


Mixed Effects Models for Complex Data

Mixed Effects Models for Complex Data

Author: Lang Wu

Publisher: CRC Press

Published: 2009-11-11

Total Pages: 431

ISBN-13: 9781420074086

DOWNLOAD EBOOK

Although standard mixed effects models are useful in a range of studies, other approaches must often be used in correlation with them when studying complex or incomplete data. Mixed Effects Models for Complex Data discusses commonly used mixed effects models and presents appropriate approaches to address dropouts, missing data, measurement errors, censoring, and outliers. For each class of mixed effects model, the author reviews the corresponding class of regression model for cross-sectional data. An overview of general models and methods, along with motivating examples After presenting real data examples and outlining general approaches to the analysis of longitudinal/clustered data and incomplete data, the book introduces linear mixed effects (LME) models, generalized linear mixed models (GLMMs), nonlinear mixed effects (NLME) models, and semiparametric and nonparametric mixed effects models. It also includes general approaches for the analysis of complex data with missing values, measurement errors, censoring, and outliers. Self-contained coverage of specific topics Subsequent chapters delve more deeply into missing data problems, covariate measurement errors, and censored responses in mixed effects models. Focusing on incomplete data, the book also covers survival and frailty models, joint models of survival and longitudinal data, robust methods for mixed effects models, marginal generalized estimating equation (GEE) models for longitudinal or clustered data, and Bayesian methods for mixed effects models. Background material In the appendix, the author provides background information, such as likelihood theory, the Gibbs sampler, rejection and importance sampling methods, numerical integration methods, optimization methods, bootstrap, and matrix algebra. Failure to properly address missing data, measurement errors, and other issues in statistical analyses can lead to severely biased or misleading results. This book explores the biases that arise when naïve methods are used and shows which approaches should be used to achieve accurate results in longitudinal data analysis.


Finite Mixture Models

Finite Mixture Models

Author: Geoffrey McLachlan

Publisher: John Wiley & Sons

Published: 2004-03-22

Total Pages: 419

ISBN-13: 047165406X

DOWNLOAD EBOOK

An up-to-date, comprehensive account of major issues in finitemixture modeling This volume provides an up-to-date account of the theory andapplications of modeling via finite mixture distributions. With anemphasis on the applications of mixture models in both mainstreamanalysis and other areas such as unsupervised pattern recognition,speech recognition, and medical imaging, the book describes theformulations of the finite mixture approach, details itsmethodology, discusses aspects of its implementation, andillustrates its application in many common statisticalcontexts. Major issues discussed in this book include identifiabilityproblems, actual fitting of finite mixtures through use of the EMalgorithm, properties of the maximum likelihood estimators soobtained, assessment of the number of components to be used in themixture, and the applicability of asymptotic theory in providing abasis for the solutions to some of these problems. The author alsoconsiders how the EM algorithm can be scaled to handle the fittingof mixture models to very large databases, as in data miningapplications. This comprehensive, practical guide: * Provides more than 800 references-40% published since 1995 * Includes an appendix listing available mixture software * Links statistical literature with machine learning and patternrecognition literature * Contains more than 100 helpful graphs, charts, and tables Finite Mixture Models is an important resource for both applied andtheoretical statisticians as well as for researchers in the manyareas in which finite mixture models can be used to analyze data.


Statistical Inference as Severe Testing

Statistical Inference as Severe Testing

Author: Deborah G. Mayo

Publisher: Cambridge University Press

Published: 2018-09-20

Total Pages: 503

ISBN-13: 1108563309

DOWNLOAD EBOOK

Mounting failures of replication in social and biological sciences give a new urgency to critically appraising proposed reforms. This book pulls back the cover on disagreements between experts charged with restoring integrity to science. It denies two pervasive views of the role of probability in inference: to assign degrees of belief, and to control error rates in a long run. If statistical consumers are unaware of assumptions behind rival evidence reforms, they can't scrutinize the consequences that affect them (in personalized medicine, psychology, etc.). The book sets sail with a simple tool: if little has been done to rule out flaws in inferring a claim, then it has not passed a severe test. Many methods advocated by data experts do not stand up to severe scrutiny and are in tension with successful strategies for blocking or accounting for cherry picking and selective reporting. Through a series of excursions and exhibits, the philosophy and history of inductive inference come alive. Philosophical tools are put to work to solve problems about science and pseudoscience, induction and falsification.


Statistical and Inductive Inference by Minimum Message Length

Statistical and Inductive Inference by Minimum Message Length

Author: C.S. Wallace

Publisher: Springer Science & Business Media

Published: 2005-05-26

Total Pages: 456

ISBN-13: 9780387237954

DOWNLOAD EBOOK

The Minimum Message Length (MML) Principle is an information-theoretic approach to induction, hypothesis testing, model selection, and statistical inference. MML, which provides a formal specification for the implementation of Occam's Razor, asserts that the ‘best’ explanation of observed data is the shortest. Further, an explanation is acceptable (i.e. the induction is justified) only if the explanation is shorter than the original data. This book gives a sound introduction to the Minimum Message Length Principle and its applications, provides the theoretical arguments for the adoption of the principle, and shows the development of certain approximations that assist its practical application. MML appears also to provide both a normative and a descriptive basis for inductive reasoning generally, and scientific induction in particular. The book describes this basis and aims to show its relevance to the Philosophy of Science. Statistical and Inductive Inference by Minimum Message Length will be of special interest to graduate students and researchers in Machine Learning and Data Mining, scientists and analysts in various disciplines wishing to make use of computer techniques for hypothesis discovery, statisticians and econometricians interested in the underlying theory of their discipline, and persons interested in the Philosophy of Science. The book could also be used in a graduate-level course in Machine Learning and Estimation and Model-selection, Econometrics and Data Mining. C.S. Wallace was appointed Foundation Chair of Computer Science at Monash University in 1968, at the age of 35, where he worked until his death in 2004. He received an ACM Fellowship in 1995, and was appointed Professor Emeritus in 1996. Professor Wallace made numerous significant contributions to diverse areas of Computer Science, such as Computer Architecture, Simulation and Machine Learning. His final research focused primarily on the Minimum Message Length Principle.


Bayesian inference with INLA

Bayesian inference with INLA

Author: Virgilio Gomez-Rubio

Publisher: CRC Press

Published: 2020-02-20

Total Pages: 330

ISBN-13: 1351707205

DOWNLOAD EBOOK

The integrated nested Laplace approximation (INLA) is a recent computational method that can fit Bayesian models in a fraction of the time required by typical Markov chain Monte Carlo (MCMC) methods. INLA focuses on marginal inference on the model parameters of latent Gaussian Markov random fields models and exploits conditional independence properties in the model for computational speed. Bayesian Inference with INLA provides a description of INLA and its associated R package for model fitting. This book describes the underlying methodology as well as how to fit a wide range of models with R. Topics covered include generalized linear mixed-effects models, multilevel models, spatial and spatio-temporal models, smoothing methods, survival analysis, imputation of missing values, and mixture models. Advanced features of the INLA package and how to extend the number of priors and latent models available in the package are discussed. All examples in the book are fully reproducible and datasets and R code are available from the book website. This book will be helpful to researchers from different areas with some background in Bayesian inference that want to apply the INLA method in their work. The examples cover topics on biostatistics, econometrics, education, environmental science, epidemiology, public health, and the social sciences.


Linear Mixed Models for Longitudinal Data

Linear Mixed Models for Longitudinal Data

Author: Geert Verbeke

Publisher: Springer Science & Business Media

Published: 2009-04-28

Total Pages: 578

ISBN-13: 1441902996

DOWNLOAD EBOOK

This book provides a comprehensive treatment of linear mixed models for continuous longitudinal data. Next to model formulation, this edition puts major emphasis on exploratory data analysis for all aspects of the model, such as the marginal model, subject-specific profiles, and residual covariance structure. Further, model diagnostics and missing data receive extensive treatment. Sensitivity analysis for incomplete data is given a prominent place. Most analyses were done with the MIXED procedure of the SAS software package, but the data analyses are presented in a software-independent fashion.


Non-Standard Parametric Statistical Inference

Non-Standard Parametric Statistical Inference

Author: Russell Cheng

Publisher: Oxford University Press

Published: 2017-09-15

Total Pages: 432

ISBN-13: 0192518313

DOWNLOAD EBOOK

This book discusses the fitting of parametric statistical models to data samples. Emphasis is placed on: (i) how to recognize situations where the problem is non-standard when parameter estimates behave unusually, and (ii) the use of parametric bootstrap resampling methods in analyzing such problems. A frequentist likelihood-based viewpoint is adopted, for which there is a well-established and very practical theory. The standard situation is where certain widely applicable regularity conditions hold. However, there are many apparently innocuous situations where standard theory breaks down, sometimes spectacularly. Most of the departures from regularity are described geometrically, with only sufficient mathematical detail to clarify the non-standard nature of a problem and to allow formulation of practical solutions. The book is intended for anyone with a basic knowledge of statistical methods, as is typically covered in a university statistical inference course, wishing to understand or study how standard methodology might fail. Easy to understand statistical methods are presented which overcome these difficulties, and demonstrated by detailed examples drawn from real applications. Simple and practical model-building is an underlying theme. Parametric bootstrap resampling is used throughout for analyzing the properties of fitted models, illustrating its ease of implementation even in non-standard situations. Distributional properties are obtained numerically for estimators or statistics not previously considered in the literature because their theoretical distributional properties are too hard to obtain theoretically. Bootstrap results are presented mainly graphically in the book, providing an accessible demonstration of the sampling behaviour of estimators.