Subset and Sample Selection for Graphical Models: Gaussian Processes, Ising Models and Gaussian Mixture Models
Author: Satyaki Mahalanabis
Publisher:
Published: 2012
Total Pages: 234
ISBN-13:
DOWNLOAD EBOOK"Probabilistic Graphical Models are a popular method of representing complex joint distributions in which stochastic dependence between subsets of random variables is expressed in terms of a graph. In many scenarios, random samples from a graphical model are only partially observed - either only a few random variables may be observed for any sample, or the values of some variables (called hidden variables) are missing from each sample unless explicitly queried. This dissertation considers the following general problem: how to select a small subset of variables to observe for a given sample (called subset selection) or a small subset of samples for which to observe the hidden variables (called sample selection) so as to accurately predict the value of the unobserved variables? We investigate this question for 3 widely-studied classes of graphical models: Gaussian Processes, Ising Models, and Gaussian Mixture Models. We prove that nding an optimal subset selection strategy is NP-hard even for a restricted class of Gaussian Processes, called Gaussian Free Fields (GFF). We give a dynamic programming algorithm for Gaussian Processes on bounded tree-width graphs, which yields a fully polynomial time approximation scheme for the case of GFFs on such graphs. For general Gaussian Processes on bounded tree-width graphs, our algorithm's running time depends polynomially on the condition number of the covariance matrix. We also give a greedy constant-factor approximation algorithm for GFFs on arbitrary graphs. We consider both adaptive and non-adaptive subset selection for Ising Models. For the simple 1-dimensional ferromagnetic Ising Model, we demonstrate that adaptive strategies outperform non-adaptive strategies, and give a simple adaptive strategy whose error is at most a constant times that of the optimal adaptive strategy for the same observation budget. We prove that it is NP-hard to compute an optimal non-adaptive strategy for ferromagnetic Ising Models on general graphs. For mixture models, we dene a 'maximum-a-posteriori' oracle and discuss how it diers from other oracle models. Then we demonstrate the advantage provided by this oracle by giving an algorithm which estimates the parameters of a mixture of high-dimensional spherical Gaussians under a weaker separation condition and more eciently than known unsupervised algorithms"--Leaves iv-v.