Bayesian Nonparametric Latent Variable Models

Download Bayesian Nonparametric Latent Variable Models PDF/ePub or read online books in Mobi eBooks. Click Download or Read Online button to get Bayesian Nonparametric Latent Variable Models book now. This website allows unlimited access to, at the time of writing, more than 1.5 million titles, including hundreds of thousands of titles in various foreign languages.
Bayesian Nonparametric Latent Variable Models

One of the important problems in machine learning is determining the complexity of the model to learn. Too much complexity leads to overfitting, which finds structures that do not actually exist in the data, while too low complexity leads to underfitting, which means that the expressiveness of the model is insufficient to capture all the structures present in the data. For some probabilistic models, the complexity depends on the introduction of one or more latent variables whose role is to explain the generative process of the data. There are various approaches to identify the appropriate number of latent variables of a model. This thesis covers various Bayesian nonparametric methods capable of determining the number of latent variables to be used and their dimensionality. The popularization of Bayesian nonparametric statistics in the machine learning community is fairly recent. Their main attraction is the fact that they offer highly flexible models and their complexity scales appropriately with the amount of available data. In recent years, research on Bayesian nonparametric learning methods have focused on three main aspects: the construction of new models, the development of inference algorithms and new applications. This thesis presents our contributions to these three topics of research in the context of learning latent variables models. Firstly, we introduce the Pitman-Yor process mixture of Gaussians, a model for learning infinite mixtures of Gaussians. We also present an inference algorithm to discover the latent components of the model and we evaluate it on two practical robotics applications. Our results demonstrate that the proposed approach outperforms, both in performance and flexibility, the traditional learning approaches. Secondly, we propose the extended cascading Indian buffet process, a Bayesian nonparametric probability distribution on the space of directed acyclic graphs. In the context of Bayesian networks, this prior is used to identify the presence of latent variables and the network structure among them. A Markov Chain Monte Carlo inference algorithm is presented and evaluated on structure identification problems and as well as density estimation problems. Lastly, we propose the Indian chefs process, a model more general than the extended cascading Indian buffet process for learning graphs and orders. The advantage of the new model is that it accepts connections among observable variables and it takes into account the order of the variables. We also present a reversible jump Markov Chain Monte Carlo inference algorithm which jointly learns graphs and orders. Experiments are conducted on density estimation problems and testing independence hypotheses. This model is the first Bayesian nonparametric model capable of learning Bayesian learning networks with completely arbitrary graph structures.
Bayesian Nonparametric Statistics

This up-to-date overview of Bayesian nonparametric statistics provides both an introduction to the field and coverage of recent research topics, including deep neural networks, high-dimensional models and multiple testing, Bernstein-von Mises theorems and variational Bayes approximations, many of which have previously only been accessible through research articles. Although Bayesian posterior distributions are widely applied in astrophysics, inverse problems, genomics, machine learning and elsewhere, their theory is still only partially understood, especially in complex settings such as nonparametric or semiparametric models. Here, the available theory on the frequentist analysis of posterior distributions is outlined in terms of convergence rates, limiting shape results and uncertainty quantification. Based on lecture notes for a course given at the St-Flour summer school in 2023, the book is aimed at researchers and graduate students in statistics and probability.
Bayesian Nonparametrics for Causal Inference and Missing Data

Bayesian Nonparametrics for Causal Inference and Missing Data provides an overview of flexible Bayesian nonparametric (BNP) methods for modeling joint or conditional distributions and functional relationships, and their interplay with causal inference and missing data. This book emphasizes the importance of making untestable assumptions to identify estimands of interest, such as missing at random assumption for missing data and unconfoundedness for causal inference in observational studies. Unlike parametric methods, the BNP approach can account for possible violations of assumptions and minimize concerns about model misspecification. The overall strategy is to first specify BNP models for observed data and then to specify additional uncheckable assumptions to identify estimands of interest. The book is divided into three parts. Part I develops the key concepts in causal inference and missing data and reviews relevant concepts in Bayesian inference. Part II introduces the fundamental BNP tools required to address causal inference and missing data problems. Part III shows how the BNP approach can be applied in a variety of case studies. The datasets in the case studies come from electronic health records data, survey data, cohort studies, and randomized clinical trials. Features • Thorough discussion of both BNP and its interplay with causal inference and missing data • How to use BNP and g-computation for causal inference and non-ignorable missingness • How to derive and calibrate sensitivity parameters to assess sensitivity to deviations from uncheckable causal and/or missingness assumptions • Detailed case studies illustrating the application of BNP methods to causal inference and missing data • R code and/or packages to implement BNP in causal inference and missing data problems The book is primarily aimed at researchers and graduate students from statistics and biostatistics. It will also serve as a useful practical reference for mathematically sophisticated epidemiologists and medical researchers.