Speaker: Trevor Campbell, MIT
Location: 60 Fifth Avenue 150
Date: March 23, 2018, 11 a.m.
Host: Rob Fergus
The automation of posterior inference in Bayesian data analysis has enabled experts and nonexperts alike to use more sophisticated models, engage in faster exploratory modeling and analysis, and ensure experimental reproducibility. However, standard automated posterior inference algorithms are not tractable at the scale of massive modern datasets, and modifications to make them so are typically model-specific, require expert tuning, and can break theoretical guarantees on inferential quality. This talk will instead take advantage of data redundancy to shrink the dataset itself as a preprocessing step, forming a "Bayesian coreset." The coreset can be used in a standard inference algorithm at significantly reduced cost while maintaining theoretical guarantees on posterior approximation quality. The talk will include an intuitive formulation of Bayesian coreset construction as sparse vector sum approximation, an automated coreset construction algorithm that takes advantage of this formulation, strong theoretical guarantees on posterior approximation quality, and applications to a variety of real and simulated datasets.
Trevor Campbell is a postdoctoral associate in the Computer Science and Artificial Intelligence Laboratory (CSAIL) at MIT. He received his Ph.D. working in the Laboratory for Information and Decision Systems (LIDS) at MIT. Prior to his graduate studies, he received his B.A.Sc. in Engineering Science from the University of Toronto. His research focuses on scalable, automated Bayesian inference with theoretical guarantees, Bayesian modeling and inference for streaming, distributed data, and Bayesian nonparametrics.
Refreshments will be offered starting 15 minutes prior to the scheduled start of the talk.