6.437 Inference and Information
Introduction to principles of Bayesian and non-Bayesian statistical inference. Hypothesis testing and parameter estimation, sufficient statistics; exponential families. EM agorithm. Log-loss inference criterion, entropy and model capacity. Kullback-Leibler distance and information geometry. Asymptotic analysis and large deviations theory. Model order estimation; nonparametric statistics. Computational issues and approximation techniques; Monte Carlo methods. Selected topics such as universal inference and learning, and universal features and neural networks.
Lecture occurs 9:30 AM to 11:00 AM on Tuesdays and Thursdays in 32-155.
This class counts for a total of 12 credits.
You can find more information at the MIT + 6.437 - Google Search site.
© Copyright 2015 Yasyf Mohamedali