Course Calendar
The acronyms which identify the primary course readings are defined below.
For readings marked with a bold (C), critical comments are due before the class meeting.
Date | Topic | Assigned Readings | Supplemental Readings | Lecture Materials |
---|---|---|---|---|
1/27 | Course Overview | Jordan 2004 | K&F: 1, 2.1, 2.2 | Sudderth: slides |
2/01 | Directed Graphical Models | K&F: 3.1 to 3.3 | Charniak 1991 K&F: 5.4, 6.2, 6.4.1 |
Sudderth |
2/03 | Undirected Graphical Models | K&F: 4.1 to 4.3 W&J: 2.1 to 2.4 |
K&F: 4.4 ES: 2.2 |
Sudderth: slides |
2/08 | Variable Elimination, Message Passing |
ES: 2.3.2, pp. 69-75 K&F: 9.2, 9.3, 10.1, 10.2.1 |
K&F: 9.1, 9.4 W&J: 2.5 |
Sudderth: slides |
2/10 | Message Passing, Junction Trees |
Kschischang et al. 2001 (C) Cowell 1999, pp. 9-37 (C) |
K&F: 10.2 to 10.4 Paskin 2003 Aji, McEliece 2000 |
Pacheco: notes Black: slides |
2/15 | Directed Gaussian Graphical Models, Kalman Filter |
K&F: 7.1, 7.2, 15.4.1 Roweis, Ghahramani 1999 (C) |
Willsky 2002 | Sudderth Kim: slides |
2/17 | Kalman Filter, Gaussian Markov Random Fields |
Sudderth 2002(a), Sec. 2.1-2.2 Szeliski 1990 (C) |
K&F: 7.3 Cowell 1999, pp. 38-42 |
Sudderth Sun: slides |
2/22 | No Class | |||
2/24 | Exponential Families, Maximum Likelihood, EM Algorithm |
W&J: 3, 6.1, 6.2 Neal, Hinton 1999 (C) Heckerman 1999, Sec. 3 to 6 |
ES: 2.1.1, 2.3.3 K&F: 8.2, 8.3, 17.1 K&F: 17.2, 19.1, 19.2 |
Sudderth Zuffi: slides |
3/01 | Variational Methods, Mean Field |
W&J: 3.6, 5.1 to 5.4 Winn, Bishop 2005 (C) |
K&F: 11.5 ES: 2.3.1 |
Sudderth Miller: slides |
3/03 | Bayesian Estimation, Variational Bayes |
ES: 2.1 W&J: 6.3 Blei, Ng, Jordan 2003 (C) |
K&F: 17.3, 17.4, 19.3 Jaakkola, Jordan 2000 |
Sudderth Santhanam: slides |
3/08 | No Class | |||
3/10 | No Class | |||
3/15 | Loopy Belief Propagation |
W&J: 4.1 to 4.1.5 Yedidia et al. 2002 |
K&F: 11.1 to 11.3 ES: 2.3.2 |
Sudderth: slides |
3/17 | Generalized BP, Convergence of Loopy BP |
Yedidia et al. 2005 (C) Ihler et al. 2005(a) (C) |
W&J: 4.2 Mooij, Kappen 2007 |
Mayer: slides Wei: slides |
3/22 | Gaussian Belief Propagation |
Weiss, Freeman 2001 Malioutov et al. 2006 (C) |
Sudderth 2002(a), Sec. 2.3 | Sudderth: slides Ulusoy: slides |
3/24 | Undirected Parameter Estimation, Reweighted BP |
K&F: 20.1 to 20.4 W&J: 7.1 to 7.2.1 Wainwright 2006 (C) |
W&J: 7.2 to 7.5 K&F: 20.5 |
Sudderth Klein: slides |
3/29 | Spring Break | |||
3/31 | Spring Break | |||
4/05 | MAP Estimation: Max-Product, Graph Cuts |
W&J: 8.1 to 8.3 Boykov et al. 2001 |
K&F: 13.1 to 13.3, 13.6 Greig et al. 1989 |
Sudderth Tsoli: slides |
4/07 | Reweighted Max-Product, Linear Programming |
W&J: 8.4 Yanover et al. 2006 (C) |
K&F: 13.4 to 13.5 Weiss et al. 2007 |
Sudderth Wang: slides |
4/12 | Sequential Importance Sampling, Particle Filters |
K&F: 12.1-12.2 Cappe et al. 2007 (C) |
ES: 2.4-2.4.2, 3.1 K&F: 15.3 |
Sudderth Jones: slides |
4/14 | Sequential Monte Carlo |
Doucet et al. 2000 Hamze, de Freitas 2006 (C) |
Del Moral et al. 2006 NIPS09: de Freitas, Doucet |
Sudderth: slides Feldman |
4/19 | Markov Chain Monte Carlo |
Andrieu et al. 2003 K&F: 12.3 |
MacKay 1999 ES: 2.4.3-2.4.4 |
Sudderth: slides |
4/21 | Blocked and Collapsed Gibbs Samplers |
Scott 2002 (C) Griffiths, Steyvers 2004 (C) Griffiths et al. 2005 |
Geman, Geman 1984 | Ural: slides Mason: slides |
4/26 | Nonparametric Belief Propagation |
ES: 3.2 to 3.5 | ES: 4 Ihler et al. 2005(b) |
Sudderth: slides |
4/28 | Expectation Propagation |
W&J: 4.3 Murphy, Weiss 2001 Heskes, Zoeter 2002 (C) |
K&F: 11.4 Minka 2001 Sudderth 2002(b) |
Sudderth: slides Buller: slides |
5/03 | Bayesian Nonparametrics |
ES: 2.5 Neal 2000 |
Jordan 2005 Ghahramani 2005 Teh 2009 |
Sudderth |
5/05 | Dirichlet Process Mixture Models |
Blei, Jordan 2006 (C) Teh et al. 2006 (C) |
Teh, Jordan 2010 | Ghosh: slides Swanson: slides |
5/10 | Final Project Presentations |
Readings
Primary Resources
K&F: D. Koller & N. Friedman, Probabilistic Graphical Models: Principles and Techniques. MIT Press, 2009.
W&J: M. Wainwright & M. Jordan, Graphical Models, Exponential Families, and Variational Inference. Foundations and Trends in Machine Learning, vol. 1, pp. 1-305, 2008.
ES: E. Sudderth, Graphical Models for Visual Object Recognition and Tracking, Chapter 2: Nonparametric and Graphical Models. Doctoral Thesis, Massachusetts Institute of Technology, May 2006.
Survey, Tutorial, & Research Articles
S. Aji & R. McEliece, The Generalized Distributive Law. IEEE Transactions on Information Theory, vol. 46, pp. 325-343, 2000.
C. Andrieu, N. de Freitas, A. Doucet, & M. Jordan, An Introduction to MCMC for Machine Learning. Machine Learning, vol. 50, pp. 5-43, 2003.
D. Blei, A. Ng, & M. Jordan, Latent Dirichlet Allocation. Journal of Machine Learning Research, vol. 3, pp. 993-1022, 2003.
D. Blei & M. Jordan, Variational Inference for Dirichlet Process Mixtures. Bayesian Analysis, vol. 1, pp. 121-144, 2006.
Y. Boykov, O. Veksler, & R. Zabih, Fast Approximate Energy Minimization via Graph Cuts. IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 23, pp. 1-18, 2001.
O. Cappe, S. Godsill, & E. Moulines, An Overview of Existing Methods and Recent Advances in Sequential Monte Carlo. Proceedings of the IEEE, vol. 95, pp. 899-924, 2007.
E. Charniak, Bayesian Networks without Tears. AI Magazine, vol. 12, pp. 50-63, 1991.
R. Cowell, Introduction to Inference for Bayesian Networks, Advanced Inference in Bayesian Networks. Learning in Graphical Models, M. Jordan, MIT Press, 1999.
P. Del Moral, A. Doucet, & A. Jasra, Sequential Monte Carlo Samplers. Journal of the Royal Statistical Society B, vol. 68, pp. 411-436, 2006.
A. Doucet, N. de Freitas, K. Murphy, & S. Russell, Rao-Blackwellised Particle Filtering for Dynamic Bayesian Networks. Uncertainty in Artificial Intelligence 16, pp. 176-183, 2000.
B. Frey & N. Jojic, A Comparison of Algorithms for Inference and Learning in Probabilistic Graphical Models. IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 27, pp. 1392-1416, 2005.
S. Geman & D. Geman, Stochastic Relaxation, Gibbs Distributions, and the Bayesian Restoration of Images. IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 6, pp. 721-741, 1984.
D. Greig, B. Porteous, & A. Seheult, Exact Maximum A Posteriori Estimation for Binary Images. Journal of the Royal Statistical Society B, vol. 51, pp. 271-279, 1989.
T. Griffiths & M. Steyvers, Finding Scientific Topics. Proceedings of the National Academy of Sciences, vol. 101, pp. 5228-5235, 2004.
T. Griffiths, M. Steyvers, D. Blei, & J. Tenenbaum, Integrating Topics and Syntax. Neural Information Processing Systems 17, pp. 537-544, 2005.
F. Hamze & N. de Freitas, Hot Coupling: A Particle Approach to Inference and Normalization on Pairwise Undirected Graphs of Arbitrary Topology. Neural Information Processing Systems 18, pp. 491-498, 2006.
D. Heckerman, A Tutorial on Learning with Bayesian Networks. Learning in Graphical Models, M. Jordan, MIT Press, 1999.
T. Heskes & O. Zoeter, Expectation Propagation for Approximate Inference in Dynamic Bayesian Networks. Uncertainty in Artificial Intelligence 18, pp. 216-223, 2002.
A. Ihler, J. Fisher, & A. Willsky, Loopy Belief Propagation: Convergence and Effects of Message Errors. Journal of Machine Learning Research, vol. 6, pp. 905-936, 2005 (a).
A. Ihler, J. Fisher, R. Moses, & A. Willsky, Nonparametric Belief Propagation for Self-Localization of Sensor Networks. IEEE Journal on Selected Areas in Communications, vol. 23, pp. 809-819, 2005 (b).
T. Jaakkola & M. Jordan, Bayesian Parameter Estimation via Variational Methods. Statistics and Computing, vol. 10, pp. 25-37, 2000.
M. Jordan, Graphical Models. Statistical Science, vol. 19, pp. 140-155, 2004.
F. Kschischang, B. Frey, & H.-A. Loeliger, Factor Graphs and the Sum-Product Algorithm. IEEE Transactions on Information Theory, vol. 47, pp. 498-519, 2001.
D. MacKay, Introduction to Monte Carlo Methods. Learning in Graphical Models, M. Jordan, MIT Press, 1999.
D. Malioutov, J. Johnson, & A. Willsky, Walk-Sums and Belief Propagation in Gaussian Graphical Models. Journal of Machine Learning Research, vol. 7, pp. 2031-2064, 2006.
T. Minka, Expectation Propagation for Approximate Bayesian Inference. Uncertainty in Artificial Intelligence 17, pp. 362-369, 2001.
J. Mooij & H. Kappen, Sufficient Conditions for Convergence of the Sum-Product Algorithm. IEEE Transactions on Information Theory, vol. 53, pp. 4422-4437, 2007.
K. Murphy & Y. Weiss, The Factored Frontier Algorithm for Approximate Inference in DBNs. Uncertainty in Artificial Intelligence 17, pp. 378-385, 2001.
R. Neal & G. Hinton, A View of the EM Algorithm that Justifies Incremental, Sparse, and Other Variants. Learning in Graphical Models, M. Jordan, MIT Press, 1999.
R. Neal, Markov Chain Sampling Methods for Dirichlet Process Mixture Models. Journal of Computational and Graphical Statistics, vol. 9, pp. 249-265, 2000.
S. Roweis & Z. Ghahramani, A Unifying Review of Linear Gaussian Models. Neural Computation, vol. 11, pp. 305-345, 1999.
S. Scott, Bayesian Methods for Hidden Markov Models: Recursive Computing in the 21st Century. Journal of the American Statistical Association, vol. 97, no. 457, 2002.
E. Sudderth, Embedded Trees: Estimation of Gaussian Processes on Graphs with Cycles, Chapter 2: Background. Masters Thesis, Massachusetts Institute of Technology, Feb. 2002 (a).
E. Sudderth, Expectation Propagation. Massachusetts Institute of Technology, 2002 (b).
R. Szeliski, Bayesian Modeling of Uncertainty in Low-Level Vision. International Journal of Computer Vision, vol. 5, pp. 271-301, 1990.
Y. Teh, M. Jordan, M. Beal, & D. Blei, Hierarchical Dirichlet Processes. Journal of the American Statistical Association, vol. 101, no. 476, 2006.
Y. Teh & M. Jordan, Hierarchical Bayesian Nonparametric Models with Applications. To appear in Bayesian Nonparametrics in Practice, Cambridge University Press, 2010.
M. Wainwright, T. Jaakkola, & A. Willsky, Tree-Based Reparameterization Framework for Analysis of Sum-Product and Related Algorithms. IEEE Transactions on Information Theory, vol. 49, pp. 1120-1146, 2003.
M. Wainwright, T. Jaakkola, & A. Willsky, A New Class of Upper Bounds on the Log Partition Function. IEEE Transactions on Information Theory, vol. 51, pp. 2313-2335, 2005.
M. Wainwright, Estimating the "Wrong" Graphical Model: Benefits in the Computation-Limited Setting. Journal of Machine Learning Research, vol. 7, pp. 1829-1859, 2006.
Y. Weiss & W. Freeman, Correctness of Belief Propagation in Gaussian Graphical Models of Arbitrary Topology. Neural Computation, vol. 13, pp. 2173-2200, 2001.
Y. Weiss, C. Yanover, & T. Meltzer, MAP Estimation, Linear Programming and Belief Propagation with Convex Free Energies. Uncertainty in Artificial Intelligence 23, pp. 416-425, 2007.
J. Winn & C. Bishop, Variational Message Passing. Journal of Machine Learning Research, vol. 6, pp. 661-694, 2005.
A. Willsky, Multiresolution Markov Models for Signal and Image Processing. Proceedings of the IEEE, vol. 90, pp. 1396-1458, 2002.
C. Yanover, T. Meltzer, & Y. Weiss, Linear Programming Relaxations and Belief Propagation - An Empirical Study. Journal of Machine Learning Research, vol. 7, pp. 1887-1907, 2006.
J. Yedidia, W. Freeman, & Y. Weiss, Understanding Belief Propagation and its Generalizations. Exploring Artificial Intelligence in the New Millennium, Morgan Kaufmann, 2002.
J. Yedidia, W. Freeman, & Y. Weiss, Constructing Free Energy Approximations and Generalized Belief Propagation Algorithms. IEEE Transactions on Information Theory, vol. 51, pp. 2282-2312, 2005.