Course Calendar

The acronyms which identify the primary course readings are defined below.

Date Topic Primary Readings Supplemental Readings Materials
1/24 Course Overview Jordan 2004 Murphy Tutorial slides
1/29 Directed Graphical Models MJ: 2.1, 2.3
ES: 2.2
Charniak 1991 slides
1/31 Undirected Graphical Models MJ: 2.2
ES: 2.2
W&J: 2.1-2.4 slides
2/05 Inference in Graphical Models,
Elimination Algorithms
MJ: 3
ES: 2.2.5
W&J: 2.5 slides
2/07 Message Passing, Belief
Propagation, & Factor Graphs
MJ: 4.1-4.2
ES: 2.3.2, pp. 69-75
Kschischang et al. 2001
Aji, McEliece 2000
slides
2/12 Bayes & Frequentist Learning
from Complete Observations
MJ: 5.1-5.2
MJ: 9.1-9.2
Heckerman 1999: 1-5 slides
2/14 Exponential Families,
Conjugate Priors
MJ: 8.1, 19
ES: 2.1
W&J: 3.1-3.5 slides
2/19 Holiday: No Lecture      
2/21 Expectation Maximization (EM)
in Directed Graphical Models
MJ: 11, 12
ES: 2.3.3
Neal, Hinton 1999
Heckerman 1999: 6
slides
2/26 EM Algorithm Continued,
Learning in Undirected Graphs
MJ: 11, 12
MJ: 9.3
W&J: 6.1-6.2 slides
2/28 Junction Tree Algorithm MJ: 17 Paskin 2003
Cowell 1999, pp. 9-37
slides
3/05 Undirected Gaussian
Markov Random Fields
Sudderth 2002, Chap. 2 MJ: 13
Szeliski 1990
slides
3/07 State Space Models,
Kalman Filter, Gaussian BP
MJ: 15
Weiss, Freeman 2001: 1-2
MJ: 14
Roweis, Ghahramani 1999
slides
3/12 Learning in Gaussian Models,
Non-Gaussian Inference
MJ: 14
ES: 3.1
Ghahramani, Hinton 1996 slides
3/14 Monte Carlo, Rejection
& Importance Sampling
Andrieu et al. 2003: 1-2
ES: 2.4
MacKay 1999: 1-3 slides
3/19 Particle Filters,
Sequential Monte Carlo
Cappe et al. 2007
Andrieu et al. 2003: 4.3
Hamze, de Freitas 2006
Del Moral et al. 2006
slides
3/21 Markov Chain Monte Carlo,
Metropolis & Gibbs Samplers
Andrieu et al. 2003: 3 MacKay 1999: 4-7
Geman, Geman 1984
slides
3/26 Spring Break: No Lecture      
3/28 Spring Break: No Lecture      
4/02 MCMC Mixing & Diagnostics,
Collapsed Gibbs Samplers
ES: 2.4.4
Andrieu et al. 2003: 3
Griffiths, Steyvers 2004 slides
4/04 Mean Field Methods,
Variational Learning
Jordan et al. 1999: 4, 6
W&J: 5, 6.3
Winn, Bishop 2005
Blei, Ng, Jordan 2003
slides
4/09 Variational Learning,
Blocked Gibbs Samplers
ES: 2.3 Scott 2002 slides
4/11 Structured Variational Methods,
Bethe Approximations, Loopy BP
Yedidia et al. 2002 W&J: 4.1 slides
4/16 Convexity and Duality
in Variational Methods
W&J: 3.4-3.6, 5.3-5.4 ES: 2.3 slides
4/18 Reparameterization & Loopy BP,
Reweighted Sum-Product
W&J: 4.1, 7.1-7.2 Ihler et al. 2005
Mooij, Kappen 2007
slides
4/23 Reweighted BP,
Convex Learning Surrogates
W&J: 7.4-7.5 Wainwright 2006 slides
4/25 Conditional Random Fields,
MAP Estimation & Max-Product
Sutton, McCallum 2012
MJ: 4.3
Aji, McEliece 2000 slides
4/30 Reweighted Max-Product,
Course Review & Outlook
W&J: 8.1-8.4 Yanover et al. 2006
Greig et al. 1989
slides
5/02 Project Help Session      
5/07 Project Presentations      
5/13 Project Reports Due      

Readings

Primary Resources

MJ: M. Jordan, An Introduction to Probabilistic Graphical Models. Printed reader available at the Metcalf Copy Center.

W&J: M. Wainwright & M. Jordan, Graphical Models, Exponential Families, and Variational Inference. Foundations and Trends in Machine Learning, vol. 1, pp. 1-305, 2008.

ES: E. Sudderth, Graphical Models for Visual Object Recognition and Tracking, Chapter 2: Nonparametric and Graphical Models. Doctoral Thesis, Massachusetts Institute of Technology, May 2006.

Surveys & Tutorials

S. Aji & R. McEliece, The Generalized Distributive Law. IEEE Transactions on Information Theory, vol. 46, pp. 325-343, 2000.

C. Andrieu, N. de Freitas, A. Doucet, & M. Jordan, An Introduction to MCMC for Machine Learning. Machine Learning, vol. 50, pp. 5-43, 2003.

O. Cappe, S. Godsill, & E. Moulines, An Overview of Existing Methods and Recent Advances in Sequential Monte Carlo. Proceedings of the IEEE, vol. 95, pp. 899-924, 2007.

E. Charniak, Bayesian Networks without Tears. AI Magazine, vol. 12, pp. 50-63, 1991.

R. Cowell, Introduction to Inference for Bayesian Networks, Advanced Inference in Bayesian Networks. Learning in Graphical Models, M. Jordan, MIT Press, 1999.

B. Frey & N. Jojic, A Comparison of Algorithms for Inference and Learning in Probabilistic Graphical Models. IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 27, pp. 1392-1416, 2005.

D. Heckerman, A Tutorial on Learning with Bayesian Networks. Learning in Graphical Models, M. Jordan, MIT Press, 1999.

M. Jordan, Z. Ghahramani, T. Jaakkola, & L. Saul, An Introduction to Variational Methods for Graphical Models. Machine Learning, vol. 37, pp. 183-233, 1999.

M. Jordan, Graphical Models. Statistical Science, vol. 19, pp. 140-155, 2004.

F. Kschischang, B. Frey, & H.-A. Loeliger, Factor Graphs and the Sum-Product Algorithm. IEEE Transactions on Information Theory, vol. 47, pp. 498-519, 2001.

D. MacKay, Introduction to Monte Carlo Methods. Learning in Graphical Models, M. Jordan, MIT Press, 1999.

R. Neal & G. Hinton, A View of the EM Algorithm that Justifies Incremental, Sparse, and Other Variants. Learning in Graphical Models, M. Jordan, MIT Press, 1999.

S. Roweis & Z. Ghahramani, A Unifying Review of Linear Gaussian Models. Neural Computation, vol. 11, pp. 305-345, 1999.

E. Sudderth, Embedded Trees: Estimation of Gaussian Processes on Graphs with Cycles, Chapter 2: Background. Masters Thesis, Massachusetts Institute of Technology, Feb. 2002.

C. Sutton & A. McCallum, An Introduction to Conditional Random Fields. Foundations and Trends in Machine Learning, vol. 4, pp. 267-373, 2012.

Y. Teh & M. Jordan, Hierarchical Bayesian Nonparametric Models with Applications. To appear in Bayesian Nonparametrics in Practice, Cambridge University Press, 2010.

A. Willsky, Multiresolution Markov Models for Signal and Image Processing. Proceedings of the IEEE, vol. 90, pp. 1396-1458, 2002.

J. Yedidia, W. Freeman, & Y. Weiss, Understanding Belief Propagation and its Generalizations. Exploring Artificial Intelligence in the New Millennium, Morgan Kaufmann, 2002.

Research Articles

D. Blei, A. Ng, & M. Jordan, Latent Dirichlet Allocation. Journal of Machine Learning Research, vol. 3, pp. 993-1022, 2003.

Y. Boykov, O. Veksler, & R. Zabih, Fast Approximate Energy Minimization via Graph Cuts. IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 23, pp. 1-18, 2001.

P. Del Moral, A. Doucet, & A. Jasra, Sequential Monte Carlo Samplers. Journal of the Royal Statistical Society B, vol. 68, pp. 411-436, 2006.

A. Doucet, N. de Freitas, K. Murphy, & S. Russell, Rao-Blackwellised Particle Filtering for Dynamic Bayesian Networks. Uncertainty in Artificial Intelligence 16, pp. 176-183, 2000.

S. Geman & D. Geman, Stochastic Relaxation, Gibbs Distributions, and the Bayesian Restoration of Images. IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 6, pp. 721-741, 1984.

D. Greig, B. Porteous, & A. Seheult, Exact Maximum A Posteriori Estimation for Binary Images. Journal of the Royal Statistical Society B, vol. 51, pp. 271-279, 1989.

T. Griffiths & M. Steyvers, Finding Scientific Topics. Proceedings of the National Academy of Sciences, vol. 101, pp. 5228-5235, 2004.

F. Hamze & N. de Freitas, Hot Coupling: A Particle Approach to Inference and Normalization on Pairwise Undirected Graphs of Arbitrary Topology. Neural Information Processing Systems 18, pp. 491-498, 2006.

A. Ihler, J. Fisher, & A. Willsky, Loopy Belief Propagation: Convergence and Effects of Message Errors. Journal of Machine Learning Research, vol. 6, pp. 905-936, 2005.

T. Jaakkola & M. Jordan, Bayesian Parameter Estimation via Variational Methods. Statistics and Computing, vol. 10, pp. 25-37, 2000.

D. Malioutov, J. Johnson, & A. Willsky, Walk-Sums and Belief Propagation in Gaussian Graphical Models. Journal of Machine Learning Research, vol. 7, pp. 2031-2064, 2006.

J. Mooij & H. Kappen, Sufficient Conditions for Convergence of the Sum-Product Algorithm. IEEE Transactions on Information Theory, vol. 53, pp. 4422-4437, 2007.

K. Murphy & Y. Weiss, The Factored Frontier Algorithm for Approximate Inference in DBNs. Uncertainty in Artificial Intelligence 17, pp. 378-385, 2001.

S. Scott, Bayesian Methods for Hidden Markov Models: Recursive Computing in the 21st Century. Journal of the American Statistical Association, vol. 97, no. 457, 2002.

R. Szeliski, Bayesian Modeling of Uncertainty in Low-Level Vision. International Journal of Computer Vision, vol. 5, pp. 271-301, 1990.

M. Wainwright, T. Jaakkola, & A. Willsky, Tree-Based Reparameterization Framework for Analysis of Sum-Product and Related Algorithms. IEEE Transactions on Information Theory, vol. 49, pp. 1120-1146, 2003.

M. Wainwright, T. Jaakkola, & A. Willsky, A New Class of Upper Bounds on the Log Partition Function. IEEE Transactions on Information Theory, vol. 51, pp. 2313-2335, 2005.

M. Wainwright, Estimating the "Wrong" Graphical Model: Benefits in the Computation-Limited Setting. Journal of Machine Learning Research, vol. 7, pp. 1829-1859, 2006.

Y. Weiss & W. Freeman, Correctness of Belief Propagation in Gaussian Graphical Models of Arbitrary Topology. Neural Computation, vol. 13, pp. 2173-2200, 2001.

Y. Weiss, C. Yanover, & T. Meltzer, MAP Estimation, Linear Programming and Belief Propagation with Convex Free Energies. Uncertainty in Artificial Intelligence 23, pp. 416-425, 2007.

J. Winn & C. Bishop, Variational Message Passing. Journal of Machine Learning Research, vol. 6, pp. 661-694, 2005.

C. Yanover, T. Meltzer, & Y. Weiss, Linear Programming Relaxations and Belief Propagation - An Empirical Study. Journal of Machine Learning Research, vol. 7, pp. 1887-1907, 2006.

J. Yedidia, W. Freeman, & Y. Weiss, Constructing Free Energy Approximations and Generalized Belief Propagation Algorithms. IEEE Transactions on Information Theory, vol. 51, pp. 2282-2312, 2005.