Graphical Models

Course Material

Lectures

All lectures are given from the slides provided here. Please print your own copies before the lectures using the 2up versions provided). The lecture itself will sometimes digress to web-based material and Java applets for demonstrations. These will usually be listed on this page, but not on the PDF slides.
Lecture 1: Introduction: Graphs and Probability

Review of some famous graphical models including circuits, fault trees, entity-attribute diagrams, and ising lattices. Review of graphs including definitions for directed and undirected graphs, relations such as parents and ancestors, and connectivity notions. Review of probability as frequency or belief in both discrete and continuous contexts. We may play with the Change of variables applet and the Integration applet to better understand the implications of working with continuous domains. Review of modelling as a key part of intelligent systems design. Examples of using graphs to model dependence, causality and relevance. [Aug. 24th, Lecture 1 in PDF, in Postscript, 2-up, gzipped ]

Preparatory reading: Please review the material under graphs from Math World (know these definitions), the Wikipedia (only graph coloring should be understood, all others briefly browse) and the Ising Model. Review course notes of Three Concepts: Probability (in particular, Set 2 and 3 from the Material).
Lecture 2: Independence Modelling

Functional independence in optimization as a method for task simplification. Graph partitioning as a means of problem decomposition. Probabilistic independence as functional independence in probability space. Undirected graphs, relevance, and independence defined as set separation. Clique trees as undirected graphs on variable sets, and a normalised form for them called Almond trees which are a tree-structured generalization of binary independence. Directed graphs, dependence, and independence defined on the underlying undirected graph. Soundness and completeness between graphical independence and functional independence forms. A catalogue of graphical forms for time-series, spatial modelling, diagnosis and fault detection and their independence properties. [Sept. 30th, Lecture 2 in PDF, in Postscript, 2-up, gzipped ]

Preparatory reading: Please review Shachter's manuscript "Examples without Numbers" (local copy) and glance briefly at Scheines' tutorial on d-separation. You don't have to understand d-separation, because its difficult and we'll use a Lauritzen-style definition instead, but have a quick look at it. Part I of Bishop's tutorial (see below) is good material for this lecture. A demo will also be giving using B-Course.
Lecture 3: Exact Inference

Given a graphical model on discrete variables of low dimension, and a probability formula, evaluate the probability, or find the maximum likelihood assignment for the free variables. We will consider these two problems in a general sense. We will develop an exact algorithm for the computation as nested summation on a clustered graph using variable elimination as the basic step. This turns out to create a junction tree which we can turn into an Almond tree, the general independence structure from the last lecture. Thus we will consider finding good junction trees, and their complexity. We will not cover implementation details for the algorithm used to speed it up in practice. [Oct. 1st, Lecture 3 in PDF, in Postscript, 2-up, gzipped]
Preparatory reading: TBD
Lecture 4: Approximate Inference

We will consider some basic types of approximate inference on graphs including Gibbs sampling, rejection sampling importance sampling. For Gibbs sampling, will will review OpenBugs, maintained at the local Statistics Department. Gibbs sampling will be compared with local search, a popular method in optimization and constraint satisfaction. [Oct. 6th, Lecture 4 in PDF, in Postscript, 2-up, gzipped]

Preparatory reading: We will look at examples from OpenBugs, given in Examples:Examples Volume I, Pump and Dogs. The Ising Model is another example of Gibbs, in fact it is the inverse of 2-coloring.
Lecture 5: Basic Learning and Graphs

We will consider a simple case of learning: proportions using the Beta Coin Experiment applet from the Prob./Stats Object Library. This example is covered in detail because it forms the basis of many algorithms. This introduces the notions of maximum likelihood, maximum aposterior, expected value, sufficient statistics, exponential family, the trade-off between complexity and predictive accuracy, and illustrates the problems with priors. The notions of maximum likelihood, maximum aposterior, sufficient statistics, exponential family, and problems with priors will be covered in more detail, making use of graphs as a tool to illustrate the variations.

Preparatory reading: Notes of Three Concepts: Probability Set 4 on proportions. Here is a presentation version.
Lecture 6: Selected Applications and Review

Please bring you questions about the assignments. We will review any material necessary so that everyone has the background they need to complete things. Some applications may also be covered.

Preparatory reading: None for this week.


Literature, further reading

There is no need to purchase any books. We will supply copies of any necessary material prior to lectures. Material available online will sometimes be in Adobe's PDF or Postscript format so you should be able to view these formats. Moreover, some of the demonstrations require Java capable browsers. Other than the use of these long-time standard formats, we make efforts to avoid online material that requires special browsers (e.g., thus material converted from Powerpoint is avoided). Note when we say something is a reference book, we mean you should not use it to learn the material, but only to check details or find more advanced material.

Graphs

Probability

Graphical Models

References

Only use as references if needed.

Toolboxes/Software

Things we use and recommend: Other good software.

Graphical Models
2002