Description
Aims:
The module will present the foundations of approximate inference and learning in probabilistic graphical models (e.g., Bayesian networks and Markov networks), with particular focus on models composed from conditional exponential family distributions. Both stochastic (Monte Carlo) methods and deterministic approximations will be covered. The methods will be discussed in relation to practical problems in real-world inference in Machine Learning, including problems in tracking and learning.
Intended learning outcomes:
On successful completion of the module, a student will be able to:
- Understand how to derive and implement state-of-the-art approximate inference techniques and be able to make contributions to research in this area.
Indicative content:
The following are indicative of the topics the module will typically cover:
- Nonlinear, hierarchical (deep), and distributed models.
- Independent component analysis, Boltzmann machines, Dirichlet topic models, manifold discovery.
- Mean-field methods, variational approximations and variational Bayes.
- Expectation propagation.
- Loopy belief propagation, the Bethe free energy and extensions.
- Convex methods and convexified bounds.
- Monte-Carlo methods: including rejection and importance sampling, Gibbs, Metropolis-Hastings, anealed importance sampling, Hamiltonian Monte-Carlo, slice sampling, sequential Monte-Carlo (particle filtering).
- Other topics as time permits.
Requisites:
To be eligible to select this module as an optional or elective, a student must: (1) be registered on a programme and year of study for which it is formally available; and (2) have taken Probabilistic and Unsupervised Learning (COMP0086).
Module deliveries for 2024/25 academic year
Last updated
This module description was last updated on 19th August 2024.
Ìý