[Machine Learning: A very quick introduction.]
[Introduction to a similar course, written by Amos Storkey at the University of Edinburgh.]
Serial no. | Topics | Lecture nos. | PRML Chapters | Other References | Slides | Exercises |
1 | Introduction: Nature of Intelligence and Learning; Revision of Probability Theory, Distributions, Bayesian Inference | 1–8 | 1, 2 | Review of Probability Theory; The Multivariate Gaussian Distribution | Problem Set 1 [Data: P3.mat] [Solutions] Problem Set 2 [Solutions] |
|
2 | Basics of Learning: Tasks, Models, Features | 8–9 | Flach Chapter 1 | Very quick introduction | ||
3 | Supervised Learning: Linear Regression Models | 10–17 | 1, 3 | Regression notes | PRML Chapter 1; 'Beauty' in Scientific Theories; PRML Chapter 3 | |
4 | Supervised Learning: Linear Classifiers, the Perceptron and Logistic Regression | 18–21 | 4 | Linear models for classification | ||
5 | Supervised Learning: Neural Networks, Deep Learning | 22–23; 27–28 | 5 | ANN tutorial; Geoff Hinton lecture on deep learning; Convolutional neural nets: tutorial, demos; Sparse autoencoder notes; Baby Human video clip on face recognition; Backpropagation with max pooling notes | Google's unsupervised deep learning | |
6 | Supervised Learning: Kernels, Support Vector Machines | 23–26 | 6, 7 | SVM notes; SVM tutorial; Properties of kernels; Platt's SMO paper; SVM demo | ||
7 | Learning Theory | 29–32 | ||||
8 | Unsupervised Learning: Dimensionality Reduction, Factor/Component Analysis | 33–34 | 12 | PCA notes; PCA tutorial | PCA |