S.No.
|
Topics
|
Lectures
|
Instructor
|
References/Notes
|
0
|
Introduction to Machine Learning
|
01-02
|
SDR
|
|
|
Flavours of Machine Learning: Unsupervised, Supervised,
Reinforcement, Hybrid models. Decision Boundaries: crisp, and
non-crisp, optimisation problems.
|
02 Jan (Tue) {lecture#01}
|
SDR
|
MS-Teams folder:
slides_k_means_em1_02jan24.pdf,
video_k_means_em1_02jan24.mp4
|
|
Introduction (contd).
|
03 Jan (Wed) {lecture#02}
|
SDR
|
MS-Teams folder:
slides_k_means_em2_03jan24.pdf,
video_k_means_em2_03jan24.mp4
|
1
|
Unsupervised Learning:
K-Means, Gaussian Mixture Models, EM
|
02-07
|
SDR
|
[Bishop Chap.9],
[Do: Gaussians],
[Do: More on Gaussians],
[Ng: K-Means],
[Ng: GMM],
[Ng: EM],
[Smyth:
EM]
|
|
The K-Means algorithm: Introduction. Algorithms: history,
flavours. A mathematical formulation of the K-Means algorithm.
The Objective function to minimise.
The K-Means algorithm:
The Objective function to minimise.
The basic K-Means algorithm, computation complexity issues: each
step, overall.
|
03 Jan (Wed) {lecture#02}
|
SDR
|
MS-Teams folder:
slides_k_means_em2_03jan24.pdf,
video_k_means_em2_03jan24.mp4
|
|
Limitations of K-Means.
K-Means: Alternate formulation with a distance threshold.
An introduction to Gaussian Mixture Models.
The Bayes rule, and Responsibilities. Maximum Likelihood
Estimation. Parameter estimation for a mixture of Gaussians,
starting with a simple 1-D single Gaussian case. ML-Estimation:
|
05 Jan (Fri) {lecture#03}
|
SDR
|
MS-Teams folder:
video_k_means_em3_05jan24.mp4,
slides_k_means_em3_05jan24.pdf
|
|
ML-Estimation: the simple case of one 1-D Gaussian, to the
general case of K D-dimensional Gaussians.
The Mahalanobis Distance.
Getting stuck, using Lagrange Multipliers.
|
09 Jan (Tue) {lecture#04}
|
SDR
|
MS-Teams folder:
video_k_means_em4_09jan23.mp4,
slides_k_means_em4_09jan23.pdf
|
|
The EM Algorithm for Gaussian Mixtures.
Application: Assignment 1:
The Stauffer and Grimson Adaptive
Background Subtraction Algorithm.
|
10 Jan (Wed) {lecture#05}
|
SDR
|
MS-Teams folder:
video_k_means_em5_10jan24.mp4,
slides_k_means_em5_10jan24.pdf
|
|
The Stauffer and Grimson Adaptive
Background Subtraction Algorithm (contd).
|
12 Jan (Fri) {lecture#06}
|
SDR
|
MS-Teams folder:
video_k_means_em6_12jan24.mp4,
slides_k_means_em6_12jan24.pdf,
|
|
The Stauffer and Grimson Adaptive
Background Subtraction Algorithm (contd).
|
16 Jan (Tue) {lecture#07}
|
SDR
|
MS-Teams folder:
video_k_means_em7_eigen1_16jan24.mp4,
slides_k_means_em7_16jan24.pdf
|
2
|
Unsupervised Learning:
EigenAnalysis:
PCA, LDA and Subspaces
|
07-10
|
SDR
|
[Ng: PCA],
[Ng: ICA],
[Burges: Dimension Reduction],
[Bishop Chap.12]
|
|
Introduction to Eigenvalues and Eigenvectors.
Properties of Eigenvalues and Eigenvectors.
|
16 Jan (Tue) {lecture#07}
|
SDR
|
MS-Teams folder:
video_k_means_em7_eigen1_16jan24.mp4,
slides_eigen1_16jan24.pdf
|
|
Properties of Eigenvalues and Eigenvectors (contd).
|
17 Jan (Wed) {lecture#08}
|
SDR
|
MS-Teams folder:
video_eigen2_17jan24.mp4, slides_eigen2_17jan24.pdf
|
|
Gram-Schmidt Orthogonalisation: an introduction.
The KL Transform.
|
19 Jan (Fri) {lecture#09}
|
SDR
|
MS-Teams folder:
video_eigen3_19jan24.mp4,
slides_eigen3_19jan24.pdf
|
|
The SVD and its properties (contd).
Application: Assignment 2: Eigenfaces
and Fisherfaces
|
20 Jan (Sat) {lecture#10}
|
SDR
|
MS-Teams folder:
video_eigen4_linear1_20jan24.mp4,
slides_eigen4_20jan24.pdf
|
3
|
Linear Models for Regression, Classification
|
10-15
|
SDR
|
[Bishop Chap.3],
[Bishop Chap.4],
[Ng: Supervised, Discriminant Analysis],
[Ng: Generative]
|
|
General introduction to Regression and Classification.
|
20 Jan (Sat) {lecture#10}
|
SDR
|
MS-Teams folder:
video_eigen4_linear1_20jan24.mp4,
slides_linear1_20jan24.pdf
|
|
General introduction to Regression and Classification (contd).
Linearity and restricted non-linearity.
|
23 Jan (Tue) {lecture#11}
|
SDR
|
MS-Teams folder:
video_linear2_23jan24.mp4,
slides_linear2_23jan24.pdf
|
|
Maximum Likelihood and Least Squares. The Moore-Penrose
Pseudo-inverse.
|
24 Jan (Wed) {lecture#12}
|
SDR
|
MS-Teams folder:
video_linear3_24jan24.mp4,
slides_linear3_24jan24.pdf,
|
|
Regularised Least Squares.
Classification: Three Approaches.
Two generalisations of linearity: restricted non-linearity.
Discriminant functions for 2 and K classes.
|
26 Jan (Fri) {lecture#13}
|
SDR
|
MS-Teams folder:
video_linear4_26jan24.mp4,
slides_linear4_26jan24.pdf,
|
|
An elegant formulation for a K-class discriminant.
Physical significance of an optimality formulation.
Fisher's Linear Discriminant.
A few furtive attempts at a derivation, followed by a constrained
optimisation approach.
|
30 Jan (Tue) {lecture#14}
|
SDR
|
MS-Teams folder:
video_linear5_30jan24.mp4,
slides_linear5_30jan24.pdf
|
|
Fisher's Linear Discriminant
|
31 Jan (Wed) {lecture#15}
|
SDR
|
[Online-only class: MS-Teams: 07:00am-08:00am]
MS-Teams folder:
video_linear6_svm1_31jan24.mp4,
slides_linear6_31jan24.pdf
|
4
|
SVMs and Kernels
|
15-23
|
SDR
|
[Bishop Chap.7],
[Alex: SVMs],
[Ng: SVMs],
[Burges: SVMs],
[Bishop Chap.6]
|
|
Introduction to SVMs: an overview.
|
31 Jan (Wed) {lecture#15}
|
SDR
|
[Online-only class: MS-Teams: 07:00am-08:00am]
MS-Teams folder:
video_linear6_svm1_31jan24.mp4,
slides_svm1_31jan24.pdf
|
|
---
|
02 Feb (Fri) {lecture#xx}
|
SDR
|
No class. Make-up class (online-only) on 26 Jan (Fri).
|
|
The basic SVM formulation (contd).
The concept of a margin. One inequality for two conditions.
The three important lines: y = 0 as the decision boundary,
and the y = +1 and y = -1 lines, and their physical
significance. An elegant formulation with a particular
formulation for the margin. Finding the `golden' regions: the
formulation in terms of a line in the `golden' region.
|
06 Feb (Tue) {lecture#16}
|
SDR
|
MS-Teams folder:
video_svm2_06feb24.mp4,
slides_svm2_06feb24.pdf,
|
|
The single-inequation characterisation of the `golden' regions.
Building a constrained optimisation formulation for the SVM.
This is one formualation which was historically the first, and
happens to be incredibly elegant. Building the primal
optimisation problem, getting the dual problem. The essence of
the kernel trick: the vaDA and the doughnut example.
|
07 Feb (Wed) {lecture#17}
|
SDR
|
MS-Teams folder:
video_svm3_08feb24.mp4,
lecture_notes_svm3_08feb24.pdf
|
|
Langange's theory for multiple equality/inequality constraints:
development, and some practical take-home points. The KKT
conditions and SVMs.
|
09 Feb (Fri) {lecture#18}
|
SDR
|
MS-Teams folder:
video_svm4_09feb24.mp4,
lecture_notes_svm4_09feb24.pdf
|
|
Plain-vanilla computation of SVM parameters from a.
The physical significance of the particular SVM formulation chosen.
The soft-margin SVM. The penalty function, and its physical
significance.
|
13 Feb (Tue) {lecture#19}
|
SDR
|
MS-Teams folder:
video_svm5_13feb24.mp4,
lecture_notes_svm5_13feb24.pdf
|
|
The physical significance of the penalty function (contd).
Taking stock.
The basic soft-margin SVM formulation.
The primal problem, the dual, and box constraints.
|
14 Feb (Tue) {lecture#20}
|
SDR
|
MS-Teams folder:
video_svm6_14feb24.mp4,
slides_svm6_14feb24.pdf
|
|
Introduction to kernels. What is a kernel? Basic properties. Why
does one use a kernel? Some examples of kernels.
Kernels in regression.
|
16 Feb (Fri) {lecture#21}
|
SDR
|
MS-Teams folder:
video_kernel1_16feb24.mp4,
lecture_notes_kernel1_16feb24.pdf
slides_kernel1_16feb24.pdf,
|
---
|
Mid-Term
|
Mid-Term: 26 Feb (Mon), 08:00am-10:00pm, LH-308, LH-310
(Please check your room and seat number
here)
|
---
|
|
|
Kernels in regression (contd).
|
04 Mar (Mon) {lecture#22}
|
SDR
|
MS-Teams folder:
video_kernel2_04mar24.mp4,
lecture_notes_kernel2_04mar24.pdf
|
|
Properties of kernels.
|
05 Mar (Tue) {lecture#23}
|
SDR
|
MS-Teams folder:
video_kernel3_nn1_05mar23.mp4,
lecture_notes_kernel3_05mar23.pdf
|
5
|
Neural Networks
|
23-xx
|
SDR
|
[Bishop Chap.5], [DL Chap.6], [DL Chap.9]
|
|
The Perceptron: A linear classifier. A non-neural connotation,
and a neural one. The history of neural networks: Rosenblatt,
Minsky, and the XOR problem.
|
05 Mar (Tue) {lecture#23}
|
SDR
|
MS-Teams folder:
video_kernel3_nn1_05mar24.mp4,
slides_nn1_05mar24.pdf
|
|
The Perceptron learning criterion. Iterative Weight Update.
A Multi-Layer Perceptron: basic structure and notations.
A note about activation functions.
|
06 Mar (Wed) {lecture#24}
|
SDR
|
MS-Teams folder:
video_nn2_06mar24.mp4,
slides_nn2_06mar24.pdf
|
|
More about activation functions.
One way of implementing an XOR function: the kernel
interpretation of the hidden layer.
|
12 Mar (Tue) {lecture#25}
|
SDR
|
video_nn3_12mar24.mp4,
slides_nn3_12mar24.pdf,
lecture_notes_nn3_12mar24.pdf
|
|
A discussion on activation functions.
A (failed) XOR attempt using regression.
|
13 Mar (Wed) {lecture#26}
|
SDR
|
MS-Teams folder:
video_nn4_13mar24.mp4,
slides_nn4_13mar24.pdf,
lecture_notes_nn4_13mar24.pdf
|
|
Another XOR implementation, this time, a completely hand-crafted
`compact' example, with a ReLU activation function.
Another piece of basic philosophy, when designing neural networks.
Yet another XOR implementation, hand-crafted again, with a
sigmoid activation function.
|
15 Mar (Fri) {lecture#27}
|
SDR
|
MS-Teams folder:
video_nn5_15mar24.mp4,
slides_nn5_15mar24.pdf,
lecture_notes_nn5_15mar24.pdf
|
|
The previous XOR implementation (contd).
The build-up to Backpropagation: factorisation basics.
The same philosophy is there in probability, and differentiation!
Partial derivatives and Taylor's series.
|
19 Mar (Tue) {lecture#28}
|
SDR
|
MS-Teams folder:
video_nn6_19mar24.mp4,
lecture_notes_nn6_19mar24.pdf
|
|
The build-up to Backpropagation: factorisation basics. The case
of 3 or more variables, and having another variable, on which all
of the previous variables depend. Extending to itself to multiple
variables. Illustration of the chain rule.
Backpropagation. Why is this needed? The basic philosophy. A
simple example. The first two steps.
|
20 Mar (Wed) {lecture#29}
|
SDR
|
MS-Teams folder:
video_nn7_20mar24.mp4
lecture_notes_nn7_20mar24.pdf
|
|
Backpropagation (contd). Alternatives.
|
22 Mar (Fri) {lecture#30}
|
SDR
|
MS-Teams folder:
video_nn8_22mar24.mp4,
lecture_notes_nn8_22mar24.pdf
|
|
A summary of some intuitive results of a neural network as a
function approximator.
Empirical observations, in going from a multi-layer perceptron to
a deep neural network. Visualising weights as an image, the
concept of local receptive fields, the significance of the first
layer of connections as local 2-D image derivative operators.
Why the term `Convolutional Neural Network' is an oxymoron.
|
23 Mar (Sat) {lecture#31}
|
SDR
|
[Online-only class: MS-Teams: 08:00am-09:00am]
MS-Teams folder:
video_nn9_23mar24.mp4,
lecture_notes_nn9_23mar24.pdf
|
|
Why does one want to relate the inner product `correlation'
operation to convolution? The well-accepted Linear Shift
Invariant Systems theory in Electrical Engineering, which carries
over to other disciplines as well. Examples.
|
24 Mar (Sun) {lecture#32}
|
SDR
|
[Online-only class: MS-Teams: 08:00am-09:00am]
MS-Teams folder:
video_nn10_24mar24.mp4,
lecture_notes_nn10_24mar24.pdf
|
|
The physical significance of Linear Shift Invariant (LSI)
systems, and why Convolution is a fundamental operation for LSI
systems.
Characteristics of Deep Neural Networks: Local Receptive Fields,
Strided Convolutions, Weight/Parameter Sharing.
|
02 Apr (Tue) {lecture#33}
|
SDR
|
MS-Teams folder:
video_nn11_02apr24.mp4,
slides_nn11_02apr24.pdf,
lecture_notes_nn11_02apr24.pdf
|
|
Another feature of deep networks: pooling.
LeNet-5 (1989): the first two layers.
|
03 Apr (Wed) {lecture#34}
|
SDR
|
MS-Teams folder:
video_nn12_03apr24.mp4,
lecture_notes_nn12_03apr24.pdf
|
|
LeNet-5 (1989) (contd).
|
05 Apr (Fri) {lecture#35}
|
SDR
|
MS-Teams folder:
video_nn13_05apr24.mp4,
lecture_notes_nn13_05apr24.pdf
|
|
More features of deep networks: pooling,
local response normalisation (AlexNet, 2012), Batch Normalisation (2015),
Residual Connections (ResNet, 2015), Dropout (AlexNet, 2012).
|
09 Apr (Tue) {lecture#36}
|
SDR
|
MS-Teams folder:
video_nn14_09apr24.mp4,
lecture_notes_nn14_09apr24.pdf
|
|
AlexNet and CaffeNet (2012): some details. The second prominent
deep learning architecture.
|
10 Apr (Wed) {lecture#37}
|
SDR
|
MS-Teams folder:
video_nn15_10apr24.mp4,
lecture_notes_nn15_10apr24.pdf
|
|
AlexNet and CaffeNet (2012) (contd).
VGG16 and VGG19 (2014): simpler constant-sized convolutions leads
to fewer parameters, covering the same area of interest (albeit
in a filtered way). 16-19 layers in place of the 5 of AlexNet.
|
12 Apr (Fri) {lecture#38}
|
SDR
|
MS-Teams folder:
video_nn16_12apr24.mp4,
lecture_notes_nn16_12apr24.pdf
|
|
---
|
16 Apr (Tue) {lecture#xx}
|
SDR
|
No class. Make-up class (online-only) on 23 Mar (Sat).
|
|
The VGG architecture parameter saving examples.
Motivation for ResNet.
The Inception architecture.
|
19 Apr (Fri) {lecture#39}
|
SDR
|
MS-Teams folder:
video_nn17_19apr24.mp4, lecture_notes_nn17_19apr24.pdf
|
|
---
|
20 Apr (Sat) {lecture#xx}
|
SDR
|
No class. Make-up class (online-only) on 21 Apr (Sun).
|
|
A gentle introduction to GANs.
|
21 Apr (Sun) {lecture#40}
|
SDR
|
[Online-only class: MS-Teams: 08:00am-09:00am]
MS-Teams folder:
video_nn18_21apr24.mp4,
slides_nn18_21apr24.pdf
|
|
A gentle introduction to GANs (contd).
Recurrent Neural Networks.
|
23 Apr (Tue) {lecture#41}
|
SDR
|
MS-Teams folder:
video_nn19_23apr24.mp4,
slides_nn19_23apr24.pdf,
lecture_notes_nn19_23apr24.pdf
|
|
Recurrent Neural Networks (contd).
|
24 Apr (Wed) {lecture#42}
|
SDR
|
MS-Teams folder:
video_nn20_24apr24.mp4, lecture_notes_nn20_24apr24.pdf
|
|
--
|
27 Apr (Sat) {lecture#xx}
|
SDR
|
No class. Make-up class (online-only) on 24 Mar (Sun).
|
---
|
End-Term
|
01 May (Wed) 08:00am-10:00am, LH-310
(Please check your seat number
here)
|
---
|
---
|
xx
|
Mathematical Basics for Machine Learning
|
xx-xx
|
xx
|
[Burges: Math for ML],
[Do,
Kolter: Linear Algebra Notes],
|