The course is devoted to modern generative models in the application to computer vision.
We will study 4 main classes of generative models:
- autoregressive models,
- latent variable models,
- normalization flow models,
- adversarial models.
Special attention is paid to the properties of various classes of generative models, their interrelationships, theoretical prerequisites and methods of quality assessment.
The aim of the course is to introduce the student to widely used advanced methods of deep learning.
The course is accompanied by practical tasks that allow you to understand the principles of the considered models.
Lecture | Date | Description | Lecture | Seminar |
---|---|---|---|---|
1 | February, 11 | Logistics. Motivation. Autoregressive models (MADE, WaveNet, PixelCNN, PixelCNN++). | slides | slides |
2 | February, 18 | Bayesian Framework. Latent Variable Models. Variational lower bound. EM-algorithm. | slides | slides |
3 | February, 25 | ELBO, Reparametrization trick, Variational Autoencoder. MLE vs MAP. VAE drawbacks. | slides | slides |
4 | March, 5 | Mean-field approximation and EM-algorithm. Flow models definition. Forward and reverse KL divergence. | slides | slides |
5 | March, 11 | Flow models (Planar flows, NICE, RealNVP, Glow). Flows in variational inference. | slides | slides |
6 | March, 18 | Autoregressive flows (MAF, IAF). Flow KL duality. Uniform dequantization. | slides | slides |
7 | March, 25 | Variational dequantization. IWAE. ELBO surgery. | slides | slides |
8 | April, 1 | VampPrior + Autoregressive prior. Posterior collapse. Disentanglement learning (beta-VAE). | slides | slides |
9 | April, 8 | Disentanglement learning (DIP-VAE + summary). Likelihood-free learning. GAN theorem. | slides | slides |
10 | April, 15 | Gan problems: vanishing gradients + mode collapse. KL vs JSD. DCGAN. Wasserstein GAN. | slides | slides |
11 | April, 22 | WGAN-GP. Spectral Normalization GAN. f-divergence minimization. GAN evaluation (Inception score, FID) | slides | slides |
12 | April, 29 | GAN evaluation (Precision-Recall). GAN models (Self-Attention GAN, BigGAN, PGGAN, StyleGAN). AVB. | slides | slides |
13 | May, 6 | Neural ODE. Continuous-in-time NF (FFJORD). Discrete VAE (Gumbel-Softmax trick, VQ-VAE, VQ-VAE-2, DALL-E). | slides | slides |
Homework | Date | Deadline | Description | Link |
---|---|---|---|---|
1 | February, 11 | February, 28 | Hostogram + MADE (practice). | |
2 | February, 28 | March, 14 | Autoregressive models + LVM (theory). | |
3 | March, 14 | March, 28 | VAE + RealNVP (practice). | |
4 | March, 28 | April, 11 | AR flow + VAE with AR prior (practice). | |
5 | April, 11 | April, 25 | Sylvester Flow + ELBO MI + IW dequantization + LSGAN (theory). | |
6 | April, 25 | May, 9 | GAN + WGAN-GP. |