236759: Generative AI – Diffusion Models (Winter 2024/2025)

********

Note: This course has been shifted to the 2025′ Spring Semester

********

 

 

Generative AI – Diffusion Models (236759)

Winter 2024/2025

Last update: August 12th 2024

General Comment: Read this page carefully – we have changed the format of the course, and its grading policy

Lecturer:

  • Prof. Michael Elad (elad@cs.technion.ac.il)

Teachning Assistant:

  • Noam Elata (noamelata@campus.technion.ac.il)

Credit:

  • 2 Points

Time and Place: 

  • Day and Time: TBD
  • The meetings will be virtual/frontal: TBD
  • Class location: TBD

Format: 

  • Registration should be done directly with the secretariat (and not through the lecturer)
  • This course might be given in English (depends on the participants)
  • The course is  open to both undergraduate and graduate students, as long as you have the pre-requisites.

Pre-requisites:

  • 236200 or 236201 (Introduction to Data Processing and Representation) or 046200 (Processing and Analysis of Images). Alternative courses on computer vision or image processing are sufficient as well.
  • You should have some background in deep-learning, either through an advanced course on this topic or a project.

Parallel Courses:

  • 048954 (Statistical Methods in Image Processing) is NOT considered as parellel, but students should avoid double-deeping their projects

Course Description:

A fascinating topic in deep learning in the last decade deals with the creation of information from `thin-air’, such as the synthesis of images, video clips, creation of music clips and more. For the most part, this topic is discussed in conjunction with learning machines called GANs, although other algorithmic tools have also been harnessed for this task, such as VAE, energy-based methods, normalized flow, and more. A real revolution is taking place these days with the introduction of diffusion methods as an alternative to all of the above. These methods are based on an iterative approach in which a Gaussian noise vector is gradually converted into a vector from the expected distribution law, and this is done with the help of the score function of the distribution law. As it turns out, this function can be approximated very accurately by a plain image denoiser – a simple algorithm designed to remove Gaussian noise from an image. This makes diffusion methods accessible and easy to train and activate. In this course, we intend to review this approach of diffusion methods, by going over the mathematical foundations underlying the proposed algorithms, familiarization with a variety of diffusion methods, various applications that rely on this technique for solving inversion problems, conditional sampling, and more.

Course Structure:

  • After few (6-7) lectures by the lecturer, we will concentrate on students’ lectures (~20 minutes each) on their assigned projects, covering various recent papers in this domain.
  • Each student participating in the course will be assigned with one paper. The project itself will include reading the paper and understanding it’s content, preparing a slide show to present it, and then presenting these slides to the class.

Grading Policy:

  • 50% of the grade will be dictated by the quality of the presentation created and the lecture given.
  • 50% of the grade will be dictated by a final exam.

Choosing a Project’s Paper:

  • Please choose your paper for the project from THIS LIST (not updated)
  • Projects can be done in singles. If the number ofparticipants istoo high, we will allow working in pairs.
  • Note: You have to notify the teaching Noam of the paper you have chosen. If it has been already taken, you will have to choose again

Tentative Syllabus: 

1. Introduction
– General Description and Administration
– Review of Basic Mathematical Tool

2. Background
– A Prior for Images–How and Why
– Evolution of Priors in Image Processing–Classical Era
– Priors in Image Processing: The Era of Deep-Learning

3. Introduction to Diffusion
– Say Hello to the Score Function
– Image Denoisers
– RED & PnP
– Langevin Dynamics
– Diffusion Models–Introduction

4. Alternative Diffusion Models
– DDPM: Forward Path
– DDPM: Reverse Path
– DDPM: The Formally Introduced Reverse Path
– Related Topic: Probability estimation

5. Acceleration Methods
– Denoising Diffusion GANs
– Nested Diffusion
– Denoising Diffusion Implicit Models
– Other Acceleration methods

5. Guided Diffusion
– The Concept of Guidance
– Classifier Guidance
– Classifier Free Guidance

6. Diffusion for Inverse Problems
– Inverse Problems (IP)–Some Fundamentals
– Diffusion Models for IP-Conceptual Solutions
– Diffusion Models for IP–Diving into The Bayesian Approach
– Recent Work of Relevance

7. Diffusion: Applications [optional]

Preliminary Resources:

People to Follow:

  • Stefano Ermon (Stanford CS)
  • Yang Song (OpenAI)
  • Arash Vahdat (NVIDIA)
  • Diederick P. Kingma (Google)
  • Prafulla Dhariwal (OpenAI)
  • Tim Salimans (Google)

Past Video Recordings (from 2023/2024 semester):

  • The 1st meeting – A general description of the course, a review of basic mathematical tools, The importance of  the prior in image processing
  • The 2nd meeting – Continuing the background chapter, describing the deep-learning era with a focus on image samplers
  • The 3rd meeting – Introducing the score-function, its relation to denoisers, and then moving to the Langevin dynamics and diffusion models relying on it
  • The 4th meeting – Denoising Diffusion Probabilistic Models (DDPM) – a thorough derivation of the reverse path, and likelihood evaluations
  • The 5th meeting – Acceleration methods – DDIM, Denoising Diffusion GANs, Nested Diffusion and more (multi-scale, consistency, latent diffusion)
  • The 6th meeting – Guided Diffusion via Classifier- and Classifier-Free Guidance, introduction to Text-2-Image methods and diving in on IMAGEN
  • The 7th meeting – Solving inverse problems with diffusion models – the Bayesian approach and alternatives
  • The 8th meeting – SDE/ODE interpretation of diffusion models (by Noam Elata) and a lecture by students on “Inverse Heat Diffusion”
  • The 9th meeting – 4 students’ presentations
  • The 10th meeting – 4 students’ presentations
  • The 11th meeting – 5 students’ presentations
  • The 12th meeting – 3 students’ presentations