Deep Learning Theory

About the Course

In recent years, deep learning has made great progress. It is widely used in many fields, such as computer vision, speech recognition, natural language processing, bioinformatics, etc., and also has a significant impact on basic science fields, such as applied and computational mathematics, statistical physics, computational chemistry and materials science, life science, etc. However, it should be noted that deep learning as a black-box model, its ability is explored through a large number of experiments. The theory of deep learning has gradually attracted the attention of many researchers, and has made progress in many aspects.

This course is closely around the latest development of deep learning theory. It intends to teach mathematical models, theories, algorithms and numerical experiments related to a series of basic problems from multiple perspectives of deep learning theory. This course is designed for doctoral, postgraduate and senior undergraduate students in all majors, who have basic knowledge of machine learning and deep learning.

The topics and the corresponding material are as follows:

  1. Introduction to Deep Learning material
  2. Approximation Theory of Neural Networks material
  3. Algorithmic Regularization material
  4. Inductive Biases due to Algorithmic Regularization material
  5. Tractable Landscapes for Nonconvex Optimization material
  6. Neural Tangent Kernel material
  7. Multilayer Convolutional Sparse Coding material
  8. Information Bottleneck and Others material
  9. The Fragility of Neural Networks material
  10. Neural ODEs material
  11. Neural Networks Learn the Geodesic Curve in the Wasserstein Space material
  12. Generative Model:GAN, VAE and Roundtrip material
  13. Interpretability of Neural Networks, Frequency principle material

Prerequisites

Mathematical Analysis, Linear Algebra, Mathematical Statistics, Numerical Optimization, Matrix Theory, Fundamentals of Machine Learning and Deep Learning

Introduction to Deep Learning

Approximation Theory of Neural Networks

Algorithmic Regularization

Inductive Biases due to Algorithmic Regularization

Tractable Landscapes for Nonconvex Optimization

Neural Tangent Kernel

Multilayer Convolutional Sparse Coding

Information Bottleneck and Others

The Fragility of Neural Networks

Neural ODEs

Neural Networks Learn the Geodesic Curve in the Wasserstein Space

Generative Model:GAN, VAE and Roundtrip

Interpretability of Neural Networks, Frequency Principle