CS261 Deep Generative Models - Understanding Deep Learning
Table of contents
The textbook is understanding deep learning. https://udlbook.github.io/udlbook/
CS261 Generative Model - Understanding Deep Learning
- Preface
- Acknowledgements
- 1 Introduction
- 2 Supervised learning
- 3 Shallow neural networks
- 4 Deep neural networks
- 5 Loss functions
- 6 Fitting models
- 7 Gradients and initialization
- 8 Measuring performance
- 9 Regularization
- 10 Convolutional networks
- 10 Convolutional networks (continued)
- 11 Residual networks
- 12 Transformers
- 12.1 Processing text data
- 12.2 Dot-product self-attention
- 12.3 Extensions to dot-product self-attention
- 12.4 Transformers
- 12.5 Transformers for natural language processing
- 12.6 Encoder model example: BERT
- 12.7 Decoder model example: GPT-3
- 12.8 Encoder-decoder model example: machine translation
- 12.9 Transformers for long sequences
- 12.10 Transformers for images
- 12.11 Summary
- 13 Graph neural networks
- 13.1 What is a graph?
- 13.2 Graph representation
- 13.3 Graph neural networks, tasks, and loss function
- 13.4 Graph convolutional networks
- 13.5 Example: graph classification
- 13.6 Inductive vs. transductive modeling
- 13.7 Example: node classification
- 13.8 Layers for graph convolutional networks
- 13.9 Edge graphs
- 13.10 Summary
- 14 Unsupervised learning
- 15 Generative Adversarial Networks
- 15 Generative Adversarial Networks (continued)
- 16 Normalizing flows
- 17 Variational autoencoders
- 18 Diffusion models
- 19 Reinforcement learning
- 20 Why does deep learning work?
- 21 Deep learning and ethics
- Appendices
- A Notation
- B Mathematics
- C Probability
- Bibliography
- Index