MENU

Fun & Interesting

MIT 6.S191 (2024): Recurrent Neural Networks, Transformers, and Attention

Alexander Amini 282,707 11 months ago
Video Not Working? Fix It Now

MIT Introduction to Deep Learning 6.S191: Lecture 2 Recurrent Neural Networks Lecturer: Ava Amini 2024 Edition For all lectures, slides, and lab materials: http://introtodeeplearning.com Lecture Outline 0:00​ - Introduction 3:42​ - Sequence modeling 5:30​ - Neurons with recurrence 12:20 - Recurrent neural networks 14:08 - RNN intuition 17:14​ - Unfolding RNNs 19:54 - RNNs from scratch 22:41 - Design criteria for sequential modeling 24:24 - Word prediction example 31:50​ - Backpropagation through time 33:40 - Gradient issues 37:15​ - Long short term memory (LSTM) 40:00​ - RNN applications 44:00- Attention fundamentals 46:46 - Intuition of attention 49:13 - Attention and search relationship 51:22 - Learning attention with neural networks 57:45 - Scaling attention and applications 1:00:08 - Summary Subscribe to stay up to date with new deep learning lectures at MIT, or follow us @MITDeepLearning on Twitter and Instagram to stay fully-connected!!

Comment