MENU

Fun & Interesting

NLP Demystified 13: Recurrent Neural Networks and Language Models

Future Mojo 11,378 3 years ago
Video Not Working? Fix It Now

Course playlist: https://www.youtube.com/playlist?list=PLw3N0OFSAYSEC_XokEcX8uzJmEZSoNGuS We'll learn how to get computers to generate text through a technique called recurrence. We'll also look at the weaknesses of the bag-of-words approaches we've seen so far, how to capture the information in word order, and in the demo, we'll build a part-of-speech tagger and text-generating language model. Colab notebook: https://colab.research.google.com/github/futuremojo/nlp-demystified/blob/main/notebooks/nlpdemystified_recurrent_neural_networks.ipynb Timestamps 00:00:00 Recurrent Neural Networks 00:00:23 The problem with bag-of-words techniques 00:02:28 Using recurrence to process text as a sequence 00:07:53 Backpropagation with RNNs 00:12:03 RNNs vs other sequence processing techniques 00:13:08 Introducing Language Models 00:14:37 Training RNN-based language models 00:17:40 Text generation with RNN-based language models 00:19:44 Evaluating language models with Perplexity 00:20:54 The shortcomings of simple RNNs 00:22:48 Capturing long-range dependencies with LSTMs 00:27:20 Multilayer and bidirectional RNNs 00:29:58 DEMO: Building a Part-of-Speech Tagger with a bidirectional LSTM 00:42:22 DEMO: Building a language model with a stacked LSTM 00:58:04 Different RNN setups This video is part of Natural Language Processing Demystified --a free, accessible course on NLP. Visit https://www.nlpdemystified.org/ to learn more.

Comment