Natural Language Processing with Deep Learning
What is this course about?
Natural language processing (NLP) is one of the most important technologies of the information age, and a crucial part of artificial intelligence. Applications of NLP are everywhere because people communicate almost everything in language: web search, advertising, emails, customer service, language translation, virtual agents, medical reports, etc. In recent years, Deep Learning approaches have obtained very high performance across many different NLP tasks, using single end-to-end neural models that do not require traditional, task-specific feature engineering. In this course, students will gain a thorough introduction to cutting-edge research in Deep Learning for NLP. Through lectures, assignments and a final project, students will learn the necessary skills to design, implement, and understand their own neural network models. This year, CS224n will be taught for the first time using PyTorch rather than TensorFlow (as in previous years).
Schedule
Lecture slides will be posted here shortly before each lecture. If you wish to view slides further in advance, refer to last year's slides, which are mostly similar.
The lecture notes are updated versions of the CS224n 2017 lecture notes (viewable here) and will be uploaded a few days after each lecture. The notes (which cover approximately the first half of the course content) give supplementary detail beyond the lectures.
This schedule is subject to change.
Date | Description | Course Materials | Events | Deadlines |
---|---|---|---|---|
Tue Jan 8 | Introduction and Word Vectors
[slides] [video] [notes] Gensim word vectors example: [code] [preview] |
Suggested Readings:
|
Assignment 1 out
[code] [preview] |
|
Thu Jan 10 | Word Vectors 2 and Word Senses
[slides] [video] [notes] |
Suggested Readings: Additional Readings: | ||
Fri Jan 11 | Python review session
[slides] |
1:30 - 2:50pm Skilling Auditorium [map] |
||
Tue Jan 15 | Word Window Classification, Neural Networks, and Matrix Calculus
[slides] [video] [matrix calculus notes] [notes (lectures 3 and 4)] |
Suggested Readings: Additional Readings: |
Assignment 2 out
[code] [handout] |
Assignment 1 due |
Thu Jan 17 | Backpropagation and Computation Graphs
[slides] [video] [notes (lectures 3 and 4)] |
Suggested Readings: | ||
Tue Jan 22 | Linguistic Structure: Dependency Parsing
[slides] [scrawled-on slides] [video] [notes] |
Suggested Readings: | Assignment 3 out
[code] [handout] |
Assignment 2 due |
Thu Jan 24 | The probability of a sentence? Recurrent Neural Networks and Language Models
[slides] [video] [notes (lectures 6 and 7)] |
Suggested Readings:
|
||
Tue Jan 29 | Vanishing Gradients and Fancy RNNs
[slides] [video] [notes (lectures 6 and 7)] |
Suggested Readings:
|
Assignment 4 out
[code] [handout] [Azure Guide] [Practical Guide to VMs] |
Assignment 3 due |
Thu Jan 31 | Machine Translation, Seq2Seq and Attention
[slides] [video] [notes] |
Suggested Readings:
|
||
Tue Feb 5 |
Practical Tips for Final Projects
[slides] [video] [notes] |
Suggested Readings:
|
||
Thu Feb 7 | Question Answering and the Default Final Project [slides] [video] [notes] |
Project Proposal out
[instructions] Default Final Project out [handout] [code] |
Assignment 4 due | |
Tue Feb 12 | ConvNets for NLP [slides] [video] [notes] |
Suggested Readings: | ||
Thu Feb 14 | Information from parts of words: Subword Models
[slides] [video] |
Suggested readings:
|
Assignment 5 out
[original code (requires Stanford login) / public version] [handout] |
Project Proposal due |
Tue Feb 19 | Modeling contexts of use: Contextual Representations and Pretraining
[slides] [video] |
Suggested readings:
|
||
Thu Feb 21 |
Transformers and Self-Attention For Generative Models
(guest lecture by Ashish Vaswani and Anna Huang) [slides] [video] |
Suggested readings: | ||
Fri Feb 22 |
Project Milestone out
[instructions] |
Assignment 5 due | ||
Tue Feb 26 |
Natural Language Generation
[slides] [video] |
|||
Thu Feb 28 | Reference in Language and Coreference Resolution
[slides] [video] |
|||
Tue Mar 5 | Multitask Learning: A general model for NLP? (guest lecture by Richard Socher)
[slides] [video] |
Project Milestone due | ||
Thu Mar 7 |
Constituency Parsing and Tree Recursive Neural Networks
[slides] [video] [notes] |
Suggested Readings: | ||
Tue Mar 12 |
Safety, Bias, and Fairness (guest lecture by Margaret Mitchell)
[slides] [video] |
|||
Thu Mar 14 |
Future of NLP + Deep Learning
[slides] [video] |
|||
Sun Mar 17 | Final Project Report due [instructions] | |||
Wed Mar 20 | Final project poster session
[details] |
5:15 - 8:30pm McCaw Hall at the Alumni Center [map] |
Project Poster/Video due [instructions] |