Table of contents
Schedule
Course Overview & Language Modeling Basics (August 26)
Slides:
Transformers / Pretraining / Finetuning
Reading Material
Attention is all you need (2017) [link]
BERT, Pre-training of Deep Bidirectional Transformers for Language Understanding [link]
Optional readings:
The Illustrated Transformer [link]
The Illustrated BERT, ELMo, and co. (How NLP Cracked Transfer Learning) [link]
T5 [link]
The Illustrated GPT2 [link]
What Language Model Architecture and Pretraining Objective Work Best for Zero-Shot Generalization? [link]
BART [link]
RoBERTa, A Robustly Optimized BERT Pretraining Approach [link]