Generative Modelling for Natural Language Processing

Table of Contents

Instructors

Syllabus


Session 1: Word Embeddings

Lecture and Lab on 09/01/2026

References:

  • Advances in Pre-Training Distributed Word Representations (Mikolov, Tomas and Grave, Edouard and Bojanowski, Piotr and Puhrsch, Christian and Joulin, Armand, 2018) pdf
  • Neural Word Embedding as Implicit Matrix Factorization (Levy, Omer and Goldberg, Yoav, 2014) pdf
  • Closed-Form Training Dynamics Reveal Learned Features and Linear Structure in Word2vec-Like Models (Karkada, Dhruva and Simon, James B. and Bahri, Yasaman and DeWeese, Michael R., 2025)

Session 2: Language Models

Lecture/Lab on 16/01/2026


Session 3: Attention in RNN and Transformer Language Models

Lecture/Lab on 23/01/2026


Session 4: Chatbots fine-tuning RHLF and DPO

Lecture/Lab on 30/01/2026

References:

  • Reinforcement Learning from Human Feedback (Nathan Lambert, 2024)
  • Reinforcement Learning: An Introduction (Sutton, Richard S. and Barto, Andrew G., 2018 )
  • Learning to summarize from human feedback (Nisan Stiennon and Long Ouyang and Jeff Wu and Daniel M. Ziegler and Ryan Lowe and Chelsea Voss and Alec Radford and Dario Amodei and Paul F. Christiano, 2020)
  • Simple statistical gradient-following algorithms for connectionist reinforcement learning (Williams, R. J., 1992)
  • Direct Preference Optimization: Your Language Model is Secretly a Reward Model (Rafael Rafailov and Archit Sharma and Eric Mitchell and Christopher D Manning and Stefano Ermon and Chelsea Finn, 2023)

Session 5: Deep Latent Variable Models for Texts

Lecture/Lab on 06/02/2026


Exam 27/02/2026


References

  • Advances in Pre-Training Distributed Word Representations (Mikolov, Tomas and Grave, Edouard and Bojanowski, Piotr and Puhrsch, Christian and Joulin, Armand, 2018) pdf
  • Neural Word Embedding as Implicit Matrix Factorization (Levy, Omer and Goldberg, Yoav, 2014) pdf
  • Reinforcement Learning from Human Feedback (Nathan Lambert, 2024)
  • Reinforcement Learning: An Introduction (Sutton, Richard S. and Barto, Andrew G., 2018 )
  • Learning to summarize from human feedback (Nisan Stiennon and Long Ouyang and Jeff Wu and Daniel M. Ziegler and Ryan Lowe and Chelsea Voss and Alec Radford and Dario Amodei and Paul F. Christiano, 2020)
  • Simple statistical gradient-following algorithms for connectionist reinforcement learning (Williams, R. J., 1992)
  • Direct Preference Optimization: Your Language Model is Secretly a Reward Model (Rafael Rafailov and Archit Sharma and Eric Mitchell and Christopher D Manning and Stefano Ermon and Chelsea Finn, 2023)

Author: root

Created: 2026-01-09 Fri 12:44