Post

Programme 2025

Programme 2025

Overview

The Oxford LLMs Programme is a five-day workshop that combines lectures, coding sessions, collaborative projects, and guest talks. Participants learn how large language models work, how to adapt them to new domains, how to evaluate them, and how to deploy them safely and fairly.

Daily Structure

Each day follows the same four-part structure:

  1. Lecture (90 min): Core concepts and methods
  2. Coding Session (90 min): Hands-on exercises using Python and Jupyter notebooks
  3. Collaborative Research (120 min): Small teams work on a real research question, aiming for a peer-review write-up
  4. Guest Talk (60 min): A social scientist presents their LLM research and answers questions

Main Topics

1) Foundations of LLMs

In this module, participants will explore training objectives for language models. They will dive into transformer architecture, examining how self-attention enables models to process and generate coherent text. Finally, the session covers in context-learning and model reasoning, demonstrating how large models can perform complex tasks without explicit fine-tuning.

2) LLM Post-Training

This part focuses on post-training methods, such as continued pre-training, fine-tuning, and reinforcement learning to specialise models for target tasks. Participants will discuss data preparation, understanding how dataset quality impacts performance. The module concludes with an overview of frameworks for efficient model updates.

3) Evaluating LLMs

Here, we examine classic tests, including Turing-style probes, to assess basic language understanding. Next, we cover benchmarks for generalization and robustness. The session ends with metrics for real-world performance and reliability, teaching how to measure model behavior on downstream tasks and in production scenarios.

4) Safety and Fairness in LLMs

Participants will learn bias detection and mitigation techniques. We explore alignment strategies that steer models toward desired behaviors. The module wraps up with a discussion of ethical and societal considerations, highlighting best practices for responsible deployment.

What You Will Gain

  • A clear understanding of transformer-based LLMs
  • Practical experience with main tools and libraries
  • Hands-on research skills in computational social science
  • Insight into current LLM applications in political science, sociology, history, economics, and related fields

Preparation

Participants should be ready to code in Python and have familiarity with basic NLP concepts. All materials, datasets, and example code will be provided ahead of time.

For anyone who needs a refresher on the basics, we also provide a short self-study pack of readings and coding exercises to complete before the workshop.

This post is licensed under CC BY 4.0 by the author.