A focused tutorial introducing the foundations of LLM alignment and providing a conceptual and practical understanding of values, safety, reasoning, and pluralism. Through intuitive explanations, worked examples, and case studies, we explore how to reason about alignment goals and critically evaluate existing methods.
Macquarie University, Australia
Macquarie University, Australia
Macquarie University, Australia
Macquarie University, Australia
Macquarie University, Australia
Macquarie University, Australia
Macquarie University, Australia
Large Language Models (LLMs) have demonstrated remarkable capabilities, yet their reliability and alignment with human expectations remain unresolved challenges. This tutorial introduces the foundations of alignment and provides participants with a conceptual and practical understanding of the field. We cover how values, safety, reasoning, and pluralism shape alignment objectives, using real-world examples and case studies. The goal is to help attendees reason about alignment goals, understand the behaviour of existing methods, and critically evaluate their strengths and limitations.
The tutorial is accessible without prior alignment experience, though familiarity with core ML and NLP concepts is helpful. No coding is required — all examples are presented through visual illustrations and guided reasoning.
Setting the stage: why alignment matters for LLMs, tutorial goals, and a quick tour of the concepts we will cover across values, safety, reasoning, and pluralism.
Human preference modelling, value alignment, and alignment objectives. We discuss how reward models, instruction tuning, and role prompts encode human preferences—and where these approaches break down.
Safety taxonomies and concrete failure modes (harmful content, hallucinations, misuse). Practical safety-alignment methods and trade-offs between helpfulness and harm reduction.
How to align LLMs for diverse users and contexts. Cultural norms, pluralistic value sets, role-driven prompts, and cases where a single global alignment target is insufficient.
Summary of practical lessons, recommended design patterns for real-world alignment, and open research challenges for attendees who want to explore the field further.
Macquarie University, Australia
📧 usman.naseem@mq.edu.au
School of Computing, Macquarie University, Sydney, NSW 2113, Australia