Project IV 2025-2026

How Large Language Models Reason

Description

While Large Language Models (LLMs) dazzle us with their human-like reasoning—from cracking puzzles to constructing logical arguments—the question remains: "How exactly do they reason?". In this project students will uncover the sophisticated reinforcement learning techniques and other approaches, wich are fueling the reasoning in LLM models. Students will engage critically with core ideas, experiment with prompting methods, explore reinforcement learning from human feedback, and deepen their understanding of both theoretical principles and practical AI innovations driving today's improvements.

Prerequisites

Strong knowledge of Python. Familiarity with basic probability theory.

Some background material