Research Cluster: Tech Accountability | Adam Tauman Kalai | Why Language Models Hallucinate

Date: 
Sun, 28/12/202510:30
Adam Tauman Kalai

Research Cluster: Tech Accountability

 

Lecturer: 

Dr. Adam Tauman Kalai (OpenAI)

Title: 

Why Language Models Hallucinate

Abstract: 

Large language models sometimes generate statements that are plausible but factually incorrect—a phenomenon commonly called "hallucination." We argue that these errors are not mysterious failures of architecture or reasoning, but rather predictable consequences of standard training and evaluation incentives.

We show (i) that hallucinations can be viewed as classification errors: when pretrained models cannot reliably distinguish a false statement from a true one, they may produce the false option rather than saying I don't know; (ii) that optimization of benchmark performance encourages guessing rather than abstaining, since most evaluation metrics penalize expressing uncertainty; and (iii) that a possible mitigation path lies in revising existing benchmarks to reward calibrated abstention, thus realigning incentives in model development.

Joint work with Santosh Vempala (Georgia Tech) and Ofir Nachum & Edwin Zhang (OpenAI).

Short Bio:

Adam Tauman Kalai is a Research Scientist at OpenAI, specializing in AI Safety and Ethics. His research interests also include algorithms, AI theory, and game theory. Adam earned his BA from Harvard University and his PhD from Carnegie Mellon University, after which he served as an Assistant Professor at TTIC and Georgia Tech and a Senior Principal Researcher at Microsoft Research New England. He is also a member of Project CETI’s science team and a recipient of the Majulook prize.

Location: 

130 Hall, Feldman Building, Edmond J. Safra campus.