Talks.cam will close on 1 July 2026, further information is available on the UIS Help Site
 

University of Cambridge > Talks.cam > Isaac Newton Institute Seminar Series > Estimating and Calibrating Uncertainty in LLMs

Estimating and Calibrating Uncertainty in LLMs

Add to your list(s) Download to your calendar using vCal

If you have a question about this talk, please contact nobody.

RCLW04 - Early Career Pioneers in Uncertainty Quantification and AI for Science

Large Language Models (LLMs) often sound certain when they are wrong. With rapid adoption in critical sectors like healthcare, scientific discovery and research automation, where the consequences of errors can be substantial, reliable Uncertainty Quantification (UQ) becomes crucial for safe and trustworthy deployment. In this talk, we will first review the current landscape of UQ methods for LLMs and discuss the trade-offs between calibration, discrimination, and compute. We’ll then look at how generation temperature (the scalar that controls generative diversity) affects calibration at a semantic level. We’ll introduce a semantic calibration framework and show that simple post-hoc temperature scaling markedly improves meaning-level calibration and selective prediction across open- and closed-book question-answering tasks.

This talk is part of the Isaac Newton Institute Seminar Series series.

Tell a friend about this talk:

This talk is included in these lists:

Note that ex-directory lists are not shown.

 

© 2006-2025 Talks.cam, University of Cambridge. Contact Us | Help and Documentation | Privacy and Publicity