University of Cambridge > Talks.cam > Isaac Newton Institute Seminar Series > Assessing the calibration of multi-horizon, probabilistic forecasts

Assessing the calibration of multi-horizon, probabilistic forecasts

Add to your list(s) Download to your calendar using vCal

If you have a question about this talk, please contact nobody.

RCLW02 - Calibrating prediction uncertainty : statistics and machine learning perspectives

We propose a new method for assessing calibration based on revisions of probabilistic forecasts. Such revisions refer to how forecasts for an outcome change as the forecasts are updated over time. We begin by extending the definition of a forecast revision from point forecasts to probabilistic forecasts. Then we show that if the forecasts are well calibrated (efficient) then the revisions form a sequence of independent uniform random variables. This motivates new statistical tests to detect miscalibration.

This talk is part of the Isaac Newton Institute Seminar Series series.

Tell a friend about this talk:

This talk is included in these lists:

Note that ex-directory lists are not shown.

 

© 2006-2025 Talks.cam, University of Cambridge. Contact Us | Help and Documentation | Privacy and Publicity