University of Cambridge > Talks.cam > Isaac Newton Institute Seminar Series > Testing and improving the robustness of amortized Bayesian inference for cognitive models

Testing and improving the robustness of amortized Bayesian inference for cognitive models

Add to your list(s) Download to your calendar using vCal

If you have a question about this talk, please contact nobody.

RCLW03 - Accelerating statistical inference and experimental design with machine learning

Author: Yufei Wu, Stefan Radev, Francis Tuerlinckx This paper addresses the challenge of achieving robust and reliable parameter estimation in cognitive models when contaminations and outliers are present. Studies have demonstrated that amortized Bayesian inference (ABI) has the potential to facilitate inference of arbitrarily complex cognitive models via simulations and neural networks. However, it remains unclear whether ABI is robust to outliers and how it can be robustified. Our findings show that by simulating a more realistic data-generation process that incorporates contaminants and using the contaminated data to train the neural networks, the inference with ABI achieves a high degree of robustness while maintaining efficiency. These results are significant because they provide a powerful yet practical solution to handle outliers in empirical research with ABI .

This talk is part of the Isaac Newton Institute Seminar Series series.

Tell a friend about this talk:

This talk is included in these lists:

Note that ex-directory lists are not shown.

 

© 2006-2025 Talks.cam, University of Cambridge. Contact Us | Help and Documentation | Privacy and Publicity