COOKIES: By using this website you agree that we can place Google Analytics Cookies on your device for performance monitoring. |
University of Cambridge > Talks.cam > Artificial Intelligence Research Group Talks (Computer Laboratory) > Towards Robust and Reliable Model Explanations
Towards Robust and Reliable Model ExplanationsAdd to your list(s) Download to your calendar using vCal
If you have a question about this talk, please contact Mateja Jamnik. As machine learning black boxes are increasingly being deployed in domains such as healthcare and criminal justice, there is growing emphasis on building tools and techniques for explaining these black boxes in an interpretable manner. Such explanations are being leveraged by domain experts to diagnose systematic errors and underlying biases of black boxes. In this talk, I will present some of our recent research that sheds light on the vulnerabilities of popular post hoc explanation techniques such as LIME and SHAP , and also introduce novel methods to address some of these vulnerabilities. More specifically, I will first demonstrate that these methods are brittle, unstable, and are vulnerable to a variety of adversarial attacks. Then, I will discuss two solutions to address some of the vulnerabilities of these methods – (i) a framework based on adversarial training that is designed to make post hoc explanations more stable and robust to shifts in the underlying data; (ii) a Bayesian framework that captures the uncertainty associated with post hoc explanations and in turn allows us to generate explanations with user specified levels of confidences. I will conclude the talk by discussing results from real world datasets to both demonstrate the vulnerabilities in post hoc explanation techniques as well as the efficacy of our aforementioned solutions. BIO: Hima Lakkaraju is an Assistant Professor at Harvard University focusing on explainability, fairness, and robustness of machine learning models. She has also been working with various domain experts in criminal justice and healthcare to understand the real world implications of explainable and fair ML. Hima has recently been named one of the 35 innovators under 35 by MIT Tech Review, and has received best paper awards at SIAM International Conference on Data Mining (SDM) and INFORMS . She has given invited workshop talks at ICML , NeurIPS, AAAI , and CVPR , and her research has also been covered by various popular media outlets including the New York Times, MIT Tech Review, TIME , and Forbes. For more information, please visit: https://himalakkaraju.github.io/ This talk is part of the Artificial Intelligence Research Group Talks (Computer Laboratory) series. This talk is included in these lists:
Note that ex-directory lists are not shown. |
Other listsYoung Nanoscientist India Award Winner's Lecture sponsored by Oxford Instruments Armourers and Brasiers Cambridge Forum Lenguaje puroOther talksComputational image analysis for coeliac disease diagnosis change of room: FW26 Break-out Groups Continued (Room Stays Open Until 8pm) Promising ARMv8/RISC-V relaxed memory Break-out Groups |