University of Cambridge > > Engineering Safe AI > An introduction to adversarial attacks and defences

An introduction to adversarial attacks and defences

Add to your list(s) Download to your calendar using vCal

If you have a question about this talk, please contact Adrià Garriga Alonso.

AI safety is not limited to RL settings. For example, we can use machine learning algorithms to design spam filters, yet attackers can still “reverse-engineer” our defence to send us junk emails. Autonomous driving systems based on computer vision techniques are also vulnerable to attacks, for instance, attackers can carefully apply a sticker to a stop sign in order to fool the vision system of the car. In this talk I will briefly discuss the mathematical framework of these attack techniques (specifically on image classifiers) and defence techniques against them.

Slides available here:

This talk is part of the Engineering Safe AI series.

Tell a friend about this talk:

This talk is included in these lists:

Note that ex-directory lists are not shown.


© 2006-2024, University of Cambridge. Contact Us | Help and Documentation | Privacy and Publicity