COOKIES: By using this website you agree that we can place Google Analytics Cookies on your device for performance monitoring. |
University of Cambridge > Talks.cam > Machine Learning @ CUED > Fairness for Sequential Decision Making Algorithms
Fairness for Sequential Decision Making AlgorithmsAdd to your list(s) Download to your calendar using vCal
If you have a question about this talk, please contact Adrian Weller. Fairness considerations in settings where decisions are made by supervised learning algorithms (e.g. criminal risk assessment) has received considerable attention, recently. As the fairness literature continues to expand mostly around this canonical learning task, it is important to recognize that many real-world applications of ML fall outside the category of supervised, one-shot learning. In this presentation, I will talk about two scenarios in which algorithmic decisions are made in a sequential manner and over time. I will argue that in such settings, being fair—-at a minimum—-requires decisions to be “consistent” across individuals who arrive at different time steps, that is, similar individuals must be treated similarly. I will then talk about how such consistency constraints affect learning. In the first part of the talk, I will introduce a generic sequential decision making framework, in which at each time step the learning algorithm receives data corresponding to a new individual (e.g. a new job application) and must make an irrevocable decision about him/her (e.g. whether to hire the applicant) based on observations it has made so far. I propose a general framework for post-processing predictions made by a black-box learning model, so that the resulting sequence of outcomes is guaranteed to be consistent. I show, both theoretically and via simulations, that imposing consistency constraints will not significantly slow down learning. In the second part of the talk, I will focus on fairness considerations in a particular type of market—-namely, combinatorial prediction markets—-where traders can submit limit orders on various security bundles, and the market maker is tasked with executing these orders in a fair manner. The main challenge in running such market is that executing one order can potentially change the price of every other order in the book. I define the notion of a “fair trading path”, which at a high level guarantees similar orders are executed similarly: no order executes at a price above its limit, and every order executes when its market price falls below its limit price. I present a market algorithm that respects these fairness conditions, and evaluate it using real combinatorial predictions made during the 2008 U.S. Presidential election. This talk is part of the Machine Learning @ CUED series. This talk is included in these lists:
Note that ex-directory lists are not shown. |
Other listsIsotope Coffee: Geochemistry and Petrology Seminars Department of Earth Sciences Type the title of a new list here Druggability and the Genome (EBI, Hinxton, 4th February 2008)Other talksParticle Physics & CERN Past and prospective transitions in the UK gas industry – some interdisciplinary insights Hacking the gene expression machinery for genome defense Contestation, contingency, and justice in the Nordic low-carbon energy transition St Catharine’s Political Economy Seminar -‘De-unionization and the growing Inequality’ |