The Fairness cluster brought together a variety of perspectives on defining and achieving fairness goals in automated decision-making systems. Such systems are commonly used for binary classification tasks — predicting recidivism, creditworthiness, hirability, etc. of individuals. Individual fairness notions demand that similar individuals be treated similarly by the classification system. Group fairness notions seek to achieve some measure of statistical parity for protected groups vis-à-vis the general population. To overcome shortcomings in these definitions, intermediate notions of fairness such as multicalibration and multimetric fairness have been defined to protect all sufficiently large, computable groups. Much work needs to be done in fine-tuning and applying these definitions to new scenarios.
While we have a good understanding of fairness for one binary classifier, real-world systems involve multiple classifiers classifying individuals in parallel (college admissions, ads shown) or in a pipeline (college admission followed by employment). Work on developing appropriate notions of fairness in these settings is in its infancy and will be further developed by this program.
In all these settings, we seek not only to design fair(er) decision procedures but also to understand computational and informational limitations that prevent us from doing so. Such negative results tell us what assumptions about the model need to change to achieve fairness and drive us to define approximate notions of fairness that can be achieved.
We also viewed fairness through an economic lens, understanding the causes for rational agents to be unfair and the costs of incentivizing such agents to behave fairly.
Long-term visitors to the cluster were primarily theoretical computer scientists who had been working on such questions. The cluster included two workshops. The first brought together scholars from the humanities, social sciences, law, and medicine to discuss phenomena of interest to their fields from the point of view of fairness. This provided theoretical computer scientists with a rich source of important problems to think about. The second workshop was more typical — with presentations by long-term visitors and people invited just for this workshop — on technical results on fairness.
This program was supported in part by the Patrick J. McGovern Foundation.
sympa [at] lists.simons.berkeley.edu (body: subscribe%20fairness2019announcements%40lists.simons.berkeley.edu) (Click here to subscribe to our announcements email list for this program).
Long-Term Participants (including Organizers):
Visiting Graduate Students and Postdocs:
Those interested in participating in this program should send an email to the organizers at this fairness2019 [at] lists.simons.berkeley.edu (at this address).