In recent years the issue of fairness (or lack thereof) in machine learning models has become a far more prominent issue. A big catalyst for this shift arose from discussions surrounding the ProPublica analysis [1, 2] and subsequent counter arguments [3] of Equivant’s (formerly Northpointe) COMPAS pre-trial risk assessment tool. Analysis showed that African-American defendants were more likely to be incorrectly labelled as higher-risk of recidivism and white defendants more likely to be incorrectly labelled as lower-risk. However the model was well-calibrated across different populations (see e.g. [4, 5, 6] for various discussions).
The fact there existed these two seemingly contradictory viewpoints gave rise to important theoretical questions.
-
Which notion of fairness is correct?
-
If one notion is fair, why is the other not?
-
Can the predictive model be made to satisfy both notions?
Answering these questions has led to the so-called impossibility theorem [3, 4], which states that, in general, it is impossible for a predictive model to simultaneously satisfy multiple established definitions of fairness.
Below I attempt to use Adam Pearce’s nice interactive example [7] (which I recommend viewing prior to reading) regarding testing for medical conditions within different age groups to give a (hopefully) simple mathematical illustration of some of the key points. I will do this following a causal data generation viewpoint put forth by Chiappa and Isaac [5].