You are on page 1of 6

AEA Papers and Proceedings 2018, 108: 22–27

https://doi.org/10.1257/pandp.20181018

ADVANCES IN BIG DATA RESEARCH IN ECONOMICS

Algorithmic Fairness†

By Jon Kleinberg, Jens Ludwig, Sendhil Mullainathan, and Ashesh Rambachan*

The growing use of algorithms in social and that the risk tool used in one Florida county
economic life has raised a concern: that they was in fact discriminatory (Angwin et al. 2016).
may inadvertently discriminate against certain This widely-read article helped further elevate
groups. For example, one recent study found concerns about fairness within the policy and
that natural language processing algorithms can research communities alike, with subsequent
embody basic gender biases, such as associating work showing that the trade-offs are more subtle
the word nurse more closely with the word she than was initially apparent.1
than with the word he (Caliskan, Bryson, and These concerns have led to a large litera-
Narayanan 2017). Because the data used to train ture that tries to “blind” the algorithm to race
these algorithms are themselves tinged with ste- to avoid exacerbating existing unfairnesses in
reotypes and past discrimination, it is natural to society. Numerous studies (many of them in
worry that biases are being “baked in.” computer science) have pointed out that this
We consider this problem in the context of requires more than just excluding race from the
a specific but important case, one that is par- predictor, since protected features such as race
ticularly amenable to economic analysis: using could be reconstructed from other features. To
algorithmic predictions to guide decisions solve this “reconstruction problem,” procedures
(Kleinberg et al. 2015). For example, predic- have been proposed such as pre-processing the
tions about a defendant’s safety risk or flight data to orthogonalize the explanatory variables
risk are increasingly being proposed as a means (“inputs”) or outcomes to race, or modifying the
to guide judge decisions about whether to grant loss function the algorithm seeks to optimize to
bail. Discriminatory predictions in these cases penalize race disparities in outcomes.
could have large consequences. One can easily We argue that this perspective about how to
imagine how this could happen since recidivism promote algorithmic fairness, while intuitive, is
predictions will be polluted by the fact that past misleading and in fact may do more harm than
arrests themselves may be racially biased. In good. We develop a simple conceptual frame-
fact, a recent ProPublica investigation argued work that models how a social planner who
cares about equity should form predictions from
data that may have potential racial biases. Our
* Kleinberg: Cornell University, Ithaca, NY 14853 primary result is exceedingly simple, yet often
(email: kleinber@cs.cornell.edu); Ludwig: University of overlooked: a preference for fairness should
Chicago, 1155 East 60th Street, Chicago, IL 60637, and
NBER (email: jludwig@uchicago.edu); Mullainathan: not change the choice of estimator. Equity
Harvard University, 1805 Cambridge Street, Cambridge,
MA 02138, and NBER (email: mullain@fas.harvard.edu);
Rambachan: Harvard University, 1805 Cambridge Street, Angwin et al. (2016) argued the risk tool they examined
1 

Cambridge, MA 02138 (email: asheshr@g.harvard.edu). was biased because African Americans are more likely to be
Thanks to the NSF for support to Rambachan, to the Simons mis-classified as higher risk, while whites were more likely
Foundation for support of Kleinberg, and to the Arnold to be mis-classified as lower risk. Kleinberg, Mullainathan,
Foundation, Susan and Tom Dunn, and Ira Handler for pro- and Raghavan (2017) and Chouldechova (2017) note that
viding support to the University of Chicago Crime Lab and this finding is an unavoidable consequence for calibrated risk
Urban Labs. Any errors and all opinions are of course our tools in the presence of differences in offending rates across
own. groups, unless we have tools that are perfectly predictive of
† 
Go to https://doi.org/10.1257/pandp.20181018 to visit risk. An alternative measure of fairness used in Kleinberg
the article page for additional materials and author disclo- et al. (2018) focuses on a quantity derived from the actual
sure statement(s). decision outcome: detention rates to African Americans. 
22
VOL. 108 ALGORITHMIC FAIRNESS 23

p­ references can change how the estimated pre- she applies some procedure to historical data
diction function is used (such as setting a dif- to form a predictor of college success and uses
ferent threshold for different groups) but the that predictor to decide on future admissions.
estimated prediction function itself should not Individuals are described by ​(Y,  X, R)​ , where ​Y​
change. Absent legal constraints, one should is eventual (measured) college success, X ​ ​is a set
include variables such as gender and race for of academic variables that are observed at time
fairness reasons. As we show in an empirical of admissions, and R ​  ∈ {0, 1}​is the applicant’s
example below, the inclusion of such variables race (with ​R = 1​if the individual is from the
can increase both equity and efficiency. minority group of interest).
Our argument collects together and builds on We consider an efficient planner and an equi-
existing insights to contribute to how we should table planner. The efficient planner maximizes
think about algorithmic fairness.2 This argument an objective function ​ϕ(S)​that depends only on
is not specific to machine learning—indeed the the predicted performance of the set S​ ​of admit-
argument is cleanest, and presented here, within ted students.
the context of unbiased estimators. Additional We assume that the efficient planner applies
issues arise beyond those we consider here for an estimator to a given dataset on individuals
high-dimensional estimation procedures that consisting of ​(​Yi​​​, ​Xi​​​, ​Ri​​​)​to produce a predictor
trade off bias and variance to maximize out-of- ˆ​​ f ​  (X, R)​
. We say this objective function ϕ is
sample (OOS) prediction accuracy, which we ­compatible with the prediction function ˆ ​​f ​​ if the
discuss separately in other work. following natural monotonicity condition holds:
We empirically illustrate this point for the case If ​S​and S​ ​′ are two sets of students of the same
of using predictions of college success to make size, sorted in descending order of predicted
admissions decisions. Using nationally repre-
performance ˆ​​ f ​  (X, R)​ , and the predicted perfor-
sentative data on college students, we underline
how the inclusion of a protected variable—race mance ˆ​​ f ​  (X, R)​of the ​i th​student in ​S​is at least
in our application—not only improves predicted as large as the predicted performance of the ​i  th​
GPAs of admitted students (efficiency), but student in ​​S ′​​ for all ​i​ , then ​ϕ(S ) ≥ ϕ(S′)​.3 Given
also can improve outcomes such as the fraction an objective function and a compatible predic-
of admitted students who are black (equity). tor, the efficient planner has a simple rule. For a
The reason for this result is extremely simple. desired number of admitted students K ​ ​ , the effi-
Equity preferences involve increasing the frac- cient planner simply admits the set S​ ​ consisting
tion of black applicants admitted. Within that of the ​K​students with the highest ˆ​​ f  ​  (X, R)​ values.
set, society is still served best by ranking as well Now consider an equitable planner, who has
as possible using the best possible predictions. preferences over both grades and the racial
Forming the best predictions possible aids both composition of the admitted class. They seek
equity and efficiency. to maximize ​ϕ(S )  + γ (S)​ , where ​ϕ​is compat-
ible with ˆ​​ f  ​​as before, and γ ​  (S)​is monotonically
I.  Conceptual Framework increasing in the number of students in ​S​ who
have R ​  =  1​ (and thus belong to the minority
As an illustrative case of our framework, sup- group).4
pose we are interested in a social planner that How should the equitable planner solve her
is trying to make college admissions decisions optimization problem? The following theorem
based on anticipated college success. To do so,
A familiar case of compatibility is when ​ϕ​ is simply the
3 

sum of performance of admitted students and OLS is applied


2 
See, for example, the excellent discussions of existing to the data to produce an unbiased estimator, such that for an
algorithmic fairness research in Barocas and Selbst (2016) individual ​i​  , ​​Yi​​​  = ​
ˆ f ​  ( ​Xi​​​, ​Ri​​​ )  + ​ϵ​i​​​ , where ​​ϵ​i​​​is mean zero and
and Dwork et al. (2012, 2017), whose arguments are consis- orthogonal to ​(​Xi​​​, ​Ri​​)​. 
4 
tent with the view we take in the present work. The impor- The analysis that follows extends directly to the more
tance of within-group rankings for affirmative action has general case in which the equitable planner seeks to maxi-
been noted by Fryer and Loury (2013). Similarly, Corbett- mize ​ψ(S )  + γ (S)​ , where ​ψ​is compatible with f​ ​but may be
Davies et al. (2017) show that several notions of algorithmic different from the efficient planner’s function ​ϕ​. We use the
fairness in the context of criminal justice rely on the use of case in which the efficient and equitable planners have the
race-specific decision rules.  same ϕ in our exposition. 
24 AEA PAPERS AND PROCEEDINGS MAY 2018

shows that they should also rank by ˆ​​ f  ​​ , just as the Y​. This happens exactly when we feel there
efficient planner, but change the cutoffs used for is disadvantage—when individuals of ​ R =  1​
admission for the two groups defined by ​R  =  0​ have a different process than those with ​R =  0​.
and ​R  =  1​. In this case, access to R​ ​improves prediction
quality.
THEOREM 1: For some choice of K ​​ 0​ ​​​ and ​​K​1​​​ with​​
K​0​​  + ​K​1​​  = K​ , the equitable planner’s problem II.  Further Implications
can be optimized by choosing the ​​K​0​​​ applicants
The theorem in the previous section helps
in the ​R = 0​group with the highest ˆ​​ f  ​  (X, R)​ , and
parse some common reasons we worry that the
the ​​K1​ ​​​ applicants in the ​R  = 1​group with the data may bake in bias. First, we may worry that
highest ˆ​​ f  ​  (X, R)​. the inputs (​X​  ) are biased. In the college con-
text, America’s history of segregated schools
We can sketch the proof of this theorem as may affect the degree to which minority stu-
follows. Let ​​S​​  ∗​​be any set of ​K​applicants that dents are comparably prepared to succeed in
maximizes ​ϕ(S ) + γ (S)​ , and partition S​​ ​​ ∗​​ into college. One consequence could be that with
the applicants S​ ​​ 0∗​  ​​in the ​R = 0​group and ​​S​ 1∗​  ​​ in fewer inputs, black students are less prepared
the ​R = 1​group. Let ​​K​0​​  = | ​S0​  ∗​  ​  |​ and ​​K1​ ​​  =  | ​S1​  ∗​  ​  |​; for college, so that ​E [ Y  | R = 1 ] < E [ Y  | R = 0]​.
let ​​S0​  +​  ​​ be the ​​K​0​​​applicants in the R ​  =  0​ group Implicit in our theorem is that the solution here
is to set a different threshold for admissions.
with the highest  ​​f ​  (X, R)​ , and let S​ ​​ 1+​  ​​ be the ​​K1​ ​​​
ˆ
Another concern may be that, even for the same
applicants in the R ​   =  1​group with the highest level of preparation to succeed in college, black
ˆ​​ f ​  (X, R)​. Write ​​S​​  +​  = ​S0​  +​  ​  ∪ ​S1​  +​  ​​. One can prove applicants appear worse on observed inputs.
that if we sort ​​S​​  +​​ and ​​S​​  ∗​​in descending order of For example, they may receive less coaching
predicted performance, the i​  th​student in S​​ ​​ +​​ has on how to take standardized tests like the SAT.
predicted performance at least as large as the i​   th​ Yet this scenario implies that ​f (X, R)​ crucially
student in S​​ ​​ ∗​​. Hence by the compatibility of ϕ depends on ​R​. For the algorithm to account for
and ​f​ , we have ​ϕ(​S​​  +​) ≥ ϕ(​S​​  ∗​)​. Since ​​S​​  +​​ and ​​S​​  ∗​​ this sort of bias, it must know ​R​. If white stu-
have the same number of members with R ​  =  1​ dents are given more SAT prep, then the same
by construction, we also have γ ​  (​S​​  +​) = γ (​S​​  ∗​)​; SAT score implies higher college success
thus ​ϕ(​S​​  +​) + γ (​S​​  +​) ≥ ϕ(​S​​  ∗​) + γ (​S​​  ∗​)​  , and so​​ for a black student than a white one; that is,
S​​  +​​is a set maximizing the equitable planner’s ​E [ Y | X, R = 1 ] > E [ Y | X, R = 0]​. A prediction
objective function and satisfying the conditions function can only uncover this if it is allowed to
of the theorem. put different “slopes” on the SAT score for white
and black candidates.5 As a result, racial equity
Put in another way, given the equita-
is promoted by giving the algorithm access to
ble ­ planner’s preferences, they still wish to
race.
rank-order individuals within each group
­
Second, we may also worry that the ​Y​ vari-
using the same estimate of expected perfor-
ables themselves are biased. When we measure​
mance ˆ​​ f  ​   (X, R)​. Y​using college GPA, biases faced in the college
The intuition behind this Theorem is simple. experience are reproduced in Y ​​. Specifically,
The efficient planner only values ranking on the there may be a true Y​​  ​​ ∗​​ (such as actual learning
best possible prediction of output. An equitable in college) which determines ϕ , even though
planner, conditional on the fraction of minority we only measure Y ​ ​ (grades). We may fear that
students admitted, cares about this as well. ​E [ Y − ​Y​​  ∗​  | R = 1 ] < E [ Y − ​Y​​  ∗​  |R = 0]​. Even
Since the fraction of admitted students that are here, as long as ˆ​​ f ​​is compatible to ϕ in the
minorities can always be altered by changing sense above, the theorem says the equitable and
the thresholds used for admission, the equitable efficient planners should use the same ˆ​​ f ​​. The
planner should use the same prediction function
as the efficient planner. Implicit in this theorem
is that the use of race will always be strictly 5 
This discussion subsumes one where there are true X​​  ​​ ∗​​  ,
improving for the equitable planner’s objective and the X ​​ ∗​​. This type of
​ ​are racially biased measures of X​​ 
function as long as race is useful for ­predicting​ mis-measurement will have the same effect. 
VOL. 108 ALGORITHMIC FAIRNESS 25

i­ntuition is again straightforward. As long as ​Y​ Error rate as percent black admits varies
and ​​Y​​  ∗​​are monotonically related, ranking on ˆ​​ f ​​
17 Blind
remains the best strategy even when you care

Percent with GPA below 2.75


Orthog.
about equity. The ​Y − ​Y​​  ∗​​bias can be accounted Aware
for by setting a different threshold for the dis- 16 Eff. blind
Eff. orthog.
criminated group. Eff. aware
15

III. Data 14

We rely on the public-use version of the US 13


Department of Education’s National Education
Longitudinal Study of 1988 (NELS:88). This 12
dataset captured information for a nationally 5 10 15 20
representative sample of students who entered
Percent of admits that are black
eighth grade in the fall of 1988, and who were
then followed up in 1990, 1992, 1994, and Figure 1. Racial Composition/Grades Curves
finally in 2000 (when respondents were in their
mid-20s).6
The decision we examine is college admis- cent​ , as are most of our measures of high school
sion, specifically whether to admit a student to a outcomes.8
four-year college or university. We limit our anal-
ysis sample to those who were followed through IV.  Empirical Results
the 2000 survey wave and had ever attended a
four-year institution. To simplify, we focus just We summarize our main results by show-
on two groups: non-Hispanic white students ing what would happen to college admissions
(​N = 4,274​) and black students (​N = 469​). outcomes for both the efficient planner and
We assume the admit decision is based on ­equitable planner using different candidate pre-
predicted student performance (college grade diction functions. The binary outcome we pre-
point average). Predictors taken from the 1988, dict has ​Y = 1​if the student’s college GPA is​
1990, and 1992 waves of NELS data include <2.75​. We consider predictions from a simple
(besides race) high school grades, course taking unbiased algorithm (ordinary least squares), one
patterns, extracurricular activities, and student version of which is blinded to race altogether,
performance on the standardized achievement one of which pre-processes the inputs to make
tests that NELS administered to students in them orthogonal to race, and one of which is
four core subject areas: math, reading, science, made race-aware by interacting race with the
and social studies.7 Consistent with previous various predictors described above.9
studies, college graduation rates are higher for For starters, Figure 1 shows the efficient plan-
white students than black students in our sam- ner would choose to use the race-aware predic-
ple (​67.4  percent​ versus ​50.9  percent​). College tor. The cross in the figure shows what would
grades such as share earning a GPA of at least happen if the efficient planner selected the top​
2.75 are also higher on average for white than 50 percent​of four-year college students in the
black students, ​ 82.2 percent​versus ​ 69.5 per- NELS using the race-aware predictor (that is, by
rank-ordering all four-year students in the NELS
by the predicted outcome from the race-aware
predictor, then selecting the top half). Just under​
6 
The NELS provides sampling weights to account for the
8 
fact that not all baseline respondents were eligible for fol- We do not use data on the sociodemographics of the stu-
low-up waves. We present unweighted results below, but our dent’s family or school in any analyses. 
9 
findings are not sensitive to using the weights.  Results are qualitatively similar when we instead use a
7 
The public-use NELS tells us whether a student took machine learning algorithm (random forest), or use different
the SAT or ACT, but not their score, so we use these tests outcome measures for college performance such as varying
as a proxy.  the GPA threshold or an indicator for college completion. 
26 AEA PAPERS AND PROCEEDINGS MAY 2018

13 percent​of students admitted using the race-


aware algorithm would go on to get GPAs below​ A10
2.75​ , at least a full percentage point lower than

Race aware, deciles of pred. prob


A9
if the efficient planner had instead ­rank-ordered
students for admission using either the race- A8
blind algorithm or the predictor that first orthog- Value
A7
onalizes inputs to race (the circle and triangle 0.04
A6
in Figure 1, respectively). Because the efficient 0.03
planner cares only about efficiency (i.e., ­location A5 0.02

along the y-axis), using the race-aware predictor A4


0.01

dominates. 0.00
A3
More interesting is our evidence that the equi-
table planner (who uses a different threshold to A2
admit white versus black students, to promote A1
fairness) would also wish to use the race-aware
B1 B2 B3 B4 B5 B6 B7 B8 B9 B10
predictor. Varying the threshold used for black Race blind, deciles of pred. prob
students changes the share of the admitted​
50 percent​of students who are black (shown on Figure 2. Heatmap of Rankings of Black Applicants
the x-axis): for a given predictor the lower the by Predicted Probability of GPA < 2.75, Using Race-
threshold used for black students, the larger the Aware Versus Race-Blind Algorithms
share of admitted students who are black, but
the higher the share of admitted students who
achieve a GPA < ​ 2.75​. So, for a given predic-
tor, the curve shows that the possible combina- the rank-ordering within the pool of black appli-
tions of diversity and college achievement that cants and form a decision rule that dominates
can be achieved has a positive slope. The race- those based on race-blind predictors.
aware predictor dominates the other prediction
functions, even for the equitable planner. For V. Conclusion
any given level of diversity among admitted
students, using the race-aware predictor leads Concern about the potential fairness conse-
to the smallest share of admitted students with quences of algorithmic decision-aids is under-
low grades. Equivalently, for any given level of standable and plays an important role in debates
achievement among admitted students, using the about their wide-scale adoption. Our central
race-aware predictor would lead to admission of argument is that across a wide range of estima-
relatively more black applicants. tion approaches, objective functions, and defi-
Figure 2 shows why this is so. The race-blind nitions of fairness, the strategy of blinding the
predictor mis-ranks black students. This “heat- algorithm to race inadvertently detracts from
map” shows the distribution of black students fairness.
in our NELS sample across predicted-outcome
deciles according to the race-blind (x-axis) or
race-aware (y-axis) predictors. For example, REFERENCES
consider the group of students in the southeast
part of Figure 2. The race-blind predictor clas- Angwin, Julia, Jeff Larson, Surya Mattu, and
sifies them as being in the ninth decile of pre- Lauren Kirchner. 2016. “Machine Bias.”
dicted probability of receiving a GPA ​<2.75​  , ProPublica, May 23. https://www.pro-
but, according to the race-aware predictor, they publica.org/article/machine-bias-risk-assess-
are actually members of the lowest decile. ments-in-criminal-sentencing.
This mis-ranking stems from the differences Barocas, Solon, and Andrew D. Selbst. 2016. “Big
in the slopes for the relationship between col- Data’s Disparate Impact.” California Law
lege grades and different predictors (such as Review 104: 671–732.
high school achievement tests) for black versus Caliskan, Aylin, Joanna J. Bryson, and Arvind
white students. Only by giving the algorithm Narayanan. 2017. “Semantics Derived Auto-
access to race can we account for this, improve matically from Language Corpora Contain
VOL. 108 ALGORITHMIC FAIRNESS 27

Human-Like Biases.” Science 356 (6334): Classifiers for Fair and Efficient Machine
183–86. Learning. arXiv 1707.06613.
Chouldechova, Alexandra. 2017. “Fair Predic- Fryer, Roland G., Jr., and Glenn C. Loury. 2013.
tion with Disparate Impact: A Study of Bias in “Valuing Diversity.” Journal of Political Econ-
Recidivism Prediction Instruments.” Big Data omy 121 (4): 747–74.
5 (2): 153–63. Kleinberg, Jon, Jens Ludwig, Sendhil Mullaina-
Corbett-Davies, Sam, Emma Pierson, Avi Feller, than, and Ziad Obermeyer. 2015. “Prediction
Sharad Goel, and Aziz Huq. 2017. “Algorithmic Policy Problems.” American Economic Review
Decision Making and the Cost of F­ airness.” 105 (5): 491–95.
Proceedings of the 23rd ACM SIGKDD Inter- Kleinberg, Jon, Sendhil Mullainathan, and Man-
national Conference on Knowledge Discovery ish Raghavan. 2017. “Inherent Trade-Offs in
and Data Mining: 797–806. the Fair Determination of Risk Scores.” Pro-
Dwork, Cynthia, Moritz Hardt, Toniann Pitassi, ceedings of the 8th Conference on Innovation
Omer Reingold, and Richard Zemel. 2012. in Theoretical Computer Science: 43:1–43:23.
“Fairness through Awareness.” Proceedings of Kleinberg, Jon, Himabindu Lakkaraju, Jure
the 3rd Innovations in Theoretical Computer Leskovec, Jens Ludwig, and Sendhil Mullain-
Science Conference: 214–26. athan. 2018. “Human Decisions and Machine
Dwork, Cynthia, Nicole Immorlica, Adam Tauman Predictions.” Quarterly Journal of Economics
Kalai, and Max Leiserson. 2017. “­Decoupled 133 (1): 237–93.

You might also like