Resolving the Disparate Impact of Uncertainty: Affirmative Action vs. Affirmative Information
←
→
Page content transcription
If your browser does not render page correctly, please read the page content below
Resolving the Disparate Impact of Uncertainty: Affirmative Action vs. Affirmative Information Claire Lazar Reich 1 MIT Economics & Statistics clazar@mit.edu arXiv:2102.10019v1 [stat.ML] 19 Feb 2021 Abstract Algorithmic risk assessments hold the promise of greatly advancing accurate decision-making, but in practice, multiple real-world examples have been shown to distribute errors disproportionately across demographic groups. In this paper, we characterize why error disparities arise in the first place. We show that predictive uncertainty often leads classifiers to systematically disadvantage groups with lower- mean outcomes, assigning them smaller true and false positive rates than their higher-mean counterparts. This can occur even when prediction is group-blind. We prove that to avoid these error imbalances, individuals in lower-mean groups must either be over-represented among positive classifications or be assigned more accurate predictions than those in higher-mean groups. We focus on the latter condition as a solution to bridge error rate divides and show that data acquisition for low-mean groups can increase access to opportunity. We call the strategy “affirmative information” and compare it to traditional affirmative action in the classification task of identifying creditworthy borrowers. 1 Introduction Algorithmic risk scores are increasingly helping lenders, physicians, and judges make high-impact decisions in the presence of uncertainty. Yet even the most advanced algorithms cannot eliminate the uncertainty surrounding these decisions [9], and in recent years there has been growing awareness of real-world cases where the burden of prediction error falls more heavily on some populations than others [2, 8, 14]. In light of these issues, a growing branch of research in fair prediction has produced methods to modify algorithms so that they yield equal error rates across subgroups [10, 16, 13]. In this paper, we show why error rate disparities across groups arise in the first place and characterize necessary conditions to correct them. We consider screening decisions in which applicants are either granted or denied an opportunity, such as a loan, based on predictions of their eventual outcomes, such as loan repayment. We provide empirical and theoretical evidence showing that applicants from lower-mean groups are often more likely than those from higher-mean groups to be misclassified as negative and mistakenly denied a loan. Conversely, those from higher-mean groups are relatively more likely to be misclassified as positive. Based on these error patterns, individuals from lower-mean groups have lower probability of accessing loans conditional on their true repayment ability. The imbalance can persist even when the predictions are group-blind, and it tends to be exacerbated when they are imprecise. We therefore call this phenomenon “the disparate impact of uncertainty.” We prove that to eliminate group error disparities in the presence of uncertainty, at least one of the following must be true: (1) members of the lower-mean groups must be over-represented among 1 I am grateful to my advisors Anna Mikusheva, Iván Werning, and David Autor for sharing invaluable insights at each stage of the project. I also thank Tom Brennan, Suhas Vijaykumar, Victor Chernozhukov, Ben Bernanke, Ben Deaner, David Hughes, David Sontag, and Frank Schilbach for helpful and inspiring discussions. This work was supported by the National Science Foundation Graduate Research Fellowship under Grant No. 1122374.
positive classifications, or (2) their predictions must be more accurate than those of higher-mean groups. We pose the enforcement of the latter condition as a promising alternative to affirmative action: rather than lowering the admissions criteria for lower-mean groups, it is possible to increase access to opportunity by acquiring data about lower-mean applicants that aligns predictions closer to their outcomes. We call this strategy “affirmative information” and compare it to traditional affirmative action in an empirical credit lending application. Such a strategy agrees with the insight in [5] that data collection is an integral part of fair prediction that deserves greater attention as an alternative to imposing model constraints. In contrast to our proposal of data enrichment, many algorithms in use today seek to achieve fairness through the omission of data. We show that contrary to intuition, omitting group identifiers and sensitive features from data generally does not resolve the disparate impact of uncertainty. In fact, by increasing predictive uncertainty, data omission may exacerbate disparities. We therefore support the growing evidence in economics and computer science showing that blind algorithms are a misguided approach to fair prediction. Within this literature, [12, 7, 15] produce theoretical results demonstrating that fairness considerations are best imposed after estimating the most accurate predictions, rather than before, and [11, 13] demonstrate this principle empirically in real-world settings. While we focus on the fairness condition of error rate equality in this paper, it is important to acknowledge that there are many additional components of algorithmic fairness [3]. Among them is accuracy on the individual level [6]. To avoid compromising this criterion, we do not propose intentionally reducing the accuracy of predictions for higher-mean groups in order to eliminate error disparities. Rather, we raise the option of basing predictions on all immediately available data for all applicants, as advocated in [11], supplemented with further data acquisition efforts to better identify the qualified individuals from lower-mean protected groups. We believe approaches like these can maximize the gains of algorithmic decision-making while ensuring they are broadly shared. The paper proceeds in four sections. We empirically illustrate the disparate impact of uncertainty in Section 2. In Section 3 we present general theoretical results that characterize this disparate impact and identify conditions necessary to overcome it. To provide intuition for these results, we restructure and analyze a canonical economic model of statistical discrimination [1]. Section 4 empirically compares two methods that aim to overcome the disparate impact of uncertainty: affirmative action and affirmative information. We show that in settings with selective admissions standards, affirmative information can work like affirmative action to raise the probability that creditworthy individuals in low-mean groups are granted loans. Unlike affirmative action, it can simultaneously yield higher return to the decision-maker and create natural incentives to admit more applicants. 2 Motivation: a statistical phenomenon about blind prediction Motivating this study is the observation that predictive algorithms can systematically favor individuals from high-mean groups even when they do not discriminate by group. We illustrate this fact empirically, showing that a credit-lending algorithm based only on financial features will lead to greater true and false positive rates for high educated applicants than for less educated applicants. Since it is uncontroversial to include financial characteristics in credit lending predictions, this exercise will suggest that the disparities stem from unobservable variation in outcomes rather than observable variation. Later sections lay out the theory that explains this finding. We consider a lender that uses risk scores to classify loan applications into approvals and denials. We suppose the designer of these risk scores wants to provide high-quality predictions that are fair to demographic subgroups, one highly educated H (more than a high school degree) and another less educated L (at most a high school degree). The lender’s notion of fairness is to avoid discrimination by purging the dataset of sensitive characteristics, including education group. Therefore the dataset is restricted only to financial features deemed relevant for loan repayment. We simulate this scenario using the U.S. Census Survey of Income and Program Participation and predict the outcome of paying rent, mortgage, and utilities in 2016 using survey responses from two years prior. We construct risk scores with a logistic regression on just three predictors: total assets, total debt, and monthly income. The risk scores designed in this experiment can be converted to binary classifications at a number of cutoffs. The results are illustrated in Figure 1. In panel (a), we see that less-educated individuals are 2
(a) (b) Figure 1: (a) Members of the higher-mean group are more likely to be classified as creditworthy. (b) Members of the higher-mean group are more likely to be classified as creditworthy, even when we condition on their true label. While the predictions are group-blind and there was no disparate treatment, there is disparate impact. less likely to be classified as creditworthy at all cutoffs. This is not necessarily unfair, as perhaps the scores have picked up on legitimate differences in repayment abilities. However, in panel (b), we show that less-educated individuals are also less likely to be classified as creditworthy even when we condition on their eventual repayment outcome. That is, they have higher true and false positive rates than less educated applicants. In practice, this means that a creditworthy applicant from H has a greater probability of being granted a loan than one from L; similarly, a non-creditworthy applicant from H also has a greater probability of being granted a loan than one from L. In this application we blinded the algorithm to education level and other sensitive characteristics in an effort to avoid disparate treatment. However, there is disparate impact nonetheless. We next show why a blind algorithm can systematically favor individuals from high-mean groups and what is required to avoid it if not data omission. 3 Theory 3.1 Characterizing the disparate impact of uncertainty In this section we describe conditions that are necessary for classifiers to yield equal true and false positive rates to groups with different mean outcomes. We start by briefly introducing our notation and basic definitions. Let Y ∈ {0, 1} be a binary outcome variable and G ∈ {L, H} be a protected attribute differentiating two groups. Each group has a unique fixed base rate µG ≡ E[Y |G] , where 0 < µL < µH < 1. We consider binary predictions ŷ of Y formulated with incomplete information about applicants, so that there is at least some predictive uncertainty: Definition 1. Predictive uncertainty is present if each group’s true and false positive rates lie in the open interval (0, 1), in particular TPRG = P[ŷ = 1|Y = 1, G], FPRG = P [ŷ = 1|Y = 0, G] ∈ (0, 1). At least one positive and one negative individual from each group will be misclassified. We will formulate conditions under which ŷ assigns positive classifications systematically more often to members of the high-mean group H, conditional on the true type Y . First, we observe that in order to achieve balance in both error rates, the TPR and FPR, members of L must be over-represented among the positive classifications. In particular, L must comprise a strictly greater portion of the positive classifications ŷ = 1 than it does positive types Y = 1. Proposition 1. If ŷ yields TPRL ≥ TPRH and FPRL ≥ FPRH then it must be that L members are over-represented among positive classifications. That is, P[G = L|ŷ = 1] > P[G = L|Y = 1]. Proof in appendix. 3
Over-representation is typically achieved in settings where admission standards are low for all groups, or in settings where those standards are lowered specifically for members of L. We can think of this as including traditional affirmative action. Relaxing the constraint of equal FPRs in Proposition 1 enables an alternative solution: improve the accuracy of L predictions, and through better targeting, increase their group TPR to match that of the higher-mean group. The accuracy gain may simultaneously lower FPRL further below FPRH , but in many settings there is no need to equate group FPRs anyway. For example, in credit lending, we may wish for a creditworthy borrower to have equal opportunity to receive a loan regardless of group (achieving equal TPRs). Yet equating group probabilities for non-creditworthy borrowers to secure loans may be unimportant (equal FPRs) [10]. The following theorem establishes that in order to achieve a TPR for the L group at least as large as for the H group, it is required that ŷ over-represents L members among positive classifications or assigns more accurate classifications for the L members. By accuracy we refer to the groups’ positive and negative predictive values, PPVG ≡ P[Y = 1|ŷ = 1, G] and NPVG ≡ P[Y = 0|ŷ = 0, G]. Theorem 1. In the presence of algorithmic uncertainty and base rate differences, one of the following must be true in order for ŷ to achieve TPRL ≥ TPRH : (i) Members of the L group are over-represented among positive classifications. That is, P[G = L|ŷ = 1] > P[G = L|Y = 1] (ii) The classifications of the L group should be more informative of the true outcomes than those of the H group. In particular NPVL > NPVH and PPVL ≥ PPVH . Proof in appendix. The theorem shows that if individuals in L are not over-represented among positive classifications, then uncertainty in prediction imparts a disparate impact. Given the same level of accuracy for the two groups, the L group will be guaranteed to face a lower TPR than the H group. It may seem that omitting sensitive features from a dataset would be a plausible solution to provide equal opportunity to members of different groups, and that strategy is indeed the predominant fairness approach used in today’s screening algorithms. However, omitting data does not resolve the tension posed by Theorem 1. In particular, the omission of data does not necessarily lead individuals from low-mean groups to be over-represented among positive classifications. The reason is that variation in outcomes across groups may still be at least partially explained by the predictors allowed in the model. In credit lending, for example, individuals from a lower-mean group may have lower levels of income, higher existing debt, or fewer assets. A model based on these variables assigns classifications according to the variation in them, and that may prevent L from being over-represented among positive classifications. If the L group’s classifications are no more accurate than those of H, then creditworthy individuals from L will be denied at higher rates than those from H. We construct a model to illustrate this phenomenon in the next section, where we assume that variation in group outcomes is explained by the predictors underlying a risk score. This will allow us to isolate the disparate effects of uncertainty and demonstrate the improvements from raising the accuracy of L predictions. 3.2 Modeling the mechanism of the disparate impact In this section, we present a model to explore why the presence of uncertainty imparts a disparate impact on groups with different mean outcomes. We consider a lender who uses data to predict applicants’ repayment ability. The signals they derive are constructed without regard to group membership, but the financial data they are based on does explain the group variation in outcomes.1 As a result, we show that predictive uncertainty contracts each group’s scores toward the group mean of ability. Scores for lower-mean groups will contract toward a lower mean, and in general the conditional distribution of scores given true ability will vary by group, causing systematic differences in true positive rates to emerge. 1 For example, suppose that low income is negatively correlated with repayment ability, and members of a disadvantaged group tend to have lower incomes. 4
Consider a setting where loan applicants each have some repayment ability A, and applicants with A > 0 are “creditworthy” and would repay a loan if granted one. A lender seeks to estimate A using financial data X available about the applicants. At no point does the lender observe protected group membership G. However, the data X does explain the variation in repayment ability across those groups. In particular we assume A and G are mean independent given X 2 , so E[A|X, G] = E[A|X]. The lender estimates E[A|X] in the model A = E[A|X] + ε, (1) where ε is a residual that is uncorrelated with any function of X. For simplicity we assume the lender successfully constructs the Bayes optimal risk score S = E[A|X], the best prediction given the available data. We can prove that those scores continue to serve the decision-maker as the best prediction regardless of whether group membership is observed: Proposition 2. The lender’s risk scores S = E[A|X] will satisfy E[A|S, G] = S. That is, an applicants’ expected ability given their score will equal their score, regardless of group membership. Proof in appendix. We know from Proposition 2 that the lender’s best available prediction of an applicant’s true ability is given to her by the score. If she imposes a cutoff rule to classify individuals, then she will choose the same cutoff for each group regardless of whether she observes group affiliation. The decision-rule is Accept ≡ 1{S > cutoff}. (2) The choice of cutoff depends on the lender’s relative valuation of false positive and false negative classifications. In a setting where false negatives are more costly, the cutoff will be lower, and in a setting where false positives are more costly, the cutoff will be higher [13]. We call low-cutoff settings “lemon-dropping markets” and high-cutoff settings “cherry-picking markets,” following the terminology in [4]. As the scores guide the decision-maker’s classifications, we can abstract away from the data under- lying them. Our model will be seen to reframe the canonical Aigner and Cain (1977) model for statistical discrimination [1], treating the risk score as a “signal” that a decision-maker receives about an applicants’ ability. At the end of the section, we clarify how our model compares to the original in Aigner and Cain (1977). We now can express an applicant’s repayment ability as a sum of two parts, the signal S estimated with available data, and an unobservable error ε. For a member of group G we write AG = SG + εG . (3) By Proposition 2, εG is mean-zero. Invoking a similar assumption as Aigner and Cain (1977), we suppose that SG and εG are bivariate normal. Since SG and εG are uncorrelated, this implies they are independent. Our model is therefore specified by three random variables: SG ∼ N (µG , σS2 G ), εG ∼ N (0, σε2G ), and AG ∼ N (µG , σS2 G + σε2G ). While we know that the score satisfies E[AG |SG ] = SG for both groups, we will now show that E[SG |AG ] 6= AG . The latter represents the expected score that a decision-maker will face for an applicant with some fixed true ability, and we will see that it varies by group. Individuals in low-mean groups will tend to display worse observable scores, even conditioned on their true ability. To compute the expected score, we use the fact that when variables X1 and X2 are bivariate normally distributed, then the distribution of X1 given X2 = x2 has conditional expectation 1 ,X2 ) E[X1 ] + cov(X σ2 (x2 − E[X2 ]). In our model, then, for both G ∈ {L, H}, X2 cov(SG , AG ) E[SG |AG = a] = E[SG ] + 2 (a − E[AG ]) σA G cov(SG , SG + εG ) = µG + (a − µG ) σS2 G + σε2G σS2 G = (1 − γG )µG + γG a, where γG ≡ σS2 G + σε2G 2 This is implied by the stronger condition (A ⊥ G)|X. 5
Therefore, the expected signal is a weighted average of the true ability and the group mean. The weight γG on the true ability reflects the portion of AG ’s total variance that is captured by the observable score SG , and the weight 1 − γG on the mean reflects the unexplained variance due to the presence of uncertainty. As γG increases, the score is expected to strongly track true ability. As γG decreases, however, the scores are expected to bunch more tightly about the group mean. An immediate consequence is that an individual from L with true ability a can generally be expected to have a different score than an individual from H with ability a: E[S|A = a, L] = (1 − γL )µL + γL a, (4) E[S|A = a, H] = (1 − γH )µH + γH a. (5) We next consider two cases, one demonstrating the disparate impact of uncertainty and the second correcting for it. The first case equates the observed variance in ability (γL = γH ), while the second explores the effect of increasing the observed variance for the L group (γL > γH ). In each case, the best unbiased estimate of ability for each group is given by the risk score. However, in the latter case, the estimate is made more precise for L. 3.2.1 Equally informative signal First consider a setting with mean differences and equal variances: µL < µH and γ L = γ H . We depict expected scores conditional on true underlying ability in Figure 2a. For each level of ability, individuals of the H group are expected to have higher observable scores. The gap between their expected scores and those of the L group is seen to increase in the differences between their means (µH − µL ) and the extent of the uncertainty, (1 − γ). Since decision-makers use the same cutoff rule (2) for members of both groups, it seems natural that the systematic disparities among scores would propagate into systematic disparities among classifications, that is, the ultimate lending decisions. We formally show this by deriving the conditional distribution of S given creditworthiness, which we defined as A > 0 early in the section. 3 . The distributions are plotted in Figure 2b and 2c with two examples of decision-maker cutoffs 4 . Because the true positive rate is the portion of creditworthy individuals ultimately granted a loan, it is straightforward to visualize it on the plot: each group’s true positive rate is given by the portion under its curve that is shaded, corresponding to those applicants with a score above the cutoff. We observe that the true positive rates are higher for the H group in both lemon-dropping and cherry-picking markets, but that the effect is most stark when the decision-maker cherry-picks applicants at the top of the distribution (Figure 2c). 3.2.2 More informative L signal Next consider the case where the score captures a greater portion of the variance in ability for the L group than for the H group, that is when γ L > γ H . Then imbalances in the true positive rate will be alleviated. The expected scores conditional on true underlying ability are plotted in Figure 2d, where we see that at low levels of underlying ability, L members have lower expected scores than H members, whereas they have higher expected scores at higher levels of ability. Meanwhile, the distributions of the scores of creditworthy individuals are plotted in Figures 2e and 2f 5 . For the particular signal variance depicted, the shaded portion under each groups’ curve is approximately the same, signifying that the groups’ true positive rates are approximately the same. We further consider the effect of signal strength on true positive rate imbalances in Figure 3. There, we vary the relative signal strength γL /γH of the L group versus the H group on the x-axis, and plot the resulting ratio of group true positive rates for various mean disparities across groups. In cherry-picking markets, all mean disparities require L to have a greater signal strength to achieve a true positive rate at least as large as that of the H group. The effect is present but relatively muted in lemon-dropping markets, in which so many applicants are accepted from both groups that both groups’ true positive rates are generally high. 3 The conditional distribution of S|A > 0 is normal. Derivation details for its mean and variance are provided in the appendix. 4 The plots in Figures 2b and 2c are constructed using parameter values (µH , σS2 H , σU 2 H ) = (0, 1, 1) and 2 2 (µL , σSL , σUL ) = (−1, 1, 1). The lemon-dropping and cherry-picking cutoffs depicted are −1 and +1, respectively. 5 The plots in Figures 2e and 2f are constructed by changing the L variances to (σS2 L , σU2 L ) = (1.6, 0.4). 6
Case I: Equal Variances Case II: Unequal Variances (a) (d) (b) (e) (c) (f) Figure 2: Distributions of scores in the parametric model. Plots in left column illustrate the case with equal variance scores, and plots in right column are based on higher-variance scores for the L group. In (a), we see that expected scores conditional on true underlying ability are systematically higher for H members. In (b) and (c), we plot the distribution of scores for creditworthy applicants (A > 0) in a lemon-dropping and cherry-picking market, respectively. The group-specific true positive rates are given by the portion under each curve that is shaded. Members of H have higher true positive rates, and the difference is particularly stark in the cherry-picking example. The imbalance is alleviated when the L scores become more informative, as seen in (d)-(f). 7
smaller μL smaller μL Figure 3: Ratios of true positive rates as a function of relative signal strengths. The mean of group H is held fixed at 0, while the mean of L is set to -0.5, -1.0, and -1.5. Bigger disparities in group means are associated with bigger disparities in true positive rates. In cherry-picking markets, L group always needs strictly stronger signal than H group to achieve the same true positive rate as the H group. 3.2.3 Comparison to Aigner and Cain (1977) The canonical Aigner and Cain (1977) model of statistical discrimination showed that when a rational decision-maker is presented with applicants’ noisy unbiased signals for ability, S = A + ε, then there is an incentive to weigh applicants’ group membership G alongside those signals. The decision-maker will assign equal outcomes to applicants with equal expected ability, but as a result of their consideration of group membership, outcomes for applicants with equal true ability will systematically differ by group. Our model shows how the disparate impact documented in Aigner and Cain (1977) can persist even when a decision-maker is blind to group membership. Group means enter our model not through the decision-maker’s interpretation of S, but rather through the construction of S itself. We first adjust the meaning of the signal S in the Aigner and Cain (1977) model. Rather than treating S as a noisy measurement of ability, like an exam, we instead suppose that it is the best estimate of ability given observables. Therefore it becomes natural for us to consider a risk score S = E[A|X] estimated from the model A = S + ε, instead of Aigner and Cain’s S = A + ε. 6 We further suppose that the signal S is constructed with observables that fully explain the group variation in ability. As a consequence, we prove that the best estimate for ability A is given by the score S regardless of whether group G is observed; unlike in Aigner and Cain (1977), the rational decision-maker has no incentive to update the ability estimate given group information. Barring the decision-maker in our model from considering group membership does not alleviate disparities and in fact has no effect. If both groups’ signals are equally informative, then the differences in true positive rates identified by Aigner and Cain (1977) will persist. We show that reducing uncertainty, however, does alleviate the disparities. If both groups’ signals are made more informative, then each groups’ scores will more closely track true ability rather than bunch around group means and disparities will narrow. Meanwhile, if care is taken to improve the precision of signals for the low-mean group in particular, then the disparity can be completely eliminated. 6 Our signal is distributed as a mean-preserving contraction of the true ability, whereas the signal in Aigner and Cain (1977) is distributed as a mean-preserving spread of the ability. 8
(a) (b) (c) Figure 4: Comparisons between affirmative information and affirmative action. (a) Affirmative information corresponds to higher portions of loans repaid at every cutoff. (b) Before the lender responds to intervention, affirmative action yields greater TPRs for the lower-mean group. (c) Next we model TPRs after the lender’s endogenous response to interventions. When the original blind prediction is replaced by affirmative action, the TPR of the higher-mean group falls and the TPR of the lower-mean group increases modestly. Meanwhile, when blind prediction is replaced by affirmative information, the TPR of the higher-mean group is roughly unchanged and the TPR of the lower-mean group increases. For all but the lowest values of k, affirmative information is associated with higher TPRs for both groups than affirmative action. 4 Empirical Application We return to our empirical example from Section 2, where we demonstrated how a set of credit risk scores can lead to disparate outcomes across education groups. Despite the fact that the scores were constructed without group identifiers, they yielded lower true and false positive rates to less-educated applicants at every cutoff. In this section, we compare two strategies for raising the true positive rate of the less-educated group: affirmative action and affirmative information. In our setting, we use the U.S. Census Survey of Income and Program Participation (SIPP), a nationally-representative panel survey of the civilian population. We label our outcome of “creditwor- thiness” as the reported ability to pay rent, mortgage, and utilities in 2016, and use survey response from two years prior to predict that outcome, as in [13]. Of the overall sample, we call the 55% who studied beyond a high school education group H and those with at most a high school degree group L. Mean outcomes vary by group: about 93% of H members report to have repaid their obligations, compared with 89% of the L members. Our risk scores are trained with a logistic regression on a subset of the data and then tested on a separate holdout set. We suppose a lender is presented with these scores and uses them to assign loan denials and approvals, without observing individuals’ group membership. The lender approves loans for anyone whose score corresponds to a positive expected return, and therefore imposes a cutoff rule that depends on their relative cost of false positive and false negative errors. When false positives are more costly, the cutoff is greater. As in Section 2, at every cutoff, blind prediction based only on total debt, total assets, and monthly income yields classifications with smaller TPR for the L group. We first consider how affirmative action can correct the error disparity by over-representing individuals from L among the positive classifications. For each of the lender’s intended cutoffs, we lower the corresponding cutoff for L members until their TPR reaches the level of the H applicants. In Figure 4a, we can see that this scheme reduces the portion of loans subsequently repaid to the lender. However, in Figure 4b, we can see that it yields group TPRs that are approximately equal. As an alternative, we consider an affirmative information scheme that aims to raise the TPR of individuals in L by improving accuracy for the L group. We suppose the decision-maker has taken effort to gather additional data about the L applicants, and to simulate that we train new L scores using the much richer set of variables available in SIPP, spanning household financial details as well as more sensitive features such as food security, family structure, and participation in social programs. Due to the large number of variables, we use LASSO to estimate risk scores. The MSE for L scores drops from .098 in the Blind Prediction scheme to .092 in the affirmative information scheme. Unlike 9
affirmative action, we impose the same cutoffs for each groups’ set of scores: the original scores for H and the data-enriched scores for L. The affirmative information scheme corresponds to a higher portion of loans repaid than affirmative action at all cutoffs, as seen in Figure 4a. Meanwhile, its immediate effect on the L group TPR depends on the cutoff. In Figure 4b we see that at many values, affirmative information raises the TPR above its original value, though not as much as affirmative action. However, this comparison does not account for how a profit-maximizing lender would respond to the interventions. One can imagine, for example, that requiring the lender to reduce an optimal cutoff for a substantial portion of the population would incentivize them to re-optimize and raise cutoffs overall. To identify the lender response to affirmative action and affirmative information, we consider a profit function that generates cutoff decision rules: TP − kFP, where TP is the number of creditworthy applicants given loans, FP is the number of defaulters given loans, and k is a constant representing the relative cost of false positive classifications. We simulate the lender response by considering different values of k in the profit function. For each value, we use the training data scores and labels to determine the decision-maker’s optimal cutoffs in the three schemes. Blind prediction and affirmative information are each associated with one optimal cutoff per k, while affirmative action is associated with two optimal group-specific cutoffs per k among all pairs that yield equal TPRs. Then for each optimal decision-rule, we compute the corresponding group TPRs in the holdout data and plot them in Figure 4c. Figure 4c illustrates the lasting impacts on TPRs from each intervention. We see that unlike in panel (b), affirmative action does not raise the TPR of L applicants all the way to the original levels of the H applicants. Instead it leads the two groups’ TPRs to meet in the middle, modestly raising the TPR of L while reducing that of H. In addition, it declines the fastest, eliminating any potential for profit and closing the market starting at k ≈ 45. By comparison, the affirmative information scheme leaves the H TPR essentially unchanged, while raising the TPR of L applicants above the level of affirmative action for all but the lowest values of k. There are two noteworthy caveats to affirmative information. The first is that affirmative information requires data acquisition, which is generally costly. The second is that it may be less effective in lemon-dropping settings in which cutoffs are well below the mean, as seen in Figure 4c at the lowest values of k. In particular, increasing accuracy in low-cutoff settings can have a counter effect of reducing the over-representation of L members among positive classifications. This can be explained by considering the effect of data enrichment on the spread of L scores. When decision cutoffs are high, increasing the spread has two effects that move the TPR in the same direction: accuracy increases, and more individuals are pushed above the cutoff than before. Yet when decision cutoffs are low, there are two opposing forces acting on the TPR: though accuracy increases, more individuals are pushed below the cutoff than before. Overall, this empirical example has demonstrated the potential advantages of affirmative information as a tool to broaden access to opportunity. Particularly in settings with selective admissions criteria, employing affirmative information rather than affirmative action can yield TPRs that are higher for both groups. In addition, depending on the cost of the data acquisition, affirmative information may yield relatively higher payoffs to the lender. Targeted data enrichment can therefore create natural incentives to broaden access to opportunities. 5 Conclusion In the presence of uncertainty, we have shown that classifiers tend to distribute Type I and Type II errors unevenly among groups whose mean outcomes differ. They tend to assign more false negatives to individuals in lower-mean groups, and more false positives to those in higher-mean groups. As a result, individuals in lower-mean groups are less likely to be classified as positive conditional on their true labels. We show both empirically and in a theoretical model that this phenomenon can persist even if sensitive data are removed from the prediction process. Instead of data omission, we show that one of two conditions is required to eliminate differences in groups’ true positive rates: either applicants from lower-mean groups are over-represented among positive classifications, or their classifications are more accurate than those of higher-mean groups. The latter condition suggests that data enrichment for individuals in lower-mean protected groups is a 10
promising alternative to data omission. We call the strategy “affirmative information” and suggest that it can improve both efficiency and equity. References [1] Dennis J Aigner and Glen G Cain. Statistical theories of discrimination in labor markets. Ilr Review, 30(2):175–187, 1977. [2] Julia Angwin, Jeff Larson, Surya Mattu, and Lauren Kirchner. Machine bias. ProPublica, May, 23:2016, 2016. [3] Solon Barocas, Moritz Hardt, and Arvind Narayanan. Fairness and Machine Learning. fairml- book.org, 2019. http://www.fairmlbook.org. [4] Vojtěch Bartoš, Michal Bauer, Julie Chytilová, and Filip Matějka. Attention discrimination: Theory and field experiments with monitoring information acquisition. American Economic Review, 106(6):1437–75, 2016. [5] Irene Chen, Fredrik D Johansson, and David Sontag. Why is my classifier discriminatory? In Advances in Neural Information Processing Systems, pages 3539–3550, 2018. [6] Cynthia Dwork, Moritz Hardt, Toniann Pitassi, Omer Reingold, and Richard Zemel. Fairness through awareness. In Proceedings of the 3rd innovations in theoretical computer science conference, pages 214–226, 2012. [7] Cynthia Dwork, Nicole Immorlica, Adam Tauman Kalai, and Max Leiserson. Decoupled classi- fiers for group-fair and efficient machine learning. In Conference on Fairness, Accountability and Transparency, pages 119–133, 2018. [8] Virginia Eubanks. A child abuse prediction model fails poor families. Wired Magazine, 2018. [9] Lex Fridman and Michael I. Jordan. Machine learning, recommender systems, and the future of AI. Lex Fridman Podcast, 2020. [10] Moritz Hardt, Eric Price, and Nati Srebro. Equality of opportunity in supervised learning. In Advances in neural information processing systems, pages 3315–3323, 2016. [11] Jon Kleinberg, Jens Ludwig, Sendhil Mullainathan, and Ashesh Rambachan. Algorithmic fairness. In AEA papers and proceedings, volume 108, pages 22–27, 2018. [12] Jon Kleinberg and Sendhil Mullainathan. Simplicity creates inequity: implications for fairness, stereotypes, and interpretability. In Proceedings of the 2019 ACM Conference on Economics and Computation, pages 807–808, 2019. [13] Claire Lazar Reich and Suhas Vijaykumar. A possibility in algorithmic fairness: calibrated scores for fair classifications. arXiv e-prints, arXiv:2002.07676v2, 2020. [14] Ziad Obermeyer, Brian Powers, Christine Vogeli, and Sendhil Mullainathan. Dissecting racial bias in an algorithm used to manage the health of populations. Science, 366(6464):447–453, 2019. [15] Ashesh Rambachan, Jon Kleinberg, Sendhil Mullainathan, and Jens Ludwig. An economic approach to regulating algorithms. Technical report, National Bureau of Economic Research, 2020. [16] Muhammad Bilal Zafar, Isabel Valera, Manuel Gomez Rodriguez, and Krishna P Gummadi. Fairness beyond disparate treatment & disparate impact: Learning classification without dis- parate mistreatment. In Proceedings of the 26th international conference on world wide web, pages 1171–1180, 2017. 11
6 Appendix Proof of Proposition 1 Proof. Proof by contradiction. Suppose that L members are not over-represented among positive classifications, so P[G = L|ŷ = 1] ≤ P[G = L|Y = 1]. Also denote the group-specific positive classification rates as µ̂G = P[ŷ = 1|G]. Then, we can show that µ̂H µH ≥ (6) µ̂L µL and µ̂H > µ̂L (7) In particular, we can prove (6) by observing the equivalence of the following statements P[G = L|ŷ = 1] ≤ P[G = L|Y = 1] P[ŷ = 1|G = L]P[G = L] P[Y = 1|G = L]P[G = L] ≤ P[ŷ = 1] P[Y = 1] P[ŷ = 1|G = L] P[Y = 1|G = L] ≤ P[ŷ = 1] P[Y = 1] µ̂L µL ≤ µ̂L P[G = L] + µ̂H P[G = H] µL P[G = L] + µH P[G = H] µ̂L P[G = L] + µ̂H P[G = H] µL P[G = L] + µH P[G = H] ≥ µ̂L µL µ̂H P[G = H] µH P[G = H] P[G = L] + ≥ P[G = L] + µ̂L µL µ̂H P[G = H] µH P[G = H] ≥ µ̂L µL µ̂H µH ≥ µ̂L µL And from this and the assumption that µH > µL the inequality (7) also follows. We will use (6) and (7) to arrive at a contradiction when the true and false positive rates are at least as big for the L group as for the H group. First, using Bayes’ rule, we rewrite the following group-specific error rates in terms of the group-specific positive predictive values P P VA , rates of positive classification µ̂A , and base rates µA . P[Y = 1|G, ŷ = 1]P[ŷ = 1|G] µ̂G T P RG = P[ŷ = 1|G, Y = 1] = = P P VG (8) P[Y = 1|G] µG P[Y = 0|G, ŷ = 1]P[ŷ = 1|G] µ̂G F P RG = P[ŷ = 1|G, Y = 0] = = (1 − P P VG ) (9) P[Y = 0|G] 1 − µG (10) Then for the F P RL to be at least as big as F P RH , it must be that µ̂L µ̂H (1 − P P VL ) ≥ (1 − P P VH ) (11) 1 − µL 1 − µH 1 − P P VL µ̂H (1 − µL ) ≥ (12) 1 − P P VH µ̂L (1 − µH ) 12
According to (7), the RHS of (12) is strictly greater than 1. Therefore it must be that P P VL < P P VH . (13) Meanwhile, for T P RL to be at least as big as T P RH , µ̂L µ̂H P P VL ≥ P P VH (14) µL µH P P VL µ̂H µL ≥ (15) P P VH µH µ̂L µ̂H P P VL µ̂L ≥ µH (16) P P VH µL Due to (6), the RHS is weakly greater than 1. Therefore, it must be that P P VL ≥ P P VH which contradicts with (13). Proof of Theorem 1 Proof. Consider a classifier ŷ for which T P RL ≥ T P RH . As seen in the proof of Proposition 1, T P RG = P P VG µ̂µG G , so we can rewrite the inequality as µ̂L µ̂H P P VL ≥ P P VH (17) µL µH First we suppose condition (i) does not hold and show that condition (ii) then must hold for (17) to be true. We will use our result in the proof of Proposition 1 where we showed that if L members are not over-represented among positive classifications, it necessarily follows that (6) and (7). Notice that (17) can be rewritten as P P VL µµHL ≥ P P VH µ̂µ̂HL , which combined with (6) implies that P P VL ≥ P P VH . Meanwhile to prove the N P V inequality of (ii), notice that if T P RL ≥ T P RH then it follows the groups’ false negative rates must satisfy F N RL ≤ F N RH . Rewriting the group-A F N R using Bayes’ rule gives F N RG = P[ŷ = 0|G, Y = 1] = P[Y =1|G,ŷ=0]P[ŷ=0|G] P[Y =1|G] = (1 − N P VG ) 1−µ̂G µG . So we get the inequality 1 − µ̂L 1 − µ̂H (1 − N P VL ) ≤ (1 − N P VH ) (18) µL µH 1 − N P VL (1 − µ̂H ) µL ≤ (19) 1 − N P VH (1 − µ̂L ) µH Using (7) and the assumption µL < µH , the RHS is strictly less than 1. Therefore on the LHS it must be that N P VL > N P VH . Therefore we have shown that if (i) does not hold then (ii) must hold. Next we show that if (ii) does not hold then (i) must for T P RL ≥ T P RH to be true. We consider the cases P P VL < P P VH and N P VL ≤ N P VH separately: µ̂H µ̂L Case I: Suppose P P VL < P P VH . Then according to (17), it must be that µH < µL , or equivalently µ̂H µH µ̂L < µL . This necessarily implies that (i) holds and the L group is over-represented among positive classifications, P[G = L|ŷ = 1] > P[G = L|Y = 1]. To see why, note that in the proof for Proposition 1, we showed that the negation P[G = L|ŷ = 1] ≤ P[G = L|Y = 1] is equivalent to (6), the negation of µ̂µ̂HL < µµHL . Case II: Suppose N P VL ≤ N P VH . Then according to (19), 13
(1 − µ̂H ) µL ≥1 (20) (1 − µ̂L ) µH (1 − µ̂H ) µH ≥ (21) (1 − µ̂L ) µL Since µH > µL , the RHS is strictly greater than 1 implying (1−µ̂ H) (1−µ̂L ) > 1 and so µ̂H < µ̂L . This clearly means that L members are over-represented among positive classifications and (i) holds. One way to see this is to note that µ̂H < µ̂L implies µ̂µ̂HL < µµHL , which is the negation of (6), a condition equivalent to P[G = L|ŷ = 1] ≤ P[G = L|Y = 1] as seen in the proof of Proposition 1. Proof of Proposition 2 Proof. We will use a lemma in Mitchell et al. (2019), based on the law of iterated expectations: any three random variables W, V, Z satisfy E[W |V, Z] = E[W |E[W |V, Z], Z]. Applied to our setting, we have E[A|X, G] = E[A|E[A|X, G], G] by lemma (22) = E[A|E[A|X], G] by mean independence E[A|X, G] = E[A|X] (23) = E[A|S, G] by definition of S (24) (25) From our conditional mean independence assumption, we also know that the LHS is equal to E[A|X] = S. Combining with the RHS, we have S = E[A|S, G]. Derivation of S|A > 0 We can derive the mean and variance for the distribution of X|Y > c where (X, Y ) are bivariate normal. To find the mean, use the law of iterated expectations: E[X|Y > c] = E[E[X|Y ]|Y > c] (26) The inner expectation E[X|Y = y] is: y − µy E[X|Y = y] = µx + cov(X, Y )( ) (27) σy2 Take the expectation of this given Y > c: E[Y |Y > c] − µy E[E[X|Y ]|Y > c] = µx + cov(X, Y )( ) (28) σy2 c−µy φ( σy ) (µy + σy ( c−µ )) − µy 1−Φ( σy y ) = µx + cov(X, Y ) (29) σy2 c−µy φ( σy ) = µx + cov(X, Y ) c−µ (30) σy (1 − Φ( σy y )) c−µy ρσx φ( σy ) = µx + c−µy (31) 1− Φ( σy ) 14
And we get c−µy ρσx φ( σy ) E[X|Y > c] = µx + c−µ (32) 1− Φ( σy y ) The variance can be computed similarly, using Var(X|Y > c) = E[X 2 |Y > c] − E[X|Y > c]2 . Solving and simplifying with Mathematica gives: c−µy c−µy c−µy ρ2 φ( σy )[σy φ( σy ) − (c − µy )(1 − Φ( σy ))] V ar[X|Y > c] = σx2 (1 − c−µy 2 ) (33) σy (1 − Φ( σy )) 15
You can also read