Week 4: Rights, Policies, and Fairness in AI

DSAN 5450: Data Ethics and Policy
Spring 2026, Georgetown University

Class Sessions
Author
Affiliation

Jeff Jacobs

Published

Wednesday, February 4, 2026

Open slides in new window →

Fair Machine Learning? In This Economy?

  1. A cool algorithm ✅ (other DSAN classes)
  2. [Possibly benign, possibly biased] Training data ✅ (HW1)
  3. \(\Longrightarrow\) Exploitation of below-minimum-wage human labor 😞🤐 (Dube et al. 2020, like and subscribe yall ❤️)

Part 3: The “Training Data Bottleneck”

With so much technical progress […] why is there so little real enterprise success? The answer all too often is that many enterprises continue to be bottlenecked by one key ingredient: the large amounts of labeled data [needed] to train these new systems.

Human Labor

Gray and Suri (2019)

Computer Scientists Being Responsible at Georgetown!

  • (PS… UMD CS class of 2013 extremely overrepresented here 😜 go Terps )

What Comes With Human Labels? Human Biases!

Crawford and Paglen (2019)

Crawford and Paglen (2019)

AI Machine Go Brrr

Image tagged in ImageNet with the label “BOLSHEVIK”. From Crawford and Paglen (2019)

\(\leadsto\)

Biases In Our Brains → Biases in Our Models → Material Effects

  • “Reification”: Pretentious word for an important phenomenon, whereby talking about something (e.g., race) as if it was real ends up leading to it becoming real (having real impacts on people’s lives)1

On average, being classified as White as opposed to Coloured would have more than quadrupled a person’s income. (Pellicer and Ranchhod 2023)

Apartheid South Africa passport Source

Johannesburg Star, 21 March 1986

Reification in Science

“Intelligence” Testing

Gould (1980)
More Generally
  • Goodhart’s Law: “When a measure becomes a target, it ceases to be a good measure”
  • Cat-and-mouse game between goals (🚩) and ways of measuring progress towards goals (also 🚩)

Metaethics

A scary-sounding word that just means:

“What we talk about when we talk about ethics”,

in contrast to

“What we talk about when we talk about [insert particular ethical framework here]”

Reflective Equilibrium

  • Most criticisms of any framework boil down to, “great in theory, but doesn’t work in practice”
  • The way to take this seriously: reflective equilibrium
  • Introduced by Rawls (1951), but popularized by Rawls (1971)

From Awad et al. (2022)

Normative vs. Descriptive Quick Recap

  • Languages are arbitrary conventions for communication
  • Ethical systems build on this language to non-arbitrarily mark out things that are good/bad
  • Society wouldn’t be too different if we “shuffled” words (we’d just vibrate our vocal chords differently), but would be very different if we “shuffled” good/bad labeling

Hare (1952)

Changing Normative Values: Capitalism \(\leftrightarrow\) “Protestant Ethic”

  • Big changes in history are associated with changes in this good/bad labeling!
  • Max Weber (second most cited sociologist ever*): Protestant value system enabled capitalist system by relabeling what things are good vs. bad (Weber 1904):

Jesus said to his disciples, “Truly, I say to you, only with difficulty will a rich person enter the kingdom of heaven. Again I tell you, it is easier for a camel to go through the eye of a needle than for a rich person to enter the kingdom of God.” (Matthew 19:23-24)

Oh, were we loving God worthily, we should have no love at all for money! (St. Augustine 1874, pg. 28)

 

*(…but remember: REIFICATION!)

The earliest capitalists lacked legitimacy in the moral climate in which they found themselves. One of the means they found [to legitimize their behavior] was to appropriate the evaluative vocabulary of Protestantism. (Skinner 2012, pg. 157)

Calvinism added [to Luther’s doctrine] the necessity of proving one’s faith in worldly activity, [replacing] spiritual aristocracy of monks outside of/above the world with spiritual aristocracy of predestined saints within it. (pg. 121).

Contemporary Example: Palestine

  • Very few of the relevant empirical facts are in dispute, since opening of crucial archives to three so-called “New Historians” in the 1980s. So why do people still argue?
  • Ilan Pappe, one of these historians, concluded from this material that:
    • The Israeli state was built upon a massive ethnic cleansing, and
    • Is not morally justifiable (Pappe 2006)

The immunity Israel has received over the last fifty years encourages others, regimes and oppositions alike, to believe that human and civil rights are irrelevant in the Middle East. The dismantling of the mega-prison in Palestine will send a different, and more hopeful, message.

  • Utilitarian “wrapped” around Kantian
  • Benny Morris, another of these historians, concluded that:
    • The Israeli state was built upon a massive ethnic cleansing, and
    • Is morally justifiable (Morris 1987)

A Jewish state would not have come into being without the uprooting of 700,000 Palestinians. Therefore it was necessary to uproot them. There was no choice but to expel that population. It was necessary to cleanse the hinterland and cleanse the border areas and cleanse the main roads.

  • Straightforwardly utilitarian!

Individual Ethics \(\rightarrow\) Social Ethics

Standard Counterargument to Consequentialism

Millions are kept permanently happy, on the one simple condition that a certain lost soul on the far-off edge of things should lead a life of lonely torture (James 1891)

  • Modern example: people “out there” suffer so we can have iPhones, etc.

Le Guin (1973)

One Solution: Individual Rights

  • Rights are vetoes which individuals can use to cancel out collective/institutional decisions which affect them (key example for us: right to privacy)
  • Rawls/liberalism: individual rights are lexically prior to “efficiency” and/or distributional concerns
  • Why the buzzword “lexically”? Enter (non-scary) math!
  • We can put lowercase letters of English alphabet in an order: \(\texttt{a} \prec \texttt{b} \prec \cdots \texttt{z}\)
  • We can put capital letters of English alphabet in an order: \(\texttt{A} \prec \texttt{B} \prec \cdots \prec \texttt{Z}\)
  • What if we need to sort stuff with both types? We can decide that capital letters are lexically prior to lowercase letters, giving us a combined ordering:

Dworkin (1977)

\[ \boxed{\texttt{A} \prec \texttt{B} \prec \cdots \prec \texttt{Z} \prec \texttt{a} \prec \texttt{b} \prec \cdots \prec \texttt{z}} \]

Lexical Ordering (I Tricked You 😈)

  • You thought I was just talking about letters, but they’re actually variables: capital letters are rights, lowercase letters are distributive principles

\[ \underbrace{\texttt{A} \prec \texttt{B} \prec \cdots \prec \texttt{Z}}_{\mathclap{\substack{\text{Individual Rights} \\ \text{Basic Goods}}}} \phantom{\prec} \prec \phantom{\prec} \underbrace{\texttt{a} \prec \texttt{b} \prec \cdots \prec \texttt{z}}_{\mathclap{\substack{\text{Distributive Principles} \\ \text{Money and whatnot}}}} \]

Better Metaphor Than Letters

  • Letters are where Rawls gets “lexically prior” from, but letters are total orderings (we know where every letter “stands” in relation to every other letter)
  • Better metaphor: UDC (2yr school) rule for allocating basketball tickets:

\[ \text{Seniors} \prec \text{Juniors} \]

  • \(\Rightarrow\) If you’re a Senior, whether at “top” or “bottom” of a ranking of Seniors, you still (by Constitution) have priority over all Juniors…
  • Why is this more helpful? Because we don’t need to define the rankings within classes to know the rankings between classes in this case

Implications for “Fair” AI?

  • This system of (Constitutionally-established) rights, as vetos over utilitarian decisions of legislator (Congress, Parliament, etc.), came about* to protect individuals against opaque structures with authority to make decisions that affect these individuals’ lives
  • …AI is an opaque structure with authority to make decisions that affect individuals’ lives 🤔

\[ \underbrace{\texttt{A} \prec \texttt{B} \prec \cdots \prec \texttt{Z}}_{\mathclap{\substack{\text{Individual Rights} \\ \text{Constraints on objective functions}}}} \phantom{\prec} \phantom{\prec} \prec \phantom{\prec} \phantom{\prec} \underbrace{\texttt{a} \prec \texttt{b} \prec \cdots \prec \texttt{z}}_{\mathclap{\substack{\text{Distributive Principles} \\ \text{Maximized objective functions}}}} \]

Psychological Synthesis: Two-Level Utilitarianism

  • It would be exhausting to compute Nash equilibrium strategies for every scenario
  • Instead, we can develop heuristics that work for most cases, then reevaluate and update when we encounter tough cases
  • (Brings us back to reflective equilibrium!)

Kahneman (2011)

Individual vs. Social Morality

  • That was all already hard enough, to reason about individual morality
  • Now add in the fact that we live in a society 😰
  • Things that happen depend not only on our choices but also the choices of others

Enter Game Theory

  • A tool for analyzing how individual choices + choices of others \(\rightarrow\) outcomes!
  • Example: You (\(A\)) and a friend (\(B\)) committed a robbery, and you’re brought into the police station for questioning.
  • You’re placed in separate rooms, and each of you is offered a plea deal: if you testify while your partner stays silent, you go free and they go to jail for 3 years.
  • Otherwise, if you both stay silent, they have very little evidence and can only jail you for 1 year
  • However, there’s a catch: if you both confess, you both get two years in jail, since they now have maximal evidence

Individual Decision-Making

  • Let’s think through \(A\)’s best responses to the possible choices \(B\) could make:
  • If \(B\) stays silent, what is \(A\)’s best option?
    • Staying silent results in 1 year of jail
    • Testifying results in 0 years of jail
    • So it is better to testify
  • If \(B\) testifies, what is \(A\)’s best option?
    • Staying silent results in 3 years of jail
    • Testifying results in 2 years of jail
    • So it is better to testify
  • The result: regardless of what \(B\) does, \(A\) is better off testifying!

The Social Outcome

  • The game is symmetric, so the same logic applies for \(B\)
  • Conclusion: the outcome of the game will be \(s^* = (\text{Testify}, \text{Testify})\)
  • This is called a Nash equilibrium: no player \(i\) can make themselves better off by deviating from \(s_i\)

How Do We Fix This? Conventions!

  • We encounter this type of problem every day if we drive! You (\(A\)) and another driver (\(B\)) arrive at an intersection:
\(B\)
Stop Drive
\(A\) Stop \(-1,-1\) \(-3,\phantom{-}0\)
Drive \(\phantom{-}0, -3\) \(-10,-10\)
  • If both stop, we’re mostly bored: \(u_A = -1\)
  • If we stop and the other person drives, we’re mad that they got to go and we didn’t: \(u_A = -3\)
  • If both drive, we crash: \(u_A = -10\)

Without A Convention

  • We’re “frozen”: this game has no unique Nash equilibrium, so we cannot say (on the basis of individual rationality) what will happen!
  • Without a convention: power/aggression takes over. “War of all against all”, only the strong survive, etc. (life is “nasty, brutish, and short”)
\(B\)
Stop Drive
\(A\) Stop \({\color{orange}\cancel{\color{black}-1}},{\color{lightblue}\cancel{\color{black}-1}}\) \(\boxed{-3},\boxed{0}\)
Drive \(\boxed{0}, \boxed{-3}\) \({\color{orange}\cancel{\color{black}-10}},{\color{lightblue}\cancel{\color{black}-10}}\)
  • If \(A\)’s aggression is \(\Pr(s_A = \textsf{Drive}) = X \sim \mathcal{U}[0,1]\), \(B\)’s aggression is \(\Pr(s_B = \textsf{Drive}) = Y \sim \mathcal{U}[0,1]\), what happens at individual and societal levels?

\[ \begin{align*} \mathbb{E}[u_A] = \mathbb{E}[u_B] &= \int_{0}^{1}\int_{0}^{1}\left(x - 2y -8xy - 1\right)dy \, dx = -3.5 \\ \underbrace{\mathbb{E}\mkern-3mu\left[u_A + u_B\right]}_{\mathclap{\text{Utilitarian Social Welfare}}} &= -3.5 \end{align*} \]

The Convention of Traffic Lights

  • If we don’t want a world where \(\text{Happiness}(i) \propto \Pr(i \text{ more aggro than }j)\), we can introduce traffic lights:
  • Now in “correlated equilibrium”, where we ensure* coordinated \(\Pr((\textsf{Drive}, \textsf{Stop})) = 0.5\), \(\Pr((\textsf{Stop}, \textsf{Drive})) = 0.5\)
  • \(\mathbb{E}[u_A] = (0.5)(0) + (0.5)(-3) = -1.5\)
  • \(\mathbb{E}[u_B] = (0.5)(-3) + (0.5)(0) = -1.5\)
\(B\)
Stop Drive
\(A\) Stop \({\color{orange}\cancel{\color{black}-1}},{\color{lightblue}\cancel{\color{black}-1}}\) \(\boxed{-3},\boxed{0}\)
Drive \(\boxed{0}, \boxed{-3}\) \({\color{orange}\cancel{\color{black}-10}},{\color{lightblue}\cancel{\color{black}-10}}\)
  • Empirical (anthropological) findings across literally thousands of different cultures throughout the world: people are willing to give up rewards to ensure fairness (see, e.g., Henrich et al. (2001))

*(through, for example, traffic laws: equal in theory… In practice? Another story)

So How Should We Make/Choose Conventions?

  • Hobbes (1668): Only way out of “war of all against all” is to surrender all power to one sovereign (the Leviathan)
  • Rousseau (1762): Social contract
  • [Big big ~200 year gap here… can you think of why? Hint: French Revolution in 1789]
  • Rawls (1971): Social contract behind the “veil of ignorance”
    • If we didn’t know where we were going to end up in society, how would we set it up?

Rawls’ Veil of Ignorance

  • Probably the most important tool for policy whitepapers!
  • “Justice as fairness” (next week: fairness in AI 😜)
  • We don’t know whether we’ll be \(A\) or \(B\) in the intersection game, so we’d choose the traffic light!
  • More profoundly: We don’t know what race, gender, class, ethnicity, sexuality, disability status we’ll have; We don’t know whether we’ll be Israeli or Palestinian; we don’t know whether we’ll own means of production or own only our labor power (and thus have to sell it on a market to survive)… 🤔

Nuts and Bolts for Fairness

One Final Reminder

  • Industry rule #4080: Cannot “prove” \(q(x) = \text{``Algorithm }x\text{ is fair''}\)! Only \(p(x) \implies q(y)\):

\[ \underbrace{p(x)}_{\substack{\text{Accept ethical} \\ \text{framework }x}} \implies \underbrace{q(y)}_{\substack{\text{Algorithms should} \\ \text{satisfy condition }y}} \]

  • Before: possible ethical frameworks (values for \(x\))
  • Now: possible fairness criteria (values for \(y\))

Categories of Fairness Criteria

Roughly, approaches to fairness/bias in AI can be categorized as follows:

Fairness
Context-Free
  • Single-Threshold Fairness
  • Equal Prediction
  • Equal Decision
Context-Sensitive
  • Fairness via Similarity Metric(s)
  • Causal Definitions
  • [Today] Context-Free Fairness: Easier to grasp from CS/data science perspective; rooted in “language” of Machine Learning (you already know much of it, given DSAN 5000!)
  • But easy-to-grasp notion \(\neq\) “good” notion!
  • Your job: push yourself to (a) consider what is getting left out of the context-free definitions, and (b) the loopholes that are thus introduced into them, whereby people/computers can discriminate while remaining “technically fair”

Laws: Often Perfectly “Technically Fair”

Ah, la majestueuse égalité des lois, qui interdit au riche comme au pauvre de coucher sous les ponts, de mendier dans les rues et de voler du pain!

(Ah, the majestic equality of the law, which prohibits rich and poor alike from sleeping under bridges, begging in the streets, and stealing loaves of bread!)

Anatole France, Le Lys Rouge (France 1894)

Context-Free Fairness

From Introduction to Formal Languages and Automata, Simon Fraser University (2006). This figure summarizes the Chomsky Hierarchy of Languages, developed by Noam Chomsky, who also has a lot to say about Ethics and Policy!

The Brogrammer’s Criterion

df.drop(columns=["race"], inplace=True)
  • Racism solved, folks! 🥳🎊🎉 End of the course, have a great rest of your data science career ✌️

(No) Fairness Through Unawareness

  • HW1: Using tiny sample (\(N < 10K\)) of Florida voter registrations… RandomForestClassifier (default settings, no hyperparameter tuning, no cross-validation, no ensembling with other methods) will predict self-reported race with \(>90\%\) accuracy (in balanced sample) from just surname and county of residence
    • Can reach \(70\text{-}75\%\) with just surname or just county of residence
  • Also in HW1: Facebook ad matching service provides over 1,000 different user attributes for (micro)targeting

From Datta et al. (2017)

To Make It Even More Concrete…

  • Bloomberg analysis of neighborhoods with same-day delivery from Amazon:

Figure 2: Figures from Ingold and Soper (2016)

We Can Do (A Bit) Better…

  • Use random variables to model inferences made by an algorithm (or a human!)
  • \(\implies\) fairness by statistically equalizing loan rejections, error rate, etc. between groups
  • Obvious societal drawback: equality does not ameliorate the effects of past injustices (see: police contact vs. trust-in-government plot from last week)
    • This one we saw coming, given “context-free” nature!
  • Less obvious mathematical drawback: impossibility results (because algebra 😳)
    • Roughly: can’t satisfy [more than] two statistical fairness criteria at once; similar to how setting \(\Pr(X) = p\) also determines \(\Pr(\text{not }X) = 1 - p\), or how plugging \(x = 3\) into \(x + y = 5\) leaves only one possibility \(y = 2\)
  • BUT, “impossibility” \(\neq\) impossibility: (a) one criteria may be “all you need” in given setting; (b) can derive more robust measures by “relaxing” confusion-matrix measures

What the Fairness Measures Will Feel Like For Now

  • (They will get more robust and will incorporate context soon, I promise!)

Who Remembers… 🎉Confusion Matrices!!!🎉

  • Terrifyingly higher stakes than in DSAN 5000, however!
  • Now \(D = 1\) could literally mean “shoot this person” or “throw this person in jail for life”

From Mitchell et al. (2021)

Our First Fairness Measure (Finally)!

  • Standard across the literature: Random Variable \(A_i\) “encoding” membership in protected/sensitive group. In HW1, for example:

\[ A_i = \begin{cases} 0 &\text{if }i\text{ self-reported ``white''} \\ 1 &\text{if }i\text{ self-reported ``black''} \end{cases} \]

  • Notice: choice of mapping into \(\{0, 1\}\) here non-arbitrary!

  • We want our models/criteria to be descriptively but also normatively robust; e.g.:

  • If (antecedent I hold, though majority in US do not) one believes that ending (much less repairing) centuries of unrelenting white supremacist violence here might require asymmetric race-based policies,

  • Then our model should allow different normative labels and differential weights on

    \[ \begin{align*} \Delta &= (\text{Fairness} \mid A = 1) - (\text{Fairness} \mid A = 0) \\ \nabla &= (\text{Fairness} \mid A = 0) - (\text{Fairness} \mid A = 1) \end{align*} \]

    despite the descriptive fact that \(\Delta = -\nabla\).

Where Descriptive and Normative Become Intertwined

  • Allowing this asymmetry is precisely what enables bring descriptive facts to bear on normative concerns!
  • Mathematically we can always “flip” the mapping from racial labels into \(\{0, 1\}\)
  • But this (in a precise, mathematical sense: namely, isomorphism) implies that we’re treating racial categorization as the same type of phenomenon as driving on left or right side of road (see: prev slides on why we make the descriptive vs. normative distinction)
  • (See also: Sweden’s Dagen H!)

“Fairness” Through Equalized Positive Rates

\[ \Pr(D = 1 \mid A = 0) = \Pr(D = 1 \mid A = 1) \]

  • This works specifically for discrete, binary-valued categories
  • For general attributes (whether discrete or continuous!), generalizes to:

\[ D \perp A \iff \Pr(D = d, A = a) = \Pr(D = d)\Pr(A = a) \]

“Fairness” Through Equalized Error Rates

  • Equalized False Positive Rate:

\[ \Pr(D = 1 \mid Y = 0, A = 0) = \Pr(D = 1 \mid Y = 0, A = 1) \]

  • Equalized False Negative Rate:

\[ \Pr(D = 0 \mid Y = 1, A = 0) = \Pr(D = 0 \mid Y = 1, A = 1) \]

  • For general (non-binary) attributes: \((D \perp A) \mid Y\):

\[ \Pr(D = d, A = a \mid Y = y) = \Pr(D = d \mid Y = y)\Pr(A = a \mid Y = y) \]

NO MORE EQUATIONS! 😤

  • Depending on your background and/or learning style (say, visual vs. auditory), you may be able to look at equations on previous slides and “see” what they’re “saying”
  • If your brain works similarly to mine, however, your eyes glazed over, you began dissociating, planning an escape route, looking for open ventilation ducts, etc.
  • If you’re in the latter group, welcome to the cult of Probabilistic Graphical Models 😈

Baby Steps: A Real-World Confusion Matrix

Labeled Low-Risk Labeled High-Risk
Didn’t Do More Crimes True Negative False Positive
Did More Crimes False Negative True Positive
  • What kinds of causal connections and/or feedback loops might there be between our decision variable (low vs. high risk) and our outcome variable (did vs. didn’t do more crimes)
  • What types of policy implications might this process have, after it “runs” for several “iterations”?
  • Why might some segments of society, with some shared ethical framework(s), weigh the “costs” of false negatives and false positives differently from other segments of society with different shared ethical framework(s)?
  • (Non-rhetorical questions!)

Tumbling into Fairness

“Repetition is the mother of perfection” - Dwayne Michael “Lil Wayne” Carter, Jr.

So How Should We Make/Choose Conventions?

  • Hobbes (1668): Only way out of “war of all against all” is to surrender all power to one sovereign (the Leviathan)
  • Rousseau (1762): Social contract
  • [Big big ~200 year gap here… can you think of why? Hint: French Revolution in 1789]
  • Rawls (1971): Social contract behind the “veil of ignorance”
    • If we didn’t know where we were going to end up in society, how would we set it up?

Rawls’ Veil of Ignorance

  • Probably the most important tool for policy whitepapers!
  • “Justice as fairness” (next week: fairness in AI 😜)
  • We don’t know whether we’ll be \(A\) or \(B\) in the intersection game, so we’d choose the traffic light!
  • More profoundly: We don’t know what race, gender, class, ethnicity, sexuality, disability status we’ll have; We don’t know whether we’ll be Israeli or Palestinian; we don’t know whether we’ll own means of production or own only our labor power (and thus have to sell it on a market to survive)… 🤔

Nuts and Bolts for Fairness

One Final Reminder

  • Industry rule #4080: Cannot “prove” \(q(x) = \text{``Algorithm }x\text{ is fair''}\)! Only \(p(x) \implies q(y)\):

\[ \underbrace{p(x)}_{\substack{\text{Accept ethical} \\ \text{framework }x}} \implies \underbrace{q(y)}_{\substack{\text{Algorithms should} \\ \text{satisfy condition }y}} \]

  • Before: possible ethical frameworks (values for \(x\))
  • Now: possible fairness criteria (values for \(y\))

Categories of Fairness Criteria

Roughly, approaches to fairness/bias in AI can be categorized as follows:

Fairness
Context-Free
  • Single-Threshold Fairness
  • Equal Prediction
  • Equal Decision
Context-Sensitive
  • Fairness via Similarity Metric(s)
  • Causal Definitions
  • [Today] Context-Free Fairness: Easier to grasp from CS/data science perspective; rooted in “language” of Machine Learning (you already know much of it, given DSAN 5000!)
  • But easy-to-grasp notion \(\neq\) “good” notion!
  • Your job: push yourself to (a) consider what is getting left out of the context-free definitions, and (b) the loopholes that are thus introduced into them, whereby people/computers can discriminate while remaining “technically fair”

Laws: Often Perfectly “Technically Fair”

Ah, la majestueuse égalité des lois, qui interdit au riche comme au pauvre de coucher sous les ponts, de mendier dans les rues et de voler du pain!

(Ah, the majestic equality of the law, which prohibits rich and poor alike from sleeping under bridges, begging in the streets, and stealing loaves of bread!)

Anatole France, Le Lys Rouge (France 1894)

Context-Free Fairness

From Introduction to Formal Languages and Automata, Simon Fraser University (2006). This figure summarizes the Chomsky Hierarchy of Languages, developed by Noam Chomsky, who also has a lot to say about Ethics and Policy!

The Brogrammer’s Criterion

df.drop(columns=["race"], inplace=True)
  • Racism solved, folks! 🥳🎊🎉 End of the course, have a great rest of your data science career ✌️

(No) Fairness Through Unawareness

  • HW1: Using tiny sample (\(N < 10K\)) of Florida voter registrations… RandomForestClassifier (default settings, no hyperparameter tuning, no cross-validation, no ensembling with other methods) will predict self-reported race with \(>90\%\) accuracy (in balanced sample) from just surname and county of residence
    • Can reach \(70\text{-}75\%\) with just surname or just county of residence
  • End-of-HW1 discussion: Facebook ad matching service provides over 1,000 different user attributes for (micro)targeting

From Datta et al. (2017)

We Can Do (A Bit) Better…

  • Use random variables to model inferences made by an algorithm (or a human!)
  • \(\implies\) fairness by statistically equalizing loan rejections, error rate, etc. between groups
  • Obvious societal drawback: equality does not ameliorate the effects of past injustices (see: police contact vs. trust-in-government plot from last week)
    • This one we saw coming, given “context-free” nature!
  • Less obvious mathematical drawback: impossibility results (because algebra 😳)
    • Roughly: can’t satisfy [more than] two statistical fairness criteria at once; similar to how setting \(\Pr(X) = p\) also determines \(\Pr(\neg X) = 1 - p\), or how plugging \(x = 3\) into \(x + y = 5\) leaves only one possibility \(y = 2\)
  • BUT, “impossibility” \(\neq\) impossibility: (a) one criteria may be “all you need” in given setting; (b) can derive more robust measures by “relaxing” confusion-matrix measures

What the Fairness Measures Will Feel Like For Now

  • (They will get more robust and will incorporate context soon, I promise!)

Who Remembers… 🎉Confusion Matrices!!!🎉

  • Terrifyingly higher stakes than in DSAN 5000, however!
  • Now \(D = 1\) could literally mean “shoot this person” or “throw this person in jail for life”

From Mitchell et al. (2021)

Baby Steps: A Real-World Confusion Matrix

Labeled Low-Risk Labeled High-Risk
Didn’t Do More Crimes True Negative False Positive
Did More Crimes False Negative True Positive
  • What kinds of causal connections and/or feedback loops might there be between our decision variable (low vs. high risk) and our outcome variable (did vs. didn’t do more crimes)
  • What types of policy implications might this process have, after it “runs” for several “iterations”?
  • Why might some segments of society, with some shared ethical framework(s), weigh the “costs” of false negatives and false positives differently from other segments of society with different shared ethical framework(s)?
  • (Non-rhetorical questions!)

Categories of Fairness Criteria

Roughly, approaches to fairness/bias in AI can be categorized as follows:

Fairness
Context-Free
  • Single-Threshold Fairness
  • Equal Prediction
  • Equal Decision
Context-Sensitive
  • Fairness via Similarity Metric(s)
  • Causal Definitions
  • [Today] Context-Free Fairness: Easier to grasp from CS/data science perspective; rooted in “language” of Machine Learning (you already know much of it, given DSAN 5000!)
  • But easy-to-grasp notion \(\neq\) “good” notion!
  • Your job: push yourself to (a) consider what is getting left out of the context-free definitions, and (b) the loopholes that are thus introduced into them, whereby people/computers can discriminate while remaining “technically fair”

Laws: Often Perfectly “Technically Fair”

Ah, la majestueuse égalité des lois, qui interdit au riche comme au pauvre de coucher sous les ponts, de mendier dans les rues et de voler du pain!

(Ah, the majestic equality of the law, which prohibits rich and poor alike from sleeping under bridges, begging in the streets, and stealing loaves of bread!)

Anatole France, Le Lys Rouge (France 1894)

(No) Fairness Through Unawareness

  • HW1: Using tiny sample (\(N < 10K\)) of Florida voter registrations… RandomForestClassifier (default settings, no hyperparameter tuning, no cross-validation, no ensembling with other methods) will predict self-reported race with \(>90\%\) accuracy (in balanced sample) from just surname and county of residence
    • Can reach \(70\text{-}75\%\) with just surname or just county of residence
  • Also in HW1: Facebook ad matching service provides over 1,000 different user attributes for (micro)targeting

From Datta et al. (2017)

Last One I Promise

Predicting self-reported whiteness with 70% accuracy

Last One I Promise

Predicting self-reported non-whiteness with 90% accuracy

We Can Do (A Bit) Better…

  • Use random variables to model inferences made by an algorithm (or a human!)
  • \(\implies\) fairness by statistically equalizing loan rejections, error rate, etc. between groups
  • Obvious societal drawback: equality does not ameliorate the effects of past injustices (see: police contact vs. trust-in-government plot from last week)
    • This one we saw coming, given “context-free” nature!
  • Less obvious mathematical drawback: impossibility results (because algebra 😳)
    • Roughly: can’t satisfy [more than] two statistical fairness criteria at once; similar to how setting \(\Pr(X) = p\) also determines \(\Pr(\text{not }X) = 1 - p\), or how plugging \(x = 3\) into \(x + y = 5\) leaves only one possibility \(y = 2\)
  • BUT, “impossibility” \(\neq\) impossibility: (a) one criteria may be “all you need” in given setting; (b) can derive more robust measures by “relaxing” confusion-matrix measures

Definitions and (Impossibility) Results

(tldr:)

  • We have information \(X_i\) about person \(i\), and
  • We’re trying to predict a binary outcome \(Y_i\) involving \(i\).
  • So, we use ML to learn a risk function \(r: \mathcal{R}_{X_i} \rightarrow \mathbb{R}\), then
  • Use this to make a binary decision \(\widehat{Y}_i = \mathbf{1}[r(X_i) > t]\)

Protected/Sensitive Attributes

  • Standard across the literature: Random Variable \(A_i\) “encoding” membership in protected/sensitive group. In HW1, for example:

\[ A_i = \begin{cases} 0 &\text{if }i\text{ self-reported ``white''} \\ 1 &\text{if }i\text{ self-reported ``black''} \end{cases} \]

  • Notice: choice of mapping into \(\{0, 1\}\) here non-arbitrary!

  • We want our models/criteria to be descriptively but also normatively robust; e.g.:

  • If (antecedent I hold, though majority in US do not) one believes that ending (much less repairing) centuries of unrelenting white supremacist violence here might require asymmetric race-based policies,

  • Then our model should allow different normative labels and differential weights on

    \[ \begin{align*} \Delta &= (\text{Fairness} \mid A = 1) - (\text{Fairness} \mid A = 0) \\ \nabla &= (\text{Fairness} \mid A = 0) - (\text{Fairness} \mid A = 1) \end{align*} \]

    despite the descriptive fact that \(\Delta = -\nabla\).

Where Descriptive and Normative Become Intertwined

  • Allowing this asymmetry is precisely what enables bring descriptive facts to bear on normative concerns!
  • Mathematically we can always “flip” the mapping from racial labels into \(\{0, 1\}\)
  • But this (in a precise, mathematical sense: namely, isomorphism) implies that we’re treating racial categorization as the same type of phenomenon as driving on left or right side of road (see: prev slides on why we make the descriptive vs. normative distinction)
  • (See also: Sweden’s Dagen H!)

Lab Time!

“Fairness” Through Equalized Positive Rates (EPR)

 

\[ \boxed{\Pr(D = 1 \mid A = 0) = \Pr(D = 1 \mid A = 1)} \]

  • This works specifically for discrete, binary-valued categories
  • For general attributes (whether discrete or continuous!), generalizes to:

\[ \boxed{D \perp A} \iff \Pr(D = d, A = a) = \Pr(D = d)\Pr(A = a) \]

Imagine you learn that a person received a scholarship (\(D = 1\)); [with equalized positive rates], this fact would give you no knowledge about the race (or sex, or class, as desired) \(A\) of the individual in question. (DeDeo 2016)

Achieving Equalized Positive Rates

  • The good news: if we want this, there is a closed-form solution: take your datapoints \(X_i\) and re-weigh each point to obtain \(\widetilde{X}_i = w_iX_i\), where

    \[ w_i = \frac{\Pr(Y_i = 1)}{\Pr(Y_i = 1 \mid A_i = 1)} \]

    and use derived dataset \(\widetilde{X}_i\) to learn \(r(X)\) (via ML algorithm)… Why does this work?

  • Let \(\mathcal{X}_{\text{fair}}\) be the set of all possible reweighted versions of \(X_i\) ensuring \(Y_i \perp A_i\). Then

\[ \widetilde{X}_i = \min_{X_i' \in \mathcal{X}_{\text{fair}}}\textsf{distance}(X_i', X_i) = \min_{X_i' \in \mathcal{X}_{\text{fair}}}\underbrace{KL(X_i' \| X_i)}_{\text{Relative entropy!}} \]

  • The bad news: nobody in the fairness in AI community read DeDeo (2016), which proves this using information theory? Idk. It has a total of 22 citations 😐

“Fairness” Through Equalized Error Rates

  • Equalized positive rates didn’t take outcomes \(Y_i\) into account…

    • (Even if \(A_i = 1 \Rightarrow Y_i = 1, A_i = 0 \Rightarrow Y_i = 0\), we’d have to choose \(\widehat{Y}_i = c\))
  • This time, we consider the outcome \(Y\) that

  • Equalized False Positive Rate (EFPR):

\[ \Pr(D = 1 \mid Y = 0, A = 0) = \Pr(D = 1 \mid Y = 0, A = 1) \]

  • Equalized False Negative Rate (EFNR):

\[ \Pr(D = 0 \mid Y = 1, A = 0) = \Pr(D = 0 \mid Y = 1, A = 1) \]

  • For general (non-binary) attributes: \((D \perp A) \mid Y\):

\[ \Pr(D = d, A = a \mid Y = y) = \Pr(D = d \mid Y = y)\Pr(A = a \mid Y = y) \]

⚠️ LESS EQUATIONS PLEASE! 😤

  • Depending on your background and/or learning style (say, visual vs. auditory), you may be able to look at equations on previous two slides and “see” what they’re “saying”
  • If your brain works similarly to mine, however, your eyes glazed over, you began dissociating, planning an escape route, etc., the moment \(> 2\) variables appeared
  • If you’re in the latter group, welcome to the cult of Probabilistic Graphical Models 😈

Your first PGM, illustrating hypothesized causal relationships between three random variables \(Y\) (outcome), \(D\) (decision), and \(A\) (protected attribute). The \(Y\) node is shaded to indicate that it is an observed value in our model, rendering the unobserved values \(D\) and \(A\) independent conditional on it. If I was elected Emperor of Math, equations would be abolished in favor of PGMs.

 

NoteEqualized False Negative/Positive Rates
  • Our first measure that 🥳🎉matches a principle of justice in society!!!🕺🪩
  • Blackstone’s Ratio: “It is better that ten guilty persons escape, than that one innocent suffers.” (Blackstone 1769)
  • (…break time!)

Back to Equalized Error Rates

  • Blackstone’s Ratio: “It is better that ten guilty persons escape, than that one innocent suffers.” (Blackstone 1769)
  • Mathematically \(\Rightarrow \text{Cost}(FPR) = 10\cdot \text{Cost}(FNR)\)
  • Legally \(\Rightarrow\) beyond reasonable doubt standard for conviction
  • EFPR \(\iff\) rates of false conviction should be the same for everyone, including members of different racial groups.
    • Violated when black people are disproportionately likely to be incorrectly convicted, as if a lower evidentiary standard were applied to black people.

One Final Context-Free Criterion: Calibration

  • A risk function \(r(X)\) is calibrated if

\[ \Pr(Y = 1 \mid r(X) = v_r) = v_r \]

  • (Sweeping a lot of details under the rug), I see this one as: the risk function “tracks” real-world probabilities
  • Then, \(r(X)\) is calibrated by group if

\[ \Pr(Y = y \mid r(X) = v_r, A = a) = v_r \]

Impossibility Results

  • tldr: We cannot possibly achieve all three of equalized positive rates (often also termed “anti-classification”), classification parity, and calibration (regardless of base rates)
  • More alarmingly: We can’t even achieve both classification parity and calibration, except in the special case of equal base rates

“Impossibility” vs. Impossibility

  • Sometimes “impossibility results” are, for all intents and purposes, mathematical curiosities: often there’s some pragmatic way of getting around them
  • Example: “Arrow’s Impossibility Theorem”
    • [In theory] It is mathematically impossible to aggregate individual preferences into societal preferences
    • [The catch] True only if people are restricted to ordinal preferences: “I prefer \(x\) to \(y\).” No more information allowed
    • [The way around it] Allow people to indicate the magnitude of their preferences: “I prefer \(x\) 5 times more than \(y\)
  • In this case, though, there are direct and (often) unavoidable real-world barriers that fairness impossibility imposes 😕

Arrow’s Impossibility Theorem

  • Aziza, Bogdan, and Charles are competing in a fitness test with four events. Goal: determine who is most fit overall
Run Jump Hurdle Weights
Aziza 10.1” 6.0’ 40” 150 lb
Bogdan 9.2” 5.9’ 42” 140 lb
Charles 10.0” 6.1’ 39” 145 lb
  • We can rank unambiguously on individual events: Jump: Charles \(\succ_J\) Aziza \(\succ_J\) Bogdan
  • Now, axioms for aggregation:
    • \(\text{WP}\) (Weak Pareto Optimality): if \(x \succ_i y\) for all events \(i\), \(x \succ y\)
    • \(\text{IIA}\) (Independence of Irrelevant Alternatives): If a fourth competitor enters, but Aziza and Bogdan still have the same relative standing on all events, their relative standing overall should not change
  • Long story short: only aggregation that can satisfy these is “dictatorship”: choose one event, give it importance of 100%, the rest have importance 0% 😰

ProPublica vs. Northpointe

  • This is… an example with 1000s of books and papers and discussions around it! (A red flag 🚩, since, obsession with one example may conceal much wider range of issues!)
  • But, tldr, Northpointe created a ML algorithm called COMPAS, used by court systems all over the US to predict “risk” of arrestees
  • In 2016, ProPublica published results from an investigative report documenting COMPAS’s racial discrimination, in the form of violating equal error rates between black and white arrestees
  • Northpointe responded that COMPAS does not discriminate, as it satisfies calibration
  • People have argued about who is “right” for 8 years, with some progress, but… not a lot

So… What Do We Do?

  • One option: argue about which of the two definitions is “better” for the next 100 years (what is the best way to give food to the poor?)

It appears to reveal an unfortunate but inexorable fact about our world: we must choose between two intuitively appealing ways to understand fairness in ML. Many scholars have done just that, defending either ProPublica’s or Northpointe’s definitions against what they see as the misguided alternative. (Simons 2023)

  • Another option: study and then work to ameliorate the social conditions which force us into this realm of mathematical impossibility (why do the poor have no food?)

The impossibility result is about much more than math. [It occurs because] the underlying outcome is distributed unevenly in society. This is a fact about society, not mathematics, and requires engaging with a complex, checkered history of systemic racism in the US. Predicting an outcome whose distribution is shaped by this history requires tradeoffs because the inequalities and injustices are encoded in data—in this case, because America has criminalized Blackness for as long as America has existed.

Why Not Both??

  • On the one hand: yes, both! On the other hand: fallacy of the “middle ground”
  • We’re back at descriptive vs. normative:
    • Descriptively, given 100 values \(v_1, \ldots, v_{100}\), their mean may be a good way to summarize, if we have to choose a single number
    • But, normatively, imagine that these are opinions that people hold about fairness.
    • Now, if it’s the US South in 1860 and \(v_i\) represents person \(i\)’s approval of slavery, from a sample of 100 people, then approx. 97 of the \(v_i\)’s are “does not disapprove” (Rousey 2001) — in this case, normatively, is the mean \(0.97\) the “correct” answer?
  • We have another case where, like the “grass is green” vs. “grass ought to be green” example, we cannot just “import” our logical/mathematical tools from the former to solve the latter! (However: this does not mean they are useless! This is the fallacy of the excluded middle, sort of the opposite of the fallacy of the middle ground)
  • This is why we have ethical frameworks in the first place! Going back to Rawls: “97% of Americans think black people shouldn’t have rights”  \(\nimplies\)“black people shouldn’t have rights”, since rights are a primary good

References

Awad, Edmond, Sydney Levine, Michael Anderson, Susan Leigh Anderson, Vincent Conitzer, M. J. Crockett, Jim A. C. Everett, et al. 2022. “Computational Ethics.” Trends in Cognitive Sciences 26 (5): 388–405. https://doi.org/10.1016/j.tics.2022.02.009.
Blackstone, William. 1769. Commentaries on the Laws of England, Volume 2: A Facsimile of the First Edition of 1765-1769. University of Chicago Press. https://books.google.com?id=wTwpCgAAQBAJ.
Crawford, Kate, and Trevor Paglen. 2019. “Excavating AI: The Politics of Training Sets for Machine Learning.” AI Now Institute, NYU. https://excavating.ai.
Datta, Anupam, Matt Fredrikson, Gihyuk Ko, Piotr Mardziel, and Shayak Sen. 2017. “Proxy Non-Discrimination in Data-Driven Systems.” July 25, 2017. https://doi.org/10.48550/arXiv.1707.08120.
DeDeo, Simon. 2016. “Wrong Side of the Tracks: Big Data and Protected Categories.” June 23, 2016. https://doi.org/10.48550/arXiv.1412.4643.
Dube, Arindrajit, Jeff Jacobs, Suresh Naidu, and Siddharth Suri. 2020. “Monopsony in Online Labor Markets.” American Economic Review: Insights 2 (1): 33–46. https://doi.org/10.1257/aeri.20180150.
Dworkin, Ronald. 1977. Taking Rights Seriously. A&C Black. https://books.google.com?id=Ud_UAAAAQBAJ.
Fields, Barbara J., and Karen E. Fields. 2012. Racecraft: The Soul of Inequality in American Life. Verso Books. https://books.google.com?id=I2vnDwAAQBAJ.
France, Anatole. 1894. Le Lys Rouge (The Red Lily). G. Wells. https://books.google.com?id=sLLFz0rSakgC.
Gould, Stephen Jay. 1980. The Mismeasure of Man (Revised and Expanded). W. W. Norton & Company. https://books.google.com?id=RTjfmxTpsVsC.
Gray, Mary L., and Siddharth Suri. 2019. Ghost Work: How to Stop Silicon Valley from Building a New Global Underclass. HarperCollins. https://books.google.com?id=FFJkDwAAQBAJ.
Hare, R. M. 1952. The Language of Morals. OUP Oxford. https://books.google.com?id=q4oSDAAAQBAJ.
Henrich, Joseph, Robert Boyd, Samuel Bowles, Colin Camerer, Ernst Fehr, Herbert Gintis, and Richard McElreath. 2001. “In Search of Homo Economicus: Behavioral Experiments in 15 Small-Scale Societies.” American Economic Review 91 (2): 73–78. https://doi.org/10.1257/aer.91.2.73.
Hobbes, Thomas. 1668. Leviathan: With Selected Variants from the Latin Edition of 1668. Hackett Publishing.
Ingold, David, and Spencer Soper. 2016. “Amazon Doesn’t Consider the Race of Its Customers. Should It?” Bloomberg, April 21, 2016. http://www.bloomberg.com/graphics/2016-amazon-same-day/.
James, William. 1891. “The Moral Philosopher and the Moral Life.” International Journal of Ethics 1 (3): 330–54. https://www.jstor.org/stable/2375309.
Kahneman, Daniel. 2011. Thinking, Fast and Slow. Farrar, Straus and Giroux. https://books.google.com?id=ZuKTvERuPG8C.
Le Guin, Ursula K. 1973. The Ones Who Walk Away from Omelas: A Story. HarperCollins. https://books.google.com?id=fZayDQAAQBAJ.
Mitchell, Shira, Eric Potash, Solon Barocas, Alexander D’Amour, and Kristian Lum. 2021. “Algorithmic Fairness: Choices, Assumptions, and Definitions.” Annual Review of Statistics and Its Application 8 (1): 141–63. https://doi.org/10.1146/annurev-statistics-042720-125902.
Morris, Benny. 1987. The Birth of the Palestinian Refugee Problem, 1947-1949. Cambridge University Press. https://books.google.com?id=lP7MRAAACAAJ.
Pappe, Ilan. 2006. The Ethnic Cleansing of Palestine. Simon and Schuster. https://books.google.com?id=Xhq9DwAAQBAJ.
Pellicer, Miquel, and Vimal Ranchhod. 2023. “Understanding the Effects of Racial Classification in Apartheid South Africa.” Journal of Development Economics 160 (January): 102998. https://doi.org/10.1016/j.jdeveco.2022.102998.
Rawls, John. 1951. “Outline of a Decision Procedure for Ethics.” The Philosophical Review 60 (2): 177–97. https://doi.org/10.2307/2181696.
———. 1971. A Theory of Justice: Original Edition. Harvard University Press. https://books.google.com?id=tUEO9SuNG1oC.
Rousey, Dennis C. 2001. “Friends and Foes of Slavery: Foreigners and Northerners in the Old South.” Journal of Social History 35 (2): 373–96. https://www.jstor.org/stable/3790193.
Rousseau, Jean-Jacques. 1762. The Social Contract. Geneva: J. M. Dent. https://books.google.com?id=G1MOAQAAIAAJ.
Simons, Josh. 2023. Algorithms for the People: Democracy in the Age of AI. Princeton University Press. https://books.google.com?id=hIKIEAAAQBAJ.
Skinner, Quentin. 2012. Visions of Politics, Volume 1: Regarding Method. Cambridge: Cambridge University Press.
St. Augustine. 1874. The Works of Aurelius Augustine: Lectures or Tractates on the Gospel According to St. John, v. 2. T. & T. Clark. https://books.google.com?id=XlRHAQAAMAAJ.
Weber, Max. 1904. The Protestant Ethic and the Spirit of Capitalism. Courier Corporation. https://books.google.com?id=ez7CAgAAQBAJ.

Appendix: Slightly Deeper Dive into Descriptive vs. Normative Judgements

Descriptive (Is) Normative (Ought)
Grass is green (true) Grass ought to be green (?)
Grass is blue (false) Grass ought to be blue (?)

Easy Mode: Descriptive Judgements

How did you acquire the concept “red”?

  • People pointed to stuff with certain properties and said “red” (or “rojo” or “红”), as pieces of an intersubjective communication system
  • These descriptive labels enable coordination, like driving on left or right side of road!
  • Nothing very profound or difficult in committing to this descriptive coordination: “for ease of communication, I’ll vibrate my vocal chords like this (or write these symbols) to indicate \(x\), and vibrate them like this (or write these other symbols) to indicate \(y\)
  • Linguistic choices, when it comes to description, are arbitrary*: Our mouths can make these sounds, and each language is a mapping: [combinations of sounds] \(\leftrightarrow\) [things]

*(Tiny text footnote: Except for, perhaps, a few fun but rare onomatopoetic cases)

What Makes Ethical Judgements “More Difficult”?

How did you acquire the concept “good”?

  • People pointed to actions with certain properties and said “good” (and pointed at others and said “bad”), as part of instilling values in you
  • “Grass is green” just links two descriptive referents together, while “Honesty is good” takes the descriptive concept “honesty” and links it with the normative concept “good”
  • In doing this, parents/teachers/friends are doing way more than just linking sounds and things in the world (describing): they are also prescribing rules of moral conduct!
  • Normative concepts go beyond “mere” communication: course of your life / future / [things that matter deeply to people] differ if you act on one set of norms vs. another
  • \(\implies\) Ethics centrally involves non-arbitrarily-chosen commitments!

Footnotes

  1. Fields and Fields (2012), for example, coined “racecraft” to describe reification of blackness in US… more on this later!↩︎