AI Ethics, Impossibility Theorems and ... - Chris Stucchio · Chris Stucchio Director of Data...

Preview:

Citation preview

AI Ethics, Impossibility Theorems and Tradeoffs

Chris StucchioDirector of Data Science, Simplhttps://chrisstucchio.com@stucchio

Simplest exampleSupermarket theft prevention algorithm:

1. Make a spreadsheet of item SKU, shrinkage (theft) rate and price

2. Sort list by shrinkage*price.3. Put anti-theft devices on the

SKUs with the highest rates of shrinkage.

sku shrinkage price =b*c

abc123 0.17 $7.24 1.23

def456 0.06 $12.53 0.752

ghi789 0.08 $8.29 0.66

jkl012 0.09 $4.50 0.40

mno234 0.16 $0.99 0.16

Simplest exampleSupermarket theft prevention algorithm:

1. Make a spreadsheet of item SKU, shrinkage (theft) rate and price

2. Sort list by shrinkage*price.3. Put anti-theft devices on the

SKUs with the highest rates of shrinkage.

Whoops!

The plastic box is an anti-theft device which rings an alarm if taken from the store.

Simplest exampleWhy this is bad

- Likely makes black customers feel offended.

- The inconvenience of a slower checkout has a disparate impact (i.e. black customers, who are mostly not stealing, face the inconvenience more)

- Perpetuates racist stereotypes (which the data suggests have an element of truth).

Why this is good

- Reducing theft lowers prices for all customers.

- Without effective anti-theft measures, shops may stop carrying frequently stolen products.

- Resources (anti-theft devices, checkout time) are limited and must be allocated wisely.

- Better to inconvenience 10% of customers than 100%.

Fundamental conflict in AI Ethics

This talk is NOT about...

CheerleedingLots of people in Silicon Valley think there’s a single clear answer.

Many talks are little more than telling the audience this single clear answer.

This talk takes no ethical position - it just tells you which ethical positions you cannot simultaneously take.

ErrorsNo algorithm is 100% accurate.

If you can improve accuracy, you should. There is no ethical question here - only a hard problem in image processing.

Fixing these problems = making more money.

"Racist Camera! No, I did not blink... I'm just Asian!" - jozjozjoz

EuropeI’m an American who lives/works in India.

My knowledge of Europe:

- GDPR is an incoherent and underspecified mess.

- You can force people to forget true facts about you.

- Too many regulatory regimes.- Delicious cheese.

Sorry!

Everything I know about Europe

Artificial IntelligenceThis talk is about decision theory.

Every ethical quandary I discuss applies to humans as well as machines.

Only benefit of human decision processes: easy obfuscation.

You can cheaply and easily run an algorithm on test data to measure an effect. You can’t do the same on, e.g., a judge or loan officer.

Classical Ethical Theories

UtilitarianismIt is bad to be murdered, raped, or sent to jail.

Utilitarianism tries to minimize the bad things in the world while maximizing the good things. In math terms, find the policy which minimizes:

Harm(policy) = A x (# of murders) + B x (years people are stuck in jail) + ...

A:B is a conversion between murders and jail time. We are indifferent between jailing someone for B years and preventing A murders.

Procedural fairnessA classical belief is that decisions should be blind to certain individual traits (t in this case):

∀t1 ∀t2 f(x, t1) == f(x, t2)

Intuitively: Me (a foreigner), my Brahmin wife, our non-Brahmin maid or Prime Minister Modi should get the same justice given the same facts.

San Francisco Ethical TheoriesEpistemic note: I am attempting to mathematically state premises, but the proponents of those premises often prefer for them to be kept informal:

“As engineers, we’re trained to pay attention to the details, think logically, challenge assumptions that may be incorrect (or just fuzzy), and so on. These are all excellent tools for technical discussions. But they can be terrible tools for discussion around race, discrimination, justice...because questioning the exact details can easily be perceived as questioning the overall validity of the effort, or the veracity of the historical context.”

- Urs Hölzle, S.V.P. at Google

Allocative FairnessImportant concept is protected class. What are these?

● In US: Blacks/Hispanics. Asians are a de-jure protected class, but de-facto not. Women, sometimes homosexuals.

● In India: Scheduled Castes and OBCs. Muslims/other religious minorities only in Tamil Nadu and Kerala.

Allocative fairness is when a certain statistic is equal across protected classes.

(Some variants choose favored classes and replace “=” with “<=”. Indian college admissions have favored castes, Americans have favored races.)

Allocative fairness, base rates and group boundariesExample:

- 25% of both Scotsmen and Englishmen get 1200 on SAT.- Cutoff for getting into college is 1200 on SAT. - No allocative harm.

No True Scotsman will score < 1200 on their SAT

- Suddenly 100% of True Scotsmen get into college vs 25% of Englishmen and 0% of False Scotsmen.

- Allocative harm is created!

Representational Fairness/Honor Culture

San Francisco Google notices nothing

Indian Google notices everything

AI may notice things we don’t want it to“Bias should be the expected result whenever even an unbiased algorithm is used to derive regularities from any data; bias is the regularities discovered.”

Semantics derived automatically from language corpora necessarily contain human biases

Core problems in AI ethics

Can’t simultaneously maximize two objectives

Constrained max <= Global max

Outcomes and protected classes are correlated

All About HyderabadThings from Hyderabad:

- Great Biryani- Dum ke Roat (best cookies in

India)- My wife- Pervasive fraud on many

lending platforms (including Simpl)

Simpl’s Underwriting AlgoSimpl (my employer) is an Indian microlending platform/payment processor.

Input data: Old data + specific fraud behavior.

Algorithm: A big, unstructured black box (think random forest or neural network).

Prediction target: 30 day delinquency, i.e. “has the user paid their bill within 30 days of the first bill due date”.

(I can’t reveal what specific fraud behavior is - think of it as something like installing the चोर App on the Evil Play Store.)

Simpl’s Underwriting AlgoIf we exclude चोर app, Hyderabad is a strong indicator of delinquency.

If we include चोर app, that is the dominant feature. It’s also highly correlated with Hyderabad and results in a very high rejection rate there.

Fact: Lending is a low margin business. Lower accuracy results in fraudsters stealing all the money.

TradeoffsUtilitarianism: We can offer loans to many Mumbaikars and Delhiites, and a smaller number of Hyderabadis. That’s a strict Pareto improvement over offering no loans to anyone.

Procedural fairness: Hyderabadis who install the चोर App (that’s “thief” in Hindi) are treated the same as Punekars who do the same (and vice versa).

Group unfairness: Our policy has a disparate impact on Hyderabadis - they get fewer loans issued.

Group reputation: We have learned a true but unflattering fact about Hyderabad: there is a disproportionate number of fraudsters there [1].

[1] Another possibility is a proportionate number of disproportionately active fraudsters.

100%This is the rejection rate for Hyderabad loan applications at many other NBFCs.

In the American context this is called redlining.

(Context: In 1934, the USA Federal Housing Association drew a red line around black neighborhoods and told banks not to issue mortgages there.)

Simpl lives in a competitive marketIf we choose to service Hyderabad with no disparities, we’ll run out of money and stop serving Hyderabad. The other NBFCs won’t.

Net result: Hyderabad is redlined by competitors and still gets no service.

Our choice: Keep the fraudsters out, utilitarianism over group rights.

A couple of weeks ago my mother in law - who lives in Hyderabad - informed me that Simpl approved her credit line.

Computational Criminology(Screenshot of ProPublica’s Article)

COMPAS Algorithm137 factors go into a black box model - age, gender, criminal history, single mother, father went to jail, number of friends who use drugs, etc.

ProPublica claims it’s “biased against blacks”.

How does COMPAS work?Dressel and Farid replicated COMPAS predictions using Logistic Regression on only 7 features: age, sex, #juvenile misdemeanors, #juvenile felonies, #adult crimes, crime degree (most recent), and crime charge (most recent).

Goal: Explainable model with same predictions as COMPAS. It’s predictions:

- 25 year old male who kidnapped and raped 6 women: high risk- 43 year old female who shoplifted a toy for her kid one Christmas: low risk

(Adding race does not significantly improve accuracy.)

Checking CalibrationProPublica checked the calibration of the algorithm, and found a disparity that was “almost statistically significant” at p=0.057.

(Flashback to CrunchConf 2015: Multiple Comparisons - Make your boss happy with false positives. Correcting ProPublica’s multiple comparisons, p=0.114.)

Conclusion: A black or white person with a risk score of 5 have equal probability of recidivism.

(Key point is cells 28-29 in their R Script.) Accuracy and Racial Biases of Recidivism Prediction Instruments, Julia J. Dressel

Distribution of scores

The “bias” comes from base ratesSimplified numbers: High risk == 60% chance of recidivism, low risk = 20%.

Black people: 60% labelled high risk * 40% chance of no recidivism/ = 24% chance of “labelled high risk, didn’t recidivate”.

White people: 30% labelled high risk * 40% chance of no recidivism = 12% chance of “labelled high risk, didn’t recidivate”.

(Not going to do a calculation with 10 deciles in a 40 minute talk.)

The “bias” comes from base ratesCalibration means: P(~recidivate|black, high risk) = P(~recidivate|white, high risk)

Group fairness means: P(high risk | black, ~recidivate) = P(high risk | ~recidivate)

Bayes theorem provides a relationship:

P(high risk | ~recidivate) = P(~recidivate | high risk) P(high risk) / P(recidivate)

The “bias” comes from base ratesP(high risk | ~recidivate) = P(~recidivate | high risk) P(high risk) / P(recidivate)

Disparity is caused by base rates being different. P(high risk) is significantly higher for blacks than whites, and as a result, P(high risk | ~recidivate) must also be higher.

If calibration is equal across groups, then by necessity false positive rates must differ if base rates do.

Impossibility Theorem

Theorem (Kleinberg, Mullainathan, Raghavan): A predictive algorithm can only be well calibrated and have equal false positive/negative rates if it achieves either perfect accuracy or base rates are equal.

Advice for journalists: If you run an analysis the way ProPublica did, you are mathematically guaranteed to get a conclusion of “bias.” There’s no risk of getting a bad story.

How to fix the disparityRacially specific thresholds: Raise the “high risk” cutoff for one group, not the other.

P(high risk | ~recidivate) = P(~recidivate | high risk) P(high risk) / P(recidivate)

Procedurally: If the high risk threshold is 3 robberies, a black criminal at this threshold is paroled while a white criminal is jailed.

Raising risk thresholds makes these terms decrease.

Which in turn makes this decrease.

Utilitarian costTheorem: For a utility function of the form U = A*Crime - B*#people jailed, the maxima is achieved when all risk thresholds are equal.

Proof: Ordinary calculus. (Technical assumption: pdf of risk scores is continuous, non-vanishing.)

Intuitive meaning: To reduce crime as much as possible, put people in jail in order of highest risk first.

Algorithmic decision making and the cost of fairness

Victimization disparityCrime victimization is disproportionately intraracial.

If crime goes up by 9% due to releasing black criminals from jail, this will have the following (ballpark) effect.

- Crime victimization will go up by about 7.5% among non-hispanic whites.

- Crime victimization will go up by about 37.5% among blacks.

[1] These stats are approximate, gained by multiplying Broward County crime increase percentages by national demographic and crime victimization statistics.

@AnechoicMedia14, avg yearly # violent crimes with injury

TradeoffsUtilitarianism: COMPAS reduces crime. Using a fairer algorithm would cause people to be raped and murdered.

Procedural fairness: COMPAS treats a black bike thief identically to a white bike thief, and a black serial killer identically to a white serial killer.

Group fairness (for victims): COMPAS reduces - but does not eliminate - racial disparities in crime victimization.

Group unfairness (for criminals): COMPAS causes a disparity in false positive rates.

Representational unfairness: COMPAS reveals that blacks are significantly more likely than whites to recidivate.

Fairness in lending

When Words Sweat: Identifying Signals for Loan Default in the Text of Loan Applications

Religious people and people with medical issues less likely to repay loans.

FICO Score- 35% payment history (or lack of payments)- 30% debt burden (how much you currently owe, relative to income, assets,

assessed limits)- 15% length of history- 10% types of credit (credit card + mortgage + consumer unsecured > only

credit card)- 10% hard pulls (consumer explicitly applying for a loan)

Key point: 100% of FICO is based on a person’s actions (ignoring identity theft).

Procedurally fair, by definition.

Equality of Opportunity in Supervised Learning

Blacks have lower FICO scores than Asians

Blacks less likely to repay than asians

Exposing tradeoffsTheorem: Using a single FICO threshold, one can only achieve the same approval rate across all groups at points x where CDF(x, A) = CDF(x, B).

Procedurally fair choice: Choose a fixed FICO cutoff, say 600. Then we reject 75% of blacks, 25% of Asians, violating the principle of group fairness.

Group fair choice: Choose a fixed approval level - say 75%, implying a risk cutoff of 600 for Asians and 410 for blacks. This violates principle of procedural fairness.

Exposing tradeoffsTheorem: Any procedurally fair loan cutoff is not utility maximizing, except at FICO=850 (the max).

At FICO=600, approx 80% of Asian borrowers will repay loans and about 60% of Black borrowers will.

Utilitarian choice is to issue loans to Asian applicants with a 590 FICO (repayment probability ~78%) over black applicants with a 600 FICO.

Exposing tradeoffsTheorem: Any procedurally fair loan cutoff is not utility maximizing, except at FICO=850 (the max).

Proof: Draw a horizontal line to intersect the point where the FICO cutoff (a vertical line) meets the lower performing graph.

Find the point where horizontal line intersects the higher performing graph. This cutoff applied to the higher group, and the original cutoff applied to the lower group, is utility maximizing.

(Technical assumption: the density of the higher performing group is non-zero in this region.)

These guys are a better credit risk

than these guys.

Predatory LendingIn 2008, the media used the term “predatory lending” to refer to making loans a bank knew would not be repaid.

Bad loans can cause many years of financial hardship, and also lower one’s credit score (exacerbating disparities).

These second order effects of attempts at fairness can drown out the first order effects, and harm those they are meant to help. (Whether this happens is a complex quantitative question.)

Delayed Impact of Fair Machine Learning

TradeoffsUtilitarianism: Racially discriminatory FICO cutoffs maximize utility.

Group fairness (today): Differently calibrated FICO cutoffs can result in either equal false positive rates, or equal loan acceptance rates.

Group fairness (tomorrow): Other differently calibrated cutoffs can reduce credit score disparities (but not simultaneously with group fairness today).

Group unfairness: Using FICO results in far fewer blacks being issued loans, and can reduce aggregate credit scores.

Representational unfairness: FICO reveals that blacks more likely to default than whites. Calibration graphs reveal that some of this default is not explained by financial history (i.e. FICO is biased in favor of blacks).

Procedural fairness: FICO score is independent of protected traits.

There is no policy choice which satisfies all ethical principles.

The laws of mathematical optimization still apply

ConclusionEarly on I said I wouldn’t be giving any ethical prescriptions.

I will, however, give one meta-ethical prescription: formalize your ethical principles as terms in your utility function or as constraints.

It is nearly certain that tradeoffs between these principles exist, and if we don’t acknowledge this, we run the risk of unknowingly engaging in bad actions.