ifair: learning individually fair representations for ...plahoti/files/ifair_icde...ifair: learning...

23
iFair: Learning Individually Fair Representations for Algorithmic Decision Making Krishna P. Gummadi 2 Preethi Lahoti 1 Gerhard Weikum 1 1 Max Planck Institute for Informatics 2 Max Planck Institute for Software Systems

Upload: others

Post on 20-Aug-2020

0 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: iFair: Learning Individually Fair Representations for ...plahoti/files/iFair_ICDE...iFair: Learning Individually Fair Representations for Algorithmic Decision Making Preethi Lahoti1

iFair: Learning Individually Fair Representationsfor Algorithmic Decision Making

Krishna P. Gummadi2Preethi Lahoti1 Gerhard Weikum1

1 Max Planck Institute for Informatics2 Max Planck Institute for Software Systems

Page 2: iFair: Learning Individually Fair Representations for ...plahoti/files/iFair_ICDE...iFair: Learning Individually Fair Representations for Algorithmic Decision Making Preethi Lahoti1

Fairness in algorithm decision making

Input data X(user records)

Learning algorithm

(e.g., classification,

recommendation)

users

non-protected attributes

(e.g., qualifications or education)

protected attributes(e.g., race or gender)

Decision making system(e.g., loans, visas, job interviews, parole request)

2

Page 3: iFair: Learning Individually Fair Representations for ...plahoti/files/iFair_ICDE...iFair: Learning Individually Fair Representations for Algorithmic Decision Making Preethi Lahoti1

COMPAS – A commercial Risk Prediction Algorithm

3

Risk Score higher for African-American defendants who did not re-offend (ProPublica / Machine Bias)

False Positive Rate

False Negative Rate

Existing Group Fairness Notions: Equalizing Group Level Statistics

(e.g., Equality of Opportunity, Parity)

Page 4: iFair: Learning Individually Fair Representations for ...plahoti/files/iFair_ICDE...iFair: Learning Individually Fair Representations for Algorithmic Decision Making Preethi Lahoti1

Group Fairness is Insufficient

Similarly qualified individualsreceive dissimilar outcomes

Group fairness notions don’t cover individual level unfairness

Top K results satisfy group fairness(fraction of protected in top 10)

4

Page 5: iFair: Learning Individually Fair Representations for ...plahoti/files/iFair_ICDE...iFair: Learning Individually Fair Representations for Algorithmic Decision Making Preethi Lahoti1

• Fairness as Group Property• Fairness of Individuals Ignored

Fairness Challenges

• Assume binary classification task • Fairness concerns ubiquitous across ML tasks

Generalizability Challenges

• Access to protected attribute required• Infeasible due to privacy and legal concerns

• Assumptions about protected attribute• Single Attribute | Binary Value | Fixed Value

Practical Challenges

Can we provide Individual Fairness while keeping utility high?

Can we be Application-Agnostic?

No explicit access to protected attribute?

Can we eliminate assumptions?

Motivation

5

Page 6: iFair: Learning Individually Fair Representations for ...plahoti/files/iFair_ICDE...iFair: Learning Individually Fair Representations for Algorithmic Decision Making Preethi Lahoti1

Learning Individually Fair Representations

• Remove protected attributes• Correlated attributes

• Identify and remove all correlations• Significant Loss in utility

• Utility vs Fairness trade-off

• Individual Fair Representation• Similar individuals should be treated similarly

[Dwork et al 2012]• Given a fairness-aware distance metric• Make similar individuals indistinguishable in

learned representation

users

non-protected attributes

protected attributes

X N x M users !"

Fairrepresentation

Input Data Output data

Retain as much utility as possible while obfuscating information about protected attribute 6

Page 7: iFair: Learning Individually Fair Representations for ...plahoti/files/iFair_ICDE...iFair: Learning Individually Fair Representations for Algorithmic Decision Making Preethi Lahoti1

Problem Formulation

Learn a Low-Rank Representation• Unsupervised probabilistic soft clustering• Learn k – representative prototypes that summarize data• Prototypes: elements of the data space that represents a group of elements

Retain Utility• Minimize reconstruction error

Individual Fairness • Obfuscate Protected Information • Make similar individuals indistinguishable in learned representation• Similar individuals: according to some fairness-aware distance metric over features

7

Page 8: iFair: Learning Individually Fair Representations for ...plahoti/files/iFair_ICDE...iFair: Learning Individually Fair Representations for Algorithmic Decision Making Preethi Lahoti1

Goal: Learn k – prototypes !" that summarize dataProbability that x$ maps to k-th prototype v&

'(," = +, - = . /() = 123 45 26,78∑:∈< 123 45 26,7:

Low-rank representation: marginalize over k prototypes (= ≅ '?@ =: B=)

C/( = D(/() =F"

'(," ⋅ !"

Utility objective: minimize data loss H2 =, B=

Unsupervised Probabilistic Soft Clustering

8

Page 9: iFair: Learning Individually Fair Representations for ...plahoti/files/iFair_ICDE...iFair: Learning Individually Fair Representations for Algorithmic Decision Making Preethi Lahoti1

Fairness-aware Distance Metric

Distance Function: Minkowski p-metrics

! "#, "% = ∑()*+ ,( "#( − "%(. */.

Goal: Learn feature weight vector ,Fairness Intuition: • Set , zero for protected attributes• learn near zero weights for correlated attributes

Fairness Objective: preserve fairness-aware distances012#3 4, 54 = ∑#,%∈7 ! "#, "% − 8! 9"#, 9"%

:

! . applies to original data records and prototypes alike8!(. ) we use simple Euclidean distance

9

Page 10: iFair: Learning Individually Fair Representations for ...plahoti/files/iFair_ICDE...iFair: Learning Individually Fair Representations for Algorithmic Decision Making Preethi Lahoti1

Optimization Problem

Learn a low-rank representation (! ≅ #$% =: (!) • Utility ~ Minimize data loss• Individual Fairness ~ preserve fairness-aware distances

Objective:1 = 2 ⋅ 14567 !, (! + : ⋅ 1;<6= !, (!

GradientDescentOptimization :• L-BFGS algorithm• Model Parameters:

• Learn k prototype vectors JK• Learn feature weight vector L

• Hyperparameters : 2, : and M10

Page 11: iFair: Learning Individually Fair Representations for ...plahoti/files/iFair_ICDE...iFair: Learning Individually Fair Representations for Algorithmic Decision Making Preethi Lahoti1

Experiments

• Does iFair reconcile utility vs individual fairness?• Application-agnostic• Learn fair representation | apply out-of-the-box ML models• Classification task :

• Compas - recidivism risk prediction | protected: race• Census income – income prediction | protected: gender• Credit scores – credit risk prediction | protected: age

• Learning-to-rank task:• Xing – candidate ranking by qualification| protected: gender• Airbnb – housing ranking by rating/price| protected: gender of host

• Inspect group fairness from empirical perspective

11

Page 12: iFair: Learning Individually Fair Representations for ...plahoti/files/iFair_ICDE...iFair: Learning Individually Fair Representations for Algorithmic Decision Making Preethi Lahoti1

Utility vs Individual Fairness

Classification task Ranking task

12

Page 13: iFair: Learning Individually Fair Representations for ...plahoti/files/iFair_ICDE...iFair: Learning Individually Fair Representations for Algorithmic Decision Making Preethi Lahoti1

Utility vs Individual Fairness

Classification task Ranking task

Original Data

13

Page 14: iFair: Learning Individually Fair Representations for ...plahoti/files/iFair_ICDE...iFair: Learning Individually Fair Representations for Algorithmic Decision Making Preethi Lahoti1

Utility vs Individual Fairness

Classification task Ranking task

Naïve Baselines

14

Page 15: iFair: Learning Individually Fair Representations for ...plahoti/files/iFair_ICDE...iFair: Learning Individually Fair Representations for Algorithmic Decision Making Preethi Lahoti1

Utility vs Individual Fairness

Classification task Ranking task

State-of-the-art

15

Page 16: iFair: Learning Individually Fair Representations for ...plahoti/files/iFair_ICDE...iFair: Learning Individually Fair Representations for Algorithmic Decision Making Preethi Lahoti1

Utility vs Individual Fairness

Classification task Ranking task

Proposed Approach

yNN measures consistency of outcome of an individual with its p-nearest neighborsAUC and MAP are measures of utility

16

Page 17: iFair: Learning Individually Fair Representations for ...plahoti/files/iFair_ICDE...iFair: Learning Individually Fair Representations for Algorithmic Decision Making Preethi Lahoti1

Utility vs Individual Fairness

Classification task Ranking task

Proposed Approach

iFair achieves the best utility-fairness tradeoff17

Page 18: iFair: Learning Individually Fair Representations for ...plahoti/files/iFair_ICDE...iFair: Learning Individually Fair Representations for Algorithmic Decision Making Preethi Lahoti1

Obfuscating Protected Information

Adversarial Classification taskInput: Learned fair representationTask: predict protected attribute

iFair outperforms state-of-the-art in obfuscating protected information

18

Page 19: iFair: Learning Individually Fair Representations for ...plahoti/files/iFair_ICDE...iFair: Learning Individually Fair Representations for Algorithmic Decision Making Preethi Lahoti1

Relation to Group fairness

Obfuscating protected information indirectly helps in group fairness

Parity: ! "# = 1 & = 0) = ! "# = 1 & = 0)EqOpp: ! "# = 1 & = 0, # = 1) = ! "# = 1 & = 0, # = 1)

Prediction independent of protected attribute

19

Page 20: iFair: Learning Individually Fair Representations for ...plahoti/files/iFair_ICDE...iFair: Learning Individually Fair Representations for Algorithmic Decision Making Preethi Lahoti1

Conclusions

• iFair – unsupervised probabilistic clustering for individually fair representations• First approach to address individual fairness in learning-to-rank• Outperforms state-of-the art on classification

• Generic and versatile • application-agnostic (e.g., classification, learning-to-rank)• Makes no assumptions about protected attributes

• Pre-processing approach• fairness as a property of the dataset• Post-processing for group fairness can still be applied

• Visit poster # tonight for discussion

20

Page 21: iFair: Learning Individually Fair Representations for ...plahoti/files/iFair_ICDE...iFair: Learning Individually Fair Representations for Algorithmic Decision Making Preethi Lahoti1

Additional slides

Page 22: iFair: Learning Individually Fair Representations for ...plahoti/files/iFair_ICDE...iFair: Learning Individually Fair Representations for Algorithmic Decision Making Preethi Lahoti1

Empirical Observation on Group Fairness

22

Page 23: iFair: Learning Individually Fair Representations for ...plahoti/files/iFair_ICDE...iFair: Learning Individually Fair Representations for Algorithmic Decision Making Preethi Lahoti1

Applying post-processing to iFair representations

Xing

Airbnb

23