+ All Categories
Home > Documents > Criteo AI Lab - Recent Results on Algorithmic Fairness and ......Criteo AI Lab, Paris, October 2,...

Criteo AI Lab - Recent Results on Algorithmic Fairness and ......Criteo AI Lab, Paris, October 2,...

Date post: 26-Feb-2021
Category:
Upload: others
View: 5 times
Download: 0 times
Share this document with a friend
26
Recent Results on Algorithmic Fairness and Meta-Learning Massimiliano Pontil Computational Statistics and Machine Learning Istituto Italiano di Tecnologia and Department of Computer Science University College London 4th Annual Machine Learning in the Real World workshop (MLiTRW) Criteo AI Lab, Paris, October 2, 2019
Transcript
Page 1: Criteo AI Lab - Recent Results on Algorithmic Fairness and ......Criteo AI Lab, Paris, October 2, 2019 Plan I Fair empirical risk minimization I Using labeled and unlabeled data I

Recent Results on Algorithmic Fairness and Meta-Learning

Massimiliano Pontil

Computational Statistics and Machine LearningIstituto Italiano di Tecnologia

andDepartment of Computer Science

University College London

4th Annual Machine Learning in the Real World workshop (MLiTRW)Criteo AI Lab, Paris, October 2, 2019

Page 2: Criteo AI Lab - Recent Results on Algorithmic Fairness and ......Criteo AI Lab, Paris, October 2, 2019 Plan I Fair empirical risk minimization I Using labeled and unlabeled data I

Plan

I Fair empirical risk minimization

I Using labeled and unlabeled data

I Multi-task approach

I Learning fair representations

I Online meta-learning

2

Page 3: Criteo AI Lab - Recent Results on Algorithmic Fairness and ......Criteo AI Lab, Paris, October 2, 2019 Plan I Fair empirical risk minimization I Using labeled and unlabeled data I

Algorithmic fairness

I Aim: ensure that learning algorithms do not treat subgroups in the population“unfairly”

I How: impose “fairness” constraints (different notions)

I Difficulty: study computationally efficient algorithms with statistical guaranteesw.r.t. both the risk and the fairness measure

Binary classification setting: let µ be a prob. distribution on X × S × {−1,+1}, whereS = {a, b} is the sensitive variable set. We wish to find a solution f ∗ of

minf ∈F

{P(f (X , S)6=Y

)s.t. “f is fair”

}

3

Page 4: Criteo AI Lab - Recent Results on Algorithmic Fairness and ......Criteo AI Lab, Paris, October 2, 2019 Plan I Fair empirical risk minimization I Using labeled and unlabeled data I

Fairness constraints(see e.g. [Hardt et al., 2016, Zafar et al., 2017])

I Equal opportunity (EO): P(f (X ,S)>0|Y=1, S=a

)= P

(f (X , S)>0|Y=1,S=b

)I Equalized odds (EOd): f (X , S) and S are conditionally independent given Y , i.e.

P(f (X , S)>0|Y=y ,S=a

)= P

(f (X ,S)>0|Y=y ,S=b

), y ∈ {−1, 1}

I Demographic parity (DP): P(f (X , S)>0|S=a

)= P

(f (X , S)>0|S=b

)I We may also loose each constraint by requiring the l.h.s. to be close to the r.h.s.

4

Page 5: Criteo AI Lab - Recent Results on Algorithmic Fairness and ......Criteo AI Lab, Paris, October 2, 2019 Plan I Fair empirical risk minimization I Using labeled and unlabeled data I

Statistical learning setting

I Let ` : R× Y → R be a loss function and let L be the associated risk:

L(f ) = E[`(f (X ),Y )], for f : X → Y

I Conditional risk of f for the positive class in group s:

L+,s(f ) = E[`(f (X ),Y )|Y = 1, S = s]

I We relax the fairness constraint by using a loss function in place of the 01-loss andintroduce a parameter ε ∈ [0, 1]. For EO, we obtain

minf ∈F

{L(f ) s.t.

∣∣L+,a(f )−L+,b(f )∣∣ ≤ ε} (1)

5

Page 6: Criteo AI Lab - Recent Results on Algorithmic Fairness and ......Criteo AI Lab, Paris, October 2, 2019 Plan I Fair empirical risk minimization I Using labeled and unlabeled data I

Fair empirical risk minimization (FERM)[Donini et al. NeurIPS 2018]

I Distribution µ is unknown and we only have a data sequence (xi , si , yi )ni=1 sampled

independently from µ. We then consider the empirical problem

minf ∈F

{L(f ) s.t.

∣∣L+,a(f )−L+,b(f )∣∣ ≤ ε} (2)

where ε is a parameter linked to ε

I Empirical risk L(f ) = 1n

n∑i=1

`(f (xi ), yi )

I Empirical risk for the positive samples in group g : L+,g (f ) = 1n+,g

∑i∈I+,g

`(yi , f (xi ))

with I+,g = {i : yi=1, si=g} and n+,g = |I+,g |, g ∈ {a, b}

6

Page 7: Criteo AI Lab - Recent Results on Algorithmic Fairness and ......Criteo AI Lab, Paris, October 2, 2019 Plan I Fair empirical risk minimization I Using labeled and unlabeled data I

Statistical analysis of FERM

We say a class of functions F is learnable (wrt. loss `) if:

supf ∈F

∣∣L(f )− L(f )∣∣ ≤ B(δ, n,F), with lim

n→∞B(δ, n,F) = 0

Proposition 1. Let δ ∈ (0, 1). If F is learnable f ∗ solves (1) and f solves (2) withε = ε+

∑g∈{a,b} B(δ, n+,g ,F) then with prob. ≥ 1− 6δ it holds simultaneously:

L(f )− L(f ∗) ≤ 2B(δ, n,F)∣∣L+,a(f )− L+,b(f )∣∣ ≤ ε+ 2

∑g∈{a,b}

B(δ, n+,g ,F)

7

Page 8: Criteo AI Lab - Recent Results on Algorithmic Fairness and ......Criteo AI Lab, Paris, October 2, 2019 Plan I Fair empirical risk minimization I Using labeled and unlabeled data I

Implications of the bound

I Bound implies that a solution f of (2) is close to a solution of f ∗ of (1) both interms of the risk and fairness constraint

I But how do we find f ? We would like to solve problem (2) for the hard(misclassification) loss:

minf ∈F

n∑i=1

1{f (xi ) 6=yi} (3)∣∣P {f (x)>0|y=1, s=a}−P {f (x)>0|y=1, s=b}∣∣ ≤ ε

I We propose to replace the hard loss in the risk with the (larger) hinge loss, and thehard loss in the fairness constraint with a linear loss

8

Page 9: Criteo AI Lab - Recent Results on Algorithmic Fairness and ......Criteo AI Lab, Paris, October 2, 2019 Plan I Fair empirical risk minimization I Using labeled and unlabeled data I

Fair learning with kernels

I Linear model f (·) = 〈w , φ(·)〉, with φ : X → H a kernel-induced feature map

I For the linear loss, the fairness constraint takes the form∣∣〈w , ua − ub〉

∣∣ ≤ ε, whereug is the barycenter of the positive points in group g :

ug =1

n+,g

∑i :∈I+,g

φ(xi ), g ∈ {a, b}

I We consider the regularized empirical risk minimization problem

minw∈H

n∑i=1

`(yi 〈w , φ(xi )〉)+λ‖w‖2 s.t.∣∣〈w , ua − ub〉

∣∣≤ε λ > 0

9

Page 10: Criteo AI Lab - Recent Results on Algorithmic Fairness and ......Criteo AI Lab, Paris, October 2, 2019 Plan I Fair empirical risk minimization I Using labeled and unlabeled data I

Form of the optimal classifier[Chzhen et al. NeurIPS 2019]

Proposition. Let η(x , s) = E [Y |X = x ,S = s] be the regression function. If for eachs ∈ {0, 1} the mapping t 7→ P (η(X ,S) ≤ t |S = s) is continuous on (0, 1), then anoptimal classifier f ∗ can be obtained for all (x , s) ∈ Rd × {a, b} as

fθ(x , a) = 1{1≤η(x ,a)(2− θP(Y=1,S=a)

)}, fθ(x , b) = 1{1≤η(x ,b)(2+ θP(Y=1,S=b)

)}

where θ ∈ [0, 2] solves the equation

EX |S=a [η(X , a)fθ(X , a)]

P (Y = 1 |S = a)=EX |S=b [η(X , b)fθ(X , b)]

P (Y = 1 |S = b).

I Similar result when S is not included as a predictor

10

Page 11: Criteo AI Lab - Recent Results on Algorithmic Fairness and ......Criteo AI Lab, Paris, October 2, 2019 Plan I Fair empirical risk minimization I Using labeled and unlabeled data I

Leveraging labeled and unlabeled

I FERM leaves open the question of designing a computationally efficient andstatistically consistent estimator for problem (*)

I Alternative method: estimate η from a labeled sample and θ from an independentunlabeled sample by minimizing the empirical difference of equal opportunity (DEO)

∆(f , µ) =

∣∣∣∣∣ EX |S=aη(X , a)fθ(X , a)

EX |S=aη(X , a)−

EX |S=bη(X , b)fθ(X , b)

EX |S=bη(X , b)

∣∣∣∣∣Theorem (informal). If η → η as n→∞, under mild additional assumptions theproposed estimator is consistent w.r.t. both accuracy and fairness:

limn,N→∞

E(Dn,DN)[∆(f , µ)] = 0 and limn,N→∞

E(Dn,DN)[R(f )] ≤ R(f ∗)

11

Page 12: Criteo AI Lab - Recent Results on Algorithmic Fairness and ......Criteo AI Lab, Paris, October 2, 2019 Plan I Fair empirical risk minimization I Using labeled and unlabeled data I

Modified validation procedure

I In experiments, we employ a two steps 10-fold CV procedure:

– Step 1: shortlist all hyperparameters with accuracy above a certain percentage (wechoose 90%) of the best accuracy

– Step 2, from the list, select the hyperparameter with highest fairness (i.e. lowest DEO)

I We compare:

– Naıve SVM, validated with a standard nested 10-fold cross validation

– SVM with the novel validation procedure

– The method by [Hardt et al., 2016] applied to the best SVM

– The method [Zafar et al., 2017] (code provided by the authors for the linear case∗)

∗Python code: https://github.com/mbilalzafar/fair-classification12

Page 13: Criteo AI Lab - Recent Results on Algorithmic Fairness and ......Criteo AI Lab, Paris, October 2, 2019 Plan I Fair empirical risk minimization I Using labeled and unlabeled data I

Experiments

Comparison between different methods. DEO is normalized in [0, 1] column-wise. The closer a

point is to the origin, the better the result

The proposed methods slightly decrease accuracy while greatly improving in the fairness measure

Code: https://github.com/jmikko/fair_ERM13

Page 14: Criteo AI Lab - Recent Results on Algorithmic Fairness and ......Criteo AI Lab, Paris, October 2, 2019 Plan I Fair empirical risk minimization I Using labeled and unlabeled data I

Taking advantage of multitask learning[Oneto et al. AIES 2019]

I We consider group specific models: f (x , s) = 〈ws , x〉and a multitask learning (MTL) formulation

minw1,...wk∈H

k∑s=1

Ls(ws) +λ

k

k∑s=1

‖ws − w0‖2 + (1− λ)‖w0‖2

I Regularization around a common mean encourages tasksimilarities

I We impose additional (linearized) fairness constraints on fand the common mean

Left: Shared model trained with MTL, with fairnessconstraint, and no sensitive feature in the predictorsvs. the group specific models trained with MTL,with fairness constraint

Right: The latter models vs. the same models whenthe sensitive feature is predicted via random forest

14

Page 15: Criteo AI Lab - Recent Results on Algorithmic Fairness and ......Criteo AI Lab, Paris, October 2, 2019 Plan I Fair empirical risk minimization I Using labeled and unlabeled data I

Learning fair representations[Oneto et al. Arxiv 2019]

I Now consider demographic parity: P(f (x) = 1|S = 0) = P(f (x) = 1|S = 1)

I Suppose f (x) = g(h(x)). If representation h : X → Rr is fair in the following sense

P(h(x) ∈ C |S = a) = P(h(x) ∈ C |S = b), ∀C ∈ Rr

then f is fair as well

I We relax this by requiring that both distributions have the same means.We let c(z) the difference of the empirical means from a dataset z

I We use multiple tasks to learn h. We illustrate the approach in the linear case,h(x) = A>x , and f (x) = b>h(x):

minA,B

{1

Tm

T∑t=1

n∑i=1

(yt,i−〈bt ,A>xt,i 〉

)2+λ

2‖A‖F‖B‖F

∣∣∣ A>c(zt) = 0, 1 ≤ t ≤ T

}

15

Page 16: Criteo AI Lab - Recent Results on Algorithmic Fairness and ......Criteo AI Lab, Paris, October 2, 2019 Plan I Fair empirical risk minimization I Using labeled and unlabeled data I

Learning fair representations (cont.)

Theorem. Let A solve the above problem and ‖A‖F = 1. Let tasks µ1, . . . , µT be i.i.d. from ameta-distribution ρ. Then, with probability at least 1− δ, the average risk of the algorithm withrepresentation A run on a random task is upper bounded

1

Tn

T∑t=1

n∑i=1

(yt,i−〈bt ,A>xt,i 〉

)2+ O

1

λ

√‖C‖∞

n

+ O

√ ln 1δ

T

and the linearized fairness constraint is bounded as

Eµ∼ρEz∼µn‖Ac(z)‖2 ≤ 1

T

T∑t=1

‖Ac(zt)‖2 + 96ln 8T 2

δ

T+ 6

√‖Σ‖∞ ln 8T 2

δ

T

16

Page 17: Criteo AI Lab - Recent Results on Algorithmic Fairness and ......Criteo AI Lab, Paris, October 2, 2019 Plan I Fair empirical risk minimization I Using labeled and unlabeled data I

Experiments

M1: Standard MTL with the fairness constraints on the outputsM2: Feed-forward neural network (FFNN) with adversarially generated representation [Madras et al. ICML 2018]

M3: Similar to M2 but with different loss function [Edwards &Storkey, ICLR 2016]

17

Page 18: Criteo AI Lab - Recent Results on Algorithmic Fairness and ......Criteo AI Lab, Paris, October 2, 2019 Plan I Fair empirical risk minimization I Using labeled and unlabeled data I

From MTL to meta-learning†

From a sequence of tasks find an algorithmwhich works well on unseen similar tasks

task 1 1 1 1 1 2 2 2 2 3 3 3 3 3 3 · · ·data 1 2 3 4 5 1 2 3 4 1 2 3 4 5 6 · · ·

I Previous work mainly focused on the batch statistical setting[Baxter, 2000, Maurer, 2009, Pentina and Lampert, 2014, Maurer et al., 2016]

I Recent interest on online meta-learning:

• Online-within-online: both tasks and within-task data arrive online[Alquier et al., 2017, Denevi et al., 2019, Khodak et al., 2019]

• Online-within-batch: tasks arrive online, their datasets in one batch[Denevi et al., 2018a, Denevi et al., 2018b, Finn et al., 2019, Bullins et al., 2019]

I Also recent interest on meta-learning with deep neural networks, e.g.[Ravi and Larochelle, 2017, Finn et al., 2017, Franceschi et al., 2018]

†Equivalent terminology: learning-to-learn or lifelong learning18

Page 19: Criteo AI Lab - Recent Results on Algorithmic Fairness and ......Criteo AI Lab, Paris, October 2, 2019 Plan I Fair empirical risk minimization I Using labeled and unlabeled data I

Meta-algorithm

A model for each task is learned byan inner algorithm, which isupdated by a meta-algorithm asthe tasks are sequentially observed

I Desiderata: memory and time efficient, and supported by learning guarantees

I Difficulty: lack of a convex meta-objective

19

Page 20: Criteo AI Lab - Recent Results on Algorithmic Fairness and ......Criteo AI Lab, Paris, October 2, 2019 Plan I Fair empirical risk minimization I Using labeled and unlabeled data I

Statistical and non-statistical settings

Let Zt = (xt,i , yt,i )ni=1 be the training sequence for the t-th task and let Z = (Zt)

Tt=1 be

the meta-sequence. We consider two settings‡

I Statistical setting [Baxter, 2000, Maurer, 2009]: the tasks are sampled from ameta-distribution ρ and we wish to bound the average excess risk

EZEµ∼ρEµ(A(Z )) = EZ

[Eµ∼ρ

[EZ∼µn Rµ

(A(Z )

)− min

w∈RdRµ(w)

]]I Non-statistical setting: we wish to bound the normalized regret across the tasks

regret(A1, ...,AT )=1

T

T∑t=1

{1

n

n∑i=1

`(〈xt,i ,wt,i 〉, yt,i

)− min

w∈Rd

1

n

n∑i=1

`(〈xt,i ,w〉, yt,i

)}

‡See [Alquier et al., 2017] for a discussion20

Page 21: Criteo AI Lab - Recent Results on Algorithmic Fairness and ......Criteo AI Lab, Paris, October 2, 2019 Plan I Fair empirical risk minimization I Using labeled and unlabeled data I

Regularizaton around a common mean – learning guarantees[Denevi et al. ICML 2019; Denevi et al. NeurIPS 2019]

We assume `(·, y) L-Lipschitz for any y ∈ Y and the inputs are bounded. Let wµ be theminimizer of the true risk for task µ

Vρ(θ) =1

2Eµ∼ρ‖wµ − θ‖2

2 θρ = argminθ∈Θ

Vρ(θ) = Eµ∼ρwµ

I Our method (from Thm. 2, tuning of λ and η)

EZ Eµ∼ρ Eµ(Aθ)≤ O

(√Vρ(θρ)

n+

√1

T

)

I Best algorithm θ = θρ

Eµ∼ρ Eµ(Aθ)≤ O

(√Vρ(θρ)

n

) I Indep. task learning (ITL) θ = 0

Eµ∼ρ Eµ(Aθ)≤ O

(√Vρ(0)

n

)21

Page 22: Criteo AI Lab - Recent Results on Algorithmic Fairness and ......Criteo AI Lab, Paris, October 2, 2019 Plan I Fair empirical risk minimization I Using labeled and unlabeled data I

Experiment

Synthetic Data Lenk Dataset

Averaged test performance of different methods on synthetic (Left) and the Lenk dataset(Right) as the number of training tasks incrementally increases.

22

Page 23: Criteo AI Lab - Recent Results on Algorithmic Fairness and ......Criteo AI Lab, Paris, October 2, 2019 Plan I Fair empirical risk minimization I Using labeled and unlabeled data I

We are hiring!

Postdoc/Researcher positions at Istituto Italiano diTecnologia in Genoa to work with me

Send me an email if interested: [email protected] info: http://tinyurl.com/MLPostDocIIT2019

23

Page 24: Criteo AI Lab - Recent Results on Algorithmic Fairness and ......Criteo AI Lab, Paris, October 2, 2019 Plan I Fair empirical risk minimization I Using labeled and unlabeled data I

References I

Alquier, P., Mai, T. T., and Pontil, M. (2017).

Regret bounds for lifelong learning.In Proceedings of the 20th International Conference on Artificial Intelligence and Statistics, volume 54 of Proceedings of Machine LearningResearch, pages 261–269.

Baxter, J. (2000).

A model of inductive bias learning.J. Artif. Intell. Res., 12(149–198):3.

Bullins, B., Hazan, E., Kalai, A., and Livni, R. (2019).

Generalize across tasks: Efficient algorithms for linear representation learning.In Algorithmic Learning Theory, pages 235–246.

Denevi, G., Ciliberto, C., Grazzi, R., and Pontil, M. (2019).

Learning-to-learn stochastic gradient descent with biased regularization.arXiv preprint arXiv:1903.10399.

Denevi, G., Ciliberto, C., Stamos, D., and Pontil, M. (2018a).

Incremental learning-to-learn with statistical guarantees.In Proc. 34th Conference on Uncertainty in Artificial Intelligence (UAI).

Denevi, G., Ciliberto, C., Stamos, D., and Pontil, M. (2018b).

Learning to learn around a common mean.In Advances in Neural Information Processing Systems, pages 10190–10200.

24

Page 25: Criteo AI Lab - Recent Results on Algorithmic Fairness and ......Criteo AI Lab, Paris, October 2, 2019 Plan I Fair empirical risk minimization I Using labeled and unlabeled data I

References II

Finn, C., Abbeel, P., and Levine, S. (2017).

Model-agnostic meta-learning for fast adaptation of deep networks.In Proceedings of the 34th International Conference on Machine Learning, volume 70 of Proceedings of Machine Learning Research, pages1126–1135. PMLR.

Finn, C., Rajeswaran, A., Kakade, S., and Levine, S. (2019).

Online meta-learning.arXiv preprint arXiv:1902.08438.

Franceschi, L., Frasconi, P., Salzo, S., Grazzi, R., and Pontil, M. (2018).

Bilevel programming for hyperparameter optimization and meta-learning.In International Conference on Machine Learning, PMLR 80, pages 568–1577.

Hardt, M., Price, E., and Srebro, N. (2016).

Equality of opportunity in supervised learning.In Advances in Neural Information Processing Systems.

Khodak, M., Balcan, M.-F., and Talwalkar, A. (2019).

Provable guarantees for gradient-based meta-learning.arXiv preprint arXiv:1902.10644.

Maurer, A. (2009).

Transfer bounds for linear feature learning.Machine Learning, 75(3):327–350.

25

Page 26: Criteo AI Lab - Recent Results on Algorithmic Fairness and ......Criteo AI Lab, Paris, October 2, 2019 Plan I Fair empirical risk minimization I Using labeled and unlabeled data I

References III

Maurer, A., Pontil, M., and Romera-Paredes, B. (2016).

The benefit of multitask representation learning.The Journal of Machine Learning Research, 17(1):2853–2884.

Pentina, A. and Lampert, C. (2014).

A PAC-Bayesian bound for lifelong learning.In International Conference on Machine Learning, pages 991–999.

Ravi, S. and Larochelle, H. (2017).

Optimization as a model for few-shot learning.In I5th International Conference on Learning Representations.

Zafar, M. B., Valera, I., Gomez Rodriguez, M., and Gummadi, K. P. (2017).

Fairness beyond disparate treatment & disparate impact: Learning classification without disparate mistreatment.In International Conference on World Wide Web.

26


Recommended