17
Active Learning with Disagreement Graphs Corinna Cortes 1 , Giulia DeSalvo 1 , Claudio Gentile 1 , Mehryar Mohri 1,2 , Ningshan Zhang 3 1 Google Research 2 Courant Institute, NYU 3 NYU ICML, June 12, 2019

Active Learning with Disagreement Graphspeople.stern.nyu.edu/nzhang/disgraph_slides.pdfExperiments 0.125 0.150 0.175 7 8 9 10 11 log 2(Number of Labels) Misclassification Loss IWAL

  • Upload
    others

  • View
    1

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Active Learning with Disagreement Graphspeople.stern.nyu.edu/nzhang/disgraph_slides.pdfExperiments 0.125 0.150 0.175 7 8 9 10 11 log 2(Number of Labels) Misclassification Loss IWAL

Active Learning with Disagreement Graphs

Corinna Cortes1, Giulia DeSalvo1, Claudio Gentile1,Mehryar Mohri1,2, Ningshan Zhang3

1 Google Research 2 Courant Institute, NYU 3 NYU

ICML, June 12, 2019

Page 2: Active Learning with Disagreement Graphspeople.stern.nyu.edu/nzhang/disgraph_slides.pdfExperiments 0.125 0.150 0.175 7 8 9 10 11 log 2(Number of Labels) Misclassification Loss IWAL

On-line Active Learning Setup

I At each round t ∈ [T ], receives unlabeled xt ∼ DX i.i.d.I Decides whether to request label:

I If label requested, receives yt .I After T rounds, returns a hypothesis hT ∈ H.

Objective:I Generalizations error:

I Accurate predictor hT : small expected loss R(hT ) = Ex,y[`(hT (x), y)

].

I Close to best-in-class h∗ = argminh∈H R(h).I Label complexity: few label requests.

Page 3: Active Learning with Disagreement Graphspeople.stern.nyu.edu/nzhang/disgraph_slides.pdfExperiments 0.125 0.150 0.175 7 8 9 10 11 log 2(Number of Labels) Misclassification Loss IWAL

Disagreement-based Active Learning

Key idea: Request label when there is some disagreement among hypotheses.Examples:

I Separable case: CAL (Cohn et al., 1994).I Non-separable case: A2 (Balcan et al., 2006), DHM (Dasgupta et al., 2008).I IWAL (Beygelzimer et al., 2009).

Can we improve upon existing disagreement-based algorithms, such as IWAL?I Better guarantees?I Leverage average disagreements?

Page 4: Active Learning with Disagreement Graphspeople.stern.nyu.edu/nzhang/disgraph_slides.pdfExperiments 0.125 0.150 0.175 7 8 9 10 11 log 2(Number of Labels) Misclassification Loss IWAL

This talk

I IWAL-D algorithm: enhanced IWAL with disagreement graph.I IZOOM algorithm: enhanced IWAL-D with zooming-in.I Better generalization and label complexity guarantees.I Experimental results.

Page 5: Active Learning with Disagreement Graphspeople.stern.nyu.edu/nzhang/disgraph_slides.pdfExperiments 0.125 0.150 0.175 7 8 9 10 11 log 2(Number of Labels) Misclassification Loss IWAL

Disagreement Graph (D-Graph)

I Vertices: hypotheses in H (a finite hypothesis set)I Edges: fully connected. The edge between h,h′ ∈ H is weighted by their

expected disagreement:

L(h,h′) = Ex∼DX

[maxy∈Y

∣∣`(h(x), y)− `(h′(x), y)∣∣].

L symmetric, ` ≤ 1⇒ L ≤ 1.I D-Graph can be accurately estimated using unlabeled data.

Page 6: Active Learning with Disagreement Graphspeople.stern.nyu.edu/nzhang/disgraph_slides.pdfExperiments 0.125 0.150 0.175 7 8 9 10 11 log 2(Number of Labels) Misclassification Loss IWAL

Disagreement Graph (D-Graph)

One favorable scenario:I Best-in-class h∗ ( ) is within an isolated cluster;I L(h,h∗) is small within the cluster.

Page 7: Active Learning with Disagreement Graphspeople.stern.nyu.edu/nzhang/disgraph_slides.pdfExperiments 0.125 0.150 0.175 7 8 9 10 11 log 2(Number of Labels) Misclassification Loss IWAL

IWAL-D Algorithm: IWAL with D-Graph

I At round t ∈ [T ], receive xt .1. Flip a coin Qt ∼ Ber(pt ), with disagreement-based bias:

pt = maxh,h′∈Ht

maxy∈Y

∣∣`(h(xt ), y)− `(h′(xt ), y)∣∣.

2. If Qt = 1, request the label yt .3. Trim the version space:

Ht+1 ={

h ∈ Ht : Lt (h) ≤ Lt (ht ) +(1 + L(h, ht )

)∆t

},

which uses importance weighted empirical risk

Lt (h) =1t

t∑s=1

Qs

ps`(h(xs), ys), ht = argmin

h∈Ht

Lt (h), ∆t = O(√

log(T |H|)t

).

I After T rounds, return hT .

Page 8: Active Learning with Disagreement Graphspeople.stern.nyu.edu/nzhang/disgraph_slides.pdfExperiments 0.125 0.150 0.175 7 8 9 10 11 log 2(Number of Labels) Misclassification Loss IWAL

IWAL-D vs. IWAL: Quantify the Improvement

Theorem (IWAL-D) With high probability,

R(hT ) ≤ R∗ +(1 + L(hT ,h∗)

)∆T ,

Ex∼DX

[pt |Ft−1

]≤ 2θ

[2R∗ + max

h∈Ht

(2 + L(h, ht−1) + L(h,h∗)

)∆t−1

].

I θ: disagreement coefficient (Hanneke, 2007).I More aggressive trimming of the version space.I Slightly better generalization guarantee and label complexity.

Page 9: Active Learning with Disagreement Graphspeople.stern.nyu.edu/nzhang/disgraph_slides.pdfExperiments 0.125 0.150 0.175 7 8 9 10 11 log 2(Number of Labels) Misclassification Loss IWAL

IWAL and IWAL-D

Problem:I Theoretical guarantees only hold for finite hypothesis sets.I Need ε-cover to extend to infinite hypothesis sets.I Expensive to construct ε-cover in practice.

Can we adaptively enrich the hypothesis set, with theoretical guarantees?

Page 10: Active Learning with Disagreement Graphspeople.stern.nyu.edu/nzhang/disgraph_slides.pdfExperiments 0.125 0.150 0.175 7 8 9 10 11 log 2(Number of Labels) Misclassification Loss IWAL

IZOOM: IWAL-D with Zooming-in

At round t ,I Request label based on dis. of (Ht )

I H′t+1 ← Trim(Ht )

I H′′t+1 ← Resample(H′t+1)

I Ht+1 ← H′t+1 ∪H′′t+1

Trim Resample

Page 11: Active Learning with Disagreement Graphspeople.stern.nyu.edu/nzhang/disgraph_slides.pdfExperiments 0.125 0.150 0.175 7 8 9 10 11 log 2(Number of Labels) Misclassification Loss IWAL

IZOOM: IWAL-D with Zooming-in

At round t ,I Request label based on dis. of (Ht )

I H′t+1 ← Trim(Ht )

I H′′t+1 ← Resample(H′t+1)

I Ht+1 ← H′t+1 ∪H′′t+1

Trim Resample

Page 12: Active Learning with Disagreement Graphspeople.stern.nyu.edu/nzhang/disgraph_slides.pdfExperiments 0.125 0.150 0.175 7 8 9 10 11 log 2(Number of Labels) Misclassification Loss IWAL

IZOOM: IWAL-D with Zooming-in

At round t ,I Request label based on dis. of (Ht )

I H′t+1 ← Trim(Ht )

I H′′t+1 ← Resample(H′t+1)

I Ht+1 ← H′t+1 ∪H′′t+1

Trim ResampleResample(H′t+1): sample new h ∈ ConvexHull(H′t+1).

I E.g., random convex combination of ht and h ∈ H′t+1.

Page 13: Active Learning with Disagreement Graphspeople.stern.nyu.edu/nzhang/disgraph_slides.pdfExperiments 0.125 0.150 0.175 7 8 9 10 11 log 2(Number of Labels) Misclassification Loss IWAL

IZOOM: IWAL-D with Zooming-in

At round t ,I Request label based on dis. of (Ht )

I H′t+1 ← Trim(Ht )

I H′′t+1 ← Resample(H′t+1)

I Ht+1 ← H′t+1 ∪H′′t+1

Trim Resample

Page 14: Active Learning with Disagreement Graphspeople.stern.nyu.edu/nzhang/disgraph_slides.pdfExperiments 0.125 0.150 0.175 7 8 9 10 11 log 2(Number of Labels) Misclassification Loss IWAL

IZOOM vs. IWAL-D

Let Ht = ∪ts=1Ht , i.e. all the hypotheses ever considered up to time t . Let

h∗t = argminh∈HtR(h).

Theorem (IZOOM) With high probability,

R(hT ) ≤ R∗T +(1 + L(hT ,h∗T )

)∆T + O( 1

T ),

Ex∼DX

[pt+1|Ft

]≤ 2θt

[2R∗t + max

h∈Ht+1

(2 + L(h, ht ) + L(h,h∗t )

)∆t]

+ O( 1T ).

I R∗t = minh∈Ht R(h) is smaller than R∗ = minh∈H0 R(h).

I More accurate hT , with fewer label requests.

Page 15: Active Learning with Disagreement Graphspeople.stern.nyu.edu/nzhang/disgraph_slides.pdfExperiments 0.125 0.150 0.175 7 8 9 10 11 log 2(Number of Labels) Misclassification Loss IWAL

Experiments

Tasks: 8 Binary classification datasets from UCI repository.I `: logistic loss rescaled to [0,1].

Baselines:I IWAL with 3,000 hypotheses.I IWAL with 12,000 hypotheses.I IZOOM with 3,000 hypotheses.

Performance measure:I 0-1 loss on test data vs. number of label requests.

Page 16: Active Learning with Disagreement Graphspeople.stern.nyu.edu/nzhang/disgraph_slides.pdfExperiments 0.125 0.150 0.175 7 8 9 10 11 log 2(Number of Labels) Misclassification Loss IWAL

Experiments

0.125

0.150

0.175

7 8 9 10 11log2(Number of Labels)

Mis

clas

sific

atio

n Lo

ss

IWAL 3000 IWAL 12000 IZOOM 3000

nomao

0.14

0.16

0.18

0.20

7 8 9 10 11log2(Number of Labels)

Mis

clas

sific

atio

n Lo

ss

IWAL 3000 IWAL 12000 IZOOM 3000

codrna

0.11

0.12

0.13

0.14

7 8 9 10 11log2(Number of Labels)

Mis

clas

sific

atio

n Lo

ss

IWAL 3000 IWAL 12000 IZOOM 3000

skin

0.36

0.38

0.40

0.42

7 8 9 10 11log2(Number of Labels)

Mis

clas

sific

atio

n Lo

ss

IWAL 3000 IWAL 12000 IZOOM 3000

covtype

0.22

0.23

0.24

0.25

7 8 9 10 11log2(Number of Labels)

Mis

clas

sific

atio

n Lo

ss

IWAL 3000 IWAL 12000 IZOOM 3000

magic04

0.18

0.20

0.22

0.24

0.26

7 8 9 10 11log2(Number of Labels)

Mis

clas

sific

atio

n Lo

ss

IWAL 3000 IWAL 12000 IZOOM 3000

a9a

Page 17: Active Learning with Disagreement Graphspeople.stern.nyu.edu/nzhang/disgraph_slides.pdfExperiments 0.125 0.150 0.175 7 8 9 10 11 log 2(Number of Labels) Misclassification Loss IWAL

Conclusion

I Key introduction and role of disagreement graph.I More favorable generalization and label complexity guarantees.I Substantial performance improvements.I Effective solutions for active learning.

Poster: Pacific Ballroom #265

KDD workshop (Alaska, August 2019) on Active Learning:Data Collection, Curation, and Labeling for Mining and Learning.