confidence regularized self-training · 2020. 9. 28. · conclusions § compared with supervised...

13
Confidence Regularized Self-Training Yang Zou Zhiding Yu Xiaofeng Liu Vijayakumar Bhagavatula Jinsong Wang

Upload: others

Post on 04-Oct-2020

0 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Confidence Regularized Self-Training · 2020. 9. 28. · Conclusions § Compared with supervised learning,self-training is an under-determined problem (EM with latent variables)

Confidence Regularized Self-Training

Yang Zou Zhiding Yu Xiaofeng Liu Vijayakumar

Bhagavatula

Jinsong Wang

Page 2: Confidence Regularized Self-Training · 2020. 9. 28. · Conclusions § Compared with supervised learning,self-training is an under-determined problem (EM with latent variables)

Source Domain (Labeled) Target Domain (Unlabeled)

Image

classification

Unsupervised Domain Adaptation (UDA)

Semantic

segmentation

Car

Adaptation

Adaptation

Page 3: Confidence Regularized Self-Training · 2020. 9. 28. · Conclusions § Compared with supervised learning,self-training is an under-determined problem (EM with latent variables)

Target Images (Cityscapes)

Deep

CNN

Source Labels (GTA5)

Source Images (GTA5)

Predictions (Cityscapes)

Pseudo-labels (Cityscapes)

GTA5 → Cityscapes

Yang Zou, Zhiding Yu et al., Unsupervised Domain Adaptation for Semantic Segmentation via Class-Balanced Self-Training, ECCV18

UDA through Iterative Deep Self-Training

Page 4: Confidence Regularized Self-Training · 2020. 9. 28. · Conclusions § Compared with supervised learning,self-training is an under-determined problem (EM with latent variables)

where: x: input sample p: class predication vector y: label vector !": pseudo-label vector

w: network parameters s: source sample index t: target sample index △$%&: probability simplex

Pseudo-label

generation

Network

retraining

Car

Person

Bus

Balanced softmax Pseudo-label !"∗ Network output

after self-training

Class-Balanced Self-Training (CBST)

Yang Zou, Zhiding Yu et al., Unsupervised Domain Adaptation for Semantic Segmentation via Class-Balanced Self-Training, ECCV18

Page 5: Confidence Regularized Self-Training · 2020. 9. 28. · Conclusions § Compared with supervised learning,self-training is an under-determined problem (EM with latent variables)

Samples from VisDA-17 (With label “Car”)

Pseudo-label Sharp output after

self-training

Pseudo-label

generation

Image label: carNetwork output before

self-trainingBackbone

Network

Network

retraining

Issues in Self-Training: Overconfident Mistakes

Car

Person

Bus

Building

Person

Building

SkyBuilding

Green: Correctly classifiedRed: Misclassified

Sample from BDD100K

Page 6: Confidence Regularized Self-Training · 2020. 9. 28. · Conclusions § Compared with supervised learning,self-training is an under-determined problem (EM with latent variables)

Car

Person

BusPseudo-label

generation

Network

retraining

where: !: regularizer weight

Label Regularized Self-Training (LR)

Balanced softmax Pseudo-label "#∗ Network output

after self-training

Page 7: Confidence Regularized Self-Training · 2020. 9. 28. · Conclusions § Compared with supervised learning,self-training is an under-determined problem (EM with latent variables)

Car

Person

BusPseudo-label

generation

Network

retraining

where: !: regularizer weight

Model Regularized Self-Training (MR)

Balanced softmax Pseudo-label "#∗ Network output

after self-training

Page 8: Confidence Regularized Self-Training · 2020. 9. 28. · Conclusions § Compared with supervised learning,self-training is an under-determined problem (EM with latent variables)

LR-Entropy (LRENT) MR-KLDiv (MRKLD)

MR-Entropy (MRENT)

MR-L2 (MRL2)

Pseudo-label solver

Regularized retraining loss

v.s. probability

Pseudo-label generation loss

v.s. probability

Proposed Confidence Regularizers

0 0.25 0.5 0.75 1y

0

1

2

3

regula

rize

d lo

ss

LRNET

=0=0.5=1

y*

0 0.25 0.5 0.75 1p

-10

2

4

6

8

10

regula

rize

d lo

ss

MRENT

=0=0.5=1

p*

0 0.25 0.5 0.75 1p

-10

2

4

6

8

10

regula

rize

d lo

ss

MRKLD

=0=0.5=1

p*

0 0.25 0.5 0.75 1p

-10

2

4

6

8

10

regula

rize

d lo

ss

MRL2

=0=0.5=1

p*

Page 9: Confidence Regularized Self-Training · 2020. 9. 28. · Conclusions § Compared with supervised learning,self-training is an under-determined problem (EM with latent variables)

Probabilistic Explanation

Proposition 1. CRST can be modeled as a regularized maximum likelihood for classification (RCML) problem

optimized via classification expectation maximization.

Convergence Analysis

Proposition 2. Given pre-determined !"′s, CRST is convergent with gradient descent for network retraining

optimization.

Theoretical Analysis

Page 10: Confidence Regularized Self-Training · 2020. 9. 28. · Conclusions § Compared with supervised learning,self-training is an under-determined problem (EM with latent variables)

Results on VisDA-17

Results on Office-31

Experiment: UDA for Image Classification

Page 11: Confidence Regularized Self-Training · 2020. 9. 28. · Conclusions § Compared with supervised learning,self-training is an under-determined problem (EM with latent variables)

Results on SYNTHIA -> Cityscapes (mIoU* - 13 class)

Results on GTA5 -> Cityscapes

Experiment: UDA for Semantic Segmentation

Page 12: Confidence Regularized Self-Training · 2020. 9. 28. · Conclusions § Compared with supervised learning,self-training is an under-determined problem (EM with latent variables)

Original Image Ground Truth Source Model CBST CRST (MRKLD)

Experiment: Qualitative Results (GTA5 -> Cityscapes)

Page 13: Confidence Regularized Self-Training · 2020. 9. 28. · Conclusions § Compared with supervised learning,self-training is an under-determined problem (EM with latent variables)

Conclusions

§ Compared with supervised learning, self-training is an under-determined problem (EM with latent variables).

§ Our work shows the importance of confidence regularizations as inductive biases to help under-constrained

problems such as unsupervised domain adaptation and semi-supervised learning.

§ CRST is still aligned with entropy minimization. The proposed confidence regularization only serves as a safety

measure to prevent over self-training/entropy minimization.

§ MR-KLD is most recommended in practice for its efficiency and good performance.

Future Works

§ This work could potentially inspire many other meaningful regularizations/inductive biases for similar problems.

Conclusions and Future Works