representation learning by learning to count - ucf crcv · 2019-03-26 · learning to count...

Post on 24-Jun-2020

0 Views

Category:

Documents

0 Downloads

Preview:

Click to see full reader

TRANSCRIPT

Representation Learning by Learning to CountPresented by: Muhammad Tayyab

1

Background

★ Supervised representation learning

Feature Extraction Classifier Dog

Object classificationObject detectionSemantic segmentation

2https://www.youtube.com/watch?v=ZmaXDb9akEI

Background

★ Supervised representation learning

Feature Extraction Classifier Dog

3

Require human annotation● Costly● Error prone● Time consuming

Background

★ Self-supervised representation learning

Feature Extraction Pseudo task Solve task

Doesn’t require human annotation● Cheap● Scalable

4

★ Good representations

Background

5

Representation hyperspace

Background

★ Representation learning

Dog

6http://cv-tricks.com/cnn/understand-resnet-alexnet-vgg-inception/

Background

★ Representation learning

Dog

7

Background

★ Representation learning

Cat

8

★ Idea: Sum of predicted patchwise feature count should be similar to the image feature count

Similarity based on counting

9

★ Formulation:○ Downsampling operator D.○ Tiling operator Tj.

Learning to Count

★ Loss

10

Learning to Count

★ Least effort bias○ Easy to satisfy loss if network always outputs zero.

11

★ Solution?○ Contrastive loss○ Learn features useful for discrimination.

Learning to Count

Contrastive loss:

12

★ Observation

Experiments

Average response of trained network on the ImageNet validation set13

Experiments

★ Transfer Learning

14

Cat

Classification

Cat

Detection Segmentation

http://cs231n.stanford.edu/slides/2017/cs231n_2017_lecture11.pdf

★ Transfer Learning

Experiments

Evaluation of transfer learning on PASCAL15

★ Transfer Learning

Experiments

16

ImageNet classification with a linear classifier

Experiments

★ Places data set○ By MIT○ 10,624,928 Images○ 434 Classes

17

○ 434 Classes

Experiments

18

Experiments

19

Places classification with a linear classifier

★ Transfer Learning

★ Ablation studies

Experiments

20

★ Ablation studies

Experiments

21

As an error metric, we use the first term in the loss function normalized by the average of the norm of the feature vector. More precisely, the error when the network is trained with the i-th downsampling style and tested on the j-th one is

★ Ablation studies

Experiments

22

Learning the downsampling style

★ Counting

Experiments

23

Image croppings of increasing size. The number of visual primitives should increase going from left to right

Counting evaluation on ImageNet

Experiments

24Examples of activating/ignored images for ImageNet test set

Least Activating Im

agesMos

t Act

ivat

ing

Imag

es

Experiments

25Examples of activating/ignored images for COCO test set

Least Activating Im

agesMos

t Act

ivat

ing

Imag

es

Experiments

26Nearest neighbor retrievals for ImageNet

Query Images Matches

Experiments

27Nearest neighbor retrievals for COCO

Query Images Matches

Experiments

28Blocks of the 8 most activating images for 4 neurons for ImageNet

Neu

ron

1N

euro

n 3

Neuron 4

Neuron 2

Experiments

29Blocks of the 8 most activating images for 4 neurons for COCO

Neu

ron

1N

euro

n 3

Neuron 4

Neuron 2

Thank you!

30

top related