fcv bio cv_cottrell

45
1 Unsupervised learning of visual Unsupervised learning of visual representations and their use in representations and their use in object & face object & face recognition recognition Gary Cottrell Gary Cottrell Chris Kanan Honghao Shan Chris Kanan Honghao Shan Lingyun Zhang Matthew Tong Lingyun Zhang Matthew Tong Tim Marks Tim Marks

Upload: zukun

Post on 14-Jul-2015

173 views

Category:

Technology


0 download

TRANSCRIPT

Page 1: Fcv bio cv_cottrell

1

Unsupervised learning of visual Unsupervised learning of visual representations and their use in representations and their use in

object & faceobject & facerecognitionrecognitionGary CottrellGary Cottrell

Chris Kanan Honghao ShanChris Kanan Honghao ShanLingyun Zhang Matthew TongLingyun Zhang Matthew Tong

Tim MarksTim Marks

Page 2: Fcv bio cv_cottrell

2

CollaboratorsCollaborators

Honghao Shan Chris Kanan

Page 3: Fcv bio cv_cottrell

3

CollaboratorsCollaborators

Tim Marks Matt TongLingyun Zhang

QuickTime and aªTIFF (LZW) decompressor

are needed to see this picture.

Page 4: Fcv bio cv_cottrell

4

Efficient Encoding of the worldEfficient Encoding of the world

Sparse Principal Components Analysis:Sparse Principal Components Analysis:A model of unsupervised learning for early A model of unsupervised learning for early perceptual processing (Honghao Shan)perceptual processing (Honghao Shan)

The model embodies three constraintsThe model embodies three constraints

• Keep as much information as possibleKeep as much information as possible

• While trying to equalize the neural responsesWhile trying to equalize the neural responses

• And minimizing the connections.And minimizing the connections.

Page 5: Fcv bio cv_cottrell

5

Trained on grayscale imagesTrained on color images

Spatial extent

Tem

pora

l ext

ent

Efficient Encoding of the world leads to magno- Efficient Encoding of the world leads to magno- and parvo-cellular response properties…and parvo-cellular response properties…

Trained on video cubes

This suggests that these cell types exist becausebecause they are useful for efficiently encoding the temporal dynamics of the world.

Midget?

Parasol?

Persistent, small

Transient, large

Page 6: Fcv bio cv_cottrell

6

Efficient Encoding of the world leads Efficient Encoding of the world leads to gammatone filters as in auditory to gammatone filters as in auditory

nerves:nerves: Using Using exactly the same algorithmexactly the same algorithm, applied to , applied to

speech, environmental sounds, etc.:speech, environmental sounds, etc.:

Page 7: Fcv bio cv_cottrell

7

Efficient Encoding of the worldEfficient Encoding of the world

A single unsupervised learning algorithm A single unsupervised learning algorithm leads toleads to Model cells with properties similar to those Model cells with properties similar to those

found in the retina when applied to natural found in the retina when applied to natural videosvideos

Models cells with properties similar to those Models cells with properties similar to those found in auditory nerve when applied to found in auditory nerve when applied to natural soundsnatural sounds

One small step towards a unified theory of One small step towards a unified theory of temporal processing.temporal processing.

Page 8: Fcv bio cv_cottrell

8

Unsupervised Learning of Unsupervised Learning of Hierarchical Representations Hierarchical Representations

(RICA 2.0; cf. Shan et al., NIPS 19)(RICA 2.0; cf. Shan et al., NIPS 19)

Recursive ICA (RICA 1.0 Recursive ICA (RICA 1.0 (Shan et al., 2008)(Shan et al., 2008)):): Alternately compress and expand Alternately compress and expand

representation using PCA and ICA;representation using PCA and ICA; ICA was modified by a component-wise ICA was modified by a component-wise

nonlinearitynonlinearity Receptive fields expanded at each ICA layerReceptive fields expanded at each ICA layer

Page 9: Fcv bio cv_cottrell

9

Unsupervised Learning of Unsupervised Learning of Hierarchical Representations Hierarchical Representations

(RICA 2.0; cf. Shan et al., NIPS 19)(RICA 2.0; cf. Shan et al., NIPS 19)

ICA was modified by a component-wise ICA was modified by a component-wise nonlinearity:nonlinearity: Think of ICA as a generative model: The Think of ICA as a generative model: The

pixels are the sum of many independent pixels are the sum of many independent random variables: Gaussian.random variables: Gaussian.

Hence ICA prefers its inputs to be Gaussian-Hence ICA prefers its inputs to be Gaussian-distributed.distributed.

We apply an inverse cumulative Gaussian to We apply an inverse cumulative Gaussian to the absolute value of the ICA components to the absolute value of the ICA components to “gaussianize” them.“gaussianize” them.

Page 10: Fcv bio cv_cottrell

10

Unsupervised Learning of Unsupervised Learning of Hierarchical Representations Hierarchical Representations

(RICA 2.0; cf. Shan et al., NIPS 19)(RICA 2.0; cf. Shan et al., NIPS 19)

Strong responses, either positive or Strong responses, either positive or negative, are mapped to the positive tail of negative, are mapped to the positive tail of the Gaussian; weak ones, to the negative the Gaussian; weak ones, to the negative tail; ambiguous ones to the center.tail; ambiguous ones to the center.

Page 11: Fcv bio cv_cottrell

11

Unsupervised Learning of Unsupervised Learning of Hierarchical Representations Hierarchical Representations

(RICA 2.0; cf. Shan et al., NIPS 19)(RICA 2.0; cf. Shan et al., NIPS 19)

RICA 2.0:RICA 2.0: Replace PCA by SPCAReplace PCA by SPCA

SPCA SPCA

Page 12: Fcv bio cv_cottrell

12

Unsupervised Learning of Unsupervised Learning of Hierarchical Representations Hierarchical Representations

(RICA 2.0; cf. Shan et al., NIPS 19)(RICA 2.0; cf. Shan et al., NIPS 19)

RICA 2.0 Results: RICA 2.0 Results: Multiple layer system withMultiple layer system with Center-surround receptive fields at the first layerCenter-surround receptive fields at the first layer Simple edge filters at the second (ICA) layerSimple edge filters at the second (ICA) layer Spatial pooling of orientations at the third (SPCA) Spatial pooling of orientations at the third (SPCA)

layer:layer:

V2-like response properties at the fourth (ICA) layerV2-like response properties at the fourth (ICA) layer

QuickTime and aª decompressor

are needed to see this picture.

Page 13: Fcv bio cv_cottrell

13

Unsupervised Learning of Unsupervised Learning of Hierarchical Representations Hierarchical Representations

(RICA 2.0; cf. Shan et al., NIPS 19)(RICA 2.0; cf. Shan et al., NIPS 19)

V2-like response properties at the V2-like response properties at the fourth (ICA) layerfourth (ICA) layer

These maps show strengths of These maps show strengths of connections to layer 1 ICA filters. connections to layer 1 ICA filters. Warm and cold colors are strong Warm and cold colors are strong +/- connections, gray is weak +/- connections, gray is weak connections, orientation connections, orientation corresponds to layer 1 orientation.corresponds to layer 1 orientation.

The left-most column displays two The left-most column displays two model neurons that show uniform model neurons that show uniform orientation preference to layer-1 orientation preference to layer-1 ICA features. ICA features.

The middle column displays model The middle column displays model neurons that have non-neurons that have non-uniform/varying orientation uniform/varying orientation preference to layer-1 ICA features. preference to layer-1 ICA features.

The right column displays two The right column displays two model neurons that have location model neurons that have location preference, but no orientation preference, but no orientation preference, to layer-1 ICA features.preference, to layer-1 ICA features.

The left two columns are consistent with Anzen, Peng, & Van Essen 2007. The

right hand column is a prediction

Page 14: Fcv bio cv_cottrell

14

Unsupervised Learning of Unsupervised Learning of Hierarchical Representations Hierarchical Representations

(RICA 2.0; cf. Shan et al., NIPS 19)(RICA 2.0; cf. Shan et al., NIPS 19)

Dimensionality Reduction & Expansion might be a general strategy of Dimensionality Reduction & Expansion might be a general strategy of information processing in the brain.information processing in the brain. The first step removes noise and reduces complexity, the second step The first step removes noise and reduces complexity, the second step

captures the statistical structure.captures the statistical structure. We showed that retinal ganglion cells and V1 complex cells may be We showed that retinal ganglion cells and V1 complex cells may be

derived from the same learning algorithm, applied to pixels in one derived from the same learning algorithm, applied to pixels in one case, and V1 simple cell outputs in the second.case, and V1 simple cell outputs in the second.

This highly simplified model of early vision is the first one that learns This highly simplified model of early vision is the first one that learns the RFs of all early visual layers, using a consistent theory the RFs of all early visual layers, using a consistent theory -- the the efficient coding theory. efficient coding theory.

We believe it could serve as a basis for more sophisticated models of We believe it could serve as a basis for more sophisticated models of early vision.early vision.

An obvious next step is to train and thus make predictions about An obvious next step is to train and thus make predictions about higher layers.higher layers.

Page 15: Fcv bio cv_cottrell

15

We showed in Shan & Cottrell (CVPR 2008) We showed in Shan & Cottrell (CVPR 2008) that we could achieve state-of-the-art face that we could achieve state-of-the-art face recognition with the non-linear ICA features recognition with the non-linear ICA features and a simple softmax output.and a simple softmax output.

We showed in Kanan & Cottrell (CVPR 2010) We showed in Kanan & Cottrell (CVPR 2010) that we could achieve state-of-the-art face and that we could achieve state-of-the-art face and object recognition with a system that used an object recognition with a system that used an ICA-based salience map, simulated fixations, ICA-based salience map, simulated fixations, non-linear ICA features, and a kernel-density non-linear ICA features, and a kernel-density memory.memory.

Here I briefly describe the latter.Here I briefly describe the latter.

Nice, but is it useful?Nice, but is it useful?

Page 16: Fcv bio cv_cottrell

16

Our attention is Our attention is automatically drawn to automatically drawn to interesting regions in interesting regions in images.images.

Our salience algorithm is Our salience algorithm is automatically drawn to automatically drawn to interesting regions in interesting regions in images.images.

These are useful locations These are useful locations for for discriminatingdiscriminating one one object (face, butterfly) from object (face, butterfly) from another.another.

One reason why this might be One reason why this might be a good idea…a good idea…

Page 17: Fcv bio cv_cottrell

17

Training Phase (learning object Training Phase (learning object appearances):appearances):

Use the salience map to decide Use the salience map to decide where where to look. to look. (We use the ICA salience map)(We use the ICA salience map)

Memorize these samplesMemorize these samples of the of the image, with labels (Bob, Carol, Ted, image, with labels (Bob, Carol, Ted, or Alice) or Alice) (We store the (compressed) ICA feature (We store the (compressed) ICA feature values)values)

Main IdeaMain Idea

Page 18: Fcv bio cv_cottrell

18

Testing Phase (recognizing Testing Phase (recognizing objects we have learned):objects we have learned):

Now, given a new face, use the salience Now, given a new face, use the salience map to decide where to look.map to decide where to look.

Compare Compare new new image samples to image samples to storedstored ones - the closest ones in memory get to ones - the closest ones in memory get to vote for their label.vote for their label.

Main IdeaMain Idea

Page 19: Fcv bio cv_cottrell

19

Stored memories of BobStored memories of AliceNew fragments

19Result: 7 votes for Alice, only 3 for Bob. It’s Alice!

Page 20: Fcv bio cv_cottrell

20

VotingVoting

The voting process is based on Bayesian The voting process is based on Bayesian updating (with Naïve Bayes).updating (with Naïve Bayes).

The size of the vote depends on the The size of the vote depends on the distance from the stored sample, using distance from the stored sample, using kernel density estimation. kernel density estimation.

Hence NIMBLE: NIM with Bayesian Hence NIMBLE: NIM with Bayesian Likelihood Estimation.Likelihood Estimation.

Page 21: Fcv bio cv_cottrell

21

The ICA features do double-duty:The ICA features do double-duty: They are They are combinedcombined to make the salience map to make the salience map

- which is used to decide where to look- which is used to decide where to look They are They are storedstored to represent the object at to represent the object at

that locationthat location

QuickTimeª and a decompressor

are needed to see this picture.

Overview of the systemOverview of the system

Page 22: Fcv bio cv_cottrell

22

Compare this to (most, not all!) Compare this to (most, not all!) computer vision systems:computer vision systems:

One pass over the image, and global One pass over the image, and global features.features.

Image Global Features

Global Classifier

Decision

NIMBLE vs. Computer VisionNIMBLE vs. Computer Vision

Page 23: Fcv bio cv_cottrell

23

QuickTime and aªH.264 decompressor

are needed to see this picture.

Page 24: Fcv bio cv_cottrell

24Belief After 1 Fixation Belief After 10 Fixations

Page 25: Fcv bio cv_cottrell

25

Human vision works in multiple environments Human vision works in multiple environments - our basic features (neurons!) don’t change - our basic features (neurons!) don’t change from one problem to the next.from one problem to the next.

We tune our parameters so that the system We tune our parameters so that the system works well on Bird and Butterfly datasets - and works well on Bird and Butterfly datasets - and then apply the system then apply the system unchangedunchanged to faces, to faces, flowers, and objectsflowers, and objects

This is very different from standard computer This is very different from standard computer vision systems, that are (usually) tuned to a vision systems, that are (usually) tuned to a particular domainparticular domain

Robust VisionRobust Vision

Page 26: Fcv bio cv_cottrell

26

Cal Tech 101: 101 Different Categories

AR dataset: 120 Different People with different lighting, expression, and accessories

Page 27: Fcv bio cv_cottrell

27

Flowers: 102 Different Flower SpeciesFlowers: 102 Different Flower Species

Page 28: Fcv bio cv_cottrell

28

~7 fixations required to achieve ~7 fixations required to achieve at least 90% of maximum at least 90% of maximum performance performance

Page 29: Fcv bio cv_cottrell

29

So, we created a simple cognitive model So, we created a simple cognitive model that uses simulated fixations to recognize that uses simulated fixations to recognize things.things. But it isn’t But it isn’t thatthat complicated. complicated.

How does it compare to approaches in How does it compare to approaches in computer vision?computer vision?

Page 30: Fcv bio cv_cottrell

30

Caveats:Caveats: As of mid-2010.As of mid-2010. Only comparing to single feature type Only comparing to single feature type

approaches (no “Multiple Kernel approaches (no “Multiple Kernel Learning” (MKL) approaches).Learning” (MKL) approaches).

Still superior to MKL with very few Still superior to MKL with very few training examples per category.training examples per category.

Page 31: Fcv bio cv_cottrell

311 5 15 30

NUMBER OF TRAINING EXAMPLES

Page 32: Fcv bio cv_cottrell

321 2 3 6 8

NUMBER OF TRAINING EXAMPLES

Page 33: Fcv bio cv_cottrell

33

QuickTime and aª decompressor

are needed to see this picture.

Page 34: Fcv bio cv_cottrell

34Again, best for single feature-type systems

and for 1 training instance better than all systems

Page 35: Fcv bio cv_cottrell

35

More neurally and behaviorally relevant More neurally and behaviorally relevant gaze control and fixation integration.gaze control and fixation integration. People don’t randomly sample images.People don’t randomly sample images.

A foveated retinaA foveated retina

Comparison with human eye movement Comparison with human eye movement data during recognition/classification of data during recognition/classification of faces, objects, etc.faces, objects, etc.

Page 36: Fcv bio cv_cottrell

36

A biologically-inspired, fixation-A biologically-inspired, fixation-based approach can work well for based approach can work well for image classification.image classification.

Fixation-based models can achieve, Fixation-based models can achieve, and even exceed, some of the best and even exceed, some of the best models in computer vision. models in computer vision.

……Especially when you don’t have a Especially when you don’t have a lot of training images.lot of training images.

Page 37: Fcv bio cv_cottrell

37

Software and Paper Available at Software and Paper Available at www.chriskanan.comwww.chriskanan.com

For more details email:For more details email:[email protected]@ucsd.edu

This work was supported by the NSF (grant

#SBE-0542013) to the Temporal Dynamics of

Learning Center.

Page 38: Fcv bio cv_cottrell

38Thanks!

Page 39: Fcv bio cv_cottrell

39

Sparse Principal Sparse Principal Components AnalysisComponents Analysis

We minimize:We minimize:

Subject to the following constraint:Subject to the following constraint:

Page 40: Fcv bio cv_cottrell

40

The SPCA model as a neural net…The SPCA model as a neural net…

It is AT that is mostly 0…

Page 41: Fcv bio cv_cottrell

41

ResultsResults

suggesting the 1/f power spectrum of images is suggesting the 1/f power spectrum of images is where this is coming from…where this is coming from…

Page 42: Fcv bio cv_cottrell

42

ResultsResults

The role of The role of λλ::

Recall this reduces the number of connections…Recall this reduces the number of connections…

Page 43: Fcv bio cv_cottrell

43

ResultsResults

The role of The role of λλ: higher : higher λλ means means fewer connections, which alters fewer connections, which alters the contrast sensitivity function the contrast sensitivity function (CSF).(CSF).

Matches recent data on Matches recent data on malnourished kids and their malnourished kids and their CSF’s: lower sensitivity at low CSF’s: lower sensitivity at low spatial frequencies, but slightly spatial frequencies, but slightly better at high than normal better at high than normal controls…controls…

Page 44: Fcv bio cv_cottrell

44

NIMBLE represents its beliefs using NIMBLE represents its beliefs using probability distributionsprobability distributions

Simple nearest neighbor density Simple nearest neighbor density estimation to estimate: estimation to estimate:

P(P(fixationfixationtt | | CategoryCategory = = kk)) Fixations are combined over fixations/Fixations are combined over fixations/

time using Bayesian updatingtime using Bayesian updating

Page 45: Fcv bio cv_cottrell

45