convolutional deep belief networks for scalable unsupervised learning of hierarchical...

18
Convolutional Deep Belief Networks for Scalable Unsupervised Learning of Hierarchical Representations Honglak Lee, Roger Grosse, Rajesh Ranganath, and Andrew Y. Ng ICML 2009 Presented by: Mingyuan Zhou Duke University, ECE September 18, 2009

Upload: elaine-butler

Post on 17-Dec-2015

216 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Convolutional Deep Belief Networks for Scalable Unsupervised Learning of Hierarchical Representations Honglak Lee, Roger Grosse, Rajesh Ranganath, and

Convolutional Deep Belief Networks for Scalable Unsupervised Learning of

Hierarchical Representations

Honglak Lee, Roger Grosse, Rajesh Ranganath, and Andrew Y. Ng

ICML 2009

Presented by: Mingyuan Zhou

Duke University, ECE

September 18, 2009

Page 2: Convolutional Deep Belief Networks for Scalable Unsupervised Learning of Hierarchical Representations Honglak Lee, Roger Grosse, Rajesh Ranganath, and

Outline

• Motivations

• Contributions

• Backgrounds

• Algorithms

• Experiment results

• Deep Vs Shallow

• Conclusions

Page 3: Convolutional Deep Belief Networks for Scalable Unsupervised Learning of Hierarchical Representations Honglak Lee, Roger Grosse, Rajesh Ranganath, and

Motivations

• To Learn hierarchical models which simultaneously represent multiple levels, e.g., pixel intensities, edges, object parts, objects, and beyond can be represented by layers from low to high.

• Combining top-down and bottom-up processing of an image.

• Limitations of deep belief networks (DBNs)

• Scaling DBNs to realistic-size images remains challenging: images are high-dimentional and objects can appear at arbitrary locations in images.

Page 4: Convolutional Deep Belief Networks for Scalable Unsupervised Learning of Hierarchical Representations Honglak Lee, Roger Grosse, Rajesh Ranganath, and

Contributions

• Convolutional RBM: feature detectors are shared among all locations in an image.

• Probabilistic max-pooling: in a probabilistic sound way allowing higher-layer units to cover larger areas of the input.

• The first translation invariant hierarchical generative model supporting both top-down and bottom-up probabilistic inference and sales to realistic image sizes.

Page 5: Convolutional Deep Belief Networks for Scalable Unsupervised Learning of Hierarchical Representations Honglak Lee, Roger Grosse, Rajesh Ranganath, and

Backgrounds: Restricted Boltzmann Machine (RBM)

(binary v)

(real-value v)

• Giving the visible layer, the hidden units are conditionally independent, and vise versa.

• Efficient block Gibbs sampling can be performed by alternately sampling each layer’s units.

• Computing the exact gradient of the log-likelihood is intractable, so the contrastive divergence approximation is commonly used.

Page 6: Convolutional Deep Belief Networks for Scalable Unsupervised Learning of Hierarchical Representations Honglak Lee, Roger Grosse, Rajesh Ranganath, and

Backgrounds: Deep belief network (DBN)

• In a DBN, two adjacent layers have a full set of connections between them, but no two units in the same layer are connected.

• A DBN can be formed by stacking RBMs.

• An efficient algorithm for training DBNs (Hinton et al., 2006): greedily training each layer, from lowest to highest, as an RBM using the previous layer's activations as inputs.

Page 7: Convolutional Deep Belief Networks for Scalable Unsupervised Learning of Hierarchical Representations Honglak Lee, Roger Grosse, Rajesh Ranganath, and

Algorithms: Convolutional RBM (CRBM)

Page 8: Convolutional Deep Belief Networks for Scalable Unsupervised Learning of Hierarchical Representations Honglak Lee, Roger Grosse, Rajesh Ranganath, and

Algorithms: Probabilistic max-pooling

Page 9: Convolutional Deep Belief Networks for Scalable Unsupervised Learning of Hierarchical Representations Honglak Lee, Roger Grosse, Rajesh Ranganath, and

Algorithms: Probabilistic max-pooling

• Each unit in a pooling layer computes the maximum activation of the units in a small region of the detection layer.

• Shrinking the representation with max-pooling allows higher-layer representations to be invariant to small translations of the input and reduces the computational burden.

• Max-pooling was intended only for feed-forward architectures. A generative model of images which supports both top-down and bottom-up inference is of interest.

Page 10: Convolutional Deep Belief Networks for Scalable Unsupervised Learning of Hierarchical Representations Honglak Lee, Roger Grosse, Rajesh Ranganath, and

Algorithms: Sparsity regulations

• Only a tiny fraction of the units should be active in relation to a given stimulus.

• Regularizing the objective function to encourage each of the hidden units to have a mean activation close to some small constant .

Page 11: Convolutional Deep Belief Networks for Scalable Unsupervised Learning of Hierarchical Representations Honglak Lee, Roger Grosse, Rajesh Ranganath, and

Algorithms: Convolutional DBN (CDBN)

• CDBN consists of several max-pooling-CRBMs stacked on top of one another.

• Once a given layer is trained, its weights are frozen, and its activations are used as input to the next layer.

Page 12: Convolutional Deep Belief Networks for Scalable Unsupervised Learning of Hierarchical Representations Honglak Lee, Roger Grosse, Rajesh Ranganath, and

Hierarchical probabilistic inference

Page 13: Convolutional Deep Belief Networks for Scalable Unsupervised Learning of Hierarchical Representations Honglak Lee, Roger Grosse, Rajesh Ranganath, and

Experimental Results: natural images

Page 14: Convolutional Deep Belief Networks for Scalable Unsupervised Learning of Hierarchical Representations Honglak Lee, Roger Grosse, Rajesh Ranganath, and

Experimental Results: image classification

Page 15: Convolutional Deep Belief Networks for Scalable Unsupervised Learning of Hierarchical Representations Honglak Lee, Roger Grosse, Rajesh Ranganath, and

Experimental Results: unsupervised learning of object parts

Page 16: Convolutional Deep Belief Networks for Scalable Unsupervised Learning of Hierarchical Representations Honglak Lee, Roger Grosse, Rajesh Ranganath, and

Experimental Results: Hierarchical probabilistic inference

Page 17: Convolutional Deep Belief Networks for Scalable Unsupervised Learning of Hierarchical Representations Honglak Lee, Roger Grosse, Rajesh Ranganath, and

Deep Vs Shallow

From Jason Weston’s slides: DEEP LEARNING VIA SEMI-SUPERVISED EMBEDDING, ICML 2009 WORKSHOP ON LEARNING FEATURE HIERARCHIES

.

From Francis Bach’s slides: Convex sparse methods for feature hierarchies, ICML 2009 WORKSHOP ON LEARNING FEATURE HIERARCHIES

Page 18: Convolutional Deep Belief Networks for Scalable Unsupervised Learning of Hierarchical Representations Honglak Lee, Roger Grosse, Rajesh Ranganath, and

Conclusions

Convolutional deep belief network:

• A scalable generative model for learning hierarchical representations from unlabeled images.

• Performing well in a variety of visual recognition tasks.