markov random fields & texture analysis - vincent cheung · vincent cheung probabilistic and...
TRANSCRIPT
-
Jan. 21, 2005
Markov Random Fields &Texture Analysis
Vincent CheungProbabilistic and Statistical Inference GroupElectrical & Computer EngineeringUniversity of Toronto
-
Jan. 21, 2005 Cheung 1
Outline
W. Freeman, E. Pasztor, andO. Carmichael, “Learning Low-Level Vision”– Markov random fields– Belief propagation
S. Zhu, Y. Wu, and D. Mumford, “MinimaxEntropy Principle and its Application toTexture Modeling”– Minimax entropy principle– Feature pursuit
-
Jan. 21, 2005 Cheung 2
Markov Random Field
An undirected graph– Nodes Variables– Edges Functions
Model the joint probability Perform inference
-
Jan. 21, 2005 Cheung 3
Markov Network for Vision
-
Jan. 21, 2005 Cheung 4
A Simple Markov Network
-
Jan. 21, 2005 Cheung 5
Belief Propagation
MAP estimates can be computed for allthe xi simultaneously using “message-passing”
Factorization structure makes thispossible
-
Jan. 21, 2005 Cheung 6
Belief Propagation Example
-
Jan. 21, 2005 Cheung 7
Belief Propagation Summary
What a node thinks another should be
What a node thinks itself is
-
Jan. 21, 2005 Cheung 8
Scene Candidate Patches
-
Jan. 21, 2005 Cheung 9
Applications (1)
Super-Resolution Shading and reflectance Motion estimation
-
Jan. 21, 2005 Cheung 10
Applications (2)
Transparency (A. Levin, A. Zomet andY. Weiss)
= +
-
Jan. 21, 2005 Cheung 11
Minimax Entropy Principle
Let I be an image Assume that observed images, Ii,
i=1,..,m, are random samples from aprobability distribution f(I)
Goal is to estimate f(I) based on theobserved images
-
Jan. 21, 2005 Cheung 12
Maximum Entropy Principle (1)
Maximize entropy to obtain the purestand simplest fusion of the observedfeatures and their statistics
Simplest model as possible
-
Jan. 21, 2005 Cheung 13
Maximum Entropy Principle (2)
s.t.
-
Jan. 21, 2005 Cheung 14
Minimum Entropy Principle
Minimize entropy to increase modelcomplexity by choosing a set of features
-
Jan. 21, 2005 Cheung 15
Feature Pursuit
Inefficient to just try all possible set of Kfeatures
Use greedy approach– Add one feature to the model at a time– Choose feature that maximally decreases the
entropy– Find feature that is most poorly modeled by the
current model and add it to it’s repertoire so thatthe model can better represent this feature.
-
Jan. 21, 2005 Cheung 16
-
Jan. 21, 2005 Cheung 17
Texture Modeling
Histogram of filters as features FRAME - Filter, Random field, And
Minimax Entropy Learn the filters Synthesize texture from the histogram
of the filter response
-
Jan. 21, 2005 Cheung 18
Texture Modeling Example (1)
Obs 0 1
2 3 6
-
Jan. 21, 2005 Cheung 19
Texture Modeling Example (2)