generative learning methods for bags of features model the probability of a bag of features given a...

Post on 21-Dec-2015

223 Views

Category:

Documents

0 Downloads

Preview:

Click to see full reader

TRANSCRIPT

Generative learning methods for bags of features

• Model the probability of a bag of features given a class

Many slides adapted from Fei-Fei Li, Rob Fergus, and Antonio Torralba

Generative methods

• We will cover two models, both inspired by text document analysis:• Naïve Bayes• Probabilistic Latent Semantic Analysis

The Naïve Bayes modelThe Naïve Bayes model

Csurka et al. 2004

• Assume that each feature is conditionally independent given the class

fi: ith feature in the imageN: number of features in the image

N

iiN cfpcffp

11 )|()|,,(

The Naïve Bayes modelThe Naïve Bayes model

Csurka et al. 2004

M

j

wnj

N

iiN

jcwpcfpcffp1

)(

11 )|()|()|,,(

• Assume that each feature is conditionally independent given the class

wj: jth visual word in the vocabularyM: size of visual vocabularyn(wj): number of features of type wj in the image

fi: ith feature in the imageN: number of features in the image

The Naïve Bayes modelThe Naïve Bayes model

Csurka et al. 2004

• Assume that each feature is conditionally independent given the class

No. of features of type wj in training images of class c

Total no. of features in training images of class c

p(wj | c) =

M

j

wnj

N

iiN

jcwpcfpcffp1

)(

11 )|()|()|,,(

The Naïve Bayes modelThe Naïve Bayes model

Csurka et al. 2004

• Assume that each feature is conditionally independent given the class

No. of features of type wj in training images of class c + 1

Total no. of features in training images of class c + M

p(wj | c) =

(Laplace smoothing to avoid zero counts)

M

j

wnj

N

iiN

jcwpcfpcffp1

)(

11 )|()|()|,,(

M

jjjc

M

j

wnjc

cwpwncp

cwpcpc j

1

1

)(

)|(log)()(logmaxarg

)|()(maxarg*

The Naïve Bayes modelThe Naïve Bayes model

Csurka et al. 2004

• Maximum A Posteriori decision:

(you should compute the log of the likelihood instead of the likelihood itself in order to avoid underflow)

The Naïve Bayes modelThe Naïve Bayes model

Csurka et al. 2004

wN

c

• “Graphical model”:

Probabilistic Latent Semantic AnalysisProbabilistic Latent Semantic Analysis

T. Hofmann, Probabilistic Latent Semantic Analysis, UAI 1999

zebra grass treeImage

= p1 + p2 + p3

“visual topics”

Probabilistic Latent Semantic AnalysisProbabilistic Latent Semantic Analysis

• Unsupervised technique• Two-level generative model: a document is a

mixture of topics, and each topic has its own characteristic word distribution

wd z

T. Hofmann, Probabilistic Latent Semantic Analysis, UAI 1999

document topic wordP(z|d) P(w|z)

Probabilistic Latent Semantic AnalysisProbabilistic Latent Semantic Analysis

• Unsupervised technique• Two-level generative model: a document is a

mixture of topics, and each topic has its own characteristic word distribution

wd z

T. Hofmann, Probabilistic Latent Semantic Analysis, UAI 1999

K

kjkkiji dzpzwpdwp

1

)|()|()|(

The pLSA modelThe pLSA model

K

kjkkiji dzpzwpdwp

1

)|()|()|(

Probability of word i in document j

(known)

Probability of word i given

topic k (unknown)

Probability oftopic k givendocument j(unknown)

The pLSA modelThe pLSA model

Observed codeword distributions

(M×N)

Codeword distributionsper topic (class)

(M×K)

Class distributionsper image

(K×N)

K

kjkkiji dzpzwpdwp

1

)|()|()|(

p(wi|dj) p(wi|zk)

p(zk|dj)

documents

wor

ds

wor

ds

topics

topi

cs

documents

=

Maximize likelihood of data:

Observed counts of word i in document j

M … number of codewords

N … number of images

Slide credit: Josef Sivic

Learning pLSA parametersLearning pLSA parameters

InferenceInference

)|(maxarg dzpzz

• Finding the most likely topic (class) for an image:

InferenceInference

)|(maxarg dzpzz

• Finding the most likely topic (class) for an image:

zzz dzpzwp

dzpzwpdwzpz

)|()|(

)|()|(maxarg),|(maxarg

• Finding the most likely topic (class) for a visual word in a given image:

Topic discovery in images

J. Sivic, B. Russell, A. Efros, A. Zisserman, B. Freeman, Discovering Objects and their Location in Images, ICCV 2005

Application of pLSA: Action recognition

Juan Carlos Niebles, Hongcheng Wang and Li Fei-Fei, Unsupervised Learning of Human Action Categories Using Spatial-Temporal Words, IJCV 2008.

Space-time interest points

Application of pLSA: Action recognition

Juan Carlos Niebles, Hongcheng Wang and Li Fei-Fei, Unsupervised Learning of Human Action Categories Using Spatial-Temporal Words, IJCV 2008.

pLSA model

– wi = spatial-temporal word

– dj = video

– n(wi, dj) = co-occurrence table (# of occurrences of word wi in video dj)

– z = topic, corresponding to an action

K

kjkkiji dzpzwpdwp

1

)|()|()|(

Probability of word i in video j(known)

Probability of word i given

topic k (unknown)

Probability oftopic k given

video j(unknown)

Action recognition example

Multiple Actions

Multiple Actions

Summary: Generative models

• Naïve Bayes• Unigram models in document analysis• Assumes conditional independence of words given class• Parameter estimation: frequency counting

• Probabilistic Latent Semantic Analysis• Unsupervised technique• Each document is a mixture of topics (image is a mixture of

classes)• Can be thought of as matrix decomposition• Parameter estimation: Expectation-Maximization

top related