text categorization and images thesis defense for carl sable committee: kathleen mckeown, vasileios...

Post on 06-Jan-2018

220 Views

Category:

Documents

0 Downloads

Preview:

Click to see full reader

DESCRIPTION

Clues for Indoor versus Outdoor: Text (as opposed to visual image features) Denver Summit of Eight leaders begin their first official meeting in the Denver Public Library, June 21. The two engines of an Amtrak passenger train lie in the mud at the edge a marsh after the train, bound for Boston from Washington, derailed on the bank of the Hackensack River, just after crossing a bridge.

TRANSCRIPT

Text Categorization

• Text categorization (TC) refers to the automatic labeling of documents, using natural language text contained in or associated with each document, into one or more pre-defined categories.

• Idea: TC techniques can be applied to image captions or articles to label the corresponding images.

Clues for Indoor versus Outdoor:Text (as opposed to visual image features)

Denver Summit of Eight leaders begin their first official meeting in the Denver Public Library, June 21.

The two engines of an Amtrak passenger train lie in the mud at the edge a marsh after the train, bound for Boston from Washington, derailed on the bank of the Hackensack River, just after crossing a bridge.

Two Paradigms of Research

• Machine learning (ML) techniques– Common in the literature– Usually involve the exploration of new algorithms

applied to bag of words representations of documents• Novel representation

– Rare in the literature– Usually more specific, but often interesting and can

lead to substantial improvement– Important for certain tasks involving images!

Contributions• General:

– An in-depth exploration of the categorization of images based on associated text

– Incorporating research into Newsblaster• Novel machine learning (ML) techniques:

– The creation of two novel TC approaches– The combination of high-precision/low-recall rules

with other systems• Novel representation:

– The integration of NLP and IR– The use of low-level image features

Framework

• Collection of Experiments– Various tasks– Multiple techniques– No clear winner for all tasks– Characteristics of tasks often dictate which

techniques work best• “No Free Lunch”

Overview

I. The Main IdeaII. Description of CorpusIII. Novel ML SystemsIV. NLP Based SystemV. High-Precision/Low-Recall RulesVI. Image FeaturesVII. NewsblasterVIII. Conclusions and Future Work

Corpus

• Raw data:– Postings from news related Usenet newsgroups– Over 2000 include embedded captioned images

• Data sets:– Multiple sets of categories representing various

levels of abstraction– Mutually exclusive and exhaustive categories

Outdoor Indoor

Events Categories

Politics Struggle

Disaster Crime Other

Subcategories for Disaster Images

Politics Struggle

Disaster Crime Other

Category F1

Politics 89%Struggle 88%Disaster 97%Crime 90%Other 59%

Affected People OtherWreckageWorkers Responding

Disaster Image Categories

Affected People

OtherWreckage

Workers Responding

Subcategories for Politics Images

Politics Struggle

Disaster Crime Other

Category F1

Politics 89%Struggle 88%Disaster 97%Crime 90%Other 59%

Meeting OtherPoliticianPhotographed

Announcement Civilians Military

Politics Image Categories

Meeting

Other

CiviliansAnnouncement

MilitaryPolitician Photographed

Collect Labels to Train Systems

Overview

I. The Main IdeaII. Description of CorpusIII. Novel ML SystemsIV. NLP Based SystemV. High-Precision/Low-Recall RulesVI. Image FeaturesVII. NewsblasterVIII. Conclusions and Future Work

Two Novel ML Approaches

• Density estimation– Applied to the results of some other system– Often improves performance– Always provides probabilistic confidence measures for

predictions• BINS

– Uses binning to estimate accurate term weights for words with scarce evidence

– Extremely competitive for two data sets in my corpus

Density Estimation

• First apply a standard system:– For each document, compute a similarity or score for

every category.– Apply to training documents as well as test documents.

• For each test document:– Find all documents from training set with similar

category scores.– Use categories of close training documents to predict

categories of test documents.

Density Estimation Example

85, 35, 25, 95, 20

100, 75, 20, 30, 5

60, 95, 20, 30, 5

90, 25, 50, 110, 25

40, 30, 80, 25, 40

80, 45, 20, 75, 10

Category score vectorsfor training documents:

Category score vectorfor test document:

20.092.5

106.4

27.491.4

36.7

Predictions:Rocchio/TF*IDF: StruggleDE: Crime (Probability .679)

100, 40, 30, 90, 10

Struggle

Politics

Disaster

Crim

e

Other

Distances:

679.07.36

14.27

10.20

17.36

10.20

1

(Crime)

(Struggle)

(Disaster)

(Struggle)

(Politics)

(Crime)

Actual Categories:

Density Estimation Significantly Improves Performancefor the Indoor versus Outdoor Data Set

65.0%

70.0%

75.0%

80.0%

85.0%

90.0%

95.0%

OverallAccuracy

Indoor F1 Outdoor F1

Density EstimationRocchio/TF*IDF

Density Estimation Slightly Degrades Performancefor the Events Data Set

30.0%40.0%50.0%60.0%70.0%80.0%90.0%

100.0%

Overall

Accu

racy

Strugg

le F1

Politic

s F1

Disaste

r F1

Crime F

1

Other F1

Density EstimationRocchio/TF*IDF

Density Estimation Sometimes Improves Performance,Always Provides Confidence Measures

70.0%

75.0%

80.0%

85.0%

90.0%

95.0%

DensityEstimation

Rocchio/TF*IDF

Indoor versus Outdoor Events: Politics, Struggle, Disaster, Crime, Other

70.0%

75.0%

80.0%

85.0%

90.0%

95.0%

Confidence Range # of Images Overall Accuracy %

High (P 0.9) 285 92.6

Medium (0.9 > P 0.7) 98 75.5

Low (0.7 > P 0.5) 62 72.6

Confidence Range # of Documents Overall Accuracy %

High (P 0.9) 301 94.4

Medium (0.9 > P 0.7) 68 79.4

Low (0.7 > P 0.5) 60 53.3

Very Low (0.5 > P) 14 42.9

Results of Density Estimation Experiments for the Events Data Set:

Results of Density Estimation Experiments for the Indoor versus Outdoor Data Set:

BINS System:Naïve Bayes + Smoothing

• Binning: based on smoothing in the speech recognition literature– Not enough training data to estimate term weights for

words with scarce evidence– Words with similar statistical features are grouped into

a common “bin”• Estimate a single weight for each bin

– This weight is assigned to all words in the bin– Credible estimates even for small (or zero) counts

Binning Uses Statistical Features of Words

Intuition Word

Indoor Category

Count

Outdoor Category

CountQuantized

IDF

Clearly Indoor

conference 14 1 4

bed 1 0 8

Clearly Outdoor

plane 0 9 5

earthquake 0 4 6

Unclearspeech 2 2 6

ceremony 3 8 5

“plane”

• Sparse data– “plane” does not occur in any Indoor training

documents– Infinitely more likely to be Outdoor ???

• Assign “plane” to bins of words with similar features (e.g. IDF, category counts)

• In first half of training set, “plane” appears in:– 9 Outdoor documents – 0 Indoor documents

Lambdas: Weights• First half of training set: Assign words to bins• Second half of training set: Estimate term weights

binword ||

)(||

1)|( docswordDF

binbinobsP

)|(log2bin binobsP

Lambdas for “plane”:4.03 times more likely in an Outdoor document

310*31.5)bin |obs( IndoorP

210*13.2)bin |obs( OutdoorP

01.2bin) |P(obs

bin) |P(obslog221 OutdoorIndoor

Binning Credible Log Likelihood Ratios

Intuition Word

Indoor minus Outdoor

Indoor Category

Count

Outdoor Category

CountQuantized

IDF

Clearly Indoor

conference 4.84 14 1 4

bed 1.35 1 0 8

Clearly Outdoor

plane -2.01 0 9 5

earthquake -1.00 0 4 6

Unclearspeech 0.84 2 2 6

ceremony -0.50 3 8 5

Lambdas Decrease with IDF

Disaster lambdas

-11-10-9-8-7-6-5-4

1 2 3 4 5 6 7 8IDF

lam

bda

count=0count=1

Methodology of BINS

• Divide training set into two halves:– First half used to determine bins for words– Second half used to determine lambdas for bins

• For each test document:– Map every word to a bin for each category– Add lambdas, obtaining a score for each category

• Switch halves of training and repeat • Combine results and assign each document to

category with highest score

Binning Improves Performancefor the Indoor versus Outdoor Data Set

70.0%

75.0%

80.0%

85.0%

90.0%

95.0%

OverallAccuracy

Indoor F1 Outdoor F1

BINSNaïve Bayes

Binning Improves Performancefor the Events Data Set

20.0%30.0%40.0%50.0%60.0%70.0%80.0%90.0%

100.0%

Overall

Accu

racy

Strugg

le F1

Politic

s F1

Disaste

r F1

Crime F

1

Other F1

BINSNaïve Bayes

BINS: Robust Version of Naïve Bayes

70.0%

75.0%

80.0%

85.0%

90.0%

95.0%

70.0%

75.0%

80.0%

85.0%

90.0%

95.0%

BINS

Naïve Bayes

Indoor versus Outdoor Events: Politics, Struggle, Disaster, Crime, Other

baseline

humans

Combining Bin Weights and Naïve Bayes Weights

• Idea:– It might be better to use the Naïve Bayes weight when

there is enough evidence for a word– Back off to the bin weight otherwise

• BINS allows combinations of weights to be used based on the level of evidence

• How can we automatically determine when to use which weights???– Entropy– Minimum Squared Error (MSE)

Can Provide File to BINS that Specifies How to Combine Weights

0

0.5

1

0

0.25

0.5

0.75

1

Based on Entropy: Based on MSE:

Use only bin weight for evidence of 0

Average bin weight and NB weight for evidence of 1

Use only NB weight for evidence of 2 or more

Appropriately Combining the Bin Weight and the Naïve Bayes Weight Leads to the Best Performance Yet

Indoor versus Outdoor Events: Politics, Struggle, Disaster, Crime, Other

70.0%

75.0%

80.0%

85.0%

90.0%

95.0%

70.0%

75.0%

80.0%

85.0%

90.0%

95.0%

BINS (Combo #2)

BINS (Combo #1)

BINS

Naïve Bayes

BINS Performs the Best of All Systems Tested

70.0%

75.0%

80.0%

85.0%

90.0%

95.0%

70.0%

75.0%

80.0%

85.0%

90.0%

95.0%Density Estimation

Rocchio/TF*IDF

BINS (Combo #2)

BINS (Combo #1)

BINS

Naïve Bayes

K-Nearest Neighbor

Naïve Bayes [R]

Rocchio/TF*IDF [R]

K-Nearest Neighbor [R]

Probabilistic Indexing [R]

Support Vector Machines [R]

Maximum Entropy [R]

Indoor versus Outdoor Events: Politics, Struggle, Disaster, Crime, Other

BINS BINSSVMs SVMs

How Can We Improve Results?

• One idea: Label more documents!– Usually works – Boring

• Another idea: Use unlabeled documents!– Easily obtainable– But can this really work??? – Maybe it can…

Binning Using Unlabeled Documents

• Apply system to unlabeled documents• Choose documents with “confident” predictions

– Each word has new feature: # of unlabeled documents containing the word that are confidently predicted to belong to each category (unlabeled category counts)

– Probably less important than regular category counts– Binning provides a natural mechanism for weighting

the new feature appropriately

Determining Confident Predictions

• BINS computes a score for each category– BINS predicts category with highest score– Confidence for predicted category is score of that category

minus score of second place category– Confidence for non-predicted category is score of that

category minus score of chosen category• Cross validation experiments can be used to determine

a confidence cutoff for each category– Maximize F for category– Beta of 1 gives precision and recall equal weight, lower beta

weights precision higher

Results for Struggle Category

0

0.2

0.4

0.6

0.8

1

-600 -300 0 300

Confidence Cutoff

Valu

e of

Met

rics

Precision

Recall

F1

F(1/3)

Use F to Optimize Confidence Cutoffs (example for a single category)

Use F to Optimize Confidence Cutoffs (important region of graph highlighted)

Results for Struggle Category

0.5

0.6

0.7

0.8

0.9

1

0 20 40 60 80 100

Confidence Cutoff

Valu

e of

Met

rics

Precision

Recall

F1

F(1/3)

Should the New Feature Matter?

zero count lambdas (IDF=8, beta=0.5)

-12

-11

-10

-9

-8

-7

-6

-5

0 1 2 3 4 5 6 7 8 9

unlabeled category count

lambd

a

DisStrPolCri

zero count lambdas (category=Disaster, beta=1.0)

-12

-11

-10

-9

-8

-7

-6

-5

0 1 2 3 4 5 6 7 8 9 10

unlabeled category count

lambd

a

IDF=5

IDF=6

IDF=7

IDF=8

Does the New Feature Help?

• No• Why???

– New features add info but make bins smaller– Perhaps more data isn’t needed in the first place

• Should more data matter?– Hard to accumulate more labeled data– Easy to try out less labeled data!

Does Size Matter?

Effect of Training Data

55.0%

60.0%

65.0%

70.0%

75.0%

80.0%

85.0%

90.0%

95.0%

10%

20%

30%

40%

50%

60%

70%

80%

90%

100%

Percentage Used

Pefo

rman

ce

IN/OUT

EVENTS

Overview

I. The Main IdeaII. Description of CorpusIII. Novel ML SystemsIV. NLP Based SystemV. High-Precision/Low-Recall RulesVI. Image FeaturesVII. NewsblasterVIII. Conclusions and Future Work

Disaster Image Categories

Affected People

OtherWreckage

Workers Responding

Performance of Standard SystemsNot Very Satisfying

52.00%

54.00%

56.00%

58.00%

60.00%

62.00%

64.00%

66.00%Density Estimation

Rocchio/TF*IDF

BINS

Naïve Bayes [R]

Rocchio/TF*IDF [R]

K-Nearest Neighbor [R]

Probabilistic Indexing [R]

Support Vector Machines [R]

Maximum Entropy [R]

Ambiguity for Disaster Images:Workers Responding vs. Affected People

Philippine rescuers carry a fire victim March 19 who perished in a blaze at a Manila disco.

Hypothetical alternative caption: A fire victim who perished in a blaze at a Manila disco is carried by Philippine rescuers March 19.

Workers Responding Affected People

Summary of Observations About Task

• Need to distinguish foreground from background, determine focus of image

• Not all words are important; some are misleading• Hypothesis: the main subject and verb are

particularly useful for this task– Problematic for bag of words approaches– Need linguistic analysis to determine predicate

argument relationships

Philippine rescuers carry a fire victim March 19 who perished in a blaze at a Manila disco.

Hypothesis: Subject and Verbare Useful Clues

Subject Verb Category Guessable?

Truck makes Wreckage No

couple mourn Affected People Yesblocks suffered Wreckage YesNAME gather Affected People No

child sleeps Affected People Yesinspectors search Workers Responding Yes

NAME observes Workers Responding No

workers confer Workers Responding Yes

child covers Affected People Yeschimney stands Wreckage Yes

Experiments with Humans Subjects: 4 Conditions

Test Hypothesis: Subject and Verb are Useful Clues

SENT: First sentence of caption

Philippine rescuers carry a fire victim March 19 who perished in a blaze at a Manila disco.

RAND: All words from first sentence in random order

At perished disco who Manila a a in 19 carry Philippine blaze victim a rescuers March fire

IDF: Top two TF*IDF words

disco rescuers

S-V: Subject and verb subject = “rescuers”, verb = “carry”

• More words are better than fewer words– SENT, RAND > S-V, IDF

• Syntax is important– SENT > RAND; S-V > IDF

Experiments with Humans Subjects: ResultsHypothesis: Subject and Verb are Useful Clues

50.0%

55.0%60.0%

65.0%70.0%

75.0%80.0%

85.0%90.0%

95.0%

SENTRANDIDFS-V Condition Average Time

(in seconds)RAND 68SENT 34IDF 22S-V 20

RAND is Very Slow!

• Perhaps human subjects unscrambled words, regaining syntactic information

Condition Average Time (in seconds)

RAND 68SENT 34IDF 22S-V 20

Using Just Two Words (S-V)Almost as Good as All the Words (Bag of Words)

52.00%

54.00%

56.00%

58.00%

60.00%

62.00%

64.00%

66.00%

SENT S-V

Density Estimation

Rocchio/TF*IDF

BINS

Naïve Bayes [R]

Rocchio/TF*IDF [R]

K-Nearest Neighbor [R]

Probabilistic Indexing [R]

Support Vector Machines [R]

Maximum Entropy [R]

Operational NLP Based System

• For each test document:– Extract subject and verb– Compare to those from training set using some method

of word-to-word similarity– Based on similarities, generate a score for every

category

Sentence POS taggerCASS shallow parser

Perl script WordNet Output

Subjects 83.9%

Verbs 80.6%

• Extract subjects and verbs from all documents in training set

Word Similarity• Examine large “extended corpus” to generate many

subject/verb pairs• Use to compute similarities:

total verbs#commonin verbs#Sim ,

21 SubSub

totalsubjects #commonin subjects #Sim ,

21 VerbVerb

totalsappearance # togethersappearance # * 2Sim , VerbSub

Choosing a Category

• For given test document d, calculate total score for every category c:

• Choose category with highest score• If subject is NAME, a bit more complicated

cc cdcd

cc cdcd

Vv vvvs

Ss svssdc

,,

,,

SimSim

SimSim|Score

The NLP Based System Beats All Others by a Considerable Margin

52.0%

54.0%

56.0%

58.0%

60.0%

62.0%

64.0%

66.0%

68.0%

SENT S-V

NLP Based System

Density Estimation

Rocchio/TF*IDF

BINS

Naïve Bayes [R]

Rocchio/TF*IDF [R]

K-Nearest Neighbor [R]

Probabilistic Indexing [R]

Support Vector Machines [R]

Maximum Entropy [R]

Politics Image Categories

Meeting

Other

CiviliansAnnouncement

MilitaryPolitician Photographed

The NLP Based System is in the Middle of the Pack for the Politics Image Data Set

35.0%

40.0%

45.0%

50.0%

55.0%

60.0%

65.0%

SENT S-V

NLP Based System

Density Estimation

Rocchio/TF*IDF

BINS

Naïve Bayes [R]

Rocchio/TF*IDF [R]

K-Nearest Neighbor [R]

Probabilistic Indexing [R]

Support Vector Machines [R]

Maximum Entropy [R]

Why is the Performance for the NLP Based System not as Strong for the Politics Image Data Set?

• A much wider range of performance scores– Range for Politics images is 36% to 64.7%– Range for Disaster images is 54% to 59.7%– The top systems are harder to beat

• Too many proper names as subjects– 60% of test instances for Politics images– Only 13% of test instances for Disaster images– For 60% of test documents, only one word (the main

verb) is being used to determine the prediction

Overview

I. The Main IdeaII. Description of CorpusIII. Novel ML SystemsIV. NLP Based SystemV. High-Precision/Low-Recall RulesVI. Image FeaturesVII. NewsblasterVIII. Conclusions and Future Work

The Original Premise

• For the Disaster image data set, the performance of the NLP based system still leaves room for improvement– NLP based system achieves 65% overall accuracy for

the Disaster image data set– Humans viewing all words in random order achieve

about 75%– Humans viewing full first sentence achieve over 90%

• Main subject and verb are particularly important, but sometimes other words might offer good clues

Higinio Guereca carries family photos he retrieved from his mobile home which was destroyed as a tornado moved through the Central Florida community, early December 27.

Choosing Indicative Words

• Let x be the number of training documents containing a word w

• Let p be the proportion of these documents that belong to category c

• If x > X and p > P then w is indicative of c• X and P can be varied to generate lists of

indicative words• Lists can be pruned manually

Selected Indicative Words for the Disaster Image Data Set

Word Indicated Category Total Count (x) Proportion (p)her Affected People 7 1.0his Affected People 7 0.86family Affected People 6 0.83relatives Affected People 6 1.0rescue Workers Responding 15 1.0search Workers Responding 9 1.0similar Other 2 1.0soldiers Workers Responding 6 1.0workers Workers Responding 12 1.0

Selected Indicative Words for the Politics Image Data Set

Word Indicated Category Total Count (x) Proportion (p)hands Meeting 10 0.90journalists Announcement 4 1.0local Civilians 4 1.0media Announcement 3 1.0presidential Politician Photographed 9 0.78press Announcement 7 0.71reporters Announcement 8 0.88meeting Meeting 15 0.73session Meeting 6 0.83victory Politician Photographed 6 0.83waves Politician Photographed 4 1.0wife Politician Photographed 6 1.0

High-Precision/Low-Recall Rules

• If a word w that indicates category c occurs in a document d, then assign d to c

• Every selected indicative word has an associated “rule” of the above form– Each rule is very accurate but rarely applicable– If only rules are used:

• most predictions will be correct (hence, high precision)• most instances of most categories will remain unlabeled

(hence, low recall)

Combining the High-Precision/Low-Recall Rules with Other Systems

• Two-pass approach:– Conduct a first-pass using the indicative words and

the high-precision/low-recall rules– For documents that are still unlabeled, fall back to

some other system• Compared to the fall back system:

– If the rules are more accurate for the documents to which they apply, overall accuracy will improve!

– Intended to improve the NLP based system, but easy to test with other systems as well

The Rules Improve Every Fall Back System for the Disaster Image Data Set

52.0%

54.0%

56.0%

58.0%

60.0%

62.0%

64.0%

66.0%

68.0%

Without Rules With Rules

NLP Based System

BINS

Naïve Bayes [R]

Rocchio/TF*IDF [R]

K-Nearest Neighbor [R]

Probabilistic Indexing [R]

Support Vector Machines [R]

Maximum Entropy [R]

The Rules Improve 7 of 8 Fall Back Systems for the Politics Image Data Set

35.0%

40.0%

45.0%

50.0%

55.0%

60.0%

65.0%

Without Rules With Rules

NLP Based System

BINS

Naïve Bayes [R]

Rocchio/TF*IDF [R]

K-Nearest Neighbor [R]

Probabilistic Indexing [R]

Support Vector Machines [R]

Maximum Entropy [R]

Overview

I. The Main IdeaII. Description of CorpusIII. Novel ML SystemsIV. NLP Based SystemV. High-Precision/Low-Recall RulesVI. Image FeaturesVII. NewsblasterVIII. Conclusions and Future Work

Low-Level Image Features

• Collaboration with Paek and Benitez– They have provided me with information,

pointers to resources, and code– I have reimplemented some of their code

• Color histograms– Based on entire images or image regions– Can be used as input to machine learning

approaches (e.g. kNN, SVMs)

Color• Three components to color

– Red, green, blue (RGB)– Hue, saturation, value (HSV)

• Can convert from RGB to HSV– Can quantize HSV triples– 18 hues * 3 saturations * 3 values + 4 grays = 166 slots

Color Histograms

• For each pixel of image, compute its quantized HSV triple

• Color histogram of image is vector such that:– There are 166 dimensions– Each dimension represents one possible HSV triple– Value of dimension is proportion of pixels with

associated HSV triple• Can be computed for image regions and

concatenated together• Can be input for machine learning techniques

Images Divided into 8 x 8 Rectangular Regions of Equal Size

Using Color Histograms to Predict Labels for the Indoor versus Outdoor Data Set

70.0%

72.0%

74.0%

76.0%

78.0%

80.0%

whole images image regions

K-Nearest Neighbor

Support VectorMachines

Combining Text and Image Features

• Combining systems has had mixed results in the TC literature, but:– Most attempts have involved systems that use the same

features (bag of words)– There is little reason to believe that indicative text is

correlated with indicative low-level image features• Most text based systems are beating the image

based systems, but:– Distance from optimal hyperplane can be used as a

confidence measure for support vector machine– Predictions with high confidence may be more accurate

than text systems

Accuracy of Support Vector Machine Approach Tends to be Higher when Confidence is Greater

Distance Cutoff Overall Accuracy % % of Images Above Cutoff3.5 --- 0.03.0 100.0 0.42.5 87.5 1.82.0 92.3 5.81.5 94.4 16.01.0 91.0 34.10.5 84.6 70.10.0 78.0 100.0

The Combination of Text and Image Beats Text Alone:Most systems show small gains, one has major improvement

77.0%

79.0%

81.0%

83.0%

85.0%

87.0%

89.0%

Text Only Text and Image

BINS

Naïve Bayes [R]

Rocchio/TF*IDF [R]

K-Nearest Neighbor [R]

Probabilistic Indexing [R]

Support Vector Machines [R]

Maximum Entropy [R]

Overview

I. The Main IdeaII. Description of CorpusIII. Novel ML SystemsIV. NLP Based SystemV. High-Precision/Low-Recall RulesVI. Image FeaturesVII. NewsblasterVIII. Conclusions and Future Work

Newsblaster Categories

Entertainment Science/Technology Sports

U.S. News World News Finance

Newsblaster

• A pragmatic showcase for NLP• My contributions:

– Extraction of images and captions from web pages

– Image browsing interface– Categorization of stories (clusters) and images– Scripts that allow users to suggest labels for

articles with incorrect predictions

Overview

I. The Main IdeaII. Description of CorpusIII. Novel ML SystemsIV. NLP Based SystemV. High-Precision/Low-Recall RulesVI. Image FeaturesVII. NewsblasterVIII. Conclusions and Future Work

Conclusions

• TC techniques can be used to categorize images• Many methods exist

– No clear winner for all tasks– BINS is very competitive– NLP can lead to substantial improvement, at least for

certain tasks– High-precision/low-recall rules are likely to improve

performance for tough tasks– Image features show promise

• Newsblaster demonstrates pragmatic benefits of my work

Future Work

• BINS– Explore additional binning features– Explore use of unlabeled data

• NLP and TC– Improve current system– Explore additional categories

• Image features– Explore additional low-level image features– Explore better methods of combining text and image

And Now the Questions…

top related