lscom

99
A Large Scale Concept Ontology for Multimedia Understanding April 2005 MITRE Milind Naphade, John R. Smith, Alexander Hauptmann, Shih-Fu Chang & Edward Chang IBM Research, Carnegie Mellon University, Columbia University & University of California at Santa Barbara [email protected] [email protected] [email protected] [email protected] [email protected] NRRC NWRRC

Upload: charliedagli

Post on 27-Jan-2015

109 views

Category:

Education


0 download

DESCRIPTION

 

TRANSCRIPT

Page 1: Lscom

A Large Scale Concept Ontology for Multimedia Understanding

April 2005

MITRE

Milind Naphade, John R. Smith, Alexander Hauptmann, Shih-Fu Chang & Edward Chang

IBM Research, Carnegie Mellon University, Columbia University & University of California at Santa Barbara [email protected] [email protected] [email protected] [email protected]

[email protected]

NRRC NWRRC

Page 2: Lscom

Page MITREPNN

L

Central Idea• Collaborative activity of three

critical communities – Users, Library Scientists and Knowledge Experts, and Technical Researchers, Algorithm, System and Solution Designers – to create a user-driven concept ontology for analysis of video broadcast news

Page 3: Lscom

Page MITREPNN

L

Users (Analysts, Broadcasters)

Intelligence Community,

BroadcastingCorporations

Central Idea• Collaborative activity of three

critical communities – Users, Library Scientists and Knowledge Experts, and Technical Researchers, Algorithm, System and Solution Designers – to create a user-driven concept ontology for analysis of video broadcast news

Page 4: Lscom

Page MITREPNN

L

Users (Analysts, Broadcasters)

Intelligence Community,

BroadcastingCorporations

Technical Researchers, Algorithm Designers & System Developers

Vision, Machine Learning, Detection

Analytics

Central Idea• Collaborative activity of three

critical communities – Users, Library Scientists and Knowledge Experts, and Technical Researchers, Algorithm, System and Solution Designers – to create a user-driven concept ontology for analysis of video broadcast news

Page 5: Lscom

Page MITREPNN

L

Knowledge Representation, Library ScientistsStandardization

Ontology Experts

Users (Analysts, Broadcasters)

Intelligence Community,

BroadcastingCorporations

Technical Researchers, Algorithm Designers & System Developers

Vision, Machine Learning, Detection

Analytics

Central Idea• Collaborative activity of three

critical communities – Users, Library Scientists and Knowledge Experts, and Technical Researchers, Algorithm, System and Solution Designers – to create a user-driven concept ontology for analysis of video broadcast news

Page 6: Lscom

Page MITREPNN

L

Lexicon and Ontology

1000 or more concepts Knowledge

Representation, Library ScientistsStandardization

Ontology Experts

Users (Analysts, Broadcasters)

Intelligence Community,

BroadcastingCorporations

Technical Researchers, Algorithm Designers & System Developers

Vision, Machine Learning, Detection

Analytics

Central Idea• Collaborative activity of three

critical communities – Users, Library Scientists and Knowledge Experts, and Technical Researchers, Algorithm, System and Solution Designers – to create a user-driven concept ontology for analysis of video broadcast news

Page 7: Lscom

Page MITREPNN

L

Problem• Users and analysts require richly annotated video content for

accomplishing required access and analysis functions over massive amount of video content.

• Big Barriers:- Research community needs to advance technology for

bridging gap from low-level features to semantics- Lack of large scale useful well-defined semantic lexicon- Lack of user-centric ontology- Lack of corpora annotated with rich lexicon- Lack of feasibility studies for any ontology if defined

• Examples:- The TRECVID lexicon defined from a frequentist

perspective. Its not user-centric.• No effort to date to design lexicon by joint partnership between

different communities (users, knowledge experts, technical)

Page 8: Lscom

Page MITREPNN

L

Workshop Goals

Page 9: Lscom

Page MITREPNN

L

Workshop Goals• Organize series of workshops that bring together three critical

communities – Users, Library Scientists and Knowledge Experts, and Technical Researchers – to create a ontology on order of 1000 concepts for analysis of video broadcast news

Page 10: Lscom

Page MITREPNN

L

Workshop Goals• Organize series of workshops that bring together three critical

communities – Users, Library Scientists and Knowledge Experts, and Technical Researchers – to create a ontology on order of 1000 concepts for analysis of video broadcast news

• Ensure impact through focused collaboration of these different communities to achieve balance of usefulness, feasibility and size

Page 11: Lscom

Page MITREPNN

L

Workshop Goals• Organize series of workshops that bring together three critical

communities – Users, Library Scientists and Knowledge Experts, and Technical Researchers – to create a ontology on order of 1000 concepts for analysis of video broadcast news

• Ensure impact through focused collaboration of these different communities to achieve balance of usefulness, feasibility and size

• Specific Tasks:

Page 12: Lscom

Page MITREPNN

L

Workshop Goals• Organize series of workshops that bring together three critical

communities – Users, Library Scientists and Knowledge Experts, and Technical Researchers – to create a ontology on order of 1000 concepts for analysis of video broadcast news

• Ensure impact through focused collaboration of these different communities to achieve balance of usefulness, feasibility and size

• Specific Tasks:- Solicit input on user needs and existing practices

Page 13: Lscom

Page MITREPNN

L

Workshop Goals• Organize series of workshops that bring together three critical

communities – Users, Library Scientists and Knowledge Experts, and Technical Researchers – to create a ontology on order of 1000 concepts for analysis of video broadcast news

• Ensure impact through focused collaboration of these different communities to achieve balance of usefulness, feasibility and size

• Specific Tasks:- Solicit input on user needs and existing practices- Analyze applications, prior work, concept modeling requirements

Page 14: Lscom

Page MITREPNN

L

Workshop Goals• Organize series of workshops that bring together three critical

communities – Users, Library Scientists and Knowledge Experts, and Technical Researchers – to create a ontology on order of 1000 concepts for analysis of video broadcast news

• Ensure impact through focused collaboration of these different communities to achieve balance of usefulness, feasibility and size

• Specific Tasks:- Solicit input on user needs and existing practices- Analyze applications, prior work, concept modeling requirements- Develop draft concept ontology for video broadcast news domain

Page 15: Lscom

Page MITREPNN

L

Workshop Goals• Organize series of workshops that bring together three critical

communities – Users, Library Scientists and Knowledge Experts, and Technical Researchers – to create a ontology on order of 1000 concepts for analysis of video broadcast news

• Ensure impact through focused collaboration of these different communities to achieve balance of usefulness, feasibility and size

• Specific Tasks:- Solicit input on user needs and existing practices- Analyze applications, prior work, concept modeling requirements- Develop draft concept ontology for video broadcast news domain- Solicit input on technical capabilities

Page 16: Lscom

Page MITREPNN

L

Workshop Goals• Organize series of workshops that bring together three critical

communities – Users, Library Scientists and Knowledge Experts, and Technical Researchers – to create a ontology on order of 1000 concepts for analysis of video broadcast news

• Ensure impact through focused collaboration of these different communities to achieve balance of usefulness, feasibility and size

• Specific Tasks:- Solicit input on user needs and existing practices- Analyze applications, prior work, concept modeling requirements- Develop draft concept ontology for video broadcast news domain- Solicit input on technical capabilities- Analyze technical capabilities for concept modeling and detection

Page 17: Lscom

Page MITREPNN

L

Workshop Goals• Organize series of workshops that bring together three critical

communities – Users, Library Scientists and Knowledge Experts, and Technical Researchers – to create a ontology on order of 1000 concepts for analysis of video broadcast news

• Ensure impact through focused collaboration of these different communities to achieve balance of usefulness, feasibility and size

• Specific Tasks:- Solicit input on user needs and existing practices- Analyze applications, prior work, concept modeling requirements- Develop draft concept ontology for video broadcast news domain- Solicit input on technical capabilities- Analyze technical capabilities for concept modeling and detection- Form benchmark and define annotation tasks

Page 18: Lscom

Page MITREPNN

L

Workshop Goals• Organize series of workshops that bring together three critical

communities – Users, Library Scientists and Knowledge Experts, and Technical Researchers – to create a ontology on order of 1000 concepts for analysis of video broadcast news

• Ensure impact through focused collaboration of these different communities to achieve balance of usefulness, feasibility and size

• Specific Tasks:- Solicit input on user needs and existing practices- Analyze applications, prior work, concept modeling requirements- Develop draft concept ontology for video broadcast news domain- Solicit input on technical capabilities- Analyze technical capabilities for concept modeling and detection- Form benchmark and define annotation tasks- Annotate benchmark dataset

Page 19: Lscom

Page MITREPNN

L

Workshop Goals• Organize series of workshops that bring together three critical

communities – Users, Library Scientists and Knowledge Experts, and Technical Researchers – to create a ontology on order of 1000 concepts for analysis of video broadcast news

• Ensure impact through focused collaboration of these different communities to achieve balance of usefulness, feasibility and size

• Specific Tasks:- Solicit input on user needs and existing practices- Analyze applications, prior work, concept modeling requirements- Develop draft concept ontology for video broadcast news domain- Solicit input on technical capabilities- Analyze technical capabilities for concept modeling and detection- Form benchmark and define annotation tasks- Annotate benchmark dataset- Perform benchmark concept modeling, detection and evaluation

Page 20: Lscom

Page MITREPNN

L

Workshop Goals• Organize series of workshops that bring together three critical

communities – Users, Library Scientists and Knowledge Experts, and Technical Researchers – to create a ontology on order of 1000 concepts for analysis of video broadcast news

• Ensure impact through focused collaboration of these different communities to achieve balance of usefulness, feasibility and size

• Specific Tasks:- Solicit input on user needs and existing practices- Analyze applications, prior work, concept modeling requirements- Develop draft concept ontology for video broadcast news domain- Solicit input on technical capabilities- Analyze technical capabilities for concept modeling and detection- Form benchmark and define annotation tasks- Annotate benchmark dataset- Perform benchmark concept modeling, detection and evaluation- Analyze concept detection performance and revise concept ontology

Page 21: Lscom

Page MITREPNN

L

Workshop Goals• Organize series of workshops that bring together three critical

communities – Users, Library Scientists and Knowledge Experts, and Technical Researchers – to create a ontology on order of 1000 concepts for analysis of video broadcast news

• Ensure impact through focused collaboration of these different communities to achieve balance of usefulness, feasibility and size

• Specific Tasks:- Solicit input on user needs and existing practices- Analyze applications, prior work, concept modeling requirements- Develop draft concept ontology for video broadcast news domain- Solicit input on technical capabilities- Analyze technical capabilities for concept modeling and detection- Form benchmark and define annotation tasks- Annotate benchmark dataset- Perform benchmark concept modeling, detection and evaluation- Analyze concept detection performance and revise concept ontology- Conduct gap analysis and identify outstanding research challenges

Page 22: Lscom

Page MITREPNN

L

Workshop Format and Duration

Page 23: Lscom

Page MITREPNN

L

Workshop Format and Duration• Propose to hold two multi-week workshops accompanied by

annotation, experimentation, and prototyping tasks

Page 24: Lscom

Page MITREPNN

L

Workshop Format and Duration• Propose to hold two multi-week workshops accompanied by

annotation, experimentation, and prototyping tasks• Focus on video broadcast news domain

Page 25: Lscom

Page MITREPNN

L

Workshop Format and Duration• Propose to hold two multi-week workshops accompanied by

annotation, experimentation, and prototyping tasks• Focus on video broadcast news domain• Workshop Organization:

Page 26: Lscom

Page MITREPNN

L

Workshop Format and Duration• Propose to hold two multi-week workshops accompanied by

annotation, experimentation, and prototyping tasks• Focus on video broadcast news domain• Workshop Organization:

- Pre-workshop 1: Call for Input on User Needs and Existing Practices

Page 27: Lscom

Page MITREPNN

L

Workshop Format and Duration• Propose to hold two multi-week workshops accompanied by

annotation, experimentation, and prototyping tasks• Focus on video broadcast news domain• Workshop Organization:

- Pre-workshop 1: Call for Input on User Needs and Existing Practices- Ontology Definition Workshop (two-weeks):

Page 28: Lscom

Page MITREPNN

L

Workshop Format and Duration• Propose to hold two multi-week workshops accompanied by

annotation, experimentation, and prototyping tasks• Focus on video broadcast news domain• Workshop Organization:

- Pre-workshop 1: Call for Input on User Needs and Existing Practices- Ontology Definition Workshop (two-weeks):

• Part 1: User Needs

Page 29: Lscom

Page MITREPNN

L

Workshop Format and Duration• Propose to hold two multi-week workshops accompanied by

annotation, experimentation, and prototyping tasks• Focus on video broadcast news domain• Workshop Organization:

- Pre-workshop 1: Call for Input on User Needs and Existing Practices- Ontology Definition Workshop (two-weeks):

• Part 1: User Needs• Part 2: Technical Analysis

Page 30: Lscom

Page MITREPNN

L

Workshop Format and Duration• Propose to hold two multi-week workshops accompanied by

annotation, experimentation, and prototyping tasks• Focus on video broadcast news domain• Workshop Organization:

- Pre-workshop 1: Call for Input on User Needs and Existing Practices- Ontology Definition Workshop (two-weeks):

• Part 1: User Needs• Part 2: Technical Analysis

- Ad hoc Tasks

Page 31: Lscom

Page MITREPNN

L

Workshop Format and Duration• Propose to hold two multi-week workshops accompanied by

annotation, experimentation, and prototyping tasks• Focus on video broadcast news domain• Workshop Organization:

- Pre-workshop 1: Call for Input on User Needs and Existing Practices- Ontology Definition Workshop (two-weeks):

• Part 1: User Needs• Part 2: Technical Analysis

- Ad hoc Tasks• Task 1: Annotation

Page 32: Lscom

Page MITREPNN

L

Workshop Format and Duration• Propose to hold two multi-week workshops accompanied by

annotation, experimentation, and prototyping tasks• Focus on video broadcast news domain• Workshop Organization:

- Pre-workshop 1: Call for Input on User Needs and Existing Practices- Ontology Definition Workshop (two-weeks):

• Part 1: User Needs• Part 2: Technical Analysis

- Ad hoc Tasks• Task 1: Annotation• Task 2: Experimentation

Page 33: Lscom

Page MITREPNN

L

Workshop Format and Duration• Propose to hold two multi-week workshops accompanied by

annotation, experimentation, and prototyping tasks• Focus on video broadcast news domain• Workshop Organization:

- Pre-workshop 1: Call for Input on User Needs and Existing Practices- Ontology Definition Workshop (two-weeks):

• Part 1: User Needs• Part 2: Technical Analysis

- Ad hoc Tasks• Task 1: Annotation• Task 2: Experimentation• Task 3: Evaluation

Page 34: Lscom

Page MITREPNN

L

Workshop Format and Duration• Propose to hold two multi-week workshops accompanied by

annotation, experimentation, and prototyping tasks• Focus on video broadcast news domain• Workshop Organization:

- Pre-workshop 1: Call for Input on User Needs and Existing Practices- Ontology Definition Workshop (two-weeks):

• Part 1: User Needs• Part 2: Technical Analysis

- Ad hoc Tasks• Task 1: Annotation• Task 2: Experimentation• Task 3: Evaluation

- Ontology Evaluation Workshop (two-weeks):

Page 35: Lscom

Page MITREPNN

L

Workshop Format and Duration• Propose to hold two multi-week workshops accompanied by

annotation, experimentation, and prototyping tasks• Focus on video broadcast news domain• Workshop Organization:

- Pre-workshop 1: Call for Input on User Needs and Existing Practices- Ontology Definition Workshop (two-weeks):

• Part 1: User Needs• Part 2: Technical Analysis

- Ad hoc Tasks• Task 1: Annotation• Task 2: Experimentation• Task 3: Evaluation

- Ontology Evaluation Workshop (two-weeks): • Part 1: Validation and Refinement

Page 36: Lscom

Page MITREPNN

L

Workshop Format and Duration• Propose to hold two multi-week workshops accompanied by

annotation, experimentation, and prototyping tasks• Focus on video broadcast news domain• Workshop Organization:

- Pre-workshop 1: Call for Input on User Needs and Existing Practices- Ontology Definition Workshop (two-weeks):

• Part 1: User Needs• Part 2: Technical Analysis

- Ad hoc Tasks• Task 1: Annotation• Task 2: Experimentation• Task 3: Evaluation

- Ontology Evaluation Workshop (two-weeks): • Part 1: Validation and Refinement• Part 2: Outstanding Challenges and Recommendations

Page 37: Lscom

Page MITREPNN

L

Workshop Format and Duration• Propose to hold two multi-week workshops accompanied by

annotation, experimentation, and prototyping tasks• Focus on video broadcast news domain• Workshop Organization:

- Pre-workshop 1: Call for Input on User Needs and Existing Practices- Ontology Definition Workshop (two-weeks):

• Part 1: User Needs• Part 2: Technical Analysis

- Ad hoc Tasks• Task 1: Annotation• Task 2: Experimentation• Task 3: Evaluation

- Ontology Evaluation Workshop (two-weeks): • Part 1: Validation and Refinement• Part 2: Outstanding Challenges and Recommendations

• Substantial off-line tasks for annotation and experimentation require

Page 38: Lscom

Page MITREPNN

L

Broadcast News Video Content Description Ontology• Why the Focus on Broadcast News Domain?

- Critical mass of users, content providers, applications- Good content availability (TRECVID, LDC, FBIS)- Shares large set of core concepts with other domains

• Ontology Formalism:- Entity-Relationship (E-R) Graphs- RDF, DAML / DAML+OIL, W3C OWL- MPEG-7, MediaNet, VEML

• Seed Representations:- TRECVID-2003 News Lexicon (Annotation Forum)- Library of Congress TGM-I- CNN, BBC Classification Systems

Broadcast News Ontology

NewsProductionGrammars

CoreVideo

BroadcastNews

ContentNews

Domain

MPEG-7 Video Annotation Tool

Page 39: Lscom

Page MITREPNN

L

Broadcast News Video Content Description Ontology• Why the Focus on Broadcast News Domain?

- Critical mass of users, content providers, applications- Good content availability (TRECVID, LDC, FBIS)- Shares large set of core concepts with other domains

• Ontology Formalism:- Entity-Relationship (E-R) Graphs- RDF, DAML / DAML+OIL, W3C OWL- MPEG-7, MediaNet, VEML

• Seed Representations:- TRECVID-2003 News Lexicon (Annotation Forum)- Library of Congress TGM-I- CNN, BBC Classification Systems

Broadcast News Ontology

NewsProductionGrammars

CoreVideo

BroadcastNews

ContentNews

Domain

Objects

Actions

Sites

Concepts

Outdoors IndoorsPerson

PeopleFace

NewsSubjectAnchor

Crowd

NewsMonolog

NewsDialog Studio

MPEG-7 Video Annotation Tool

Page 40: Lscom

Page MITREPNN

L

Broadcast News Video Content Description Ontology• Why the Focus on Broadcast News Domain?

- Critical mass of users, content providers, applications- Good content availability (TRECVID, LDC, FBIS)- Shares large set of core concepts with other domains

• Ontology Formalism:- Entity-Relationship (E-R) Graphs- RDF, DAML / DAML+OIL, W3C OWL- MPEG-7, MediaNet, VEML

• Seed Representations:- TRECVID-2003 News Lexicon (Annotation Forum)- Library of Congress TGM-I- CNN, BBC Classification Systems

Broadcast News Ontology

NewsProductionGrammars

CoreVideo

BroadcastNews

ContentNews

Domain

Objects

Actions

Sites

Concepts

Outdoors IndoorsPerson

PeopleFace

NewsSubjectAnchor

Crowd

NewsMonolog

NewsDialog Studio

MPEG-7 Video Annotation Tool

Page 41: Lscom

Page MITREPNN

L

Approach (Pre-workshop and 1st workshop)

Page 42: Lscom

Page MITREPNN

L

Approach (Pre-workshop and 1st workshop)• Pre-workshop: Call for Input

Page 43: Lscom

Page MITREPNN

L

Approach (Pre-workshop and 1st workshop)• Pre-workshop: Call for Input

- Solicit input on user needs and existing practices

Page 44: Lscom

Page MITREPNN

L

Approach (Pre-workshop and 1st workshop)• Pre-workshop: Call for Input

- Solicit input on user needs and existing practices• Ontology Definition Workshop

Page 45: Lscom

Page MITREPNN

L

Approach (Pre-workshop and 1st workshop)• Pre-workshop: Call for Input

- Solicit input on user needs and existing practices• Ontology Definition Workshop

Part 1: User Needs

Page 46: Lscom

Page MITREPNN

L

Approach (Pre-workshop and 1st workshop)• Pre-workshop: Call for Input

- Solicit input on user needs and existing practices• Ontology Definition Workshop

Part 1: User Needs• Analyze use cases, concept modeling requirements, prior lexicon and ontology work

Page 47: Lscom

Page MITREPNN

L

Approach (Pre-workshop and 1st workshop)• Pre-workshop: Call for Input

- Solicit input on user needs and existing practices• Ontology Definition Workshop

Part 1: User Needs• Analyze use cases, concept modeling requirements, prior lexicon and ontology work• Develop draft concept ontology for video broadcast news domain

Page 48: Lscom

Page MITREPNN

L

Approach (Pre-workshop and 1st workshop)• Pre-workshop: Call for Input

- Solicit input on user needs and existing practices• Ontology Definition Workshop

Part 1: User Needs• Analyze use cases, concept modeling requirements, prior lexicon and ontology work• Develop draft concept ontology for video broadcast news domain

Output: Version 1

Page 49: Lscom

Page MITREPNN

L

Approach (Pre-workshop and 1st workshop)• Pre-workshop: Call for Input

- Solicit input on user needs and existing practices• Ontology Definition Workshop

Part 1: User Needs• Analyze use cases, concept modeling requirements, prior lexicon and ontology work• Develop draft concept ontology for video broadcast news domain

Output: Version 1• Requirements and Existing Practices

Page 50: Lscom

Page MITREPNN

L

Approach (Pre-workshop and 1st workshop)• Pre-workshop: Call for Input

- Solicit input on user needs and existing practices• Ontology Definition Workshop

Part 1: User Needs• Analyze use cases, concept modeling requirements, prior lexicon and ontology work• Develop draft concept ontology for video broadcast news domain

Output: Version 1• Requirements and Existing Practices• Domain Concepts and Ontology System

Page 51: Lscom

Page MITREPNN

L

Approach (Pre-workshop and 1st workshop)• Pre-workshop: Call for Input

- Solicit input on user needs and existing practices• Ontology Definition Workshop

Part 1: User Needs• Analyze use cases, concept modeling requirements, prior lexicon and ontology work• Develop draft concept ontology for video broadcast news domain

Output: Version 1• Requirements and Existing Practices• Domain Concepts and Ontology System• Video Concept Ontology

Page 52: Lscom

Page MITREPNN

L

Approach (Pre-workshop and 1st workshop)• Pre-workshop: Call for Input

- Solicit input on user needs and existing practices• Ontology Definition Workshop

Part 1: User Needs• Analyze use cases, concept modeling requirements, prior lexicon and ontology work• Develop draft concept ontology for video broadcast news domain

Output: Version 1• Requirements and Existing Practices• Domain Concepts and Ontology System• Video Concept Ontology

Part 2: Technical Analysis

Page 53: Lscom

Page MITREPNN

L

Approach (Pre-workshop and 1st workshop)• Pre-workshop: Call for Input

- Solicit input on user needs and existing practices• Ontology Definition Workshop

Part 1: User Needs• Analyze use cases, concept modeling requirements, prior lexicon and ontology work• Develop draft concept ontology for video broadcast news domain

Output: Version 1• Requirements and Existing Practices• Domain Concepts and Ontology System• Video Concept Ontology

Part 2: Technical Analysis• Analyze technical capabilities for concept modeling and detection

Page 54: Lscom

Page MITREPNN

L

Approach (Pre-workshop and 1st workshop)• Pre-workshop: Call for Input

- Solicit input on user needs and existing practices• Ontology Definition Workshop

Part 1: User Needs• Analyze use cases, concept modeling requirements, prior lexicon and ontology work• Develop draft concept ontology for video broadcast news domain

Output: Version 1• Requirements and Existing Practices• Domain Concepts and Ontology System• Video Concept Ontology

Part 2: Technical Analysis• Analyze technical capabilities for concept modeling and detection• Form benchmark and define annotation tasks

Page 55: Lscom

Page MITREPNN

L

Approach (Pre-workshop and 1st workshop)• Pre-workshop: Call for Input

- Solicit input on user needs and existing practices• Ontology Definition Workshop

Part 1: User Needs• Analyze use cases, concept modeling requirements, prior lexicon and ontology work• Develop draft concept ontology for video broadcast news domain

Output: Version 1• Requirements and Existing Practices• Domain Concepts and Ontology System• Video Concept Ontology

Part 2: Technical Analysis• Analyze technical capabilities for concept modeling and detection• Form benchmark and define annotation tasks

Output: Version 1

Page 56: Lscom

Page MITREPNN

L

Approach (Pre-workshop and 1st workshop)• Pre-workshop: Call for Input

- Solicit input on user needs and existing practices• Ontology Definition Workshop

Part 1: User Needs• Analyze use cases, concept modeling requirements, prior lexicon and ontology work• Develop draft concept ontology for video broadcast news domain

Output: Version 1• Requirements and Existing Practices• Domain Concepts and Ontology System• Video Concept Ontology

Part 2: Technical Analysis• Analyze technical capabilities for concept modeling and detection• Form benchmark and define annotation tasks

Output: Version 1• Benchmark (Use cases, Annotation)

Page 57: Lscom

Page MITREPNN

L

Approach (Ad-hoc Tasks and 2nd workshop)

Page 58: Lscom

Page MITREPNN

L

Approach (Ad-hoc Tasks and 2nd workshop)• Ad hoc Group

Page 59: Lscom

Page MITREPNN

L

Approach (Ad-hoc Tasks and 2nd workshop)• Ad hoc Group

Task 1: Annotation

Page 60: Lscom

Page MITREPNN

L

Approach (Ad-hoc Tasks and 2nd workshop)• Ad hoc Group

Task 1: Annotation- Annotate benchmark dataset

Page 61: Lscom

Page MITREPNN

L

Approach (Ad-hoc Tasks and 2nd workshop)• Ad hoc Group

Task 1: Annotation- Annotate benchmark datasetTask 2: Experimentation

Page 62: Lscom

Page MITREPNN

L

Approach (Ad-hoc Tasks and 2nd workshop)• Ad hoc Group

Task 1: Annotation- Annotate benchmark datasetTask 2: Experimentation- Perform benchmark concept modeling and detection

Page 63: Lscom

Page MITREPNN

L

Approach (Ad-hoc Tasks and 2nd workshop)• Ad hoc Group

Task 1: Annotation- Annotate benchmark datasetTask 2: Experimentation- Perform benchmark concept modeling and detectionTask 3: Evaluation

Page 64: Lscom

Page MITREPNN

L

Approach (Ad-hoc Tasks and 2nd workshop)• Ad hoc Group

Task 1: Annotation- Annotate benchmark datasetTask 2: Experimentation- Perform benchmark concept modeling and detectionTask 3: Evaluation - Evaluation of concept detection, ontology and use of automatic detection

for use cases and evaluation

Page 65: Lscom

Page MITREPNN

L

Approach (Ad-hoc Tasks and 2nd workshop)• Ad hoc Group

Task 1: Annotation- Annotate benchmark datasetTask 2: Experimentation- Perform benchmark concept modeling and detectionTask 3: Evaluation - Evaluation of concept detection, ontology and use of automatic detection

for use cases and evaluationOutput:

Page 66: Lscom

Page MITREPNN

L

Approach (Ad-hoc Tasks and 2nd workshop)• Ad hoc Group

Task 1: Annotation- Annotate benchmark datasetTask 2: Experimentation- Perform benchmark concept modeling and detectionTask 3: Evaluation - Evaluation of concept detection, ontology and use of automatic detection

for use cases and evaluationOutput:- Benchmark v.2

Page 67: Lscom

Page MITREPNN

L

Approach (Ad-hoc Tasks and 2nd workshop)• Ad hoc Group

Task 1: Annotation- Annotate benchmark datasetTask 2: Experimentation- Perform benchmark concept modeling and detectionTask 3: Evaluation - Evaluation of concept detection, ontology and use of automatic detection

for use cases and evaluationOutput:- Benchmark v.2- Concept Detection Evaluation v.1

Page 68: Lscom

Page MITREPNN

L

Approach (Ad-hoc Tasks and 2nd workshop)• Ad hoc Group

Task 1: Annotation- Annotate benchmark datasetTask 2: Experimentation- Perform benchmark concept modeling and detectionTask 3: Evaluation - Evaluation of concept detection, ontology and use of automatic detection

for use cases and evaluationOutput:- Benchmark v.2- Concept Detection Evaluation v.1- Ontology Evaluation v.1

Page 69: Lscom

Page MITREPNN

L

Approach (Ad-hoc Tasks and 2nd workshop)• Ad hoc Group

Task 1: Annotation- Annotate benchmark datasetTask 2: Experimentation- Perform benchmark concept modeling and detectionTask 3: Evaluation - Evaluation of concept detection, ontology and use of automatic detection

for use cases and evaluationOutput:- Benchmark v.2- Concept Detection Evaluation v.1- Ontology Evaluation v.1- Query Answering Effectiveness with Automated Detection Evaluation v.1

Page 70: Lscom

Page MITREPNN

L

Approach (Ad-hoc Tasks and 2nd workshop)• Ad hoc Group

Task 1: Annotation- Annotate benchmark datasetTask 2: Experimentation- Perform benchmark concept modeling and detectionTask 3: Evaluation - Evaluation of concept detection, ontology and use of automatic detection

for use cases and evaluationOutput:- Benchmark v.2- Concept Detection Evaluation v.1- Ontology Evaluation v.1- Query Answering Effectiveness with Automated Detection Evaluation v.1

• Ontology Evaluation Workshop

Page 71: Lscom

Page MITREPNN

L

Approach (Ad-hoc Tasks and 2nd workshop)• Ad hoc Group

Task 1: Annotation- Annotate benchmark datasetTask 2: Experimentation- Perform benchmark concept modeling and detectionTask 3: Evaluation - Evaluation of concept detection, ontology and use of automatic detection

for use cases and evaluationOutput:- Benchmark v.2- Concept Detection Evaluation v.1- Ontology Evaluation v.1- Query Answering Effectiveness with Automated Detection Evaluation v.1

• Ontology Evaluation Workshop Part 1: Validation

Page 72: Lscom

Page MITREPNN

L

Approach (Ad-hoc Tasks and 2nd workshop)• Ad hoc Group

Task 1: Annotation- Annotate benchmark datasetTask 2: Experimentation- Perform benchmark concept modeling and detectionTask 3: Evaluation - Evaluation of concept detection, ontology and use of automatic detection

for use cases and evaluationOutput:- Benchmark v.2- Concept Detection Evaluation v.1- Ontology Evaluation v.1- Query Answering Effectiveness with Automated Detection Evaluation v.1

• Ontology Evaluation Workshop Part 1: Validation- Analyze evaluation of ontology, concept detection and its application to use case

answering.

Page 73: Lscom

Page MITREPNN

L

Approach (Ad-hoc Tasks and 2nd workshop)• Ad hoc Group

Task 1: Annotation- Annotate benchmark datasetTask 2: Experimentation- Perform benchmark concept modeling and detectionTask 3: Evaluation - Evaluation of concept detection, ontology and use of automatic detection

for use cases and evaluationOutput:- Benchmark v.2- Concept Detection Evaluation v.1- Ontology Evaluation v.1- Query Answering Effectiveness with Automated Detection Evaluation v.1

• Ontology Evaluation Workshop Part 1: Validation- Analyze evaluation of ontology, concept detection and its application to use case

answering.Output

Page 74: Lscom

Page MITREPNN

L

Approach (Ad-hoc Tasks and 2nd workshop)• Ad hoc Group

Task 1: Annotation- Annotate benchmark datasetTask 2: Experimentation- Perform benchmark concept modeling and detectionTask 3: Evaluation - Evaluation of concept detection, ontology and use of automatic detection

for use cases and evaluationOutput:- Benchmark v.2- Concept Detection Evaluation v.1- Ontology Evaluation v.1- Query Answering Effectiveness with Automated Detection Evaluation v.1

• Ontology Evaluation Workshop Part 1: Validation- Analyze evaluation of ontology, concept detection and its application to use case

answering.Output- Domain Concepts v.2 and Ontology System v.2

Page 75: Lscom

Page MITREPNN

L

Approach (Ad-hoc Tasks and 2nd workshop)• Ad hoc Group

Task 1: Annotation- Annotate benchmark datasetTask 2: Experimentation- Perform benchmark concept modeling and detectionTask 3: Evaluation - Evaluation of concept detection, ontology and use of automatic detection

for use cases and evaluationOutput:- Benchmark v.2- Concept Detection Evaluation v.1- Ontology Evaluation v.1- Query Answering Effectiveness with Automated Detection Evaluation v.1

• Ontology Evaluation Workshop Part 1: Validation- Analyze evaluation of ontology, concept detection and its application to use case

answering.Output- Domain Concepts v.2 and Ontology System v.2- Video Concept Ontology v.2

Page 76: Lscom

Page MITREPNN

L

Approach (Ad-hoc Tasks and 2nd workshop)• Ad hoc Group

Task 1: Annotation- Annotate benchmark datasetTask 2: Experimentation- Perform benchmark concept modeling and detectionTask 3: Evaluation - Evaluation of concept detection, ontology and use of automatic detection

for use cases and evaluationOutput:- Benchmark v.2- Concept Detection Evaluation v.1- Ontology Evaluation v.1- Query Answering Effectiveness with Automated Detection Evaluation v.1

• Ontology Evaluation Workshop Part 1: Validation- Analyze evaluation of ontology, concept detection and its application to use case

answering.Output- Domain Concepts v.2 and Ontology System v.2- Video Concept Ontology v.2Part 2: Outstanding Challenges

Page 77: Lscom

Page MITREPNN

L

Approach (Ad-hoc Tasks and 2nd workshop)• Ad hoc Group

Task 1: Annotation- Annotate benchmark datasetTask 2: Experimentation- Perform benchmark concept modeling and detectionTask 3: Evaluation - Evaluation of concept detection, ontology and use of automatic detection

for use cases and evaluationOutput:- Benchmark v.2- Concept Detection Evaluation v.1- Ontology Evaluation v.1- Query Answering Effectiveness with Automated Detection Evaluation v.1

• Ontology Evaluation Workshop Part 1: Validation- Analyze evaluation of ontology, concept detection and its application to use case

answering.Output- Domain Concepts v.2 and Ontology System v.2- Video Concept Ontology v.2Part 2: Outstanding Challenges- Conduct gap analysis and identify outstanding research challenges

Page 78: Lscom

Page MITREPNN

L

Approach (Ad-hoc Tasks and 2nd workshop)• Ad hoc Group

Task 1: Annotation- Annotate benchmark datasetTask 2: Experimentation- Perform benchmark concept modeling and detectionTask 3: Evaluation - Evaluation of concept detection, ontology and use of automatic detection

for use cases and evaluationOutput:- Benchmark v.2- Concept Detection Evaluation v.1- Ontology Evaluation v.1- Query Answering Effectiveness with Automated Detection Evaluation v.1

• Ontology Evaluation Workshop Part 1: Validation- Analyze evaluation of ontology, concept detection and its application to use case

answering.Output- Domain Concepts v.2 and Ontology System v.2- Video Concept Ontology v.2Part 2: Outstanding Challenges- Conduct gap analysis and identify outstanding research challengesOutput:

Page 79: Lscom

Page MITREPNN

L

Approach (Ad-hoc Tasks and 2nd workshop)• Ad hoc Group

Task 1: Annotation- Annotate benchmark datasetTask 2: Experimentation- Perform benchmark concept modeling and detectionTask 3: Evaluation - Evaluation of concept detection, ontology and use of automatic detection

for use cases and evaluationOutput:- Benchmark v.2- Concept Detection Evaluation v.1- Ontology Evaluation v.1- Query Answering Effectiveness with Automated Detection Evaluation v.1

• Ontology Evaluation Workshop Part 1: Validation- Analyze evaluation of ontology, concept detection and its application to use case

answering.Output- Domain Concepts v.2 and Ontology System v.2- Video Concept Ontology v.2Part 2: Outstanding Challenges- Conduct gap analysis and identify outstanding research challengesOutput:- Research Challenges v.1

Page 80: Lscom

Page MITREPNN

L

OutputDocuments

Input

Tasks

Page 81: Lscom

Page MITREPNN

L

OutputDocuments

Input

Tasks

Ontologysystems

Catalog terms

Queries

Existing practices

Queries

Use cases User needs

Appl.Analysis

Prior workAnalysis

Ex. Ontology & catalog systems: RDF Topic Maps DAML / DAML+OIL W3C OWL Dublin Core OCLC Open Directory Project MARC MODS MediaNet VEML

Ex. Catalog terms LOC TGM-I MPEG-7 classif. Schemes

Ex. Queries TRECVID BBC

Users: IC analysts Broadcasters

Ex. Usage: Searching Browsing Navigation Threading Summarization

Pre-Workshop: Call for Input

Page 82: Lscom

Page MITREPNN

L

OutputDocuments

Input

Tasks

RequirementsStudy v.1

Existing PracticesStudy v.1

OntologySystem

Study v.1

ConceptAnalysis

DomainConceptsStudy v.1

ModelingAnalysis

OntologyDesign

Video ConceptOntology

v.1

• Identifies indexing requirements for broadcast news video including example queries

• Documents existing practices for indexing broadcast news

• Identifies and defines domain concepts, terms, classification systems for broadcast news video (ex. objects, actions, sites, events)

• Specifies ontology system for broadcast news video domain

• Specifies draft lexicon and ontology for broadcast news video domain

Workshop 1: User Needs

Ontologysystems

Catalog terms

Queries

Existing practices

Queries

Use cases User needs

Appl.Analysis

Prior workAnalysis

Ex. Ontology & catalog systems: RDF Topic Maps DAML / DAML+OIL W3C OWL Dublin Core OCLC Open Directory Project MARC MODS MediaNet VEML

Ex. Catalog terms LOC TGM-I MPEG-7 classif. Schemes

Ex. Queries TRECVID BBC

Users: IC analysts Broadcasters

Ex. Usage: Searching Browsing Navigation Threading Summarization

Pre-Workshop: Call for Input

Page 83: Lscom

Page MITREPNN

L

OutputDocuments

Input

Tasks

Page 84: Lscom

Page MITREPNN

L

OutputDocuments

Input

Tasks

State-of-artTechniques

Systems &Prototypes

Technical capabilities

BenchmarkResults

Systems: Video logging Video retrieval

Ex. State-of-art techniques: Content-based search Segmentation Tracking High-level feature detection Story segmentation

Ex. Benchmarks: TRECVID

BenchmarkFormation• Defines

benchmark for concept detection for video broadcast news (dataset, detection and search tasks, metrics)

Workshop 1: Technical Analysis

TechnicalAnalysis

BenchmarkDraft v.1

Video ConceptOntologyDraft v.1

Page 85: Lscom

Page MITREPNN

L

OutputDocuments

Input

Tasks

• Identifies and maps techniques for modeling and detecting each of draft concepts

• Identifies features required for modeling each of draft concepts

AnnotationTask 1

BenchmarkDraft v.2

• Refines benchmark to include annotated ground-truth for experimentation and evaluation

Ad Hoc Task 1: Annotation

Ad Hoc Task 2,3: ExperimentationExperimentation

ConceptDetection

Evaluation v.1

ConceptModeling

v.1

FeatureExtraction

v. 1

Query Evaluation with Automatic Detection v.1

OntologyEvaluation v.1

State-of-artTechniques

Systems &Prototypes

Technical capabilities

BenchmarkResults

Systems: Video logging Video retrieval

Ex. State-of-art techniques: Content-based search Segmentation Tracking High-level feature detection Story segmentation

Ex. Benchmarks: TRECVID

BenchmarkFormation• Defines

benchmark for concept detection for video broadcast news (dataset, detection and search tasks, metrics)

Workshop 1: Technical Analysis

TechnicalAnalysis

BenchmarkDraft v.1

Video ConceptOntologyDraft v.1

Page 86: Lscom

Page MITREPNN

L

Workshop 2: Evaluation

OutputDocuments

Input

Tasks

Page 87: Lscom

Page MITREPNN

L

Workshop 2: Evaluation

OutputDocuments

Input

Tasks

OntologySystem

Study v.2

DomainConceptsStudy v.2

OntologyRe-design

Video ConceptOntology

v.2

• Revises lexicon based on performance analysis

• Revises ontology system based on performance analysis

• Refines lexicon and ontology for broadcast news video domain

RequirementsStudy v.1

ConceptDetection

Evaluation v.1

Query Evaluation with Automatic Detection v.1

OntologyEvaluation v.1

Analysis

Page 88: Lscom

Page MITREPNN

L

Workshop 2: Evaluation

OutputDocuments

Input

Tasks

GapAnalysis

ResearchChallenge v.1

• Identifies and defines technology gaps and challenges for future research

Workshop 2: Outstanding Challenges

Recommendations v.1

• Recommendations for ontology exploitation and solution design

OntologySystem

Study v.2

DomainConceptsStudy v.2

OntologyRe-design

Video ConceptOntology

v.2

• Revises lexicon based on performance analysis

• Revises ontology system based on performance analysis

• Refines lexicon and ontology for broadcast news video domain

RequirementsStudy v.1

ConceptDetection

Evaluation v.1

Query Evaluation with Automatic Detection v.1

OntologyEvaluation v.1

Analysis

Page 89: Lscom

Page MITREPNN

L

Domain and Data Sets• Candidate data set:

- TRECVID Corpus (>200 hours of video broadcast news from CNN and ABC). Has the following advantages• availability• generalization capability better with than other domains• # of research groups up to speed on this domain for tools/detectors• TREC established some benchmark and evaluation metrics already.

- Will avoid letting domain specifics influence the design of ontology to an extent where the ontology starts catering to artifacts of the BN domain.

- Will seek other sources such as FBIS, WNC etc.

• Annotation issues:- Plan to leverage prior video annotation efforts where possible (e.g.,

TRECVID annotation forum)- Hands-on annotation effort will induce discussions and requires

refinements of concepts meanings

Page 90: Lscom

Page MITREPNN

L

Evaluation Methods• Require benchmarks and metrics for evaluating:

- Utility of ontology – coverage of queries in terms of quality and quantity- Feasibility of ontology:

• Accuracy of concept detection and degree of automation (amount of training)

• Effectiveness of query systems using automatically extracted concepts

• Metrics of Retrieval Effectiveness

- Precision & Recall Curves, Average Precision, Precision at Fixed Depth

• Metrics of Lexicon Effectiveness

- Number of Use Cases that can be answered by lexicon successfully

- Mean average precision across the set of use cases

• Evaluate at multiple levels of granularity:

- Individual concept, classes, hierarchies

Page 91: Lscom

Page MITREPNN

L

Confirmed Participants – Knowledge Experts and Users

Page 92: Lscom

Page MITREPNN

L

Confirmed Participants – Knowledge Experts and UsersLibrary Sciences and Knowledge

representation (definition of lexicon):

Corrine Jorgensen, School of Information Studies, Florida State University

Barbara Tillett, Chief of Cataloging Policy and Support, Library of Congress

Jerry Hobbs, USC / ISI Michael Witbrock, Cycorp Ronald Murray, Preservation

Reformatting Division, Library of Congress

Page 93: Lscom

Page MITREPNN

L

Confirmed Participants – Knowledge Experts and UsersLibrary Sciences and Knowledge

representation (definition of lexicon):

Corrine Jorgensen, School of Information Studies, Florida State University

Barbara Tillett, Chief of Cataloging Policy and Support, Library of Congress

Jerry Hobbs, USC / ISI Michael Witbrock, Cycorp Ronald Murray, Preservation

Reformatting Division, Library of Congress

R&D Agencies John Prange, ARDA Sankar Basu, Div. of Computing and

Comm. Foundations, NSF Maria Zemankova, Div. of Inform. and

Intell. Systems., NSF

Page 94: Lscom

Page MITREPNN

L

Confirmed Participants – Knowledge Experts and UsersStandardization and Benchmarking

(theoretical and empirical evaluation):

Paul Over, NIST John Garofolo, NIST Donna Harman, NIST David Day, MITRE John R. Smith, IBM Research

Library Sciences and Knowledge representation (definition of lexicon):

Corrine Jorgensen, School of Information Studies, Florida State University

Barbara Tillett, Chief of Cataloging Policy and Support, Library of Congress

Jerry Hobbs, USC / ISI Michael Witbrock, Cycorp Ronald Murray, Preservation

Reformatting Division, Library of Congress

R&D Agencies John Prange, ARDA Sankar Basu, Div. of Computing and

Comm. Foundations, NSF Maria Zemankova, Div. of Inform. and

Intell. Systems., NSF

Page 95: Lscom

Page MITREPNN

L

Confirmed Participants – Knowledge Experts and UsersStandardization and Benchmarking

(theoretical and empirical evaluation):

Paul Over, NIST John Garofolo, NIST Donna Harman, NIST David Day, MITRE John R. Smith, IBM Research

Library Sciences and Knowledge representation (definition of lexicon):

Corrine Jorgensen, School of Information Studies, Florida State University

Barbara Tillett, Chief of Cataloging Policy and Support, Library of Congress

Jerry Hobbs, USC / ISI Michael Witbrock, Cycorp Ronald Murray, Preservation

Reformatting Division, Library of Congress User Communities (interpretation of

use cases for lexicon definition, broadcasters help getting query logs for finding useful lexical entries)

Joanne Evans, British Broadcasting Corporation

Chris Porter, Getty Images ARDA and analysts

R&D Agencies John Prange, ARDA Sankar Basu, Div. of Computing and

Comm. Foundations, NSF Maria Zemankova, Div. of Inform. and

Intell. Systems., NSF

Page 96: Lscom

Page MITREPNN

L

Confirmed Participants – Technical Team

Theoretical Analysis: (Help conduct analysis during initial lexicon and ontology design)

Milind R. Naphade, IBM Research

Ramesh Jain, Georgia Institute of Technology

Thomas Huang, UIUC Edward Delp, Purdue

University

Experimentation: (Help address evaluation issues for lexicon, ontology and concept evaluation)

Alexander Hauptmann, CMU

Alan Smeaton, Dublin City University

HongJiang Zhang, Microsoft Research

Ajay Divakaran, MERL Wessel Kraaij,

Information Systems Division, TNO TPD

Ching-Yung Lin, IBM Research

Mubarak Shah, University of Central Florida

Prototyping: (Help with prototyping tools for annotation, evaluation, querying, summarization and statistics gathering)

Shih-Fu Chang, Columbia University

Edward Chang, UCSB

Nevenka Dimitrova, Phillips Research

Rainer Lienhart, Intel Apostol Natsev, IBM

Research Tat-Seng Chua, NUS Ram Nevatia, USC John Kender,

Columbia University

Page 97: Lscom

Page MITREPNN

L

Impact and Outcome• First of a Kind Ontology of 1000 or more semantic concepts that have been

evaluated for their usability and feasibility by different communities including UC, OC, MC.

• Annotated corpus (200 hours) and ontology can be further exploited for future TRECVID, VACE, MPEG-7 activities. Core semantic primitives, that can be included in various video description standards/languages such as MPEG-7.

• Empirical and theoretical study of automatic concept detection performance for elements of this large ontology. Use of current state of the art detection wherever possible. Use of simulation where the detection is not available.

• Use cases (queries) testing and expansion into ontology• Reports documenting use cases, existing practices, research challenges and

recommendations• Prototype systems and tools for annotation, query formulation and evaluation

• Guidelines on manual and automatic multimedia query formulation techniques going from use-cases to concepts.

• Categorization of classes of concepts based on feasibility, detection performance and difficulty in automation

BOTTOMLINE: All this is driven by the user

Page 98: Lscom

Page MITREPNN

L

Summary of Key Questions• How easy was it to create annotations

- (man-hours/hr of video?)• How well does the lexicon 'partition' the collection• Given perfect annotations/classification:

- How well does the lexicon aid with queries/tasks• How good is automatic annotation of the sample collection

- What fraction of perfect annotations accuracy is obtained for the queries/tasks

• How much is automatic classification performance of a given lexical item a function of training data

- Estimate how much training data would get this lexical item to 60%, 80%, 90%, 95%?

• What lexicon changes are necessary or desirable?

Page 99: Lscom

Page MITREPNN

L

Video Event Ontology (VEO) & VEML

• A Video Event Ontology was developed in the ARDA workshop on video event ontologies for surveillance and meetings allows natural, hierarchical representation of complex spatio-temporal events common in the physical world by a composition of simpler (primitive) events

• VEML – XML-derived Video Event Markup Language used to annotate data by instantiating a class defined in that ontology. Example: We will attempt to use or adapt their notation to the extent possible

• (http://www.veml.org:8668//space/2003-10-08/StealingByBlocking.veml)• Broadcast video news ontology is likely to have little overlap with the

complex surveillance events described in the VEO, except for some basic concepts. We expect our ontology to be broader, but much shallower

• Our broadcast news ontology is largely applicable to any edited broadcast video (e.g. documentaries, talk shows, movies) and somewhat applicable to video in general (including surveillance, UAV and home videos).