part 8: relevance feedbackricci/isr/slides-2015/08...part 8: relevance feedback francesco ricci most...

39
Part 8: Relevance Feedback Francesco Ricci Most of these slides comes from the course: Information Retrieval and Web Search, Christopher Manning and Prabhakar Raghavan 1

Upload: others

Post on 26-Jul-2020

1 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Part 8: Relevance Feedbackricci/ISR/slides-2015/08...Part 8: Relevance Feedback Francesco Ricci Most of these slides comes from the course: Information Retrieval and Web Search, Christopher

Part 8: Relevance Feedback

Francesco Ricci

Most of these slides comes from the course:

Information Retrieval and Web Search, Christopher Manning and Prabhakar

Raghavan 1

Page 2: Part 8: Relevance Feedbackricci/ISR/slides-2015/08...Part 8: Relevance Feedback Francesco Ricci Most of these slides comes from the course: Information Retrieval and Web Search, Christopher

Content

p  Local methods n  Relevance feedback n  Pseudo relevance feedback n  Indirect (implicit) relevance feedback

p  Global methods n  Query expansion

p Thesauri p Automatic thesaurus generation p Query log mining

2

Page 3: Part 8: Relevance Feedbackricci/ISR/slides-2015/08...Part 8: Relevance Feedback Francesco Ricci Most of these slides comes from the course: Information Retrieval and Web Search, Christopher

Relevance Feedback

p  Relevance feedback: user feedback on relevance of docs in initial set of results n  User issues a (short, simple) query n  The user marks some results as relevant or

non-relevant n  The system computes a better query

representation of the information need based on feedback

n  Relevance feedback can go through one or more iterations

p  Idea: it may be difficult to formulate a good query when you don’t know the collection well, so iterate.

3

Page 4: Part 8: Relevance Feedbackricci/ISR/slides-2015/08...Part 8: Relevance Feedback Francesco Ricci Most of these slides comes from the course: Information Retrieval and Web Search, Christopher

Example: search images

Feedback

4

Page 5: Part 8: Relevance Feedbackricci/ISR/slides-2015/08...Part 8: Relevance Feedback Francesco Ricci Most of these slides comes from the course: Information Retrieval and Web Search, Christopher

New Interface

5

Page 6: Part 8: Relevance Feedbackricci/ISR/slides-2015/08...Part 8: Relevance Feedback Francesco Ricci Most of these slides comes from the course: Information Retrieval and Web Search, Christopher

Preference-based: DieToRecs

[Ricci et al., 2006] 6

Page 7: Part 8: Relevance Feedbackricci/ISR/slides-2015/08...Part 8: Relevance Feedback Francesco Ricci Most of these slides comes from the course: Information Retrieval and Web Search, Christopher

7

Exploratory Search: Example

liveplasma.com

Page 8: Part 8: Relevance Feedbackricci/ISR/slides-2015/08...Part 8: Relevance Feedback Francesco Ricci Most of these slides comes from the course: Information Retrieval and Web Search, Christopher

8

Critiquing

Page 9: Part 8: Relevance Feedbackricci/ISR/slides-2015/08...Part 8: Relevance Feedback Francesco Ricci Most of these slides comes from the course: Information Retrieval and Web Search, Christopher

9

Critiquing Interaction

[Pu et al., 2006]

Page 10: Part 8: Relevance Feedbackricci/ISR/slides-2015/08...Part 8: Relevance Feedback Francesco Ricci Most of these slides comes from the course: Information Retrieval and Web Search, Christopher

Key concept: Centroid

p  The centroid is the center of mass of a set of points

p  Recall that we represent documents as points in a high-dimensional space

p  Definition: Centroid

where C is a set of documents.

µ (C) =

1|C |

d

d∈C∑

10

How would you call it if C is a set of numbers?

Page 11: Part 8: Relevance Feedbackricci/ISR/slides-2015/08...Part 8: Relevance Feedback Francesco Ricci Most of these slides comes from the course: Information Retrieval and Web Search, Christopher

Example

p  The centroid is not normalized

0.2 0.4 0.6 0.8 1

1

0.8

0.6

0.4

0.2

term 1

term 2

doc 1

doc 2

doc 3

centroid

11

Page 12: Part 8: Relevance Feedbackricci/ISR/slides-2015/08...Part 8: Relevance Feedback Francesco Ricci Most of these slides comes from the course: Information Retrieval and Web Search, Christopher

The Theoretically Best Query

xx

xx

o o

o

Optimal query

x non-relevant documents o relevant documents

o

o

o

x x

xxx

x

x

x

x

x

x

x

x

12

Page 13: Part 8: Relevance Feedbackricci/ISR/slides-2015/08...Part 8: Relevance Feedback Francesco Ricci Most of these slides comes from the course: Information Retrieval and Web Search, Christopher

Rocchio Algorithm

p  The Rocchio algorithm uses the vector space model to pick a relevance feedback query

p  Rocchio seeks the query qopt that maximizes

p  Tries to separate docs marked relevant Cr and non-relevant Cnr– the solution is:

p  Problem: we don’t know the truly relevant docs.

qopt = qargmax[cos(

q, µ(Cr ))− cos(q, µ(Cnr ))]

q opt =1

Cr

d j

d j ∈Cr

∑ −1

Cnr

d j

d j ∉Cr

13

Page 14: Part 8: Relevance Feedbackricci/ISR/slides-2015/08...Part 8: Relevance Feedback Francesco Ricci Most of these slides comes from the course: Information Retrieval and Web Search, Christopher

Rocchio 1971 Algorithm (SMART)

p  Used in practice:

p  Dr = set of known relevant doc vectors p  Dnr = set of known irrelevant doc vectors

n  These are different from Cr and Cnr ! p  qm = modified query vector; q0 = original query

vector; α,β,γ: weights (hand-chosen or set empirically)

p  New query moves toward relevant documents and away from irrelevant documents.

q m = α q 0 + β

1Dr

d j

d j ∈Dr

∑ −γ1

Dnr

d j

d j ∈Dnr

14

Page 15: Part 8: Relevance Feedbackricci/ISR/slides-2015/08...Part 8: Relevance Feedback Francesco Ricci Most of these slides comes from the course: Information Retrieval and Web Search, Christopher

Relevance feedback on initial query

xx

xx

o o

o

Revised query

x known non-relevant documents o known relevant documents

o

o

o x

x

x x

xx

x

x

xx

x

x

Δ x

x

Initial query

Δ

15

Page 16: Part 8: Relevance Feedbackricci/ISR/slides-2015/08...Part 8: Relevance Feedback Francesco Ricci Most of these slides comes from the course: Information Retrieval and Web Search, Christopher

Subtleties to note

p  Tradeoff α vs. β and γ: If we have a lot of judged documents, we want a higher β and γ

p  Some weights in query vector can go negative: n  Negative term weights are ignored (set to 0)

p  Positive feedback is more valuable than negative feedback (so, set γ < β; e.g. γ = 0.25, β = 0.75) - many systems only allow positive feedback (γ=0)

p  Relevance feedback can improve recall and precision p  Relevance feedback is most useful for increasing recall

in situations where recall is important – why? n  Users can be expected to review results and to

take time to iterate – when recall is important. 16

Page 17: Part 8: Relevance Feedbackricci/ISR/slides-2015/08...Part 8: Relevance Feedback Francesco Ricci Most of these slides comes from the course: Information Retrieval and Web Search, Christopher

Relevance Feedback: Assumptions

p  A1: User has sufficient knowledge for initial query p  A2: Relevance prototypes are “well-behaved”

n  Term distribution in relevant documents will be similar

n  Term distribution in non-relevant documents will be different from those in relevant documents p Either: all relevant documents are tightly

clustered around a single prototype p Or: there are different prototypes, but they

have significant vocabulary overlap p Similarities between relevant and irrelevant

documents are small.

17

Page 18: Part 8: Relevance Feedbackricci/ISR/slides-2015/08...Part 8: Relevance Feedback Francesco Ricci Most of these slides comes from the course: Information Retrieval and Web Search, Christopher

Relevance Feedback: Problems

p  Long queries are inefficient for typical IR engine n  Long response times for user n  High cost for retrieval system n  Partial solution:

p Only reweight certain prominent terms - perhaps top 20 by term frequency

p  Users are often reluctant to provide explicit feedback

p  It’s often harder to understand why a particular document was retrieved after applying relevance feedback

p  Information needs may change during the interaction (so what?).

20

Why are we concerned about that?

Page 19: Part 8: Relevance Feedbackricci/ISR/slides-2015/08...Part 8: Relevance Feedback Francesco Ricci Most of these slides comes from the course: Information Retrieval and Web Search, Christopher

Evaluation of relevance feedback strategies p  Use q0 and compute precision and recall graph p  Use qm and compute precision recall graph

n  1) Assess on all documents in the collection p  Spectacular improvements, but … it’s cheating! p  Known relevant documents ranked higher p  Must evaluate with respect to documents not seen by

user n  2) Use documents in residual collection (all docs minus

those assessed relevant) p  Measures usually then lower than for original query p  But a more realistic evaluation p  Relative performance can be validly compared

p  Empirically, one round of relevance feedback is often very useful - two rounds is sometimes marginally useful. 21

Page 20: Part 8: Relevance Feedbackricci/ISR/slides-2015/08...Part 8: Relevance Feedback Francesco Ricci Most of these slides comes from the course: Information Retrieval and Web Search, Christopher

Evaluation of relevance feedback

p  Second method – assess only the docs not rated by the user in the first round n  Could make relevance feedback look worse

than it really is n  Can still assess relative performance of

algorithms p  Most satisfactory – use two collections each with

their own relevance assessments (i.e., split randomly the collection in two parts) n  q0 and user feedback from first collection n  qm run on second collection and measured.

22

Page 21: Part 8: Relevance Feedbackricci/ISR/slides-2015/08...Part 8: Relevance Feedback Francesco Ricci Most of these slides comes from the course: Information Retrieval and Web Search, Christopher

Evaluation: Caveat

p  True evaluation of usefulness must compare to other methods taking the same amount of time – or using similar user effort

p  Alternative to relevance feedback: user revises and resubmits query n  See next topic: query expansion

p  Users may prefer revision/resubmission to having to judge relevance of documents

p  There is no clear evidence that relevance feedback is the “best use” of the user’s time – it may be in some "situations".

23

Page 22: Part 8: Relevance Feedbackricci/ISR/slides-2015/08...Part 8: Relevance Feedback Francesco Ricci Most of these slides comes from the course: Information Retrieval and Web Search, Christopher

Managing implicit feedback in information search p  What can I derive from the fact that a user

clicked on the 2nd link?

Page 23: Part 8: Relevance Feedbackricci/ISR/slides-2015/08...Part 8: Relevance Feedback Francesco Ricci Most of these slides comes from the course: Information Retrieval and Web Search, Christopher

Learning to rank

p  If all the users click on the third link then Yahoo should provide a different ranking for that query.

Page 24: Part 8: Relevance Feedbackricci/ISR/slides-2015/08...Part 8: Relevance Feedback Francesco Ricci Most of these slides comes from the course: Information Retrieval and Web Search, Christopher

Adapting the Search Engine

p  SE could be adapted to a specific user: using just his own clicks (personalization)

p  SE could be adapted to a community: e.g., the students attending this course

p  SE could be adapted to a specific documents collection

Page 25: Part 8: Relevance Feedbackricci/ISR/slides-2015/08...Part 8: Relevance Feedback Francesco Ricci Most of these slides comes from the course: Information Retrieval and Web Search, Christopher

Search Engines Bias Users I

p  Percentage of queries where a user viewed the search result presented at a particular rank (measured with eye tracking).

Page 26: Part 8: Relevance Feedbackricci/ISR/slides-2015/08...Part 8: Relevance Feedback Francesco Ricci Most of these slides comes from the course: Information Retrieval and Web Search, Christopher

Search Engines Bias Users II

p  Blue is the normal rank

p  Red is obtained by swapping the top two results.

Percentage of queries where a user clicked the result presented at a given rank.

Page 27: Part 8: Relevance Feedbackricci/ISR/slides-2015/08...Part 8: Relevance Feedback Francesco Ricci Most of these slides comes from the course: Information Retrieval and Web Search, Christopher

Query Expansion

p  In relevance feedback, users give additional input (relevant/non-relevant) on documents, which is used to reweight terms in the query for documents

p  In query expansion, users give additional input (good/bad search term) on words or phrases n  Generally it is simpler than relevance

feedback.

33

Page 28: Part 8: Relevance Feedbackricci/ISR/slides-2015/08...Part 8: Relevance Feedback Francesco Ricci Most of these slides comes from the course: Information Retrieval and Web Search, Christopher

Example: search images

Feedback

34

Page 29: Part 8: Relevance Feedbackricci/ISR/slides-2015/08...Part 8: Relevance Feedback Francesco Ricci Most of these slides comes from the course: Information Retrieval and Web Search, Christopher

Query assist

Would you expect such a feature to increase the query volume at a search engine?

35

Page 30: Part 8: Relevance Feedbackricci/ISR/slides-2015/08...Part 8: Relevance Feedback Francesco Ricci Most of these slides comes from the course: Information Retrieval and Web Search, Christopher

How do we augment the user query?

p  Manual thesaurus n  E.g. MedLine: physician, syn: doc, doctor, MD,

medico n  Can be related queries rather than just synonyms

p  Global Analysis: static; based on all documents in collection n  Automatically derived thesaurus

p co-occurrence statistics n  Refinements based on query log mining

p Common on the web p  Local Analysis: dynamic

n  Analysis of documents in result set 36

Page 31: Part 8: Relevance Feedbackricci/ISR/slides-2015/08...Part 8: Relevance Feedback Francesco Ricci Most of these slides comes from the course: Information Retrieval and Web Search, Christopher

Example of manual thesaurus

37

Page 32: Part 8: Relevance Feedbackricci/ISR/slides-2015/08...Part 8: Relevance Feedback Francesco Ricci Most of these slides comes from the course: Information Retrieval and Web Search, Christopher

Thesaurus-based query expansion

p  For each term, t, in a query, expand the query with synonyms and related words of t from the thesaurus n  feline → feline cat

p  May weight added terms less than original query terms

p  Generally increases recall p  Widely used in many science/engineering fields p  May significantly decrease precision, particularly with

ambiguous terms n  “interest rate” → “interest rate benefit evaluate”

p  There is a high cost of manually producing a thesaurus n  And for updating it for scientific changes. 38

Page 33: Part 8: Relevance Feedbackricci/ISR/slides-2015/08...Part 8: Relevance Feedback Francesco Ricci Most of these slides comes from the course: Information Retrieval and Web Search, Christopher

WordNet p  WordNet® is a large lexical database of English

(there are also other languages) p  Nouns, verbs, adjectives and adverbs are

grouped into sets of cognitive synonyms - Synsets - each expressing a distinct concept

p  Synsets are interlinked by means of conceptual-semantic and lexical relations.

39

Page 34: Part 8: Relevance Feedbackricci/ISR/slides-2015/08...Part 8: Relevance Feedback Francesco Ricci Most of these slides comes from the course: Information Retrieval and Web Search, Christopher

Automatic Thesaurus Generation

p  Attempt to generate a thesaurus automatically by analyzing the collection of documents

p  Fundamental notion: similarity between two words (can we use Jaccard on word bigrams or Levenshtein? )

p  Definition 1: Two words are similar if they often co-occur

p  Definition 2: Two words are similar if they occur in a given grammatical relation with the same words n  You can harvest, peel, eat, prepare, etc. apples

and pears, so apples and pears must be similar p  Co-occurrence based is more robust, grammatical

relations are more accurate. Why? 40

Page 35: Part 8: Relevance Feedbackricci/ISR/slides-2015/08...Part 8: Relevance Feedback Francesco Ricci Most of these slides comes from the course: Information Retrieval and Web Search, Christopher

Co-occurrence Thesaurus

p  Simplest way to compute a thesaurus is to observe the term-term similarities in C = AAT : n  A is term-document matrix

n  Alternatively A = [wi,j ]MxN = (row normalized) weight for (ti ,dj)

p  For each ti, pick terms tj with high Cij values

ti

dj N

M

What does C = AAT contain if A is the term-doc incidence (0/1) matrix?

41

A= [wi,j ]

Page 36: Part 8: Relevance Feedbackricci/ISR/slides-2015/08...Part 8: Relevance Feedback Francesco Ricci Most of these slides comes from the course: Information Retrieval and Web Search, Christopher

Automatic Thesaurus Generation Example

42

Page 37: Part 8: Relevance Feedbackricci/ISR/slides-2015/08...Part 8: Relevance Feedback Francesco Ricci Most of these slides comes from the course: Information Retrieval and Web Search, Christopher

Automatic Thesaurus Generation Discussion

n  Quality of associations is usually a problem n  Term ambiguity may introduce irrelevant

statistically correlated terms: n  “Apple computer” → “Apple red fruit

computer” (synsets are not distinguished) n  Problems:

n  False positives: Words deemed similar that are not

n  False negatives: Words deemed dissimilar that are similar

n  Since terms are highly correlated anyway, expansion may not retrieve many additional documents. 43

Page 38: Part 8: Relevance Feedbackricci/ISR/slides-2015/08...Part 8: Relevance Feedback Francesco Ricci Most of these slides comes from the course: Information Retrieval and Web Search, Christopher

Query assist

p  Generally done by query log mining p  Recommend frequent recent queries that contain

partial string typed by user p  A ranking problem! View each prior query as a

doc – Rank-order those matching partial string …

44

Page 39: Part 8: Relevance Feedbackricci/ISR/slides-2015/08...Part 8: Relevance Feedback Francesco Ricci Most of these slides comes from the course: Information Retrieval and Web Search, Christopher

Resources

IIR Ch 9

45