query anchoring using discriminative query...

19
Query Anchoring Using Discriminative Query Models Saar Kuzi Anna Shtok Oren Kurland Technion – Israel Institute Of Technology We thank SIGIR for the conference travel grant

Upload: others

Post on 23-Apr-2020

19 views

Category:

Documents


0 download

TRANSCRIPT

Query Anchoring Using Discriminative

Query ModelsSaar Kuzi Anna Shtok Oren Kurland

Technion – Israel Institute Of Technology

We thank SIGIR for the conference travel grant

Pseudo-feedback-basedQuery ExpansionHighly ranked documents are used to induce a query model

Relying on pseudo feedback may result in query drift:

Documents in the result list could be non relevant

Relevant documents can contain non query-pertaining information(Harman ’92, He&Ounis ’09, Lv&Zhai ’09)

2

Initial Result List

Pseudo Feedback

𝐷𝑖𝑛𝑖𝑑

Query

Query AnchoringTechniques for mitigating the risk in relying on pseudo feedback

Direct:

Interpolation with a model of the original query (e.g., Zhai&Lafferty ’01, Abdul-Jaleel et al. ’04, Lv&Zhai ’09)

Using the original query model as a prior (Tao&Zhai ’04, Tao&Zhai ’06)

Indirect:

Term clipping (e.g., Zhai&Lafferty ’01, Abdul-Jaleel et al. ’04, Ye et al. ’10)

Differential impact of documents on the query model (Lavrenko&Croft β€˜01, Abdul-Jaleel et al. ’04, Lv&Zhai β€˜14)

3

Our Approach A novel indirect query anchoring approach using a new discriminativeterm-based model

An accurate term-based representation of the initial ranking

Initial Result List𝐷𝑖𝑛𝑖𝑑

Query Model

Method

4

Discriminative Query Model

Anchored Query Model

Learning-to-rank

Language Model Notation Unigram language models are used

Given text π‘₯,

𝑝𝑀𝐿𝐸 𝑑 π‘₯ ≝𝑑𝑓 𝑑 ∈ x

π‘₯

π‘π·π‘–π‘Ÿ 𝑑 π‘₯ ≝𝑑𝑓 𝑑 ∈ 𝑑 + πœ‡π‘MLE 𝑑 𝐢

πœ‡ + π‘₯

Two language models, πœƒ1 and πœƒ2, are compared using cross entropy:

𝑑 is a term, π‘₯ is the length of π‘₯, and 𝐢 is the collection of documents

5

𝐢𝐸(𝑝(βˆ™ |πœƒ1)||𝑝 βˆ™ πœƒ2 ) = βˆ’

𝑑

𝑝 𝑑 πœƒ1 log 𝑝(𝑑|πœƒ2)

Mixture Model (Zhai&Lafferty ’01)

6

C

π’…πŸ

πœ½π‘»

π’…πŸ 𝒅𝒏…

π‘‘βˆˆπ·π‘–π‘›π‘–π‘‘

π‘‘βˆˆπ‘‘

𝑑𝑓 𝑑 ∈ 𝑑 log( 1 βˆ’ 𝛾 𝑝 𝑑 πœƒπ‘‡ + 𝛾𝑝 𝑑 𝐢 ))

𝑝 𝑑 𝑀𝑀 ≝ πœ†π‘π‘€πΏπΈ 𝑑 π‘ž + 1 βˆ’ πœ† 𝑝(𝑑|πœƒπ‘‡π‘π‘™π‘–π‘π‘π‘’π‘‘

)

R

π’…πŸq 𝒅𝒏…

𝑝 𝑑 𝑅𝑀1 ≝

π‘‘βˆˆπ·π‘–π‘›π‘–π‘‘

π‘π·π‘–π‘Ÿ 𝑑 𝑑 𝑝(𝑑|π‘ž)

𝑝 𝑑 𝑅𝑀3 ≝ πœ†π‘π‘€πΏπΈ 𝑑 π‘ž + 1 βˆ’ πœ† 𝑝 𝑑 𝑅𝑀1𝑐𝑙𝑖𝑝𝑝𝑒𝑑

Relevance Model (Lavrenko&Croft ’01)

(Abdul-Jaleel et al. ’04)

The log-likelihood of documents in 𝐷𝑖𝑛𝑖𝑑:

Generative Query

Models

Discriminative Model

The pseudo feedback assumption: the higher a document is

ranked in the initial result list, the higher its relevance likelihood

βˆ€π‘‘π‘– ,𝑑𝑗 ∈ 𝐷𝑖𝑛𝑖𝑑, if π‘Ÿ(𝑑𝑖) > π‘Ÿ(𝑑𝑗), then 𝑑𝑖 is more likely to

be relevant than 𝑑𝑗

7

SVM-rank (Joachims ’02)

min1

2𝑀 βˆ™ 𝑀 + 𝐢

𝑖,𝑗

πœ‰π‘–,𝑗

βˆ€π‘–βˆ€π‘—. π‘Ÿ 𝑑𝑖 > π‘Ÿ 𝑑𝑗 𝑀(πœ™ 𝑑𝑖 βˆ’ πœ™ 𝑑𝑗 ) β‰₯ 1 βˆ’ πœ‰π‘–,𝑗

βˆ€π‘–βˆ€π‘—. π‘Ÿ 𝑑𝑖 > π‘Ÿ 𝑑𝑗 πœ‰π‘– ,𝑗 β‰₯ 0

πœ™ 𝑑 = (log π‘π·π‘–π‘Ÿ 𝑑1 𝑑 , . . . , π‘™π‘œπ‘” π‘π·π‘–π‘Ÿ(𝑑 𝑉 |𝑑))

𝑉 is the vocabulary used in the initial result list

8

Model Derivation

9

𝑀

𝑀+ π‘€βˆ’

πœƒπ‘€+ πœƒπ‘€βˆ’

𝐿1 Norm

Positive Anchor Model

Negative Anchor Model

NegativeComponents

Positive Components

AnchorPos Method

βˆ€π‘‘. 𝑠(𝑑) = πœ†2𝑝(𝑑|πœƒ) + πœ†3𝑝(𝑑|πœƒπ‘€+)

𝑝(𝑑|πœ—+)

𝑝(𝑑|πœƒπ΄π‘›π‘β„Žπ‘œπ‘Ÿπ‘ƒπ‘œπ‘ ) = πœ†1𝑝𝑀𝐿𝐸 (𝑑|π‘ž) + (1 βˆ’ πœ†1)𝑝(𝑑|πœ—+)

Anchoring a generative model, πœƒ, using the positive anchor model, πœƒπ‘€+

1. Term clipping

2. Sum normalization

Interpolation with the original query

model

πœ†1 + πœ†2 + πœ†3 = 1, πœ†π‘– β‰₯ 0

10

ClipNeg MethodClipping negative anchor terms

1. Setting to 0 the probabilities of negative anchor terms

2. Term clipping

3. Sum normalization

𝑝(𝑑|πœ—βˆ’)

𝑝(𝑑|πœƒπΆπ‘™π‘–π‘π‘π‘’π‘”) = πœ†π‘π‘€πΏπΈ(𝑑|π‘ž) + (1 βˆ’ πœ†)𝑝(𝑑|πœ—βˆ’)

Interpolation with the original query

model

𝑝(𝑑|πœƒ)

11

Related Work Existing query anchoring techniques (direct query anchoring, term

clipping and differential weighting)

Applying our approach on top of these yields further improvements

Methods for improving the quality of the pseudo feedback result list

(e.g., Mitra et al. ’98, Lee et al. ’08)

Our model can be induced from any ranked list

12

Related Work A supervised term classification approach (Cao et al. ’08)

Our approach is unsupervised and focuses on unigram query models

Clustering of terms in a query model (Udupa et al. ’09)

A method for cluster selection was not proposed

Fusing the result lists: the initial and the expansion-based

(Zighelnic&Kurland ’08)

Our methods operate on the model level and post better performance

13

Evaluation TREC datasets: TREC123, ROBUST, and WT10G

The initial result list is retrieved using a standard language model

method (Lafferty&Zhai ’01): βˆ’πΆπΈ(𝑝𝑀𝐿𝐸(βˆ™ |π‘ž)||π‘π·π‘–π‘Ÿ βˆ™ 𝑑 )

Baselines:

β—¦ Generative Models (RM3, MM)

β—¦ Fusion (Zighelnic&Kurland ’08)

Values of free parameters are set using leave-one-out cross validation

14

The Discriminative Model

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

1

0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1

Ke

nd

all's

tau

ROBUST

v=100

v=500

v=ALL

Ξ±

βˆ’π›ΌπΆπΈ(𝑝(βˆ™ |πœƒπ‘€+)||π‘π·π‘–π‘Ÿ βˆ™ 𝑑 ) + 1 βˆ’ 𝛼 𝐢𝐸(𝑝(βˆ™ |πœƒπ‘€βˆ’)||π‘π·π‘–π‘Ÿ βˆ™ 𝑑 )

β€’ Re-ranking an initial result list of 100 documents according to:

β€’ The positive and negative anchor models are clipped to use Ξ½ terms

16

π‘πŒπŸ (𝐀𝐏 = πŸπŸ’. πŸ–)

𝛉𝐰+ (𝐀𝐏 = πŸ’πŸ—. πŸ‘)

Query: Airport Security, ROBUST, QL(AP=24.8) Discriminative vs.

Generative

16

β€’ The discriminative model assigns high probabilities to terms with high IDF values

β€’ The generative models are much more similar to each other, with respect to the terms they promote, than they are to the discriminative model

20

22

24

26

28

30

RM3 MM

MA

P

TREC123

An

cho

rPo

s

An

cho

rPo

s

Fusi

on

Fusi

on

20

22

24

26

28

30

RM3 MMM

AP

ROBUST

Fusi

on

18

19

20

21

22

RM3 MM

MA

P

WT10G

Fusi

on

17

πœ“βˆ—βˆ—

πœ“βˆ— βˆ—

πœ“βˆ— πœ“βˆ—

πœ“

βˆ—βˆ—

AnchorPos

Statistically significant differences with:

Generative Model

Fusion

Anchoring a generative model using the positive anchor

model

πœ“

βˆ—

Fusi

on

An

cho

rPo

s

An

cho

rPo

s

Fusi

on

Fusi

on

An

cho

rPo

s

An

cho

rPo

s

Fusi

on

Fusi

on

20

22

24

26

28

30

RM3 MM

MA

P

ROBUST

20

22

24

26

28

30

RM3 MM

MA

P

TREC123

ClipNeg

18

19

20

21

22

RM3 MM

MA

P

WT10G

Clip

Ne

g

Clip

Ne

g

Clip

Ne

gC

lipN

eg

Clipping negative

anchor terms

Statistically significant differences with the generative model

18

βˆ—

βˆ—

Clip

Ne

g

Clip

Ne

g

Clip

Ne

g

Clip

Ne

g

Clip

Ne

g

βˆ—

Summary We presented a novel unsupervised pseudo-feedback-based

discriminative query model that is based on a learning-to-rank-

approach

We devised a few methods that use the discriminative model to

perform (indirect) query anchoring of existing query models

Empirical evaluation showed that using our methods can improve the

performance of highly effective generative query models

19