query anchoring using discriminative query...
TRANSCRIPT
Query Anchoring Using Discriminative
Query ModelsSaar Kuzi Anna Shtok Oren Kurland
Technion β Israel Institute Of Technology
We thank SIGIR for the conference travel grant
Pseudo-feedback-basedQuery ExpansionHighly ranked documents are used to induce a query model
Relying on pseudo feedback may result in query drift:
Documents in the result list could be non relevant
Relevant documents can contain non query-pertaining information(Harman β92, He&Ounis β09, Lv&Zhai β09)
2
Initial Result List
Pseudo Feedback
π·ππππ‘
Query
Query AnchoringTechniques for mitigating the risk in relying on pseudo feedback
Direct:
Interpolation with a model of the original query (e.g., Zhai&Lafferty β01, Abdul-Jaleel et al. β04, Lv&Zhai β09)
Using the original query model as a prior (Tao&Zhai β04, Tao&Zhai β06)
Indirect:
Term clipping (e.g., Zhai&Lafferty β01, Abdul-Jaleel et al. β04, Ye et al. β10)
Differential impact of documents on the query model (Lavrenko&Croft β01, Abdul-Jaleel et al. β04, Lv&Zhai β14)
3
Our Approach A novel indirect query anchoring approach using a new discriminativeterm-based model
An accurate term-based representation of the initial ranking
Initial Result Listπ·ππππ‘
Query Model
Method
4
Discriminative Query Model
Anchored Query Model
Learning-to-rank
Language Model Notation Unigram language models are used
Given text π₯,
πππΏπΈ π‘ π₯ βπ‘π π‘ β x
π₯
ππ·ππ π‘ π₯ βπ‘π π‘ β π + ππMLE π‘ πΆ
π + π₯
Two language models, π1 and π2, are compared using cross entropy:
π‘ is a term, π₯ is the length of π₯, and πΆ is the collection of documents
5
πΆπΈ(π(β |π1)||π β π2 ) = β
π‘
π π‘ π1 log π(π‘|π2)
Mixture Model (Zhai&Lafferty β01)
6
C
π π
π½π»
π π π πβ¦
πβπ·ππππ‘
π‘βπ
π‘π π‘ β π log( 1 β πΎ π π‘ ππ + πΎπ π‘ πΆ ))
π π‘ ππ β ππππΏπΈ π‘ π + 1 β π π(π‘|πππππππππ
)
R
π πq π πβ¦
π π‘ π π1 β
πβπ·ππππ‘
ππ·ππ π‘ π π(π|π)
π π‘ π π3 β ππππΏπΈ π‘ π + 1 β π π π‘ π π1πππππππ
Relevance Model (Lavrenko&Croft β01)
(Abdul-Jaleel et al. β04)
The log-likelihood of documents in π·ππππ‘:
Generative Query
Models
Discriminative Model
The pseudo feedback assumption: the higher a document is
ranked in the initial result list, the higher its relevance likelihood
βππ ,ππ β π·ππππ‘, if π(ππ) > π(ππ), then ππ is more likely to
be relevant than ππ
7
SVM-rank (Joachims β02)
min1
2π€ β π€ + πΆ
π,π
ππ,π
βπβπ. π ππ > π ππ π€(π ππ β π ππ ) β₯ 1 β ππ,π
βπβπ. π ππ > π ππ ππ ,π β₯ 0
π π = (log ππ·ππ π‘1 π , . . . , πππ ππ·ππ(π‘ π |π))
π is the vocabulary used in the initial result list
8
Model Derivation
9
π€
π€+ π€β
ππ€+ ππ€β
πΏ1 Norm
Positive Anchor Model
Negative Anchor Model
NegativeComponents
Positive Components
AnchorPos Method
βπ‘. π (π‘) = π2π(π‘|π) + π3π(π‘|ππ€+)
π(π‘|π+)
π(π‘|ππ΄ππβπππππ ) = π1πππΏπΈ (π‘|π) + (1 β π1)π(π‘|π+)
Anchoring a generative model, π, using the positive anchor model, ππ€+
1. Term clipping
2. Sum normalization
Interpolation with the original query
model
π1 + π2 + π3 = 1, ππ β₯ 0
10
ClipNeg MethodClipping negative anchor terms
1. Setting to 0 the probabilities of negative anchor terms
2. Term clipping
3. Sum normalization
π(π‘|πβ)
π(π‘|ππΆππππππ) = ππππΏπΈ(π‘|π) + (1 β π)π(π‘|πβ)
Interpolation with the original query
model
π(π‘|π)
11
Related Work Existing query anchoring techniques (direct query anchoring, term
clipping and differential weighting)
Applying our approach on top of these yields further improvements
Methods for improving the quality of the pseudo feedback result list
(e.g., Mitra et al. β98, Lee et al. β08)
Our model can be induced from any ranked list
12
Related Work A supervised term classification approach (Cao et al. β08)
Our approach is unsupervised and focuses on unigram query models
Clustering of terms in a query model (Udupa et al. β09)
A method for cluster selection was not proposed
Fusing the result lists: the initial and the expansion-based
(Zighelnic&Kurland β08)
Our methods operate on the model level and post better performance
13
Evaluation TREC datasets: TREC123, ROBUST, and WT10G
The initial result list is retrieved using a standard language model
method (Lafferty&Zhai β01): βπΆπΈ(πππΏπΈ(β |π)||ππ·ππ β π )
Baselines:
β¦ Generative Models (RM3, MM)
β¦ Fusion (Zighelnic&Kurland β08)
Values of free parameters are set using leave-one-out cross validation
14
The Discriminative Model
0.2
0.3
0.4
0.5
0.6
0.7
0.8
0.9
1
0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1
Ke
nd
all's
tau
ROBUST
v=100
v=500
v=ALL
Ξ±
βπΌπΆπΈ(π(β |ππ€+)||ππ·ππ β π ) + 1 β πΌ πΆπΈ(π(β |ππ€β)||ππ·ππ β π )
β’ Re-ranking an initial result list of 100 documents according to:
β’ The positive and negative anchor models are clipped to use Ξ½ terms
16
πππ (ππ = ππ. π)
ππ°+ (ππ = ππ. π)
Query: Airport Security, ROBUST, QL(AP=24.8) Discriminative vs.
Generative
16
β’ The discriminative model assigns high probabilities to terms with high IDF values
β’ The generative models are much more similar to each other, with respect to the terms they promote, than they are to the discriminative model
20
22
24
26
28
30
RM3 MM
MA
P
TREC123
An
cho
rPo
s
An
cho
rPo
s
Fusi
on
Fusi
on
20
22
24
26
28
30
RM3 MMM
AP
ROBUST
Fusi
on
18
19
20
21
22
RM3 MM
MA
P
WT10G
Fusi
on
17
πββ
πβ β
πβ πβ
π
ββ
AnchorPos
Statistically significant differences with:
Generative Model
Fusion
Anchoring a generative model using the positive anchor
model
π
β
Fusi
on
An
cho
rPo
s
An
cho
rPo
s
Fusi
on
Fusi
on
An
cho
rPo
s
An
cho
rPo
s
Fusi
on
Fusi
on
20
22
24
26
28
30
RM3 MM
MA
P
ROBUST
20
22
24
26
28
30
RM3 MM
MA
P
TREC123
ClipNeg
18
19
20
21
22
RM3 MM
MA
P
WT10G
Clip
Ne
g
Clip
Ne
g
Clip
Ne
gC
lipN
eg
Clipping negative
anchor terms
Statistically significant differences with the generative model
18
β
β
Clip
Ne
g
Clip
Ne
g
Clip
Ne
g
Clip
Ne
g
Clip
Ne
g
β
Summary We presented a novel unsupervised pseudo-feedback-based
discriminative query model that is based on a learning-to-rank-
approach
We devised a few methods that use the discriminative model to
perform (indirect) query anchoring of existing query models
Empirical evaluation showed that using our methods can improve the
performance of highly effective generative query models
19