learning to estimate query difficulty including applications to missing content detection and...

Post on 17-Jan-2018

214 Views

Category:

Documents

0 Downloads

Preview:

Click to see full reader

DESCRIPTION

3 Introduction (1/2) Many IR systems suffer from a radical variance in performance. Many IR systems suffer from a radical variance in performance. Estimating query difficulty is an attempt to quantify the quality of results returned by a given system for the query. Estimating query difficulty is an attempt to quantify the quality of results returned by a given system for the query. Reasons for query difficulty estimation Reasons for query difficulty estimation Feedback to the user Feedback to the user The user can rephrase “difficult” queries. The user can rephrase “difficult” queries. Feedback to the search engine Feedback to the search engine To invoke alternative strategies for different queries To invoke alternative strategies for different queries Feedback to the system administrator Feedback to the system administrator To identify queries related to a specific subject, and expand the document collection. To identify queries related to a specific subject, and expand the document collection. For distributed information retrieval For distributed information retrieval

TRANSCRIPT

Learning to Estimate Learning to Estimate Query DifficultyQuery Difficulty

Including Applications to Missing Content Detection and Including Applications to Missing Content Detection and Distributed Information RetrievalDistributed Information Retrieval

Elad Yom-Tov, Shai Fine, David Elad Yom-Tov, Shai Fine, David Carmel, Adam DarlowCarmel, Adam Darlow

IBM Haifa Research LabsIBM Haifa Research LabsSIGIR 2005SIGIR 2005

2

AbstractAbstract Novel learning methods are used for estimating the Novel learning methods are used for estimating the

quality of results returned by a search engine in quality of results returned by a search engine in response to a query.response to a query.

Estimation is based on the agreement between the Estimation is based on the agreement between the top results of the full query and the top results of its top results of the full query and the top results of its sub-queries.sub-queries.

Quality estimation are useful for several applications, Quality estimation are useful for several applications, including including improvement of retrievalimprovement of retrieval, detecting queries , detecting queries for whichfor which no relevant content existsno relevant content exists in the document in the document collection, and collection, and distributed information retrievaldistributed information retrieval..

3

Introduction (1/2)Introduction (1/2) Many IR systems suffer from a radical variance in Many IR systems suffer from a radical variance in

performance.performance. Estimating query difficultyEstimating query difficulty is an attempt to quantify the is an attempt to quantify the

quality of results returned by a given system for the query.quality of results returned by a given system for the query. Reasons for query difficulty estimationReasons for query difficulty estimation

Feedback to the userFeedback to the user The user can rephrase “difficult” queries.The user can rephrase “difficult” queries.

Feedback to the search engineFeedback to the search engine To invoke alternative strategies for different queriesTo invoke alternative strategies for different queries

Feedback to the system administratorFeedback to the system administrator To identify queries related to a specific subject, and expand the To identify queries related to a specific subject, and expand the

document collection.document collection. For distributed information retrievalFor distributed information retrieval

4

Introduction (2/2)Introduction (2/2) The observation and motivation:The observation and motivation:

queries answered well are those whose query terms queries answered well are those whose query terms agree on most of the returned documents.agree on most of the returned documents. Agreement is measured by the overlap between the top Agreement is measured by the overlap between the top

results.results. Difficult queries are those:Difficult queries are those:

A.A. The query terms cannot agree on top results.The query terms cannot agree on top results.B.B. Most of the terms do agree except a few Most of the terms do agree except a few outliers outliers (( 局外人局外人 ))..

A TREC query for example:A TREC query for example:““What impact has the chunnel (What impact has the chunnel ( 水底隧道水底隧道 ) ) had on the had on the

British economy and/or the life style of the British”British economy and/or the life style of the British”

5

Related Work (1/2)Related Work (1/2) In the In the RobustRobust tracktrack of TREC 2004of TREC 2004, systems are asked to , systems are asked to

rank the topics by predicted difficulty.rank the topics by predicted difficulty. The goal is eventually to use such predictions to do topic-The goal is eventually to use such predictions to do topic-

specific processing.specific processing. Prediction methods suggested by the participants:Prediction methods suggested by the participants:

Measuring clarity based on the system’s score of the top resultsMeasuring clarity based on the system’s score of the top results Analyzing the ambiguity of the query termsAnalyzing the ambiguity of the query terms Learning a predictor using old TREC topics as training dataLearning a predictor using old TREC topics as training data

(Ounis, 2004) showed that (Ounis, 2004) showed that IDF-basedIDF-based predictor is positively predictor is positively related to query precision.related to query precision.

(Diaz, 2004) used (Diaz, 2004) used temporal distributiontemporal distribution together with together with content of the documents to improve the prediction of AP content of the documents to improve the prediction of AP for a query.for a query.

6

Related Work (2/2)Related Work (2/2) The The Reliable Information AccessReliable Information Access (RIA) workshop (RIA) workshop

investigated the reasons for system performance investigated the reasons for system performance variance across queries.variance across queries. 10 failure categories were identified.10 failure categories were identified.

4 of which are due to emphasizing only partial aspects of 4 of which are due to emphasizing only partial aspects of the query.the query.

One of the conclusions of this workshop:One of the conclusions of this workshop:“…“…comparing a full topic ranking against ranking based on comparing a full topic ranking against ranking based on

only one aspect of the topic will give a measure of the only one aspect of the topic will give a measure of the importance of that aspect to the retrieved set”importance of that aspect to the retrieved set”

7

Estimating Query DifficultyEstimating Query Difficulty Query terms are defined as the Query terms are defined as the keywordskeywords and the and the

lexical affinitieslexical affinities.. Features used for learning:Features used for learning:

The overlap between each sub-query and the full queryThe overlap between each sub-query and the full query Measured by κ-statistics Measured by κ-statistics

The rounded logarithm of the document frequency, The rounded logarithm of the document frequency, log(log(DFDF), of each of the sub-queries.), of each of the sub-queries.

Two challenges for learning:Two challenges for learning: The number of sub-queries is not constant.The number of sub-queries is not constant.

A canonic representation is needed.A canonic representation is needed. The sub-queries are not ordered.The sub-queries are not ordered.

8

Query Estimator Using a Query Estimator Using a Histogram (1/2)Histogram (1/2)

The basic procedure:The basic procedure:1)1) Find the top N results for the full query and for each sub-query.Find the top N results for the full query and for each sub-query. Build a histogram of the overlaps Build a histogram of the overlaps hh((ii,,jj) to form a feature vector.) to form a feature vector.

Values of log(Values of log(DFDF) are split into 3 discrete values {0) are split into 3 discrete values {0 -- 1, 21, 2 -- 3, 43, 4++ }.}.

hh((ii,,jj) means log() means log(DFDF)) == ii & overlaps & overlaps == jj.. The rows of The rows of hh((ii,,jj) are concatenated as a feature vector.) are concatenated as a feature vector.

1)1) Compute the linear weight vector Compute the linear weight vector cc for prediction. for prediction. An example, suppose a query has 4 sub-queries:An example, suppose a query has 4 sub-queries:

log(log(DFDF((nn)))) == [0 1 1 2], overlap[0 1 1 2], overlap == [2 0 0 1][2 0 0 1]

→ h(i) = [0 0 1 2 0 0 0 1 0]

9

Query Estimator Using a Query Estimator Using a Histogram (2/2)Histogram (2/2)

Two additional featuresTwo additional features1)1) The score of the top-ranked documentThe score of the top-ranked document2)2) The number of words in the queryThe number of words in the query

Estimate the linear weight vector Estimate the linear weight vector cc (Moore-Penrose (Moore-Penrose pseudo-inverse):pseudo-inverse):

cc = ( = (HH .. HHTT))-1-1 .. HH .. ttTT

HH == the matrix with columns are feature vectors of training the matrix with columns are feature vectors of training queriesqueries

tt == a vector of the target measure (P@10 or MAP) of training a vector of the target measure (P@10 or MAP) of training queriesqueries

((HH and and tt can be modified according to the objective) can be modified according to the objective)

10

Query Estimator Using a Query Estimator Using a Modified Decision Tree (1/2)Modified Decision Tree (1/2)

Useful for sparseness, i.e. queries are too short.Useful for sparseness, i.e. queries are too short. A binary decision treeA binary decision tree Pairs of overlap and log(DF) of sub-queries form Pairs of overlap and log(DF) of sub-queries form

features.features. Each node consists of a Each node consists of a weight vectorweight vector, , thresholdthreshold, and, and

scorescore.. An example:An example:

11

Query Estimator Using a Query Estimator Using a Modified Decision Tree (2/2)Modified Decision Tree (2/2)

The concept of The concept of Random ForestRandom Forest Better decision trees can be obtained by training a Better decision trees can be obtained by training a

multitude of trees, each in a slightly different manner or multitude of trees, each in a slightly different manner or using different data.using different data.

Apply AdaBoost algo. to resample training dataApply AdaBoost algo. to resample training data

12

Experiment and Evaluation (1/2)Experiment and Evaluation (1/2) The IR system is Juru.The IR system is Juru. Two document collectionsTwo document collections

TREC-8: 528,155 documents, 200 topicsTREC-8: 528,155 documents, 200 topics WT10G: 1,692,096 documents, 100 topicsWT10G: 1,692,096 documents, 100 topics

Four-fold cross-validation,Four-fold cross-validation, Measured by Kendall’s-τcoefficientMeasured by Kendall’s-τcoefficient

13

Experiment and Evaluation (2/2)Experiment and Evaluation (2/2) Compared with some other algorithmsCompared with some other algorithms

Estimation based on the score of the top resultEstimation based on the score of the top result Estimation based on the average score of the top ten Estimation based on the average score of the top ten

resultsresults Estimation based on the standard deviation of IDF values Estimation based on the standard deviation of IDF values

of query termsof query terms Estimation based on learning a SVM for regressionEstimation based on learning a SVM for regression

14

Application 1: Improving IR Application 1: Improving IR Using Query Estimation (1/2)Using Query Estimation (1/2)

Selective automatic query expansionSelective automatic query expansion1.1. Adding terms to the query based on frequently Adding terms to the query based on frequently

appearing terms in the top retrieved documentsappearing terms in the top retrieved documents2.2. Only works for Only works for easyeasy queries queries3.3. Using the same features to train a SVM classifierUsing the same features to train a SVM classifier

Deciding which part of the topic should be usedDeciding which part of the topic should be used1.1. TREC topics contain two parts: short TREC topics contain two parts: short titletitle and longer and longer

descriptiondescription2.2. Some topics that are not answered well by the Some topics that are not answered well by the

description part are better answered by the title part.description part are better answered by the title part.3.3. Difficult topics use title part and easy topics use Difficult topics use title part and easy topics use

description.description.

15

Application 1: Improving IR Application 1: Improving IR Using Query Estimation (2/2)Using Query Estimation (2/2)

16

Application 2: Detecting Missing Application 2: Detecting Missing Content (1/2)Content (1/2)

Missing content queries (Missing content queries (MCQsMCQs) are those have no ) are those have no relevant document in the collection.relevant document in the collection.

Experiment methodExperiment method 166 MCQs are created artificially from 400 TREC queries166 MCQs are created artificially from 400 TREC queries

200 TREC topics consist of title and description.200 TREC topics consist of title and description. Ten-fold cross-validationTen-fold cross-validation A tree-based classifier is trained to classify MCQs from A tree-based classifier is trained to classify MCQs from

non-MCQs.non-MCQs. A query difficulty estimator may or may not be used as a A query difficulty estimator may or may not be used as a

pre-filter of easy queries before the MCQ classifier.pre-filter of easy queries before the MCQ classifier.

17

Application 2: Detecting Missing Application 2: Detecting Missing Content (2/2)Content (2/2)

18

Application 3: Merging the Results Application 3: Merging the Results of Distributed Retrieval (1/2)of Distributed Retrieval (1/2)

It is difficult to rerank the documents from different It is difficult to rerank the documents from different datasets since the scores are local for each specific datasets since the scores are local for each specific dataset.dataset.

CORICORI (W. Croft, 1995) is one of the state-of-the-art (W. Croft, 1995) is one of the state-of-the-art algorithm for distributed retrieval, using inference algorithm for distributed retrieval, using inference network to do collection ranking.network to do collection ranking.

Apply the estimator to this problem:Apply the estimator to this problem: A query estimator is trained for each dataset.A query estimator is trained for each dataset. The estimated difficulty is used for weighting the scores.The estimated difficulty is used for weighting the scores. These weighted scores are merged to built the final ranking.These weighted scores are merged to built the final ranking. Ten-fold cross-validationTen-fold cross-validation Only minimal information is supplied by the search engine.Only minimal information is supplied by the search engine.

19

Application 3: Merging the Results Application 3: Merging the Results of Distributed Retrieval (2/2)of Distributed Retrieval (2/2)

Selective weightingSelective weighting All queries are clustered (2-means) based on their All queries are clustered (2-means) based on their

estimations for each of the datasets.estimations for each of the datasets. In one cluster, the variance of the estimations is smallIn one cluster, the variance of the estimations is small

→ → unweighted scores are better for queries in this cluster.unweighted scores are better for queries in this cluster. The estimations of difficulty become noise when there is The estimations of difficulty become noise when there is

little variance.little variance.

20

Conclusions and Future WorkConclusions and Future Work Two methods for learning an estimator of query difficulty are Two methods for learning an estimator of query difficulty are

described.described. The learned estimator predicts the expected precision of the The learned estimator predicts the expected precision of the

query by analyzing the overlap between the results of the full query by analyzing the overlap between the results of the full query and the results of its sub-queries.query and the results of its sub-queries.

We show that such an estimator can be used to several We show that such an estimator can be used to several applications.applications.

Our results show that the quality of query prediction strongly Our results show that the quality of query prediction strongly depends on the query length.depends on the query length.

One of the future work is to look for additional features not One of the future work is to look for additional features not depend on the query length.depend on the query length.

Whether more training data can be accumulated in automatic Whether more training data can be accumulated in automatic or semi-automatic manner is left for future research.or semi-automatic manner is left for future research.

top related