direct answers for search queries in the long tail

40
Direct Answers for Search Queries in the Long Tail MIT HUMAN-COMPUTER INTERACTI l Bernstein, Jaime Teevan, Susan Dumais, ebling, and Eric Horvitz AIL and Microsoft Research

Upload: stu

Post on 23-Feb-2016

47 views

Category:

Documents


0 download

DESCRIPTION

Direct Answers for Search Queries in the Long Tail. Michael Bernstein, Jaime Teevan , Susan Dumais , Dan Liebling , and Eric Horvitz MIT CSAIL and Microsoft Research. mit human-computer interaction. Answers: Direct Search Results. Manually constructed for popular queries. - PowerPoint PPT Presentation

TRANSCRIPT

Page 1: Direct Answers for Search Queries in the Long Tail

Direct Answers for Search Queries in the Long Tail

MIT HUMAN-COMPUTER INTERACTION

Michael Bernstein, Jaime Teevan, Susan Dumais, Dan Liebling, and Eric HorvitzMIT CSAIL and Microsoft Research

Page 2: Direct Answers for Search Queries in the Long Tail

Answers: Direct Search Results

weather bostonManually constructed for popular queries

Page 3: Direct Answers for Search Queries in the Long Tail

the girl with the dragon tattoo

memorial day 2012

AAPL

good answers reduce clicks on result pages

users trigger answers repeatedly once discovered

[Chilton + Teevan 2009]

Page 4: Direct Answers for Search Queries in the Long Tail

the girl with the dragon tattoo

memorial day 2012

AAPL

only popularquery typesanswers are:- high cost- high maintenance

Page 5: Direct Answers for Search Queries in the Long Tail

Prevalence of Uncommon Searches

molasses substitutesincrease volume windows xpdissolvable stitches speeddog body temperatureCHI 2013 deadline…

No answers for many information needs

Page 6: Direct Answers for Search Queries in the Long Tail

Tail AnswersDirect results for queries in the long tail

molasses substitutes

Page 7: Direct Answers for Search Queries in the Long Tail

Tail AnswersDirect results for queries in the long tail

molasses substitutes

Tail Answers improve the search experience for less common queries, and fully compensate for poor results.

Page 8: Direct Answers for Search Queries in the Long Tail

The Long Tail of Answers#

occu

rrenc

es

Information needs

weathermovies

chi 2017 location

Hard to find structured informationNot enough query volume for dedicated teams

Tail Answers

Page 9: Direct Answers for Search Queries in the Long Tail
Page 10: Direct Answers for Search Queries in the Long Tail

Crowds can support the long tail of user goals in interactive systems.

Crowd Datasearch logs

Paid Crowdson-demand

Page 11: Direct Answers for Search Queries in the Long Tail

Tail Answers PipelineFind URLs that satisfy fact-finding information needs, then extract answers from those pages

Page 12: Direct Answers for Search Queries in the Long Tail

Tail Answers Pipeline

Extract answer content

Identify answer candidates

Filter candidates

1

2

3

Page 13: Direct Answers for Search Queries in the Long Tail

Identify Answer Candidates

IdentifyFilterExtractEvaluationDiscussion

Outli

ne

Crowd data: 75 million search sessionsAll information needs are answer candidates:queries leading to a clickthrough on a single URL

force quit mac

force quit on macs

how to force quit mac

URL

Page 14: Direct Answers for Search Queries in the Long Tail

Example Answer Candidatesforce quit mac

force quit on macshow to force quit mac

URL1

410 area code

area code 410 location

URL2

Page 15: Direct Answers for Search Queries in the Long Tail

Tail Answers Pipeline

Extract answer content

Identify answer candidates

Filter candidates

1

2

3

Page 16: Direct Answers for Search Queries in the Long Tail

Filtering Answer CandidatesFocus on fact-finding information needs [Kellar 2007]

Exclude popular but unanswerable candidates:

IdentifyFilterExtractEvaluationDiscussion

Outli

ne

radio

pandora

pandora radio log in

pandora.com

Page 17: Direct Answers for Search Queries in the Long Tail

Filtering Answer CandidatesThree filters remove answer candidates that do not address fact-finding information needs:

Navigation behaviorPages addressing search needs

Query behaviorUnambiguous needs

Answer typeSuccinct answers

Page 18: Direct Answers for Search Queries in the Long Tail

Filter by Navigation BehaviorDestination Probability for :P(session length = 2 | in trail)Probability of ending session at after clicking through from the search results

URL1

query

URL2

URL3

URL1

query

URL1

query

URL4

URL1

query

URL1 destination probability = 0.5

URL

URL

URL

Page 19: Direct Answers for Search Queries in the Long Tail

Filter by Navigation BehaviorDestination Probability Filter:URLs with low probability that searchers will end their session (Lots of back navigations, later clicks)

Focus on queries where searchers addressed an information need

Page 20: Direct Answers for Search Queries in the Long Tail

Filter by Query BehaviorWhat answers are these searchers looking for?

dissolvable stitches

732 area code

(how long they last?)(what they’re made of?)

(city and state?)(count of active phone numbers?)

Page 21: Direct Answers for Search Queries in the Long Tail

Filter by Query BehaviorA minority of searchers use question words:

Filter candidates with fewer than 1% of clickthroughs from question queries

how long dissolvable stitches last

where is 732 area code

Page 22: Direct Answers for Search Queries in the Long Tail

Filter by Answer TypeCan a concise answer address this need?Ask paid crowdsourcing workers to select:– Short: phrase or sentence

“The optimal fish frying temperature is 350°F.” – List: small set of directions or alternatives

“To change your password over Remote Desktop: 1) Click on Start > Windows Security. 2) Click the Change Password button. [...]”

– Summary: synthesize large amount of contentImpact of Budget Cuts on Teachers

Today

Page 23: Direct Answers for Search Queries in the Long Tail

Creating Tail Answers

Extract answer content

Identify answer candidates

Filter candidates

1

2

3

Page 24: Direct Answers for Search Queries in the Long Tail

Extracting the Tail AnswerWe now have answer candidates with:

Factual responsesSuccinct responses

However, the answer is buried:

IdentifyFilterExtractEvaluationDiscussion

Outli

ne

dissolvable stitchesdissolvable stitches how longdissolvable stitches absorption

Page 25: Direct Answers for Search Queries in the Long Tail

Crowdsourcing WorkflowReliably extract relevant answer from the URLvia paid crowdsourcing (CrowdFlower)

Title

Vote

Proofread Vote

Vote

Extract

[Bernstein et al. 2010, Little et al. 2010]

Page 26: Direct Answers for Search Queries in the Long Tail

Quality Challenge: Overgenerating

Typical extraction length:Lorem ipsum dolor sit amet, consectetur adipiscing elit. Curabitur nisl ligula, venenatis eget vulputate at, venenatis non sem. Pellentesque viverra metus vel orci suscipit vitae ullamcorper nisl vestibulum. Ut bibendum venenatis erat nec porttitor. Integer aliquam elit tempor tortor iaculis ut volutpat est lacinia. Aenean fringilla interdum tristique. Duis id felis sit amet libero porttitor suscipit eget vitae elit. Fusce neque augue, facilisis quis bibendum a, dapibus et felis. Aliquam at sagittis magna. Sed commodo semper tortor in facilisis. Nullam consequat quam et felis faucibus sed imperdiet purus luctus. Proin adipiscing felis ac nulla euismod ac dictum massa blandit. In volutpat auctor pharetra. Phasellus at nisl massa. Vivamus malesuada turpis a ligula lacinia ut interdum dui congue. Curabitur a molestie leo. Nulla mattis posuere sapien sit amet orci aliquam.

Page 27: Direct Answers for Search Queries in the Long Tail

Quality Challenge: Overgenerating

Extraction length with gold standards:Lorem ipsum dolor sit amet, consectetur adipiscing elit. Curabitur nisl ligula, venenatis eget vulputate at, venenatis non sem. Pellentesque viverra metus vel orci suscipit vitae ullamcorper nisl vestibulum. Ut bibendum venenatis erat nec porttitor. Integer aliquam elit tempor tortor iaculis ut volutpat est lacinia. Aenean fringilla interdum tristique. Duis id felis sit amet libero porttitor suscipit eget vitae elit. Fusce neque augue, facilisis quis bibendum a, dapibus et felis. Aliquam at sagittis magna. Sed commodo semper tortor in facilisis. Nullam consequat quam et felis faucibus sed imperdiet purus luctus. Proin adipiscing felis ac nulla euismod ac dictum massa blandit. In volutpat auctor pharetra. Phasellus at nisl massa. Vivamus malesuada turpis a ligula lacinia ut interdum dui congue. Curabitur a molestie leo. Nulla mattis posuere sapien sit amet orci aliquam.

Page 28: Direct Answers for Search Queries in the Long Tail

Tail Answer Pipeline

Extract answer content

Identify answer candidates

Filter candidates

1

2

3

Page 29: Direct Answers for Search Queries in the Long Tail

75 million search trails19,167 answer candidates

Median answer triggered once a day44 cents to create each answer

Page 30: Direct Answers for Search Queries in the Long Tail

Evaluation: Answer QualityHand-coded for correctness and writing errors (two–three redundant coders)

83% of Tail Answers had no writing errors87% of Tail Answers were completely correct or had a minor error (e.g., title != content)False positives in crowd data: dynamic web pages Identify

FilterExtractEvaluationDiscussion

Outli

ne

Page 31: Direct Answers for Search Queries in the Long Tail

Field ExperimentHow do Tail Answers impact searchers’

subjective impressions of the result page?Method:Recruit 361 users to issue queries that trigger Tail Answers to a modified version of Bing.

Page 32: Direct Answers for Search Queries in the Long Tail

Field Experiment DesignWithin-subjects 2x2 design:

Tail Answers vs. no Tail AnswersGood Ranking vs. Bad Ranking

Measurement: 7-point Likert responses1. Result page is useful2. No need to click through to a

resultAnalysis: linear mixed effects modelGeneralization of ANOVA

Page 33: Direct Answers for Search Queries in the Long Tail

Tail Answers’ Usefulness Are Comparable to Good Result Ranking

Tail Answers main effect: 0.34pts (7pt Likert)Ranking main effect: 0.68ptsInteraction effect: 1.03pts

Results

All results significant p<0.001

15

Tail AnswerNo Tail AnswerUs

eful

Page 34: Direct Answers for Search Queries in the Long Tail

Answers Make Result Clickthroughs Less Necessary

Tail Answers main effect: 1.01pts (7pt Likert)Result ranking main effect: 0.50ptsInteraction effect: 0.91pts

Results

All results significant p<0.001

15

Tail AnswerNo Tail Answer

No c

licks

Page 35: Direct Answers for Search Queries in the Long Tail

Tail Answers impact subjective ratings half as much as good ranking, and fully compensate for poor results.…but, we need to improve the trigger queries.

Page 36: Direct Answers for Search Queries in the Long Tail

Ongoing ChallengesSpreading incorrect or unverified information

IdentifyFilterExtractEvaluationDiscussion

Outli

ne

Cannibalizing pageviews from the original content pages

Page 37: Direct Answers for Search Queries in the Long Tail

Extension: A.I.-driven AnswersUse open information extraction systems to propose answers, and crowd to verifyCrowd-authored

Authored by AI, verified by crowds

Page 38: Direct Answers for Search Queries in the Long Tail

Extension: Better Result SnippetsImprove result pages for popular queriesAutomatically extracted

Crowd-authored

Page 39: Direct Answers for Search Queries in the Long Tail

Extension: Domain-Specific Answers

Design for specific information needsCrowds structuring new data types

Page 40: Direct Answers for Search Queries in the Long Tail

Direct Answers for Search Queries in the Long TailCrowd data can support many uncommon user goals in interactive systems.