recommender systems! @asai 2011

46
Lic. Ernesto Mislej [email protected] - @fetnelio ASAI 2011 – 40 JAIIO Agosto 2011 – Córdoba – Argentina 7puentes.com Recommender Systems @ASAI 2011

Upload: ernesto-mislej

Post on 06-May-2015

1.828 views

Category:

Technology


5 download

TRANSCRIPT

Page 1: Recommender Systems! @ASAI 2011

Lic. Ernesto [email protected] - @fetnelio !

ASAI 2011 – 40 JAIIO !Agosto 2011 – Córdoba – Argentina !

7puentes.com!

Recommender Systems!@ASAI 2011 !

Page 2: Recommender Systems! @ASAI 2011

Outline

• Introduction: What are recommender systems for?• Paradigms: How do they work?

• Collaborative Filtering• Content-based Filtering• Knowledge-Based Recommendations

• Summary, Issues and Challenges

Source: Tutorial: Recommender Systems, D. Jannach, G. Friedrich, IJCAI 2011.

Page 3: Recommender Systems! @ASAI 2011

Introduction

Page 4: Recommender Systems! @ASAI 2011

Problem domain

There is an extensive class of Web applications that involvepredicting user responses to options. Such a facility is called arecommendation system.

RS are software agents that elicit the interests andpreferences of individual consumers . . . and makerecommendations accordingly.They have the potential to support and improve thequality of the decision consumers make while searchingfor and selecting products online.

— Xiao & Benbasat, MISQ, 2007

Page 5: Recommender Systems! @ASAI 2011

Different perspectives

Retrieval perspective

• Reduce search costs• Provide correct proposals• Users know in advance what they want

Recommendation perspective

• Serendipity identify items from the Long Tail• Users did not know about existence

Serendipity (a.k.a chiripa) is when someone finds somethingthat they weren’t expecting to find.

Page 6: Recommender Systems! @ASAI 2011

Different perspectives cont.

Prediction perspective

• Predict to what degree users like an item• Most popular evaluation scenario in research

Interaction perspective

• Give users a good feeling• Educate users about the product domain• Convince/persuade users - explain

Conversion perspective

• Commercial situations• Increase hit, clickthrough, lookers to bookers rates• Optimize sales margins and profit

Page 7: Recommender Systems! @ASAI 2011

The Long Tail

• Recommend widely unknown items that users might actuallylike!

• 20 % of items accumulate 74 % of all positive ratings

Page 8: Recommender Systems! @ASAI 2011

RS seen as a function

Given:

• User model (e.g. ratings, preferences, demographics,situational context)

• Items (with or without description of item characteristics)

Find:

• Relevance score. Used for ranking.

Page 9: Recommender Systems! @ASAI 2011

Paradigms

Page 10: Recommender Systems! @ASAI 2011

Paradigms of recommender systems

Recommender systems reduce information overload byestimating relevance.

Page 11: Recommender Systems! @ASAI 2011

Paradigms of recommender systems

Personalized recommendations

Page 12: Recommender Systems! @ASAI 2011

Paradigms of recommender systems

Collaborative: Tell me what’s popular among my peers

Page 13: Recommender Systems! @ASAI 2011

Paradigms of recommender systems

Content-based: Show me more of the same what I’ve liked

Page 14: Recommender Systems! @ASAI 2011

Paradigms of recommender systems

Knowledge-based: Tell me what fits based on my needs

Page 15: Recommender Systems! @ASAI 2011

Paradigms of recommender systems

Hybrid: combinations of various inputs and/or composition ofdifferent mechanism

Page 16: Recommender Systems! @ASAI 2011

Collaborative Filtering

Page 17: Recommender Systems! @ASAI 2011

Collaborative Filtering (CF)

The most prominent approach to generate recommendations:• used by large, commercial e-commerce sites• well-understood, various algorithms and variations exist• applicable in many domains (book, movies, DVDs, . . . )

Approach:• use the wisdom of the crowd to recommend items

Basic assumption and idea:• Users give ratings to catalog items (implicitly or explicitly)• Customers who had similar tastes in the past, will have

similar tastes in the future

Page 18: Recommender Systems! @ASAI 2011

User-based nearest-neighbor CF

Given an active user (Alice) and an Item i not yet seen by Alice.The goal is to estimate Alice’s rating for this item, e.g., by:

• find a set of users (peers) who liked the same items as Alice inthe past and who have rated item i

• use, e.g. the average of their ratings to predict, if Alice will likeitem i

• do this for all items Alice has not seen and recommend thebest-rated

Page 19: Recommender Systems! @ASAI 2011

User-based nearest-neighbor CF

Some questions:

• How do we measure similarity?

• How many neighbors should we consider?

• How do we generate a prediction from the neighbors’ ratings?

Utility Matrix:

Item 1 Item 2 Item 3 Item 4 Item 5Alice 5 3 4 4 ?Berta 3 1 2 3 3Clara 4 3 4 3 5Diana 5 1 4 3 4Erika 1 5 5 2 1

Page 20: Recommender Systems! @ASAI 2011

Measuring user similarity

Pearson correlation:

sim(a, b) =

∑p∈P(ra,p − r̄a)(rb,p − r̄b)√∑

p∈P(ra,p − r̄a)2√∑

p∈P(rb,p − r̄b)2

a, b: usersra,p: rating of user a for item pP: set of items, rated both by a and br̄a, r̄b: user’s average ratingsPossible similarity values between -1 and 1

Item 1 Item 2 Item 3 Item 4 Item 5 SimAlice 5 3 4 4 ? -Berta 3 1 2 3 3 0.852Clara 4 3 4 3 5 0.707Diana 5 1 4 3 4 0.956Erika 1 5 5 2 1 -0.792

Page 21: Recommender Systems! @ASAI 2011

Making predictions

• A common prediction function:

pred(a, p) = r̄a +

∑b∈N sim(a, b) · (rb,p − r̄b)∑

b∈N sim(a, b)

• Calculate, whether the neighbors’ ratings for the unseenitem i are higher or lower than their average

• Combine the rating differences - use the similarity with as aweight

• Add/subtract the neighbors’ bias from the active user’saverage and use this as a prediction

Page 22: Recommender Systems! @ASAI 2011

Improving the metrics / prediction function

Not all neighbor ratings might be equally valuable:Agreement on commonly liked items is not so informative asagreement on controversial items.

Value of number of co-rated items:Use significance weighting, by e.g., linearly reducing the weightwhen the number of co-rated items is low.

Case amplification:Give more weight to very similar neighbors, i.e., where thesimilarity value is close to 1.

Neighborhood selection:Use similarity threshold or fixed number of neighbors

Page 23: Recommender Systems! @ASAI 2011

Memory-based and model-based approaches

User-based CF is said to be memory-based:• The rating matrix is directly used to find neighbors / make

predictions• does not scale for most real-world scenarios• large e-commerce sites have tens of millions of customers

and millions of items

Model-based approaches:• based on an offline pre-processing or model-learning phase• at run-time, only the learned model is used to make

predictions• models are updated / re-trained periodically

Page 24: Recommender Systems! @ASAI 2011

Item-based CF

Item-based collaborative filtering recommendation algorithms, B.Sarwar et al., WWW 2001

• Scalability issues arise with U2U if many more users thanitems (m� n,m = |users|,n = |items|)

• High sparsity leads to few common ratings between twousers.

• Basic idea: Item-based CF exploits relationships betweenitems first, instead of relationships between users

Page 25: Recommender Systems! @ASAI 2011

Item-based CF

Use the similarity between items (and not users) to makepredictions.

• Look for items that are similar to Item 5.• Take Alice’s ratings for these items to predict the rating for

Item 5.

Item 1 Item 2 Item 3 Item 4 Item 5Alice 5 3 4 4 ?Berta 3 1 2 3 3Clara 4 3 4 3 5Diana 3 3 1 5 4Erika 1 5 5 2 1

Page 26: Recommender Systems! @ASAI 2011

Item-based CF

The cosine similarity measure:• Produces better results in item-to-item filtering.• Ratings are seen as vector in n-dimensional space.• Similarity is calculated based on the angle between the

vectors.

Pre-processing for item-based filtering:• Calculate all pair-wise item similarities in advance.• The neighborhood to be used at run-time is typically rather

small, because only items are taken into account which theuser has rated.

• Item similarities are supposed to be more stable than usersimilarities.

Page 27: Recommender Systems! @ASAI 2011

More on Ratings

Explicit ratings:• Most commonly used (1 to 5, 1 to 7 Likert response scales)• Optimal scale; 10-point scale is accepted in movie domain.• Multidimensional ratings (multiple ratings per movie).• Users not always willing to rate many items.• How to stimulate users to rate more items?

Implicit ratings:• Clicks, page views, time spent on some page, demo

downloads . . .• Can be used in addition to explicit ones; question of

correctness of interpretation.

Page 28: Recommender Systems! @ASAI 2011

Data sparsity problems

Cold start problem:• How to recommend new items?• What to recommend to new users?

Straightforward approaches:• Ask/force users to rate a set of items.• Use another method (e.g., content-based, demographic or

simply non-personalized) in the initial phase

Alternatives:• Assume transitivity of neighborhoods.• Recursive CF.• Graph-based methods.

Page 29: Recommender Systems! @ASAI 2011

More model-based approaches

Plethora of different techniques proposed in the last years:

• Matrix factorization techniques, statistics.• Singular value decomposition (SVD), principal component

analysis (PCA).

• Association rule mining.• Shopping basket analysis.

• Probabilistic models.• Clustering models, Bayesian networks, probabilistic Latent

Semantic Analysis.

• Various other machine learning approaches.

Page 30: Recommender Systems! @ASAI 2011

Collaborative Filtering Issues

PROs:• Well-understood.• Works well in some domains.• No knowledge engineering required.

CONs:• Requires user community• Sparsity problems.• No integration of other knowledge sources.• No explanation of results.

Page 31: Recommender Systems! @ASAI 2011

Content-basedrecommendation

Page 32: Recommender Systems! @ASAI 2011

Content-based recommendation

Idea: CF methods do not require any information about theitems, it might be reasonable to exploit such information; andrecommend fantasy novels to people who liked fantasy novels inthe past.

What do we need:• some information about the available items such as the

genre (content).• some sort of user profile describing what the user likes (the

preferences).

The task:• learn user preferences.• locate/recommend items that are similar to the user

preferences.

Page 33: Recommender Systems! @ASAI 2011

What is the content?

The genre is actually not part of the content of a book.

• Most CB-recommendation methods originate fromInformation Retrieval (IR) field.

• The goal is to find and rank interesting text documents(news articles, web pages)

• The item descriptions are usually automatically extracted(important words)

Page 34: Recommender Systems! @ASAI 2011

Content representation and item similarities

IR methods:• Compute the similarity of an unseen item with the user

profile based on the keyword overlap

sim(a, b) =2 · |keywords(a) ∩ keywords(b)||keywords(a)|+ |keywords(b)|

• Standard measure: TF-IDF.• Vector space model. Vectors are usually long and sparse.• Remove stopwords, stemming, top-tokens . . .• Cosine similarity.

Page 35: Recommender Systems! @ASAI 2011

Recommending items

Simple method: nearest neighbors• Given a set of documents D already rated by the user

(like/dislike)• Find the n nearest neighbors of a not-yet-seen item i in D• Take these ratings to predict a rating/vote for i• (Variations: neighborhood size, lower/upper similarity

thresholds..)

• Good to model short-term interests / follow-up stories• Used in combination with method to model long-term

preferences

Page 36: Recommender Systems! @ASAI 2011

Recommending items

Query-based retrieval: Rocchio’s method• The SMART System: Users are allowed to rate

(relevant/irrelevant) retrieved documents (feedback)• Document collections D+ and D−

Probabilistic methods:• Recommendation as classical text classification problem.• 2 classes: hot/cold; simple Boolean document

representation.• calculate probability that document is hot/cold based on

Bayes theorem.

Page 37: Recommender Systems! @ASAI 2011

Limitations of CB recommendation methods

Keywords alone may not be sufficient to judge quality/relevanceof a document or web page:• up-to-dateness, usability, aesthetics, writing style• content may also be limited / too short• content may not be automatically extractable (multimedia)

Ramp-up phase required:• Some training data is still required.• Web 2.0: Use other sources to learn the user preferences

Overspecialization:• Algorithms tend to propose more of the same• Or: too similar news items

Page 38: Recommender Systems! @ASAI 2011

Knowledge-BasedRecommender Systems

Page 39: Recommender Systems! @ASAI 2011

KB Recommendation

Explicit domain knowledge:• Sales knowledge elicitation from domain experts• System mimics the behavior of experienced sales assistant• Best-practice sales interactions• Can guarantee correct recommendations (determinism)

with respect to expert knowledge

Conversational interaction strategy:• Opposed to one-shot interaction• Elicitation of user requirements• Transfer of product knowledge educating users phase

Hybridization:• E.g. merging explicit knowledge with community data• Can ensure some policies based on e.g. availability, user

context or profit margin

Page 40: Recommender Systems! @ASAI 2011

KB Recommendation

Constraint-based recommendation:• A knowledge-based RS formulated as constraint satisfaction

problem• What if no solution exists?• Find minimal relaxations

Ask user:• Computation of minimal revisions of requirements.• Do you want to relax your brand preference? Accept

Panasonic instead of Canon brand• Idea that users navigate a product space.

Page 41: Recommender Systems! @ASAI 2011

Conversational strategies

Process consisting of multiple conversational moves:• Resembles natural sales interactions• Not all user requirements known beforehand• Customers are rarely satisfied with the initial

recommendations

Different styles of preference elicitation:• Free text query interface.• Asking technical/generic properties• Images / inspiration• Proposing and Critiquing

Page 42: Recommender Systems! @ASAI 2011

Summary, Issues andChallenges

Page 43: Recommender Systems! @ASAI 2011

Summary

Collaborative Filtering (CF):PROs: Nearly no ramp-up effort, serendipity of results, learns marketsegmentsCONs: Requires some form of rating feedback, cold start for new usersand new items

Content-based:PROs: No community required, comparison between items possibleCONs: Content-descriptions necessary, cold start for new users, nosurprises

Knowledge-based:PROs: Deterministic recommendations, assured quality, no cold-start,can resemble sales dialogue.CONs: Knowledge engineering effort to bootstrap, basically static,does not react to short-term trends.

Page 44: Recommender Systems! @ASAI 2011

Issues and Challenges

Evaluating Recommender Systems

• Is a RS efficient with respect to a specific criteria like accuracy,user satisfaction, response time, serendipity, online conversion,ramp-up efforts, . . . .

• Do customers like/buy recommended items?

• Do customers buy items they otherwise would have not?

• Are they satisfied with a recommendation after purchase?

Recommendation is viewed as IR task.

• Recommendation is viewed as information retrieval task.

• Precision and Recall, F1

F1 = 2 · precision · recallprecision + recall

• Mean Absolute Error (MAE), Root Mean Square Error (RMSE).

Page 45: Recommender Systems! @ASAI 2011

Issues and Challenges

Hybridization Strategies• Idea of crossing two (or more) implementations• Parallel and Pipelined invocation of different systems• Netflix competition - stacking recommender systems

Weighted design based on > 100 predictor

Explanations in recommender systems:• The digital camera Profishot is a must-buy for you because. . .• Additional information to explain the system’s output

following some objectives• Knowledgeable explanations significantly increase the

users’s perceived utility

Page 46: Recommender Systems! @ASAI 2011

Gracias por vuestra !atención !

7puentes.com!

Lic. Ernesto [email protected] - @fetnelio !