Successfully reported this slideshow.
We use your LinkedIn profile and activity data to personalize ads and to show you more relevant ads. You can change your ad preferences anytime.
Past, Present, and Future of
Recommender Systems
Xavier Amatriain (@xamat)
A bit about me...
A bit about me...
A bit about me...
A bit about me...
A bit about me...
A bit about me...
Quora
Our Mission
“To share and grow
the world’s knowledge”
• Millions of questions & answers
• Millions of users
• Thousands of...
Demand
What we care about
Quality
Relevance
1. The Recommender Problem
2. The Netflix Prize
3. Beyond Rating Prediction
4. Lessons Learned
Index
1. The Recommender
Problem
The “Recommender problem”
● Traditional definition: Estimate a utility function that
automatically predicts how much a use...
Recommendation as data mining
The core of the Recommendation
Engine can be assimilated to a
general data mining problem
(A...
Data Mining + all those other things
● User Interface
● System requirements (efficiency, scalability, privacy....)
● Seren...
2. The Netflix Prize
What we were interested in:
▪ High quality recommendations
Proxy question:
▪ Accuracy in predicted rating
▪ Improve by 10%...
2007 Progress Prize
▪ Top 2 algorithms
▪ SVD - Prize RMSE: 0.8914
▪ RBM - Prize RMSE: 0.8990
▪ Linear blend Prize RMSE: 0....
What about the final prize ensembles?
● Offline studies showed they were too computationally
intensive to scale
● Expected...
3. Beyond Rating
Prediction
Everything is a recommendation
Evolution of the Recommender Problem
Rating Ranking Page Optimization
4.7
Context-aware
Recommendations
Context
3.1 Ranking
Ranking
● Most recommendations are presented in a sorted list
● Recommendation can be understood as a ranking problem
● Po...
Ranking by ratings
4.7 4.6 4.5 4.5 4.5 4.5 4.5 4.5 4.5 4.5
Niche titles
High average ratings… by those who would watch it
RMSE
Popularity
PredictedRating
1
2
3
4
5
Linear Model:
frank
(u,v) = w1
p(v) + w2
r(u,v) + b
FinalRanking
Example: Two feature...
Popularity
1
2
3
4
5
FinalRanking
PredictedRatingExample: Two features, linear model
Learning to rank
● Machine learning problem: goal is to construct ranking
model from training data
● Training data can be ...
Learning to rank - Metrics
● Quality of ranking measured using metrics as
o Normalized Discounted Cumulative Gain
o Mean R...
Ranking - Quora Feed
● Goal: Present most interesting
stories for a user at a given time
○ Interesting = topical relevance...
3.2 Similarity
● Displayed in
many different
contexts
○ In response to
user
actions/context
(search, queue
add…)
○ More like… rows
Simila...
Similars: Related Questions
● Given interest in question A (source) what other
questions will be interesting?
● Not only a...
Graph-based similarities
0.8
0.2
0.3
0.4
0.3
0.7
0.3
Example of graph-based similarity: SimRank
● SimRank (Jeh & Widom, 02): “two objects are
similar if they are referenced by...
Similarity ensembles
● Similarity can refer to different dimensions
○ Similar in metadata/tags
○ Similar in user play beha...
3.3 Social
Recommendations
Recommendations - Users
Goal: Recommend new users to follow
● Based on:
○ Other users followed
○ Topics followed
○ User in...
User Trust/Expertise Inference
Goal: Infer user’s trustworthiness in relation
to a given topic
● We take into account:
○ A...
Social and Trust-based recommenders
● A social recommender system recommends items that are “popular”
in the social proxim...
Other ways to use Social
● Social connections can be used in combination with
other approaches
● In particular, “friendshi...
3.4 Explore/Exploit
● One of the key issues when building any kind of
personalization algorithm is how to trade off:
○ Exploitation: Cashing i...
● Given possible strategies/candidates (slot machines) pick the arm that has
the maximum potential of being good (minimize...
Multi-armed Bandits
3.5 Page
Optimization
10,000s of
possible
rows …
10-40
rows
Variable number of
possible videos per
row (up to thousands)
1 personalized page
per...
Page Composition
From “Modeling User Attention and
Interaction on the Web” 2014 - PhD Thesis by Dmitry Lagun (Emory U.)
User Attention Modeling
From “Modeling User Attention and
Interaction on the Web” 2014 - PhD Thesis by Dmitry Lagun (Emory...
vs.Accurate Diverse
vs.Discovery Continuation
vs.Depth Coverage
vs.Freshness Stability
vs.Recommendations Tasks
Page Compo...
4. Lessons Learned
1. Implicitsignalsbeat
explicitones
(almostalways)
Implicit vs. Explicit
● Many have acknowledged
that implicit feedback is more
useful
● Is implicit feedback really always
...
● Implicit data is (usually):
○ More dense, and available for all users
○ Better representative of user behavior vs.
user ...
● However
○ It is not always the case that
direct implicit feedback correlates
well with long-term retention
○ E.g. clickb...
2.bethoughtfulaboutyour
TrainingData
Defining training/testing data
● Training a simple binary classifier for
good/bad answer
○ Defining positive and negative ...
3.YourModelwilllearn
whatyouteachittolearn
Training a model
● Model will learn according to:
○ Training data (e.g. implicit and explicit)
○ Target function (e.g. pro...
Example 2 - Quora’s feed
● Training data = implicit + explicit
● Target function: Value of showing a
story to a
user ~ wei...
4.Explanationsmightmatter
morethantheprediction
Explanation/Support for Recommendations
Social Support
5.Learntodealwith
PresentationBias
2D Navigational modeling
More likely
to see
Less likely
The curse of presentation bias
● User can only click on what you decide to show
○ But, what you decide to show is the resu...
corollary:
TheUIistheonlycommunication
channelwithwhatmattersthemost:
Users
UI->Algorithm->UI
● The UI generates the user feedback
that we will input into the algorithms
● The UI is also where the r...
Conclusions
● Recommendation is about much more than just
predicting a rating
● All forms of recommendation require of a tight
connect...
Questions?
Upcoming SlideShare
Loading in …5
×

Past, present, and future of Recommender Systems: an industry perspective

8,921 views

Published on

Keynote for the ACM Intelligent User Interface conference in 2016 in Sonoma, CA. I start with the past by talking about the Recommender Problem, and the Netflix Prize. Then I go into the Present and the Future by talking about approaches that go beyond rating prediction and ranking and by finishing with some of the most important lessons learned over the years. Throughout my talk I put special emphasis on the relation between algorithms and the User Interface.

Published in: Technology
  • 3 Days Masai Mara Adventure Camping Safari Kenya budget camping safaris Day 1: Nairobi/Masai Mara Pick up from your hotel in Nairobi at 8.00 am and head southward to the viewpoint of the Great Rift Valley. Continue to Narok a small Masaai town then to Masaai Mara for lunch followed by afternoon game viewing to see the many animals that inhabit this unique park. Dinner and overnight adventure camping. Day 2: Masaai Mara After breakfast spend full day game drives, explore the parks amazingly high concentration of game. This game reserve is Kenya's finest wildlife sanctuary including migrants, well over 450 species of animals have been recorded here although July, August and September are the months when the Mara plains are filled with migrating wildebeest and zebra, there is also resident wildlife year round. The famous Maasai tribesmen live within the dispersal area of the game reserve lunch is served at the picnic site in the park. Dinner and overnight adventure camping. Day 3: Masaai Mara/Nairobi An early morning game drive in Masai Mara followed by late breakfast at the campsite and a departure to Nairobi stopping sporadically for the breathtaking dragon adventure views of the rift valley and adjacent landforms. The tour ends in Nairobi late in the afternoon. WHAT’S INCLUDED ON THE PRICE. -Transport based on our customized seven-seater safari Vehicle with specially modified suspension, full-length photographic roof hatches and seats designed with seat belt and all clients are guaranteed window seats. - All game Park Entrance fees - All adventure Camping Fees -Full board accommodation offered in a large adventure tents, all en suite bathrooms with hot and cold shower, flash toilet. - Three meals a day while on your adventure camping Safari - Airport pick up on arrival in Nairobi. - Use of our Professional Driver Guide. - 3 litres Mineral water for drinking. Total cost is USD 295 per person, for more information, call : +254724297698 ADVENTURE DRAGON SAFARIS www.adventuredargonsafaris.com
       Reply 
    Are you sure you want to  Yes  No
    Your message goes here
  • My life is back!!! After 1 years of Broken marriage, my husband left me with two kids . I felt like my life was about to end i almost committed suicide, i was emotionally down for a very long time. Thanks to a spell caster called Dr Mohammed, which i met online. On one faithful day, as I was browsing through the internet,I came across allot of testimonies about this particular spell caster. Some people testified that he brought their Ex lover back, some testified that he restores womb,cure cancer,and other sickness, some testified that he can cast a spell to stop divorce and so on. i also come across one particular testimony,it was about a woman called Sonia,she testified about how he brought back her Ex lover in less than 2 days, and at the end of her testimony she dropped Dr Mohammed's e-mail address. After reading all these,I decided to give it a try. I contacted him via email and explained my problem to him. In just 48hours, my husband came back to me. We solved our issues, and we are even happier than before Dr Mohammed, is really a gifted man and i will not stop publishing him because he is a wonderful man... If you have a problem and you are looking for a real and genuine spell caster to solve all your problems for you. Try High monicaspiritualtemple@gmail.com anytime, he might be the answer to your problems. Here's his contact: monicaspiritualtemple@gmail.com or call him +2348134493948
       Reply 
    Are you sure you want to  Yes  No
    Your message goes here

Past, present, and future of Recommender Systems: an industry perspective

  1. 1. Past, Present, and Future of Recommender Systems Xavier Amatriain (@xamat)
  2. 2. A bit about me...
  3. 3. A bit about me...
  4. 4. A bit about me...
  5. 5. A bit about me...
  6. 6. A bit about me...
  7. 7. A bit about me...
  8. 8. Quora
  9. 9. Our Mission “To share and grow the world’s knowledge” • Millions of questions & answers • Millions of users • Thousands of topics • ...
  10. 10. Demand What we care about Quality Relevance
  11. 11. 1. The Recommender Problem 2. The Netflix Prize 3. Beyond Rating Prediction 4. Lessons Learned Index
  12. 12. 1. The Recommender Problem
  13. 13. The “Recommender problem” ● Traditional definition: Estimate a utility function that automatically predicts how much a user will like an item. ● Based on: o Past behavior o Relations to other users o Item similarity o Context o …
  14. 14. Recommendation as data mining The core of the Recommendation Engine can be assimilated to a general data mining problem (Amatriain et al. Data Mining Methods for Recommender Systems in Recommender Systems Handbook)
  15. 15. Data Mining + all those other things ● User Interface ● System requirements (efficiency, scalability, privacy....) ● Serendipity ● Diversity ● Awareness ● Explanations ● …
  16. 16. 2. The Netflix Prize
  17. 17. What we were interested in: ▪ High quality recommendations Proxy question: ▪ Accuracy in predicted rating ▪ Improve by 10% = $1million!
  18. 18. 2007 Progress Prize ▪ Top 2 algorithms ▪ SVD - Prize RMSE: 0.8914 ▪ RBM - Prize RMSE: 0.8990 ▪ Linear blend Prize RMSE: 0.88 ▪ Currently in use as part of Netflix’ rating prediction component ▪ Limitations ▪ Designed for 100M ratings, not XB ratings ▪ Not adaptable as users add ratings ▪ Performance issues
  19. 19. What about the final prize ensembles? ● Offline studies showed they were too computationally intensive to scale ● Expected improvement not worth engineering effort ● Plus…. Focus had already shifted to other issues that had more impact than rating prediction.
  20. 20. 3. Beyond Rating Prediction
  21. 21. Everything is a recommendation
  22. 22. Evolution of the Recommender Problem Rating Ranking Page Optimization 4.7 Context-aware Recommendations Context
  23. 23. 3.1 Ranking
  24. 24. Ranking ● Most recommendations are presented in a sorted list ● Recommendation can be understood as a ranking problem ● Popularity is the obvious baseline ● What about rating predictions?
  25. 25. Ranking by ratings 4.7 4.6 4.5 4.5 4.5 4.5 4.5 4.5 4.5 4.5 Niche titles High average ratings… by those who would watch it
  26. 26. RMSE
  27. 27. Popularity PredictedRating 1 2 3 4 5 Linear Model: frank (u,v) = w1 p(v) + w2 r(u,v) + b FinalRanking Example: Two features, linear model
  28. 28. Popularity 1 2 3 4 5 FinalRanking PredictedRatingExample: Two features, linear model
  29. 29. Learning to rank ● Machine learning problem: goal is to construct ranking model from training data ● Training data can be a partial order or binary judgments (relevant/not relevant). ● Resulting order of the items typically induced from a numerical score ● Learning to rank is a key element for personalization ● You can treat the problem as a standard supervised classification problem
  30. 30. Learning to rank - Metrics ● Quality of ranking measured using metrics as o Normalized Discounted Cumulative Gain o Mean Reciprocal Rank (MRR) o Fraction of Concordant Pairs (FCP) o Others… ● But, it is hard to optimize machine-learned models directly on these measures (e.g. non-differentiable) ● Recent research on models that directly optimize ranking measures
  31. 31. Ranking - Quora Feed ● Goal: Present most interesting stories for a user at a given time ○ Interesting = topical relevance + social relevance + timeliness ○ Stories = questions + answers ● ML: Personalized learning-to-rank approach ● Relevance-ordered vs time-ordered = big gains in engagement
  32. 32. 3.2 Similarity
  33. 33. ● Displayed in many different contexts ○ In response to user actions/context (search, queue add…) ○ More like… rows Similars
  34. 34. Similars: Related Questions ● Given interest in question A (source) what other questions will be interesting? ● Not only about similarity, but also “interestingness” ● Features such as: ○ Textual ○ Co-visit ○ Topics ○ … ● Important for logged-out use case
  35. 35. Graph-based similarities 0.8 0.2 0.3 0.4 0.3 0.7 0.3
  36. 36. Example of graph-based similarity: SimRank ● SimRank (Jeh & Widom, 02): “two objects are similar if they are referenced by similar objects.”
  37. 37. Similarity ensembles ● Similarity can refer to different dimensions ○ Similar in metadata/tags ○ Similar in user play behavior ○ Similar in user rating behavior ○ … ● Combine them using an ensemble ○ Weights are learned using regression over existing response ○ Or… some MAB explore/exploit approach ● The final concept of “similarity” responds to what users vote as similar
  38. 38. 3.3 Social Recommendations
  39. 39. Recommendations - Users Goal: Recommend new users to follow ● Based on: ○ Other users followed ○ Topics followed ○ User interactions ○ User-related features ○ ...
  40. 40. User Trust/Expertise Inference Goal: Infer user’s trustworthiness in relation to a given topic ● We take into account: ○ Answers written on topic ○ Upvotes/downvotes received ○ Endorsements ○ ... ● Trust/expertise propagates through the network ● Must be taken into account by other algorithms
  41. 41. Social and Trust-based recommenders ● A social recommender system recommends items that are “popular” in the social proximity of the user. ● Social proximity = trust (can also be topic-specific) ● Given two individuals - the source (node A) and sink (node C) - derive how much the source should trust the sink. ● Algorithms o Advogato (Levien) o Appleseed (Ziegler and Lausen) o MoleTrust (Massa and Avesani) o TidalTrust (Golbeck)
  42. 42. Other ways to use Social ● Social connections can be used in combination with other approaches ● In particular, “friendships” can be fed into collaborative filtering methods in different ways ○ replace or modify user-user “similarity” by using social network information ○ use social connection as a part of the ML objective function as regularizer ○ ...
  43. 43. 3.4 Explore/Exploit
  44. 44. ● One of the key issues when building any kind of personalization algorithm is how to trade off: ○ Exploitation: Cashing in on what we know about the user right now ○ Exploration: Using the interaction as an opportunity to learn more about the user ● We need to have informed and optimal strategies to drive that tradeoff ○ Solution: pick a reasonable set of candidates and show users only “enough” to gather information on them Explore/Exploit
  45. 45. ● Given possible strategies/candidates (slot machines) pick the arm that has the maximum potential of being good (minimize regret) ● Naive strategy: ε-greedy ○ Explore with a small probability ε (e.g. 5%) -> choose an arm at random ○ Exploit with a high probability (1 - ε ) (e.g. 95%) -> choose the best-known arm so far ● Translation to recommender systems ○ Choose an arm = choose an item/choose an algorithm (MAB testing) ● Thompson Sampling ○ Given a posterior distribution, sample on each iteration and choose the action that maximizes the expected reward Multi-armed Bandits
  46. 46. Multi-armed Bandits
  47. 47. 3.5 Page Optimization
  48. 48. 10,000s of possible rows … 10-40 rows Variable number of possible videos per row (up to thousands) 1 personalized page per device Page Composition
  49. 49. Page Composition From “Modeling User Attention and Interaction on the Web” 2014 - PhD Thesis by Dmitry Lagun (Emory U.)
  50. 50. User Attention Modeling From “Modeling User Attention and Interaction on the Web” 2014 - PhD Thesis by Dmitry Lagun (Emory U.)
  51. 51. vs.Accurate Diverse vs.Discovery Continuation vs.Depth Coverage vs.Freshness Stability vs.Recommendations Tasks Page Composition ● To put things together we need to combine different elements o Navigational/Attention Model o Personalized Relevance Model o Diversity Model
  52. 52. 4. Lessons Learned
  53. 53. 1. Implicitsignalsbeat explicitones (almostalways)
  54. 54. Implicit vs. Explicit ● Many have acknowledged that implicit feedback is more useful ● Is implicit feedback really always more useful? ● If so, why?
  55. 55. ● Implicit data is (usually): ○ More dense, and available for all users ○ Better representative of user behavior vs. user reflection ○ More related to final objective function ○ Better correlated with AB test results ● E.g. Rating vs watching Implicit vs. Explicit
  56. 56. ● However ○ It is not always the case that direct implicit feedback correlates well with long-term retention ○ E.g. clickbait ● Solution: ○ Combine different forms of implicit + explicit to better represent long-term goal Implicit vs. Explicit
  57. 57. 2.bethoughtfulaboutyour TrainingData
  58. 58. Defining training/testing data ● Training a simple binary classifier for good/bad answer ○ Defining positive and negative labels -> Non-trivial task ○ Is this a positive or a negative? ■ funny uninformative answer with many upvotes ■ short uninformative answer by a well-known expert in the field ■ very long informative answer that nobody reads/upvotes ■ informative answer with grammar/spelling mistakes ■ ...
  59. 59. 3.YourModelwilllearn whatyouteachittolearn
  60. 60. Training a model ● Model will learn according to: ○ Training data (e.g. implicit and explicit) ○ Target function (e.g. probability of user reading an answer) ○ Metric (e.g. precision vs. recall) ● Example 1 (made up): ○ Optimize probability of a user going to the cinema to watch a movie and rate it “highly” by using purchase history and previous ratings. Use NDCG of the ranking as final metric using only movies rated 4 or higher as positives.
  61. 61. Example 2 - Quora’s feed ● Training data = implicit + explicit ● Target function: Value of showing a story to a user ~ weighted sum of actions: v = ∑a va 1{ya = 1} ○ predict probabilities for each action, then compute expected value: v_pred = E[ V | x ] = ∑a va p(a | x) ● Metric: any ranking metric
  62. 62. 4.Explanationsmightmatter morethantheprediction
  63. 63. Explanation/Support for Recommendations Social Support
  64. 64. 5.Learntodealwith PresentationBias
  65. 65. 2D Navigational modeling More likely to see Less likely
  66. 66. The curse of presentation bias ● User can only click on what you decide to show ○ But, what you decide to show is the result of what your model predicted is good ● Simply treating things you show as negatives is not likely to work ● Better options ○ Correcting for the probability a user will click on a position -> Attention models ○ Explore/exploit approaches such as MAB
  67. 67. corollary: TheUIistheonlycommunication channelwithwhatmattersthemost: Users
  68. 68. UI->Algorithm->UI ● The UI generates the user feedback that we will input into the algorithms ● The UI is also where the results of our algorithms will be shown ● A change in the UI might require a change in algorithms and viceversa
  69. 69. Conclusions
  70. 70. ● Recommendation is about much more than just predicting a rating ● All forms of recommendation require of a tight connection with the UI ○ Capture the right kind of feedback ■ Explicit/implicit feedback ■ Correct for presentation bias ■ ... ○ Present the recommendations correctly ■ Explanations ■ Diversity ■ Exploration/Exploitation ■ ….
  71. 71. Questions?

×