Successfully reported this slideshow.
We use your LinkedIn profile and activity data to personalize ads and to show you more relevant ads. You can change your ad preferences anytime.
acm recsys 2012:recommender systems, today@neal_lathia
warning:daunting task   lookout for twitter handles
why #recsys?      information overloadmailing lists; usenet news (1992)            see: @jkonstan, @presnick
why #recsys?    information overload         filter failuremovies; books; music (~1995)
why #recsys?           information overload                 filter failure               creating valueadvertising; engage...
@dtunkelang
(1) collaborative “based on the premise that people looking forinformation should be able to make use of what   others hav...
(2) query-less“in September 2010 Schmidt said that one day the     combination of cloud computing and mobile       phones ...
(3) discovery engines“we are leaving the age of information and  entering the age of recommendation”                 (ande...
input: ratings, clicks, views              users → items      process: SVD, kNN, RBM, etc.    f(user, item) → prediction ~...
traditional problemsaccuracy, scalability, distributed computation,           similarity, cold-start, …              (dont...
acm recsys 2012:5 open problems
problem 1: predictionstemporality, multiple co-occurring objectives:  diversity, novelty, freshness, serendipity,         ...
problem 2: algorithmsmore algorithms vs. more data    vs. more rating effort
what is your algorithm doing?      f(user, item) → R  f(user, item1, item2) → R f(user, [item1...itemn]) → R              ...
problem 3: users + ratingssignals, context, groups, intents, interfaces
@xamat
problem 4: itemslifestyle, behaviours, decisions, processes,            software development
@presnick
problem 5: measurementranking metrics vs. usability testing         vs. A/B testing
Online Controlled Experiments: Introduction, Learnings, and HumblingStatisticshttp://www.exp-platform.com/Pages/2012RecSys...
3 key lessons
lesson 1: #recsys is an ensemble         ...of disciplines  statistics, machine learning,  human-computer interaction,    ...
lesson 2: return to the domainuser effort, generative models, cost of a freakommendation,   value you seek to create
@plamere
lesson 3: join the #recsys community   learn, build, research, deploy:   @MyMediaLite, @LensKitRS     @zenogantner, @eleha...
recommender systems, today@neal_lathia
Recommender Systems in 2012
Recommender Systems in 2012
Recommender Systems in 2012
Upcoming SlideShare
Loading in …5
×

Recommender Systems in 2012

5,136 views

Published on

Talk at Data Science London Meetup on Recommender Systems

Recommender Systems in 2012

  1. 1. acm recsys 2012:recommender systems, today@neal_lathia
  2. 2. warning:daunting task lookout for twitter handles
  3. 3. why #recsys? information overloadmailing lists; usenet news (1992) see: @jkonstan, @presnick
  4. 4. why #recsys? information overload filter failuremovies; books; music (~1995)
  5. 5. why #recsys? information overload filter failure creating valueadvertising; engagement; connection (today)
  6. 6. @dtunkelang
  7. 7. (1) collaborative “based on the premise that people looking forinformation should be able to make use of what others have already found and evaluated” (maltz & ehrlick)
  8. 8. (2) query-less“in September 2010 Schmidt said that one day the combination of cloud computing and mobile phones would allow Google to pass on information to users without them even typing in search queries”
  9. 9. (3) discovery engines“we are leaving the age of information and entering the age of recommendation” (anderson)
  10. 10. input: ratings, clicks, views users → items process: SVD, kNN, RBM, etc. f(user, item) → prediction ~ ratingoutput: prediction-ranked recommendations measure: |prediction – rating| 2 (prediction – rating)
  11. 11. traditional problemsaccuracy, scalability, distributed computation, similarity, cold-start, … (dont reinvent the wheel)
  12. 12. acm recsys 2012:5 open problems
  13. 13. problem 1: predictionstemporality, multiple co-occurring objectives: diversity, novelty, freshness, serendipity, explainability
  14. 14. problem 2: algorithmsmore algorithms vs. more data vs. more rating effort
  15. 15. what is your algorithm doing? f(user, item) → R f(user, item1, item2) → R f(user, [item1...itemn]) → R e.g., @alexk_z @abellogin
  16. 16. problem 3: users + ratingssignals, context, groups, intents, interfaces
  17. 17. @xamat
  18. 18. problem 4: itemslifestyle, behaviours, decisions, processes, software development
  19. 19. @presnick
  20. 20. problem 5: measurementranking metrics vs. usability testing vs. A/B testing
  21. 21. Online Controlled Experiments: Introduction, Learnings, and HumblingStatisticshttp://www.exp-platform.com/Pages/2012RecSys.aspx
  22. 22. 3 key lessons
  23. 23. lesson 1: #recsys is an ensemble ...of disciplines statistics, machine learning, human-computer interaction, social network analysis, psychology
  24. 24. lesson 2: return to the domainuser effort, generative models, cost of a freakommendation, value you seek to create
  25. 25. @plamere
  26. 26. lesson 3: join the #recsys community learn, build, research, deploy: @MyMediaLite, @LensKitRS @zenogantner, @elehack contribute, read: #recsyswiki, @alansaid
  27. 27. recommender systems, today@neal_lathia

×