SlideShare a Scribd company logo
1 of 27
Download to read offline
Local Collaborative Autoencoders
Minjin Choi1, Yoonki Jeong1, Joonseok Lee2, Jongwuk Lee1
Sungkyunkwan University (SKKU), South Korea1
Google Research, United States2
Motivation
2
Global Low-rank Assumption
➢Existing models are based on the global low-lank assumption.
• All users and items share the same latent features.
➢Limitation: some users/items may have different latent features.
3
1 1 ? ? 1
1 1 1 ? ?
1 1 ? 1 ?
? 1 1 1 ?
? 1 1 ? 1
0.5 ... 0.1
0.6 ... 0.2
0.7 ... 0.3
… ... ...
0.1 ... 0.7
0.2 0.4 1.2 … 2.3
… ... … ... ...
0.4 0.6 0.3 … 1.5
k features
k
features
Local Low-rank Assumption
➢A user-item matrix can be divided to several sub-matrices
with the local low-rank assumption.
• Each sub-matrix represents different communities.
• Local models represent various communities with different characteristics.
4
1 1 ? ? 1
1 1 1 ? ?
1 1 ? 1 ?
? 1 1 1 ?
? 1 1 ? 1
Limitation of Existing Local Models
➢If the local model is too large, it is close to the global model.
• Because LLORMA uses large local models, the local model may not
represent its unique characteristic.
• The performance gain may come from an ensemble effect.
5
1 1 ? ? 1
1 1 1 ? ?
1 1 ? 1 ?
? 1 1 1 ?
? 1 1 ? 1
1 1 ? ? 1
1 1 1 ? ?
1 1 ? 1 ?
? 1 1 1 ?
? 1 1 ? 1
Lee et al., "LLORMA: Local Low-Rank Matrix Approximation," JMLR 2016
Limitation of Existing Local Models
➢If the local model is too small, the accuracy is too low.
• Because sGLSVD uses small local models, some local model may have
insufficient training data.
6
1 1 ? ? 1
1 1 1 ? ?
1 1 ? 1 ?
? 1 1 1 ?
? 1 1 ? 1
Evangelia Christakopoulou and George Karypis, “Local Latent Space Models for Top-N Recommendation,” KDD 2018
1 1 ? ? 1
1 1 1 ? ?
1 1 ? 1 ?
? 1 1 1 ?
? 1 1 ? 1
Research Question
7
How to build coherent and accurate local models?
Our Key Contributions
➢When the local model keeps small and coherent,
we build the local model with a relatively large training data.
8
(Dataset: ML10M)
Training matrix
== Test matrix
training matrix
> test matrix
Our Key Contributions
➢Autoencoder-based models are used as the base model
to train the local model.
• They are useful for capturing non-linear and complicated patterns.
9
1 1 ? ? 1
1 1 1 ? ?
1 1 ? 1 ?
? 1 1 1 ?
? 1 1 ? 1
Sedhain et al., "AutoRec: Autoencoders Meet Collaborative Filtering," WWW 2015
2 1 3
…
2 1 3
…
…
𝒉(𝒓)
𝒓
ො
𝒓
𝑾, 𝒃
𝑾′
, 𝒃′
Autoencoder-based models take
each row as input.
Proposed Method
10
Local Collaborative Autoencoders (LOCA)
➢Overall architecture of LOCA
• Step 1: Discovering two local communities for an anchor user
• Step 2: Training a local model with an expanded community
• Step 3: Combining multiple local models
11
A local community
for an anchor user
An expanded local
community
Local model q
…
Final model
Local model 2
Local model 1
Step 1: Discovering Local Communities
➢For an anchor user, determine a local community
and expand the local community for training.
12
1.0
0.6
0.3
0.5
0.4
0.7
0.4
0.2
Calculate the similarities between the
anchor user and the other users.
…
…
1.0
0.7
0.6
1.0
0.7
0.6
0.5
0.5
0.4
0.4
Neighbors for the
local community
Expanded neighbors to
train the local community
Step 2: Training a Local Model
➢Train the local model with the expanded community.
• Use the autoencoder-based model for training the local model.
• Note: It is possible to utilize any base models, e.g., MF, AE and EASER.
• The similarities with the anchor are used for the user weights for training.
13
1.0
0.6
0.3
0.5
0.4
0.7
0.4
1 1 ? 1
1 1 1 1
1 1 ? ?
1 1 ? 1
1 ? ? ?
1 1 1 1
1 1 ? ?
…
…
…
…
…
…
…
…
Harald Steck., "Embarrassingly shallow autoencoders for sparse data,” WWW 2019.
Step 3: Combining Multiple Local Models
➢Aggregate multiple local models with the weight of
each local model.
14
Local model 1
Local model q
…
1 1 ? 1
1 1 1 1
1 1 ? ?
1 1 ? 1
1 ? ? ?
1 1 1 1
1 1 ? ?
? 1 1 ?
…
…
…
…
…
…
…
…
…
…
…
…
…
…
…
.9 .4 .5 .7
.8 .9 .6 .6
.7 .8 .5 .8
…
.7 .4 .5 .9
.7 .5 .5 .8
.8 .6 .7 .8
…
…
…
.9 .4 .5 .7
.8 .9 .6 .6
.7 .8 .5 .8
.8 .7 .5 1
.7 .4 .5 .9
.7 .5 .5 .8
.8 .6 .7 .8
.6 1 1 .5
…
…
…
…
…
…
…
…
…
…
…
…
…
…
…
…
…
…
…
…
…
Training Local Models in Detail
➢The loss function for the 𝑗-th local model
➢Aggregating all local models and a global model
15
argmin
𝜽(𝒋)
෍
𝒓𝒖∈𝑹
𝒕𝒖
(𝒋)
ℒ 𝒓𝒖, 𝑴𝒍𝒐𝒄𝒂𝒍
(𝒓𝒖; 𝜽(𝒋)
) + 𝝀𝛀(𝜽(𝒋)
)
෡
𝑹 = 𝜶𝑴𝒈𝒍𝒐𝒃𝒂𝒍 𝑹; 𝜽 𝒈 + 𝟏 − 𝜶 ෍
𝒋=𝟏
𝒒
𝒘(𝒋) ⊙ 𝑴𝒍𝒐𝒄𝒂𝒍(𝑹; 𝜽(𝒋)) ⊘ 𝒘
Parameter for the
𝒋-th local model
The 𝒋-th local model
User weight for training the 𝒋-th local model
User weight for inferring the 𝒋-th local model
The global model is used to handle the users
who are not covered by local models.
How to Choose Anchor Users
➢Random selection: Choose 𝑘 anchor users at random.
➢Need to maximize the coverage of users by local models.
• An optimal maximum coverage algorithm is the NP-hard problem.
➢Use the greedy method to maximize the coverage.
• Select the anchor user who has the most uncovered users iteratively.
16
Group1
Group 2
Group 3
Experiments
17
Experimental Setup: Dataset
➢We evaluate our model over five public datasets
with various characteristics (e.g., domain, sparsity).
18
Dataset # of users # of items # of ratings Sparsity
MovieLens 10M
(ML10M)
69,878 10,677 10,000,054 98.66%
MovieLens 20M
(ML20M)
138,493 26,744 20,000,263 99.46%
Amazon Music
(AMusic)
4,964 11,797 97,439 99.83%
Amazon Game
(AGame)
13,063 17,408 236,415 99.90%
Yelp 25,677 25,815 731,671 99.89%
Evaluation Protocol and Metrics
➢Evaluation protocol: leave-5-out
• Hold-out the last 5 interactions as the test data for each user.
➢Evaluation metrics
• Recall@100
• Measures the number of test items included in the top-N list.
• NDCG@100
• Measures the ranking of test items in the top-N list.
19
Time
Training data:
Test data:
User interaction:
Competitive Global/Local Models
➢Four autoencoder-based global models
• CDAE: a denoising autoencoder-based model with a latent user vector
• MultVAE: a VAE-based model
• RecVAE: a VAE-based model by improving MultVAE
• EASER: an item-to-item latent factor model
➢Two local models
• LLORMA: local model using MF as the base model
• sGLSVD: local model using SVD as the base model
20
Yao Wu et al., “Collaborative Denoising Auto-Encoders for Top-N Recommender Systems,” WSDM 2016
Dawen Liang et al., "Variational autoencoders for collaborative filtering,” WWW 2018.
Ilya Shenbin et al., "RecVAE: A new variational autoencoder for Top-N recommendations with implicit feedback,” WSDM 2020.
Harald Steck., "Embarrassingly shallow autoencoders for sparse data,” WWW 2019.
Lee et al., "LLORMA: Local Low-Rank Matrix Approximation," JMLR 2016
Evangelia Christakopoulou and George Karypis, “Local Latent Space Models for Top-N Recommendation,” KDD 2018
Accuracy: LOCA vs. Competing Models
21
Dataset Metric CDAE MultVAE EASER RecVAE LLORMA sGLSVD LOCAVAE LOCAEASE
ML10M
Recall@100 0.4685 0.4653 0.4648 0.4705 0.4692 0.4468 0.4865 0.4798
NDCG@100 0.1982 0.1945 0.2000 0.1996 0.2042 0.1953 0.2073 0.2049
ML20M
Recall@100 0.4324 0.4397 0.4468 0.4417 0.3355 0.4342 0.4419 0.4654
NDCG@100 0.1844 0.1860 0.1948 0.1857 0.1446 0.1919 0.1884 0.2024
Amusic
Recall@100 0.0588 0.0681 0.0717 0.0582 0.0517 0.0515 0.0748 0.0717
NDCG@100 0.712 0.0822 0.0821 0.0810 0.0638 0.0613 0.0893 0.0826
Agames
Recall@100 0.1825 0.2081 0.1913 0.1920 0.1223 0.1669 0.2147 0.1947
NDCG@100 0.0808 0.0920 0.0915 0.0849 0.0539 0.0777 0.0966 0.0922
Yelp
Recall@100 0.2094 0.2276 0.2187 0.2262 0.1013 0.1965 0.2354 0.2205
NDCG@100 0.0920 0.0982 0.0972 0.0975 0.0429 0.0857 0.1103 0.0981
Global Models Local Models Ours
➢LOCA consistently outperforms competitive
global/local models over five benchmark datasets.
Effect of the Number of Local Models
➢The accuracy of LOCA improved consistently with
an increase in the number of local models.
22
0.185
0.19
0.195
0.2
0.205
0.21
0 50 100 150 200 250 300
NDCG@100
The number of local models
ML10M
MultVAE Ensemble_VAE
LLORMA_VAE LOCA_VAE
0.078
0.08
0.082
0.084
0.086
0.088
0.09
0 50 100 150 200 250 300
NDCG@100 The number of local models
AMusic
MultVAE Ensemble_VAE
LLORMA_VAE LOCA_VAE
Effect of Anchor Selection Method
➢Our coverage-based anchor selection outperforms
the other methods in terms of accuracies and coverages.
23
0.196
0.198
0.2
0.202
0.204
0.206
0.208
50 100 150 200 250 300
NDCG@100
The number of local models
ML10M
Random K-means Ours
0.08
0.082
0.084
0.086
0.088
0.09
50 100 150 200 250 300
NDCG@100 The number of local models
AMusic
Random K-means Ours
Illustration of LOCA
➢When a user has multiple tastes, LOCA can capture
the user preference by combining different local patterns.
• For a user (66005 in ML10M) who likes Sci-Fi and Horror movies, LOCA
shows a better accuracy.
24
Recommendation Local 70 Local 179 Global Ground Truth
Top-1
Sci-Fi
Adventure
Horror
Action
Thriller
Action
Sci-Fi
Action
Top-2
Sci-Fi
Horror
Horror
Drama
Drama
Horror
Thriller
Top-3
Sci-Fi
Action
Horror
Drama
Drama
Mystery
Horror
Action
Conclusion
25
Conclusion
➢We propose a new local recommender framework,
namely local collaborative autoencoders (LOCA).
➢LOCA can handle a large number of local models effectively.
• Adopts a local model with different training/inference strategies.
• Utilizes autoencoder-based model as the base model.
• Makes use of a greedy maximum coverage method to build various
local models.
➢LOCA outperforms the state-of-the-art global and local models
over various benchmark datasets.
26
Q&A
27
Code: https://github.com/jin530/LOCA
Email: zxcvxd@skku.edu

More Related Content

What's hot

Music Recommendation 2018
Music Recommendation 2018Music Recommendation 2018
Music Recommendation 2018Fabien Gouyon
 
Recent advances in deep recommender systems
Recent advances in deep recommender systemsRecent advances in deep recommender systems
Recent advances in deep recommender systemsNAVER Engineering
 
The Science and the Magic of User Feedback for Recommender Systems
The Science and the Magic of User Feedback for Recommender SystemsThe Science and the Magic of User Feedback for Recommender Systems
The Science and the Magic of User Feedback for Recommender SystemsXavier Amatriain
 
Chapter 02 collaborative recommendation
Chapter 02   collaborative recommendationChapter 02   collaborative recommendation
Chapter 02 collaborative recommendationAravindharamanan S
 
Introduction of Deep Reinforcement Learning
Introduction of Deep Reinforcement LearningIntroduction of Deep Reinforcement Learning
Introduction of Deep Reinforcement LearningNAVER Engineering
 
NuMicro® M480 Series Introduction - Sept.2019
NuMicro® M480 Series Introduction - Sept.2019NuMicro® M480 Series Introduction - Sept.2019
NuMicro® M480 Series Introduction - Sept.2019Mason Lyu
 
Deep Learning for Recommender Systems RecSys2017 Tutorial
Deep Learning for Recommender Systems RecSys2017 Tutorial Deep Learning for Recommender Systems RecSys2017 Tutorial
Deep Learning for Recommender Systems RecSys2017 Tutorial Alexandros Karatzoglou
 
Facebook Talk at Netflix ML Platform meetup Sep 2019
Facebook Talk at Netflix ML Platform meetup Sep 2019Facebook Talk at Netflix ML Platform meetup Sep 2019
Facebook Talk at Netflix ML Platform meetup Sep 2019Faisal Siddiqi
 
Introduction to deep learning
Introduction to deep learningIntroduction to deep learning
Introduction to deep learningJunaid Bhat
 
[Phd Thesis Defense] CHAMELEON: A Deep Learning Meta-Architecture for News Re...
[Phd Thesis Defense] CHAMELEON: A Deep Learning Meta-Architecture for News Re...[Phd Thesis Defense] CHAMELEON: A Deep Learning Meta-Architecture for News Re...
[Phd Thesis Defense] CHAMELEON: A Deep Learning Meta-Architecture for News Re...Gabriel Moreira
 
Bài tập thực hành số 1
Bài tập thực hành số 1Bài tập thực hành số 1
Bài tập thực hành số 1Tran Trung Dung
 
Introduction to Recommendation Systems
Introduction to Recommendation SystemsIntroduction to Recommendation Systems
Introduction to Recommendation SystemsTrieu Nguyen
 
Recsys2016 Tutorial by Xavier and Deepak
Recsys2016 Tutorial by Xavier and DeepakRecsys2016 Tutorial by Xavier and Deepak
Recsys2016 Tutorial by Xavier and DeepakDeepak Agarwal
 
Mã hóa đường cong Elliptic
Mã hóa đường cong EllipticMã hóa đường cong Elliptic
Mã hóa đường cong EllipticLE Ngoc Luyen
 
The Case for Learned Index Structures
The Case for Learned Index StructuresThe Case for Learned Index Structures
The Case for Learned Index Structures宇 傅
 
Loss functions (DLAI D4L2 2017 UPC Deep Learning for Artificial Intelligence)
Loss functions (DLAI D4L2 2017 UPC Deep Learning for Artificial Intelligence)Loss functions (DLAI D4L2 2017 UPC Deep Learning for Artificial Intelligence)
Loss functions (DLAI D4L2 2017 UPC Deep Learning for Artificial Intelligence)Universitat Politècnica de Catalunya
 

What's hot (20)

Music Recommendation 2018
Music Recommendation 2018Music Recommendation 2018
Music Recommendation 2018
 
Recent advances in deep recommender systems
Recent advances in deep recommender systemsRecent advances in deep recommender systems
Recent advances in deep recommender systems
 
The Science and the Magic of User Feedback for Recommender Systems
The Science and the Magic of User Feedback for Recommender SystemsThe Science and the Magic of User Feedback for Recommender Systems
The Science and the Magic of User Feedback for Recommender Systems
 
Chapter 02 collaborative recommendation
Chapter 02   collaborative recommendationChapter 02   collaborative recommendation
Chapter 02 collaborative recommendation
 
Recommendation system
Recommendation systemRecommendation system
Recommendation system
 
Introduction of Deep Reinforcement Learning
Introduction of Deep Reinforcement LearningIntroduction of Deep Reinforcement Learning
Introduction of Deep Reinforcement Learning
 
NuMicro® M480 Series Introduction - Sept.2019
NuMicro® M480 Series Introduction - Sept.2019NuMicro® M480 Series Introduction - Sept.2019
NuMicro® M480 Series Introduction - Sept.2019
 
Deep Learning for Recommender Systems RecSys2017 Tutorial
Deep Learning for Recommender Systems RecSys2017 Tutorial Deep Learning for Recommender Systems RecSys2017 Tutorial
Deep Learning for Recommender Systems RecSys2017 Tutorial
 
Role of Data Science in eCommerce
Role of Data Science in eCommerceRole of Data Science in eCommerce
Role of Data Science in eCommerce
 
Facebook Talk at Netflix ML Platform meetup Sep 2019
Facebook Talk at Netflix ML Platform meetup Sep 2019Facebook Talk at Netflix ML Platform meetup Sep 2019
Facebook Talk at Netflix ML Platform meetup Sep 2019
 
Introduction to deep learning
Introduction to deep learningIntroduction to deep learning
Introduction to deep learning
 
eScience SHAP talk
eScience SHAP talkeScience SHAP talk
eScience SHAP talk
 
[Phd Thesis Defense] CHAMELEON: A Deep Learning Meta-Architecture for News Re...
[Phd Thesis Defense] CHAMELEON: A Deep Learning Meta-Architecture for News Re...[Phd Thesis Defense] CHAMELEON: A Deep Learning Meta-Architecture for News Re...
[Phd Thesis Defense] CHAMELEON: A Deep Learning Meta-Architecture for News Re...
 
Bài tập thực hành số 1
Bài tập thực hành số 1Bài tập thực hành số 1
Bài tập thực hành số 1
 
Introduction to Recommendation Systems
Introduction to Recommendation SystemsIntroduction to Recommendation Systems
Introduction to Recommendation Systems
 
Open-ended Visual Question-Answering
Open-ended  Visual Question-AnsweringOpen-ended  Visual Question-Answering
Open-ended Visual Question-Answering
 
Recsys2016 Tutorial by Xavier and Deepak
Recsys2016 Tutorial by Xavier and DeepakRecsys2016 Tutorial by Xavier and Deepak
Recsys2016 Tutorial by Xavier and Deepak
 
Mã hóa đường cong Elliptic
Mã hóa đường cong EllipticMã hóa đường cong Elliptic
Mã hóa đường cong Elliptic
 
The Case for Learned Index Structures
The Case for Learned Index StructuresThe Case for Learned Index Structures
The Case for Learned Index Structures
 
Loss functions (DLAI D4L2 2017 UPC Deep Learning for Artificial Intelligence)
Loss functions (DLAI D4L2 2017 UPC Deep Learning for Artificial Intelligence)Loss functions (DLAI D4L2 2017 UPC Deep Learning for Artificial Intelligence)
Loss functions (DLAI D4L2 2017 UPC Deep Learning for Artificial Intelligence)
 

Similar to Local collaborative autoencoders (WSDM2021)

RS in the context of Big Data-v4
RS in the context of Big Data-v4RS in the context of Big Data-v4
RS in the context of Big Data-v4Khadija Atiya
 
Using Metrics for Fun, Developing with the KV Store + Javascript & News from ...
Using Metrics for Fun, Developing with the KV Store + Javascript & News from ...Using Metrics for Fun, Developing with the KV Store + Javascript & News from ...
Using Metrics for Fun, Developing with the KV Store + Javascript & News from ...Harry McLaren
 
Lessons learnt at building recommendation services at industry scale
Lessons learnt at building recommendation services at industry scaleLessons learnt at building recommendation services at industry scale
Lessons learnt at building recommendation services at industry scaleDomonkos Tikk
 
@RISK Unchained Webinar
@RISK Unchained Webinar@RISK Unchained Webinar
@RISK Unchained WebinarAndrew Sich
 
Why biased matrix factorization works well?
Why biased matrix factorization works well?Why biased matrix factorization works well?
Why biased matrix factorization works well?Joonyoung Yi
 
Explain Yourself: Why You Get the Recommendations You Do
Explain Yourself: Why You Get the Recommendations You DoExplain Yourself: Why You Get the Recommendations You Do
Explain Yourself: Why You Get the Recommendations You DoDatabricks
 
Towards a Quality Assessment of Web Corpora for Language Technology Applications
Towards a Quality Assessment of Web Corpora for Language Technology ApplicationsTowards a Quality Assessment of Web Corpora for Language Technology Applications
Towards a Quality Assessment of Web Corpora for Language Technology ApplicationsMarina Santini
 
NLP-Focused Applied ML at Scale for Global Fleet Analytics at ExxonMobil
NLP-Focused Applied ML at Scale for Global Fleet Analytics at ExxonMobilNLP-Focused Applied ML at Scale for Global Fleet Analytics at ExxonMobil
NLP-Focused Applied ML at Scale for Global Fleet Analytics at ExxonMobilDatabricks
 
Deep Learning for Recommender Systems
Deep Learning for Recommender SystemsDeep Learning for Recommender Systems
Deep Learning for Recommender SystemsMarcel Kurovski
 
Deep Learning for Recommender Systems
Deep Learning for Recommender SystemsDeep Learning for Recommender Systems
Deep Learning for Recommender Systemsinovex GmbH
 
It Does What You Say, Not What You Mean: Lessons From A Decade of Program Repair
It Does What You Say, Not What You Mean: Lessons From A Decade of Program RepairIt Does What You Say, Not What You Mean: Lessons From A Decade of Program Repair
It Does What You Say, Not What You Mean: Lessons From A Decade of Program RepairClaire Le Goues
 
Using Bayesian Optimization to Tune Machine Learning Models
Using Bayesian Optimization to Tune Machine Learning ModelsUsing Bayesian Optimization to Tune Machine Learning Models
Using Bayesian Optimization to Tune Machine Learning ModelsScott Clark
 
Using Bayesian Optimization to Tune Machine Learning Models
Using Bayesian Optimization to Tune Machine Learning ModelsUsing Bayesian Optimization to Tune Machine Learning Models
Using Bayesian Optimization to Tune Machine Learning ModelsSigOpt
 
Machine Learning, Deep Learning and Data Analysis Introduction
Machine Learning, Deep Learning and Data Analysis IntroductionMachine Learning, Deep Learning and Data Analysis Introduction
Machine Learning, Deep Learning and Data Analysis IntroductionTe-Yen Liu
 
In the age of Big Data, what role for Software Engineers?
In the age of Big Data, what role for Software Engineers?In the age of Big Data, what role for Software Engineers?
In the age of Big Data, what role for Software Engineers?CS, NcState
 
Using SigOpt to Tune Deep Learning Models with Nervana Cloud
Using SigOpt to Tune Deep Learning Models with Nervana CloudUsing SigOpt to Tune Deep Learning Models with Nervana Cloud
Using SigOpt to Tune Deep Learning Models with Nervana CloudSigOpt
 
Moving Toward Deep Learning Algorithms on HPCC Systems
Moving Toward Deep Learning Algorithms on HPCC SystemsMoving Toward Deep Learning Algorithms on HPCC Systems
Moving Toward Deep Learning Algorithms on HPCC SystemsHPCC Systems
 
Building a Location Based Social Graph in Spark at InMobi-(Seinjuti Chatterje...
Building a Location Based Social Graph in Spark at InMobi-(Seinjuti Chatterje...Building a Location Based Social Graph in Spark at InMobi-(Seinjuti Chatterje...
Building a Location Based Social Graph in Spark at InMobi-(Seinjuti Chatterje...Spark Summit
 
Performance evaluation of IR models
Performance evaluation of IR modelsPerformance evaluation of IR models
Performance evaluation of IR modelsNisha Arankandath
 

Similar to Local collaborative autoencoders (WSDM2021) (20)

RS in the context of Big Data-v4
RS in the context of Big Data-v4RS in the context of Big Data-v4
RS in the context of Big Data-v4
 
Using Metrics for Fun, Developing with the KV Store + Javascript & News from ...
Using Metrics for Fun, Developing with the KV Store + Javascript & News from ...Using Metrics for Fun, Developing with the KV Store + Javascript & News from ...
Using Metrics for Fun, Developing with the KV Store + Javascript & News from ...
 
Lessons learnt at building recommendation services at industry scale
Lessons learnt at building recommendation services at industry scaleLessons learnt at building recommendation services at industry scale
Lessons learnt at building recommendation services at industry scale
 
Machine Learning Impact on IoT - Part 2
Machine Learning Impact on IoT - Part 2Machine Learning Impact on IoT - Part 2
Machine Learning Impact on IoT - Part 2
 
@RISK Unchained Webinar
@RISK Unchained Webinar@RISK Unchained Webinar
@RISK Unchained Webinar
 
Why biased matrix factorization works well?
Why biased matrix factorization works well?Why biased matrix factorization works well?
Why biased matrix factorization works well?
 
Explain Yourself: Why You Get the Recommendations You Do
Explain Yourself: Why You Get the Recommendations You DoExplain Yourself: Why You Get the Recommendations You Do
Explain Yourself: Why You Get the Recommendations You Do
 
Towards a Quality Assessment of Web Corpora for Language Technology Applications
Towards a Quality Assessment of Web Corpora for Language Technology ApplicationsTowards a Quality Assessment of Web Corpora for Language Technology Applications
Towards a Quality Assessment of Web Corpora for Language Technology Applications
 
NLP-Focused Applied ML at Scale for Global Fleet Analytics at ExxonMobil
NLP-Focused Applied ML at Scale for Global Fleet Analytics at ExxonMobilNLP-Focused Applied ML at Scale for Global Fleet Analytics at ExxonMobil
NLP-Focused Applied ML at Scale for Global Fleet Analytics at ExxonMobil
 
Deep Learning for Recommender Systems
Deep Learning for Recommender SystemsDeep Learning for Recommender Systems
Deep Learning for Recommender Systems
 
Deep Learning for Recommender Systems
Deep Learning for Recommender SystemsDeep Learning for Recommender Systems
Deep Learning for Recommender Systems
 
It Does What You Say, Not What You Mean: Lessons From A Decade of Program Repair
It Does What You Say, Not What You Mean: Lessons From A Decade of Program RepairIt Does What You Say, Not What You Mean: Lessons From A Decade of Program Repair
It Does What You Say, Not What You Mean: Lessons From A Decade of Program Repair
 
Using Bayesian Optimization to Tune Machine Learning Models
Using Bayesian Optimization to Tune Machine Learning ModelsUsing Bayesian Optimization to Tune Machine Learning Models
Using Bayesian Optimization to Tune Machine Learning Models
 
Using Bayesian Optimization to Tune Machine Learning Models
Using Bayesian Optimization to Tune Machine Learning ModelsUsing Bayesian Optimization to Tune Machine Learning Models
Using Bayesian Optimization to Tune Machine Learning Models
 
Machine Learning, Deep Learning and Data Analysis Introduction
Machine Learning, Deep Learning and Data Analysis IntroductionMachine Learning, Deep Learning and Data Analysis Introduction
Machine Learning, Deep Learning and Data Analysis Introduction
 
In the age of Big Data, what role for Software Engineers?
In the age of Big Data, what role for Software Engineers?In the age of Big Data, what role for Software Engineers?
In the age of Big Data, what role for Software Engineers?
 
Using SigOpt to Tune Deep Learning Models with Nervana Cloud
Using SigOpt to Tune Deep Learning Models with Nervana CloudUsing SigOpt to Tune Deep Learning Models with Nervana Cloud
Using SigOpt to Tune Deep Learning Models with Nervana Cloud
 
Moving Toward Deep Learning Algorithms on HPCC Systems
Moving Toward Deep Learning Algorithms on HPCC SystemsMoving Toward Deep Learning Algorithms on HPCC Systems
Moving Toward Deep Learning Algorithms on HPCC Systems
 
Building a Location Based Social Graph in Spark at InMobi-(Seinjuti Chatterje...
Building a Location Based Social Graph in Spark at InMobi-(Seinjuti Chatterje...Building a Location Based Social Graph in Spark at InMobi-(Seinjuti Chatterje...
Building a Location Based Social Graph in Spark at InMobi-(Seinjuti Chatterje...
 
Performance evaluation of IR models
Performance evaluation of IR modelsPerformance evaluation of IR models
Performance evaluation of IR models
 

Recently uploaded

Grafana in space: Monitoring Japan's SLIM moon lander in real time
Grafana in space: Monitoring Japan's SLIM moon lander  in real timeGrafana in space: Monitoring Japan's SLIM moon lander  in real time
Grafana in space: Monitoring Japan's SLIM moon lander in real timeSatoshi NAKAHIRA
 
Analytical Profile of Coleus Forskohlii | Forskolin .pptx
Analytical Profile of Coleus Forskohlii | Forskolin .pptxAnalytical Profile of Coleus Forskohlii | Forskolin .pptx
Analytical Profile of Coleus Forskohlii | Forskolin .pptxSwapnil Therkar
 
All-domain Anomaly Resolution Office U.S. Department of Defense (U) Case: “Eg...
All-domain Anomaly Resolution Office U.S. Department of Defense (U) Case: “Eg...All-domain Anomaly Resolution Office U.S. Department of Defense (U) Case: “Eg...
All-domain Anomaly Resolution Office U.S. Department of Defense (U) Case: “Eg...Sérgio Sacani
 
Biopesticide (2).pptx .This slides helps to know the different types of biop...
Biopesticide (2).pptx  .This slides helps to know the different types of biop...Biopesticide (2).pptx  .This slides helps to know the different types of biop...
Biopesticide (2).pptx .This slides helps to know the different types of biop...RohitNehra6
 
Recombination DNA Technology (Nucleic Acid Hybridization )
Recombination DNA Technology (Nucleic Acid Hybridization )Recombination DNA Technology (Nucleic Acid Hybridization )
Recombination DNA Technology (Nucleic Acid Hybridization )aarthirajkumar25
 
GFP in rDNA Technology (Biotechnology).pptx
GFP in rDNA Technology (Biotechnology).pptxGFP in rDNA Technology (Biotechnology).pptx
GFP in rDNA Technology (Biotechnology).pptxAleenaTreesaSaji
 
Call Girls in Mayapuri Delhi 💯Call Us 🔝9953322196🔝 💯Escort.
Call Girls in Mayapuri Delhi 💯Call Us 🔝9953322196🔝 💯Escort.Call Girls in Mayapuri Delhi 💯Call Us 🔝9953322196🔝 💯Escort.
Call Girls in Mayapuri Delhi 💯Call Us 🔝9953322196🔝 💯Escort.aasikanpl
 
SOLUBLE PATTERN RECOGNITION RECEPTORS.pptx
SOLUBLE PATTERN RECOGNITION RECEPTORS.pptxSOLUBLE PATTERN RECOGNITION RECEPTORS.pptx
SOLUBLE PATTERN RECOGNITION RECEPTORS.pptxkessiyaTpeter
 
Cultivation of KODO MILLET . made by Ghanshyam pptx
Cultivation of KODO MILLET . made by Ghanshyam pptxCultivation of KODO MILLET . made by Ghanshyam pptx
Cultivation of KODO MILLET . made by Ghanshyam pptxpradhanghanshyam7136
 
Animal Communication- Auditory and Visual.pptx
Animal Communication- Auditory and Visual.pptxAnimal Communication- Auditory and Visual.pptx
Animal Communication- Auditory and Visual.pptxUmerFayaz5
 
Nightside clouds and disequilibrium chemistry on the hot Jupiter WASP-43b
Nightside clouds and disequilibrium chemistry on the hot Jupiter WASP-43bNightside clouds and disequilibrium chemistry on the hot Jupiter WASP-43b
Nightside clouds and disequilibrium chemistry on the hot Jupiter WASP-43bSérgio Sacani
 
Orientation, design and principles of polyhouse
Orientation, design and principles of polyhouseOrientation, design and principles of polyhouse
Orientation, design and principles of polyhousejana861314
 
Boyles law module in the grade 10 science
Boyles law module in the grade 10 scienceBoyles law module in the grade 10 science
Boyles law module in the grade 10 sciencefloriejanemacaya1
 
Scheme-of-Work-Science-Stage-4 cambridge science.docx
Scheme-of-Work-Science-Stage-4 cambridge science.docxScheme-of-Work-Science-Stage-4 cambridge science.docx
Scheme-of-Work-Science-Stage-4 cambridge science.docxyaramohamed343013
 
Natural Polymer Based Nanomaterials
Natural Polymer Based NanomaterialsNatural Polymer Based Nanomaterials
Natural Polymer Based NanomaterialsAArockiyaNisha
 
Luciferase in rDNA technology (biotechnology).pptx
Luciferase in rDNA technology (biotechnology).pptxLuciferase in rDNA technology (biotechnology).pptx
Luciferase in rDNA technology (biotechnology).pptxAleenaTreesaSaji
 
Spermiogenesis or Spermateleosis or metamorphosis of spermatid
Spermiogenesis or Spermateleosis or metamorphosis of spermatidSpermiogenesis or Spermateleosis or metamorphosis of spermatid
Spermiogenesis or Spermateleosis or metamorphosis of spermatidSarthak Sekhar Mondal
 
Nanoparticles synthesis and characterization​ ​
Nanoparticles synthesis and characterization​  ​Nanoparticles synthesis and characterization​  ​
Nanoparticles synthesis and characterization​ ​kaibalyasahoo82800
 
Bentham & Hooker's Classification. along with the merits and demerits of the ...
Bentham & Hooker's Classification. along with the merits and demerits of the ...Bentham & Hooker's Classification. along with the merits and demerits of the ...
Bentham & Hooker's Classification. along with the merits and demerits of the ...Nistarini College, Purulia (W.B) India
 

Recently uploaded (20)

Grafana in space: Monitoring Japan's SLIM moon lander in real time
Grafana in space: Monitoring Japan's SLIM moon lander  in real timeGrafana in space: Monitoring Japan's SLIM moon lander  in real time
Grafana in space: Monitoring Japan's SLIM moon lander in real time
 
Analytical Profile of Coleus Forskohlii | Forskolin .pptx
Analytical Profile of Coleus Forskohlii | Forskolin .pptxAnalytical Profile of Coleus Forskohlii | Forskolin .pptx
Analytical Profile of Coleus Forskohlii | Forskolin .pptx
 
All-domain Anomaly Resolution Office U.S. Department of Defense (U) Case: “Eg...
All-domain Anomaly Resolution Office U.S. Department of Defense (U) Case: “Eg...All-domain Anomaly Resolution Office U.S. Department of Defense (U) Case: “Eg...
All-domain Anomaly Resolution Office U.S. Department of Defense (U) Case: “Eg...
 
Biopesticide (2).pptx .This slides helps to know the different types of biop...
Biopesticide (2).pptx  .This slides helps to know the different types of biop...Biopesticide (2).pptx  .This slides helps to know the different types of biop...
Biopesticide (2).pptx .This slides helps to know the different types of biop...
 
Recombination DNA Technology (Nucleic Acid Hybridization )
Recombination DNA Technology (Nucleic Acid Hybridization )Recombination DNA Technology (Nucleic Acid Hybridization )
Recombination DNA Technology (Nucleic Acid Hybridization )
 
GFP in rDNA Technology (Biotechnology).pptx
GFP in rDNA Technology (Biotechnology).pptxGFP in rDNA Technology (Biotechnology).pptx
GFP in rDNA Technology (Biotechnology).pptx
 
Call Girls in Mayapuri Delhi 💯Call Us 🔝9953322196🔝 💯Escort.
Call Girls in Mayapuri Delhi 💯Call Us 🔝9953322196🔝 💯Escort.Call Girls in Mayapuri Delhi 💯Call Us 🔝9953322196🔝 💯Escort.
Call Girls in Mayapuri Delhi 💯Call Us 🔝9953322196🔝 💯Escort.
 
SOLUBLE PATTERN RECOGNITION RECEPTORS.pptx
SOLUBLE PATTERN RECOGNITION RECEPTORS.pptxSOLUBLE PATTERN RECOGNITION RECEPTORS.pptx
SOLUBLE PATTERN RECOGNITION RECEPTORS.pptx
 
Cultivation of KODO MILLET . made by Ghanshyam pptx
Cultivation of KODO MILLET . made by Ghanshyam pptxCultivation of KODO MILLET . made by Ghanshyam pptx
Cultivation of KODO MILLET . made by Ghanshyam pptx
 
Animal Communication- Auditory and Visual.pptx
Animal Communication- Auditory and Visual.pptxAnimal Communication- Auditory and Visual.pptx
Animal Communication- Auditory and Visual.pptx
 
Nightside clouds and disequilibrium chemistry on the hot Jupiter WASP-43b
Nightside clouds and disequilibrium chemistry on the hot Jupiter WASP-43bNightside clouds and disequilibrium chemistry on the hot Jupiter WASP-43b
Nightside clouds and disequilibrium chemistry on the hot Jupiter WASP-43b
 
Orientation, design and principles of polyhouse
Orientation, design and principles of polyhouseOrientation, design and principles of polyhouse
Orientation, design and principles of polyhouse
 
Boyles law module in the grade 10 science
Boyles law module in the grade 10 scienceBoyles law module in the grade 10 science
Boyles law module in the grade 10 science
 
Scheme-of-Work-Science-Stage-4 cambridge science.docx
Scheme-of-Work-Science-Stage-4 cambridge science.docxScheme-of-Work-Science-Stage-4 cambridge science.docx
Scheme-of-Work-Science-Stage-4 cambridge science.docx
 
Natural Polymer Based Nanomaterials
Natural Polymer Based NanomaterialsNatural Polymer Based Nanomaterials
Natural Polymer Based Nanomaterials
 
Luciferase in rDNA technology (biotechnology).pptx
Luciferase in rDNA technology (biotechnology).pptxLuciferase in rDNA technology (biotechnology).pptx
Luciferase in rDNA technology (biotechnology).pptx
 
Spermiogenesis or Spermateleosis or metamorphosis of spermatid
Spermiogenesis or Spermateleosis or metamorphosis of spermatidSpermiogenesis or Spermateleosis or metamorphosis of spermatid
Spermiogenesis or Spermateleosis or metamorphosis of spermatid
 
Nanoparticles synthesis and characterization​ ​
Nanoparticles synthesis and characterization​  ​Nanoparticles synthesis and characterization​  ​
Nanoparticles synthesis and characterization​ ​
 
The Philosophy of Science
The Philosophy of ScienceThe Philosophy of Science
The Philosophy of Science
 
Bentham & Hooker's Classification. along with the merits and demerits of the ...
Bentham & Hooker's Classification. along with the merits and demerits of the ...Bentham & Hooker's Classification. along with the merits and demerits of the ...
Bentham & Hooker's Classification. along with the merits and demerits of the ...
 

Local collaborative autoencoders (WSDM2021)

  • 1. Local Collaborative Autoencoders Minjin Choi1, Yoonki Jeong1, Joonseok Lee2, Jongwuk Lee1 Sungkyunkwan University (SKKU), South Korea1 Google Research, United States2
  • 3. Global Low-rank Assumption ➢Existing models are based on the global low-lank assumption. • All users and items share the same latent features. ➢Limitation: some users/items may have different latent features. 3 1 1 ? ? 1 1 1 1 ? ? 1 1 ? 1 ? ? 1 1 1 ? ? 1 1 ? 1 0.5 ... 0.1 0.6 ... 0.2 0.7 ... 0.3 … ... ... 0.1 ... 0.7 0.2 0.4 1.2 … 2.3 … ... … ... ... 0.4 0.6 0.3 … 1.5 k features k features
  • 4. Local Low-rank Assumption ➢A user-item matrix can be divided to several sub-matrices with the local low-rank assumption. • Each sub-matrix represents different communities. • Local models represent various communities with different characteristics. 4 1 1 ? ? 1 1 1 1 ? ? 1 1 ? 1 ? ? 1 1 1 ? ? 1 1 ? 1
  • 5. Limitation of Existing Local Models ➢If the local model is too large, it is close to the global model. • Because LLORMA uses large local models, the local model may not represent its unique characteristic. • The performance gain may come from an ensemble effect. 5 1 1 ? ? 1 1 1 1 ? ? 1 1 ? 1 ? ? 1 1 1 ? ? 1 1 ? 1 1 1 ? ? 1 1 1 1 ? ? 1 1 ? 1 ? ? 1 1 1 ? ? 1 1 ? 1 Lee et al., "LLORMA: Local Low-Rank Matrix Approximation," JMLR 2016
  • 6. Limitation of Existing Local Models ➢If the local model is too small, the accuracy is too low. • Because sGLSVD uses small local models, some local model may have insufficient training data. 6 1 1 ? ? 1 1 1 1 ? ? 1 1 ? 1 ? ? 1 1 1 ? ? 1 1 ? 1 Evangelia Christakopoulou and George Karypis, “Local Latent Space Models for Top-N Recommendation,” KDD 2018 1 1 ? ? 1 1 1 1 ? ? 1 1 ? 1 ? ? 1 1 1 ? ? 1 1 ? 1
  • 7. Research Question 7 How to build coherent and accurate local models?
  • 8. Our Key Contributions ➢When the local model keeps small and coherent, we build the local model with a relatively large training data. 8 (Dataset: ML10M) Training matrix == Test matrix training matrix > test matrix
  • 9. Our Key Contributions ➢Autoencoder-based models are used as the base model to train the local model. • They are useful for capturing non-linear and complicated patterns. 9 1 1 ? ? 1 1 1 1 ? ? 1 1 ? 1 ? ? 1 1 1 ? ? 1 1 ? 1 Sedhain et al., "AutoRec: Autoencoders Meet Collaborative Filtering," WWW 2015 2 1 3 … 2 1 3 … … 𝒉(𝒓) 𝒓 ො 𝒓 𝑾, 𝒃 𝑾′ , 𝒃′ Autoencoder-based models take each row as input.
  • 11. Local Collaborative Autoencoders (LOCA) ➢Overall architecture of LOCA • Step 1: Discovering two local communities for an anchor user • Step 2: Training a local model with an expanded community • Step 3: Combining multiple local models 11 A local community for an anchor user An expanded local community Local model q … Final model Local model 2 Local model 1
  • 12. Step 1: Discovering Local Communities ➢For an anchor user, determine a local community and expand the local community for training. 12 1.0 0.6 0.3 0.5 0.4 0.7 0.4 0.2 Calculate the similarities between the anchor user and the other users. … … 1.0 0.7 0.6 1.0 0.7 0.6 0.5 0.5 0.4 0.4 Neighbors for the local community Expanded neighbors to train the local community
  • 13. Step 2: Training a Local Model ➢Train the local model with the expanded community. • Use the autoencoder-based model for training the local model. • Note: It is possible to utilize any base models, e.g., MF, AE and EASER. • The similarities with the anchor are used for the user weights for training. 13 1.0 0.6 0.3 0.5 0.4 0.7 0.4 1 1 ? 1 1 1 1 1 1 1 ? ? 1 1 ? 1 1 ? ? ? 1 1 1 1 1 1 ? ? … … … … … … … … Harald Steck., "Embarrassingly shallow autoencoders for sparse data,” WWW 2019.
  • 14. Step 3: Combining Multiple Local Models ➢Aggregate multiple local models with the weight of each local model. 14 Local model 1 Local model q … 1 1 ? 1 1 1 1 1 1 1 ? ? 1 1 ? 1 1 ? ? ? 1 1 1 1 1 1 ? ? ? 1 1 ? … … … … … … … … … … … … … … … .9 .4 .5 .7 .8 .9 .6 .6 .7 .8 .5 .8 … .7 .4 .5 .9 .7 .5 .5 .8 .8 .6 .7 .8 … … … .9 .4 .5 .7 .8 .9 .6 .6 .7 .8 .5 .8 .8 .7 .5 1 .7 .4 .5 .9 .7 .5 .5 .8 .8 .6 .7 .8 .6 1 1 .5 … … … … … … … … … … … … … … … … … … … … …
  • 15. Training Local Models in Detail ➢The loss function for the 𝑗-th local model ➢Aggregating all local models and a global model 15 argmin 𝜽(𝒋) ෍ 𝒓𝒖∈𝑹 𝒕𝒖 (𝒋) ℒ 𝒓𝒖, 𝑴𝒍𝒐𝒄𝒂𝒍 (𝒓𝒖; 𝜽(𝒋) ) + 𝝀𝛀(𝜽(𝒋) ) ෡ 𝑹 = 𝜶𝑴𝒈𝒍𝒐𝒃𝒂𝒍 𝑹; 𝜽 𝒈 + 𝟏 − 𝜶 ෍ 𝒋=𝟏 𝒒 𝒘(𝒋) ⊙ 𝑴𝒍𝒐𝒄𝒂𝒍(𝑹; 𝜽(𝒋)) ⊘ 𝒘 Parameter for the 𝒋-th local model The 𝒋-th local model User weight for training the 𝒋-th local model User weight for inferring the 𝒋-th local model The global model is used to handle the users who are not covered by local models.
  • 16. How to Choose Anchor Users ➢Random selection: Choose 𝑘 anchor users at random. ➢Need to maximize the coverage of users by local models. • An optimal maximum coverage algorithm is the NP-hard problem. ➢Use the greedy method to maximize the coverage. • Select the anchor user who has the most uncovered users iteratively. 16 Group1 Group 2 Group 3
  • 18. Experimental Setup: Dataset ➢We evaluate our model over five public datasets with various characteristics (e.g., domain, sparsity). 18 Dataset # of users # of items # of ratings Sparsity MovieLens 10M (ML10M) 69,878 10,677 10,000,054 98.66% MovieLens 20M (ML20M) 138,493 26,744 20,000,263 99.46% Amazon Music (AMusic) 4,964 11,797 97,439 99.83% Amazon Game (AGame) 13,063 17,408 236,415 99.90% Yelp 25,677 25,815 731,671 99.89%
  • 19. Evaluation Protocol and Metrics ➢Evaluation protocol: leave-5-out • Hold-out the last 5 interactions as the test data for each user. ➢Evaluation metrics • Recall@100 • Measures the number of test items included in the top-N list. • NDCG@100 • Measures the ranking of test items in the top-N list. 19 Time Training data: Test data: User interaction:
  • 20. Competitive Global/Local Models ➢Four autoencoder-based global models • CDAE: a denoising autoencoder-based model with a latent user vector • MultVAE: a VAE-based model • RecVAE: a VAE-based model by improving MultVAE • EASER: an item-to-item latent factor model ➢Two local models • LLORMA: local model using MF as the base model • sGLSVD: local model using SVD as the base model 20 Yao Wu et al., “Collaborative Denoising Auto-Encoders for Top-N Recommender Systems,” WSDM 2016 Dawen Liang et al., "Variational autoencoders for collaborative filtering,” WWW 2018. Ilya Shenbin et al., "RecVAE: A new variational autoencoder for Top-N recommendations with implicit feedback,” WSDM 2020. Harald Steck., "Embarrassingly shallow autoencoders for sparse data,” WWW 2019. Lee et al., "LLORMA: Local Low-Rank Matrix Approximation," JMLR 2016 Evangelia Christakopoulou and George Karypis, “Local Latent Space Models for Top-N Recommendation,” KDD 2018
  • 21. Accuracy: LOCA vs. Competing Models 21 Dataset Metric CDAE MultVAE EASER RecVAE LLORMA sGLSVD LOCAVAE LOCAEASE ML10M Recall@100 0.4685 0.4653 0.4648 0.4705 0.4692 0.4468 0.4865 0.4798 NDCG@100 0.1982 0.1945 0.2000 0.1996 0.2042 0.1953 0.2073 0.2049 ML20M Recall@100 0.4324 0.4397 0.4468 0.4417 0.3355 0.4342 0.4419 0.4654 NDCG@100 0.1844 0.1860 0.1948 0.1857 0.1446 0.1919 0.1884 0.2024 Amusic Recall@100 0.0588 0.0681 0.0717 0.0582 0.0517 0.0515 0.0748 0.0717 NDCG@100 0.712 0.0822 0.0821 0.0810 0.0638 0.0613 0.0893 0.0826 Agames Recall@100 0.1825 0.2081 0.1913 0.1920 0.1223 0.1669 0.2147 0.1947 NDCG@100 0.0808 0.0920 0.0915 0.0849 0.0539 0.0777 0.0966 0.0922 Yelp Recall@100 0.2094 0.2276 0.2187 0.2262 0.1013 0.1965 0.2354 0.2205 NDCG@100 0.0920 0.0982 0.0972 0.0975 0.0429 0.0857 0.1103 0.0981 Global Models Local Models Ours ➢LOCA consistently outperforms competitive global/local models over five benchmark datasets.
  • 22. Effect of the Number of Local Models ➢The accuracy of LOCA improved consistently with an increase in the number of local models. 22 0.185 0.19 0.195 0.2 0.205 0.21 0 50 100 150 200 250 300 NDCG@100 The number of local models ML10M MultVAE Ensemble_VAE LLORMA_VAE LOCA_VAE 0.078 0.08 0.082 0.084 0.086 0.088 0.09 0 50 100 150 200 250 300 NDCG@100 The number of local models AMusic MultVAE Ensemble_VAE LLORMA_VAE LOCA_VAE
  • 23. Effect of Anchor Selection Method ➢Our coverage-based anchor selection outperforms the other methods in terms of accuracies and coverages. 23 0.196 0.198 0.2 0.202 0.204 0.206 0.208 50 100 150 200 250 300 NDCG@100 The number of local models ML10M Random K-means Ours 0.08 0.082 0.084 0.086 0.088 0.09 50 100 150 200 250 300 NDCG@100 The number of local models AMusic Random K-means Ours
  • 24. Illustration of LOCA ➢When a user has multiple tastes, LOCA can capture the user preference by combining different local patterns. • For a user (66005 in ML10M) who likes Sci-Fi and Horror movies, LOCA shows a better accuracy. 24 Recommendation Local 70 Local 179 Global Ground Truth Top-1 Sci-Fi Adventure Horror Action Thriller Action Sci-Fi Action Top-2 Sci-Fi Horror Horror Drama Drama Horror Thriller Top-3 Sci-Fi Action Horror Drama Drama Mystery Horror Action
  • 26. Conclusion ➢We propose a new local recommender framework, namely local collaborative autoencoders (LOCA). ➢LOCA can handle a large number of local models effectively. • Adopts a local model with different training/inference strategies. • Utilizes autoencoder-based model as the base model. • Makes use of a greedy maximum coverage method to build various local models. ➢LOCA outperforms the state-of-the-art global and local models over various benchmark datasets. 26