Review: [SIGIR'22]Interpolative Distillation for Unifying Biased and Debiased Recommendation
1. Interpolative Distillation for Unifying Biased
and Debiased Recommendation
SIGIR’22, Sihao Ding(USTC) et al.
POSTECH DI Lab
Presenter: Changsoo Kwak
2022.5.24
1
2. Motivation
2
▪ Most recommender system’s test set for evaluate
▪ Normal biased test set(𝐷𝑏)
▪ Debiased test set (𝐷𝑑)
[1] Self-supervised Graph Learning for Recommendation, Jiancan Wu(USTC) et al, SIGIR’21
Existing models didn’t perform well on both test set
Biased or Unbiased model
Only reflect part of whole picture
3. Intuitive solution?
3
▪ Unifying 𝐷𝑏, 𝐷𝑑
▪ Usually 𝐷𝑏 ≫ |𝐷𝑑|
▪ Train two models for 𝐷𝑏, 𝐷𝑑 respectively, and ensemble
▪ Unclear that each models are strong/weak at which type of users/items
▪ Existing ensemble strategies are not tailored for win-win recommendation scenario
▪ Possible solution?
▪ Distillation!
▪ Aggregate two models at the level of user-item pair
Determine coefficient automatically for distillation
4. Proposed model(InterD)
4
Environment 𝐸 ∈ {𝑒𝑏, 𝑒𝑑}
Probability of environment given user-item pair
Existing models only consider one environment
- Only achieve good performance on one of 𝐷𝑏 or 𝐷𝑑
Predicted rating with given environment assumption
Let student model learns predicted ratings generated by
fine-grained weighted sum of prediction of pre-trained
models, considering environment
RCT: Randomized Control Trial(https://books.google.co.kr/books?id=JUTqDwAAQBAJ&pg=PA244&lpg=PA244&dq=yahoo!r3+randomized+controlled+trial&source=bl&ots=0cagKMc4KG&sig=ACfU3U3oFb-FZsxO3PuYDFYRz6gX9O97tA&hl=ko&sa=X&ved=2ahUKEwj5qp-psev3AhWim1YBHfVgC2QQ6AF6BAgDEAM#v=onepage&q=yahoo!r3%20randomized%20controlled%20trial&f=false)
In other words, the student tends to learn the easier aspects of knowledge since the smaller distance makes it easier to follow the corresponding teacher
학생 입장에서 더 쉬운 쪽(거리가 적은 쪽 teacher)을 따라가기 때문에 curriculum learning으로 볼 수도 있다?
Weight 계산에 student prediction이 들어가니까 self-paced learning으,로 볼 수도 있다?