Fairness in Search & RecSys 네이버 검색 콜로키움 김진영Jin Young Kim
검색 및 추천 시스템의 사회적 역할이 커지면서, 그 결과의 공정성 역시 최근 관심사로 대두되었다. 본 발표에서는 검색 및 추천시스템의 공정성 이슈 및 그 해법을 다룬다. 공정한 검색 및 추천 결과를 정의하는 다양한 방법, 공정성의 결여가 미치는 자원 배분 및 스테레오타이핑 문제, 그리고 검색 및 추천시스템 개발의 각 단계별로 어떤 해결책이 있는지를 최신 연구 중심으로 살펴본다. 마지막으로 실제 공정한 시스템 개발을 위한 실무적인 고려사항을 다룬다.
Slides from SIGIR 2013 talk. The full paper can be found here: http://staff.science.uva.nl/~mdr/Publications/sigir2013-metrics.pdf
ABSTRACT: In recent years many models have been proposed that are aimed at predicting clicks of web search users. In addition, some information retrieval evaluation metrics have been built on top of a user model. In this paper we bring these two directions together and propose a common approach to converting any click model into an evaluation metric. We then put the resulting model-based metrics as well as traditional metrics (like DCG or Precision) into a common evaluation framework and compare them along a number of dimensions.
One of the dimensions we are particularly interested in is the agreement between offline and online experimental outcomes. It is widely believed, especially in an industrial setting, that online A/B-testing and interleaving experiments are generally better at capturing system quality than offline measurements. We show that offline metrics that are based on click models are more strongly correlated with online experimental outcomes than traditional offline metrics, especially in situations when we have incomplete relevance judgements.
Fairness in Search & RecSys 네이버 검색 콜로키움 김진영Jin Young Kim
검색 및 추천 시스템의 사회적 역할이 커지면서, 그 결과의 공정성 역시 최근 관심사로 대두되었다. 본 발표에서는 검색 및 추천시스템의 공정성 이슈 및 그 해법을 다룬다. 공정한 검색 및 추천 결과를 정의하는 다양한 방법, 공정성의 결여가 미치는 자원 배분 및 스테레오타이핑 문제, 그리고 검색 및 추천시스템 개발의 각 단계별로 어떤 해결책이 있는지를 최신 연구 중심으로 살펴본다. 마지막으로 실제 공정한 시스템 개발을 위한 실무적인 고려사항을 다룬다.
Slides from SIGIR 2013 talk. The full paper can be found here: http://staff.science.uva.nl/~mdr/Publications/sigir2013-metrics.pdf
ABSTRACT: In recent years many models have been proposed that are aimed at predicting clicks of web search users. In addition, some information retrieval evaluation metrics have been built on top of a user model. In this paper we bring these two directions together and propose a common approach to converting any click model into an evaluation metric. We then put the resulting model-based metrics as well as traditional metrics (like DCG or Precision) into a common evaluation framework and compare them along a number of dimensions.
One of the dimensions we are particularly interested in is the agreement between offline and online experimental outcomes. It is widely believed, especially in an industrial setting, that online A/B-testing and interleaving experiments are generally better at capturing system quality than offline measurements. We show that offline metrics that are based on click models are more strongly correlated with online experimental outcomes than traditional offline metrics, especially in situations when we have incomplete relevance judgements.
UserZoom Webinar: How to Conduct Web Customer Experience BenchmarkingUserZoom
You can't manage what you can't measure, so... How do you actually measure user experience?
In this webinar we covered what, why, and how to conduct website user experience & usability benchmarking. We discussed how to effectively measure the quality of a website's user experience across various competitors, within one industry, across time, using an online quantitative research methodology commonly referred to as "unmoderated remote usability testing."
Mining Large Streams of User Data for PersonalizedRecommenda.docxARIV4
Mining Large Streams of User Data for Personalized
Recommendations
Xavier Amatriain
Netflix
[email protected]
ABSTRACT
The Netflix Prize put the spotlight on the use of data min-
ing and machine learning methods for predicting user pref-
erences. Many lessons came out of the competition. But
since then, Recommender Systems have evolved. This evo-
lution has been driven by the greater availability of different
kinds of user data in industry and the interest that the area
has drawn among the research community. The goal of this
paper is to give an up-to-date overview of the use of data
mining approaches for personalization and recommendation.
Using Netflix personalization as a motivating use case, I will
describe the use of different kinds of data and machine learn-
ing techniques.
After introducing the traditional approaches to recommen-
dation, I highlight some of the main lessons learned from
the Netflix Prize. I then describe the use of recommenda-
tion and personalization techniques at Netflix. Finally, I
pinpoint the most promising current research avenues and
unsolved problems that deserve attention in this domain.
1. INTRODUCTION
Recommender Systems (RS) are a prime example of the
mainstream applicability of large scale data mining. Ap-
plications such as e-commerce, search, Internet music and
video, gaming or even online dating make use of similar
techniques to mine large volumes of data to better match
their users’ needs in a personalized fashion.
There is more to a good recommender system than the data
mining technique. Issues such as the user interaction design,
outside the scope of this paper, may have a deep impact
on the effectiveness of an approach. But given an existing
application, an improvement in the algorithm can have a
value of millions of dollars, and can even be the factor that
determines the success or failure of a business. On the other
hand, given an existing method or algorithm, adding more
features coming from different data sources can also result
in a significant improvement. I will describe the use of data,
models, and other personalization techniques at Netflix in
section 3. I will also discuss whether we should focus on
more data or better models in section 4.
Another important issue is how to measure the success of
a given personalization technique. Root mean squared er-
ror (RMSE) was the offline evaluation metric of choice in
the Netflix Prize (see Section 2). But there are many other
relevant metrics that, if optimized, would lead to different
solutions - think, for example, of ranking metrics such as
Normalized Discounted Cumulative Gain (NDCG) or other
information retrieval ones such as recall or area under the
curve (AUC). Beyond the optimization of a given offline met-
ric, what we are really pursuing is the impact of a method on
the business. Is there a way to relate the goodness of an algo-
rithm to more customer-facing metrics such as click-through
rate (CTR) or retention? I will describe our ...
NO1 Uk best vashikaran specialist in delhi vashikaran baba near me online vas...Amil Baba Dawood bangali
Contact with Dawood Bhai Just call on +92322-6382012 and we'll help you. We'll solve all your problems within 12 to 24 hours and with 101% guarantee and with astrology systematic. If you want to take any personal or professional advice then also you can call us on +92322-6382012 , ONLINE LOVE PROBLEM & Other all types of Daily Life Problem's.Then CALL or WHATSAPP us on +92322-6382012 and Get all these problems solutions here by Amil Baba DAWOOD BANGALI
#vashikaranspecialist #astrologer #palmistry #amliyaat #taweez #manpasandshadi #horoscope #spiritual #lovelife #lovespell #marriagespell#aamilbabainpakistan #amilbabainkarachi #powerfullblackmagicspell #kalajadumantarspecialist #realamilbaba #AmilbabainPakistan #astrologerincanada #astrologerindubai #lovespellsmaster #kalajaduspecialist #lovespellsthatwork #aamilbabainlahore#blackmagicformarriage #aamilbaba #kalajadu #kalailam #taweez #wazifaexpert #jadumantar #vashikaranspecialist #astrologer #palmistry #amliyaat #taweez #manpasandshadi #horoscope #spiritual #lovelife #lovespell #marriagespell#aamilbabainpakistan #amilbabainkarachi #powerfullblackmagicspell #kalajadumantarspecialist #realamilbaba #AmilbabainPakistan #astrologerincanada #astrologerindubai #lovespellsmaster #kalajaduspecialist #lovespellsthatwork #aamilbabainlahore #blackmagicforlove #blackmagicformarriage #aamilbaba #kalajadu #kalailam #taweez #wazifaexpert #jadumantar #vashikaranspecialist #astrologer #palmistry #amliyaat #taweez #manpasandshadi #horoscope #spiritual #lovelife #lovespell #marriagespell#aamilbabainpakistan #amilbabainkarachi #powerfullblackmagicspell #kalajadumantarspecialist #realamilbaba #AmilbabainPakistan #astrologerincanada #astrologerindubai #lovespellsmaster #kalajaduspecialist #lovespellsthatwork #aamilbabainlahore #Amilbabainuk #amilbabainspain #amilbabaindubai #Amilbabainnorway #amilbabainkrachi #amilbabainlahore #amilbabaingujranwalan #amilbabainislamabad
Understanding Inductive Bias in Machine LearningSUTEJAS
This presentation explores the concept of inductive bias in machine learning. It explains how algorithms come with built-in assumptions and preferences that guide the learning process. You'll learn about the different types of inductive bias and how they can impact the performance and generalizability of machine learning models.
The presentation also covers the positive and negative aspects of inductive bias, along with strategies for mitigating potential drawbacks. We'll explore examples of how bias manifests in algorithms like neural networks and decision trees.
By understanding inductive bias, you can gain valuable insights into how machine learning models work and make informed decisions when building and deploying them.
Sachpazis:Terzaghi Bearing Capacity Estimation in simple terms with Calculati...Dr.Costas Sachpazis
Terzaghi's soil bearing capacity theory, developed by Karl Terzaghi, is a fundamental principle in geotechnical engineering used to determine the bearing capacity of shallow foundations. This theory provides a method to calculate the ultimate bearing capacity of soil, which is the maximum load per unit area that the soil can support without undergoing shear failure. The Calculation HTML Code included.
Using recycled concrete aggregates (RCA) for pavements is crucial to achieving sustainability. Implementing RCA for new pavement can minimize carbon footprint, conserve natural resources, reduce harmful emissions, and lower life cycle costs. Compared to natural aggregate (NA), RCA pavement has fewer comprehensive studies and sustainability assessments.
Saudi Arabia stands as a titan in the global energy landscape, renowned for its abundant oil and gas resources. It's the largest exporter of petroleum and holds some of the world's most significant reserves. Let's delve into the top 10 oil and gas projects shaping Saudi Arabia's energy future in 2024.
Welcome to WIPAC Monthly the magazine brought to you by the LinkedIn Group Water Industry Process Automation & Control.
In this month's edition, along with this month's industry news to celebrate the 13 years since the group was created we have articles including
A case study of the used of Advanced Process Control at the Wastewater Treatment works at Lleida in Spain
A look back on an article on smart wastewater networks in order to see how the industry has measured up in the interim around the adoption of Digital Transformation in the Water Industry.
HEAP SORT ILLUSTRATED WITH HEAPIFY, BUILD HEAP FOR DYNAMIC ARRAYS.
Heap sort is a comparison-based sorting technique based on Binary Heap data structure. It is similar to the selection sort where we first find the minimum element and place the minimum element at the beginning. Repeat the same process for the remaining elements.
Immunizing Image Classifiers Against Localized Adversary Attacksgerogepatton
This paper addresses the vulnerability of deep learning models, particularly convolutional neural networks
(CNN)s, to adversarial attacks and presents a proactive training technique designed to counter them. We
introduce a novel volumization algorithm, which transforms 2D images into 3D volumetric representations.
When combined with 3D convolution and deep curriculum learning optimization (CLO), itsignificantly improves
the immunity of models against localized universal attacks by up to 40%. We evaluate our proposed approach
using contemporary CNN architectures and the modified Canadian Institute for Advanced Research (CIFAR-10
and CIFAR-100) and ImageNet Large Scale Visual Recognition Challenge (ILSVRC12) datasets, showcasing
accuracy improvements over previous techniques. The results indicate that the combination of the volumetric
input and curriculum learning holds significant promise for mitigating adversarial attacks without necessitating
adversary training.
8. Do users scan document from top to bottom?
1. The click-through rate (CTR) of the first document is about 0.45 while the CTR
of the tenth document is well below 0.05
2. The document below a click is viewed roughly 50% of the times
17. Random Click Model (RCM)
Click-through Rate Models (CTR)
Rank-based CTR Model (RCTR)
Document-based CTR Model (DCTR)
User Browsing Model (UBM)
Position-based Model (PBM)
Dependent Click Model (DCM)
Click Chain Model (CCM)
Dynamic Bayesian Network Model (DBN)
Simplified DBN Model (SDBN)
Cascade Model (CM)
Click Model
18. Baseline model
1.Random Click Model (RCM)
Any document can be click with the same (fixed) probability
2. Click-Through Rate Models (RCTR)
the click probability depends on the rank of the document
3. Document-Based CTR Model (DCTR)
the click-through rates for each query-document pair.
subject to overfitting for the reason that some documents and/or queries were not previously
encountered in our click log
19. Position-Based Model
position-based model (PBD)
Means that a document is clicked when user
Examine and attractive with it
Examination hypothesis. The probability of a user examining a document depends heavily
on its rank or position. PBM introduces a set of examination parameters Y, one for each rank.
PBM does not depend on the events at previous ranks.
20. Cascade Model
Cascade model (CM)
Step:
1.Start from the first document
2.Examine documents one by one
3.If click, then stop
4.Otherwise, continue
21. Cascade model (CM)
In particular:
1.CM does not allow sessions with more than one click
2.CM can not explain non-linear examination patterns
22. So far,
1.CTR models
+ count clicks (simple and fast)
- do not distinguish examination and attractiveness
2. Position-based model (PBM) User browsing model
+ examination and attractiveness
- examination of a document at rank r does not depend on
examinations and clicks above r
3. Cascade model (CM) Dynamic Bayesian network
+ cascade dependency of examination at r on examinations
and clicks above r
- only one click is allowed
23. User Browsing Model
User Browsing model (UBM)
the examination probability depends not
only on the rank of a document r, but also
on the rank of the previously clicked document r’
r’ is the rank of the previously clicked document or 0 if none of them was clicked
where c0 is set to 1 for convenience
24. Dynamic Bayesian Model
Dynamic Bayesian model (DBN)
Step:
1.Start from the first document
2.Examine documents one by one
3.If click, read actual document and can be satisfied
4.If satisfied, stop
5.Otherwise,continue with fixed probability
25. Dynamic Bayesian model (DBN)
In particular:
1.Gamma is the continuation probability for a user that either did not click on a document or click
ed but was not satisfied by it
2.DBN set gamma to 1,is Simplified DBN Model (SDBN) – MLE & good performance
3.SDBN set to 1,then model become Cascade Model (CM)
26. Random Click Model (RCM)
Click-through Rate Models (CTR)
Rank-based CTR Model (RCTR)
Document-based CTR Model (DCTR)
User Browsing Model (UBM)
Position-based Model (PBM)
Dependent Click Model (DCM)
Click Chain Model (CCM)
Dynamic Bayesian Network Model (DBN)
Simplified DBN Model (SDBN)
Cascade Model (CM)
1. Maximum likelihood estimation (RCM,RCTP,DCTP,DCM,SDBN,CM)
2. Expectation maximization (UBM,PBM,CCM,DBN)
Parameter Estimation
27. Simplified DBN Model (SDBN) -- MLE
In particular:
1. SDBN assumes that a user examines all documents until the last-clicked one and then aband
ons the search. In this case, both the attractiveness A and satisfaction S of SDBN are observed.
2.吸引度A即是给定query,其ducument的点击次数和展示次数(最后一个点击或之前)之比
3.满意度S即是给定query,在其ducument的点击集合中该ducument最后一次点击的占比
29. Dynamic Bayesian model (DBN) -- EM
In particular:
1. E-step. Given three parameters,compute the posterior probabilities A,E,S, This involves the
forward-backward algorithm
2. M-step. Given the posterior probabilities, update three parameters
30. 1.The DBN outperform others
2. X-axis = 100 means those urls whose train set >= 100;more session means priors not as important. Cascade & DBN improve.
3. Navigational queries have quality of context bias, and lots of sessions. Position models suffer
Result
31. Limit:
1.Click model cannot model out of order clicks
2. Completely blind to query reformulations
3. Assumes homogeneous user population
Future research:
1. Why not learning the structure of a click model from data
instead of defining it manually
2. Interactions beyond clicks
Limitations and future research
32. [1] Anne Schuth, Floor Sietsma, Shimon Whiteson, and Maarten de Rijke. “Optimizing Base Rankers Using Clicks A
Case Study using BM25”
[2] Thorsten Joachims, Laura Granka Bing Pan, Helene Hembrooke,and Geri Gay.” Accuratelyinterpreting clickthro
ugh data as implicit feedback”
[3] Nick Craswell, Onno Zoeter, Michael Taylor, and Bill Ramsey. “An experimental comparison of click position-bi
as models”
[4] Fan Guo, Chao Liu, Anitha Kannan, Tom Minka, Michael Taylor, Yi-Min Wang, and Christos Faloutsos. ”Click chain
model in web search”
[5] Olivier Chapelle and Ya Zhang.” A dynamic bayesian network click model for web search ranking”
[6] Kevin Patrick Murphy. Machine Learning: “A Probabilistic Perspective.”
[7] Suzan Verberne, Hans van Halteren, Daphne Theijssen,” Learning to Rank QA Data”
[8] Thorsten Joachims,” Optimizing Search Engines using Clickthrough Data”
[9] Daxin Jiang, Jian Pei, Hang Li,” Mining Search and Browse Logs for Web Search: A Survey”
Reference