The document summarizes research on evaluating the user experience of recommender systems. It presents hypotheses about how personalized recommendations versus random recommendations affect user perception, choice satisfaction, and feedback behavior. An experiment tested the hypotheses using a video recommender system and found that personalized recommendations increased perceived quality and choice satisfaction, which in turn increased feedback intentions. Privacy concerns decreased feedback intentions while trust in technology reduced privacy concerns. The summarizes lessons learned and discusses areas for future work such as confirming results in other systems and incorporating additional influences.
Understanding choice overload in recommender systemsdirkheld
Even though people are attracted by large, high quality rec- ommendation sets, psychological research on choice overload shows that choosing an item from recommendation sets con- taining many attractive items can be a very difficult task. A web-based user experiment using a matrix factorization algorithm applied to the MovieLens dataset was used to investigate the effect of recommendation set size (5 or 20 items) and set quality (low or high) on perceived variety, recommendation set attractiveness, choice difficulty and sat- isfaction with the chosen item. The results show that larger sets containing only good items do not necessarily result in higher choice satisfaction compared to smaller sets, as the increased recommendation set attractiveness is counteracted by the increased difficulty of choosing from these sets. These findings were supported by behavioral measurements reveal- ing intensified information search and increased acquisition times for these large attractive sets. Important implications of these findings for the design of recommender system user interfaces will be discussed.
Understanding choice overload in recommender systemsdirkheld
Even though people are attracted by large, high quality rec- ommendation sets, psychological research on choice overload shows that choosing an item from recommendation sets con- taining many attractive items can be a very difficult task. A web-based user experiment using a matrix factorization algorithm applied to the MovieLens dataset was used to investigate the effect of recommendation set size (5 or 20 items) and set quality (low or high) on perceived variety, recommendation set attractiveness, choice difficulty and sat- isfaction with the chosen item. The results show that larger sets containing only good items do not necessarily result in higher choice satisfaction compared to smaller sets, as the increased recommendation set attractiveness is counteracted by the increased difficulty of choosing from these sets. These findings were supported by behavioral measurements reveal- ing intensified information search and increased acquisition times for these large attractive sets. Important implications of these findings for the design of recommender system user interfaces will be discussed.
Privacy in Mobile Personalized Systems - The Effect of Disclosure JustificationsBart Knijnenburg
Paper Presentation at the Workshop on Usable Privacy & Security for Mobile Devices (U-PriSM) at the Symposium On Usable Privacy and Security (SOUPS) 2012
Paper can be found here: http://appanalysis.org/u-prism/soups12_mobile-final11.pdf
Full journal paper (under review): http://bit.ly/TiiSprivacy
Explaining the User Experience of Recommender Systems with User ExperimentsBart Knijnenburg
A talk I gave at the Netflix offices on July 2nd, 2012.
Please do not use any of the slides or their contents without my explicit permission (bart@usabart.nl for inquiries).
Transcript: Selling digital books in 2024: Insights from industry leaders - T...BookNet Canada
The publishing industry has been selling digital audiobooks and ebooks for over a decade and has found its groove. What’s changed? What has stayed the same? Where do we go from here? Join a group of leading sales peers from across the industry for a conversation about the lessons learned since the popularization of digital books, best practices, digital book supply chain management, and more.
Link to video recording: https://bnctechforum.ca/sessions/selling-digital-books-in-2024-insights-from-industry-leaders/
Presented by BookNet Canada on May 28, 2024, with support from the Department of Canadian Heritage.
Securing your Kubernetes cluster_ a step-by-step guide to success !KatiaHIMEUR1
Today, after several years of existence, an extremely active community and an ultra-dynamic ecosystem, Kubernetes has established itself as the de facto standard in container orchestration. Thanks to a wide range of managed services, it has never been so easy to set up a ready-to-use Kubernetes cluster.
However, this ease of use means that the subject of security in Kubernetes is often left for later, or even neglected. This exposes companies to significant risks.
In this talk, I'll show you step-by-step how to secure your Kubernetes cluster for greater peace of mind and reliability.
Observability Concepts EVERY Developer Should Know -- DeveloperWeek Europe.pdfPaige Cruz
Monitoring and observability aren’t traditionally found in software curriculums and many of us cobble this knowledge together from whatever vendor or ecosystem we were first introduced to and whatever is a part of your current company’s observability stack.
While the dev and ops silo continues to crumble….many organizations still relegate monitoring & observability as the purview of ops, infra and SRE teams. This is a mistake - achieving a highly observable system requires collaboration up and down the stack.
I, a former op, would like to extend an invitation to all application developers to join the observability party will share these foundational concepts to build on:
State of ICS and IoT Cyber Threat Landscape Report 2024 previewPrayukth K V
The IoT and OT threat landscape report has been prepared by the Threat Research Team at Sectrio using data from Sectrio, cyber threat intelligence farming facilities spread across over 85 cities around the world. In addition, Sectrio also runs AI-based advanced threat and payload engagement facilities that serve as sinks to attract and engage sophisticated threat actors, and newer malware including new variants and latent threats that are at an earlier stage of development.
The latest edition of the OT/ICS and IoT security Threat Landscape Report 2024 also covers:
State of global ICS asset and network exposure
Sectoral targets and attacks as well as the cost of ransom
Global APT activity, AI usage, actor and tactic profiles, and implications
Rise in volumes of AI-powered cyberattacks
Major cyber events in 2024
Malware and malicious payload trends
Cyberattack types and targets
Vulnerability exploit attempts on CVEs
Attacks on counties – USA
Expansion of bot farms – how, where, and why
In-depth analysis of the cyber threat landscape across North America, South America, Europe, APAC, and the Middle East
Why are attacks on smart factories rising?
Cyber risk predictions
Axis of attacks – Europe
Systemic attacks in the Middle East
Download the full report from here:
https://sectrio.com/resources/ot-threat-landscape-reports/sectrio-releases-ot-ics-and-iot-security-threat-landscape-report-2024/
zkStudyClub - Reef: Fast Succinct Non-Interactive Zero-Knowledge Regex ProofsAlex Pruden
This paper presents Reef, a system for generating publicly verifiable succinct non-interactive zero-knowledge proofs that a committed document matches or does not match a regular expression. We describe applications such as proving the strength of passwords, the provenance of email despite redactions, the validity of oblivious DNS queries, and the existence of mutations in DNA. Reef supports the Perl Compatible Regular Expression syntax, including wildcards, alternation, ranges, capture groups, Kleene star, negations, and lookarounds. Reef introduces a new type of automata, Skipping Alternating Finite Automata (SAFA), that skips irrelevant parts of a document when producing proofs without undermining soundness, and instantiates SAFA with a lookup argument. Our experimental evaluation confirms that Reef can generate proofs for documents with 32M characters; the proofs are small and cheap to verify (under a second).
Paper: https://eprint.iacr.org/2023/1886
Le nuove frontiere dell'AI nell'RPA con UiPath Autopilot™UiPathCommunity
In questo evento online gratuito, organizzato dalla Community Italiana di UiPath, potrai esplorare le nuove funzionalità di Autopilot, il tool che integra l'Intelligenza Artificiale nei processi di sviluppo e utilizzo delle Automazioni.
📕 Vedremo insieme alcuni esempi dell'utilizzo di Autopilot in diversi tool della Suite UiPath:
Autopilot per Studio Web
Autopilot per Studio
Autopilot per Apps
Clipboard AI
GenAI applicata alla Document Understanding
👨🏫👨💻 Speakers:
Stefano Negro, UiPath MVPx3, RPA Tech Lead @ BSP Consultant
Flavio Martinelli, UiPath MVP 2023, Technical Account Manager @UiPath
Andrei Tasca, RPA Solutions Team Lead @NTT Data
PHP Frameworks: I want to break free (IPC Berlin 2024)Ralf Eggert
In this presentation, we examine the challenges and limitations of relying too heavily on PHP frameworks in web development. We discuss the history of PHP and its frameworks to understand how this dependence has evolved. The focus will be on providing concrete tips and strategies to reduce reliance on these frameworks, based on real-world examples and practical considerations. The goal is to equip developers with the skills and knowledge to create more flexible and future-proof web applications. We'll explore the importance of maintaining autonomy in a rapidly changing tech landscape and how to make informed decisions in PHP development.
This talk is aimed at encouraging a more independent approach to using PHP frameworks, moving towards a more flexible and future-proof approach to PHP development.
Privacy in Mobile Personalized Systems - The Effect of Disclosure JustificationsBart Knijnenburg
Paper Presentation at the Workshop on Usable Privacy & Security for Mobile Devices (U-PriSM) at the Symposium On Usable Privacy and Security (SOUPS) 2012
Paper can be found here: http://appanalysis.org/u-prism/soups12_mobile-final11.pdf
Full journal paper (under review): http://bit.ly/TiiSprivacy
Explaining the User Experience of Recommender Systems with User ExperimentsBart Knijnenburg
A talk I gave at the Netflix offices on July 2nd, 2012.
Please do not use any of the slides or their contents without my explicit permission (bart@usabart.nl for inquiries).
Transcript: Selling digital books in 2024: Insights from industry leaders - T...BookNet Canada
The publishing industry has been selling digital audiobooks and ebooks for over a decade and has found its groove. What’s changed? What has stayed the same? Where do we go from here? Join a group of leading sales peers from across the industry for a conversation about the lessons learned since the popularization of digital books, best practices, digital book supply chain management, and more.
Link to video recording: https://bnctechforum.ca/sessions/selling-digital-books-in-2024-insights-from-industry-leaders/
Presented by BookNet Canada on May 28, 2024, with support from the Department of Canadian Heritage.
Securing your Kubernetes cluster_ a step-by-step guide to success !KatiaHIMEUR1
Today, after several years of existence, an extremely active community and an ultra-dynamic ecosystem, Kubernetes has established itself as the de facto standard in container orchestration. Thanks to a wide range of managed services, it has never been so easy to set up a ready-to-use Kubernetes cluster.
However, this ease of use means that the subject of security in Kubernetes is often left for later, or even neglected. This exposes companies to significant risks.
In this talk, I'll show you step-by-step how to secure your Kubernetes cluster for greater peace of mind and reliability.
Observability Concepts EVERY Developer Should Know -- DeveloperWeek Europe.pdfPaige Cruz
Monitoring and observability aren’t traditionally found in software curriculums and many of us cobble this knowledge together from whatever vendor or ecosystem we were first introduced to and whatever is a part of your current company’s observability stack.
While the dev and ops silo continues to crumble….many organizations still relegate monitoring & observability as the purview of ops, infra and SRE teams. This is a mistake - achieving a highly observable system requires collaboration up and down the stack.
I, a former op, would like to extend an invitation to all application developers to join the observability party will share these foundational concepts to build on:
State of ICS and IoT Cyber Threat Landscape Report 2024 previewPrayukth K V
The IoT and OT threat landscape report has been prepared by the Threat Research Team at Sectrio using data from Sectrio, cyber threat intelligence farming facilities spread across over 85 cities around the world. In addition, Sectrio also runs AI-based advanced threat and payload engagement facilities that serve as sinks to attract and engage sophisticated threat actors, and newer malware including new variants and latent threats that are at an earlier stage of development.
The latest edition of the OT/ICS and IoT security Threat Landscape Report 2024 also covers:
State of global ICS asset and network exposure
Sectoral targets and attacks as well as the cost of ransom
Global APT activity, AI usage, actor and tactic profiles, and implications
Rise in volumes of AI-powered cyberattacks
Major cyber events in 2024
Malware and malicious payload trends
Cyberattack types and targets
Vulnerability exploit attempts on CVEs
Attacks on counties – USA
Expansion of bot farms – how, where, and why
In-depth analysis of the cyber threat landscape across North America, South America, Europe, APAC, and the Middle East
Why are attacks on smart factories rising?
Cyber risk predictions
Axis of attacks – Europe
Systemic attacks in the Middle East
Download the full report from here:
https://sectrio.com/resources/ot-threat-landscape-reports/sectrio-releases-ot-ics-and-iot-security-threat-landscape-report-2024/
zkStudyClub - Reef: Fast Succinct Non-Interactive Zero-Knowledge Regex ProofsAlex Pruden
This paper presents Reef, a system for generating publicly verifiable succinct non-interactive zero-knowledge proofs that a committed document matches or does not match a regular expression. We describe applications such as proving the strength of passwords, the provenance of email despite redactions, the validity of oblivious DNS queries, and the existence of mutations in DNA. Reef supports the Perl Compatible Regular Expression syntax, including wildcards, alternation, ranges, capture groups, Kleene star, negations, and lookarounds. Reef introduces a new type of automata, Skipping Alternating Finite Automata (SAFA), that skips irrelevant parts of a document when producing proofs without undermining soundness, and instantiates SAFA with a lookup argument. Our experimental evaluation confirms that Reef can generate proofs for documents with 32M characters; the proofs are small and cheap to verify (under a second).
Paper: https://eprint.iacr.org/2023/1886
Le nuove frontiere dell'AI nell'RPA con UiPath Autopilot™UiPathCommunity
In questo evento online gratuito, organizzato dalla Community Italiana di UiPath, potrai esplorare le nuove funzionalità di Autopilot, il tool che integra l'Intelligenza Artificiale nei processi di sviluppo e utilizzo delle Automazioni.
📕 Vedremo insieme alcuni esempi dell'utilizzo di Autopilot in diversi tool della Suite UiPath:
Autopilot per Studio Web
Autopilot per Studio
Autopilot per Apps
Clipboard AI
GenAI applicata alla Document Understanding
👨🏫👨💻 Speakers:
Stefano Negro, UiPath MVPx3, RPA Tech Lead @ BSP Consultant
Flavio Martinelli, UiPath MVP 2023, Technical Account Manager @UiPath
Andrei Tasca, RPA Solutions Team Lead @NTT Data
PHP Frameworks: I want to break free (IPC Berlin 2024)Ralf Eggert
In this presentation, we examine the challenges and limitations of relying too heavily on PHP frameworks in web development. We discuss the history of PHP and its frameworks to understand how this dependence has evolved. The focus will be on providing concrete tips and strategies to reduce reliance on these frameworks, based on real-world examples and practical considerations. The goal is to equip developers with the skills and knowledge to create more flexible and future-proof web applications. We'll explore the importance of maintaining autonomy in a rapidly changing tech landscape and how to make informed decisions in PHP development.
This talk is aimed at encouraging a more independent approach to using PHP frameworks, moving towards a more flexible and future-proof approach to PHP development.
GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using Deplo...James Anderson
Effective Application Security in Software Delivery lifecycle using Deployment Firewall and DBOM
The modern software delivery process (or the CI/CD process) includes many tools, distributed teams, open-source code, and cloud platforms. Constant focus on speed to release software to market, along with the traditional slow and manual security checks has caused gaps in continuous security as an important piece in the software supply chain. Today organizations feel more susceptible to external and internal cyber threats due to the vast attack surface in their applications supply chain and the lack of end-to-end governance and risk management.
The software team must secure its software delivery process to avoid vulnerability and security breaches. This needs to be achieved with existing tool chains and without extensive rework of the delivery processes. This talk will present strategies and techniques for providing visibility into the true risk of the existing vulnerabilities, preventing the introduction of security issues in the software, resolving vulnerabilities in production environments quickly, and capturing the deployment bill of materials (DBOM).
Speakers:
Bob Boule
Robert Boule is a technology enthusiast with PASSION for technology and making things work along with a knack for helping others understand how things work. He comes with around 20 years of solution engineering experience in application security, software continuous delivery, and SaaS platforms. He is known for his dynamic presentations in CI/CD and application security integrated in software delivery lifecycle.
Gopinath Rebala
Gopinath Rebala is the CTO of OpsMx, where he has overall responsibility for the machine learning and data processing architectures for Secure Software Delivery. Gopi also has a strong connection with our customers, leading design and architecture for strategic implementations. Gopi is a frequent speaker and well-known leader in continuous delivery and integrating security into software delivery.
Welcome to the first live UiPath Community Day Dubai! Join us for this unique occasion to meet our local and global UiPath Community and leaders. You will get a full view of the MEA region's automation landscape and the AI Powered automation technology capabilities of UiPath. Also, hosted by our local partners Marc Ellis, you will enjoy a half-day packed with industry insights and automation peers networking.
📕 Curious on our agenda? Wait no more!
10:00 Welcome note - UiPath Community in Dubai
Lovely Sinha, UiPath Community Chapter Leader, UiPath MVPx3, Hyper-automation Consultant, First Abu Dhabi Bank
10:20 A UiPath cross-region MEA overview
Ashraf El Zarka, VP and Managing Director MEA, UiPath
10:35: Customer Success Journey
Deepthi Deepak, Head of Intelligent Automation CoE, First Abu Dhabi Bank
11:15 The UiPath approach to GenAI with our three principles: improve accuracy, supercharge productivity, and automate more
Boris Krumrey, Global VP, Automation Innovation, UiPath
12:15 To discover how Marc Ellis leverages tech-driven solutions in recruitment and managed services.
Brendan Lingam, Director of Sales and Business Development, Marc Ellis
Climate Impact of Software Testing at Nordic Testing DaysKari Kakkonen
My slides at Nordic Testing Days 6.6.2024
Climate impact / sustainability of software testing discussed on the talk. ICT and testing must carry their part of global responsibility to help with the climat warming. We can minimize the carbon footprint but we can also have a carbon handprint, a positive impact on the climate. Quality characteristics can be added with sustainability, and then measured continuously. Test environments can be used less, and in smaller scale and on demand. Test techniques can be used in optimizing or minimizing number of tests. Test automation can be used to speed up testing.
GraphRAG is All You need? LLM & Knowledge GraphGuy Korland
Guy Korland, CEO and Co-founder of FalkorDB, will review two articles on the integration of language models with knowledge graphs.
1. Unifying Large Language Models and Knowledge Graphs: A Roadmap.
https://arxiv.org/abs/2306.08302
2. Microsoft Research's GraphRAG paper and a review paper on various uses of knowledge graphs:
https://www.microsoft.com/en-us/research/blog/graphrag-unlocking-llm-discovery-on-narrative-private-data/
The Metaverse and AI: how can decision-makers harness the Metaverse for their...Jen Stirrup
The Metaverse is popularized in science fiction, and now it is becoming closer to being a part of our daily lives through the use of social media and shopping companies. How can businesses survive in a world where Artificial Intelligence is becoming the present as well as the future of technology, and how does the Metaverse fit into business strategy when futurist ideas are developing into reality at accelerated rates? How do we do this when our data isn't up to scratch? How can we move towards success with our data so we are set up for the Metaverse when it arrives?
How can you help your company evolve, adapt, and succeed using Artificial Intelligence and the Metaverse to stay ahead of the competition? What are the potential issues, complications, and benefits that these technologies could bring to us and our organizations? In this session, Jen Stirrup will explain how to start thinking about these technologies as an organisation.
4. Recommender systems
Recommend items to users
based on their stated preferences
(e.g. books, movies, laptops)
Users indicate preferences
by rating presented items
(e.g. from one to five stars)
Predict the users’ rating value of new items...
then present items with the highest predicted rating
6. Two premises
Premise 1 | Users want to receive
recommendations
Do recommendations have any effect on the user experience at all?
Compare a system with vs. without recommendations
Premise 2 | Users will provide preference
feedback
Without feedback, no recommendations
What causes - and inhibits - them to do this?
Analyze users’ feedback behavior and intentions
8. Effect of
Premise 1 | Users want to receive
recommendations
Users are able to notice differences in prediction
accuracy
But... higher accuracy can lead to lower usefulness of
recommendations
Distinction between perception and evaluation
of recommendation quality
9. Constructs and
Perception
Perceived recommendation
quality
User experience
Evaluation Personalized vs.
random
H2a + Choice
satisfaction
Choice satisfaction H1 + Perceived recom-
Perceived system effectiveness mendation quality
Perceived system
H2b + e ectiveness
Questionnaires and
process data
10. Feedback
Premise 2 | Users will provide preference
feedback
Satisfaction increases feedback intentions
However, only a minority is willing to give up personal information
in return for a personalized experience (Teltzrow & Kobsa)
Privacy decreases feedback intentions
However, most people are usually or always comfortable disclosing
personal taste preferences (Ackerman et al.)
11. Constructs and
Feedback
Willingness to provide feedback
User experience
H3a
Privacy Choice
satisfaction
System-specific privacy
concerns +
Perceived system Intention to
Trust in technology e ectiveness H3b + provide feedback
Process data General trust
in technology
H4 System-specific
privacy concerns
H5
Actual feedback behavior
12. A model of user
User experience
Personalized vs.
random
H2a + Choice H3a
satisfaction
H1 + Perceived recom-
mendation quality +
Perceived system Intention to
H2b + e ectiveness H3b + provide feedback
General trust H4 System-specific H5
in technology privacy concerns
13. Experiment
Test
with actual recommender system
Two versions of the Personalized vs.
User experience
random
H2a + Choice H3a
system: H1 + Perceived recom-
satisfaction
mendation quality +
One that provides personalized Perceived system Intention to
+
recommendations H2b + e ectiveness H3b provide feedback
One that provides random clips General trust H4 System-specific H5
as ‘recommendations’ in technology privacy concerns
15. Setup
Online experiment
Conducted by EMIC in Germany,
September and October, 2009
Two slightly modified versions
of
the MSN ClipClub system
43 participants
25 in the random and 18 in the
personalized condition
65% male, all German
Average age of 31 (SD = 9.45)
16. System
Microsoft ClipClub
Lifestyle & entertainment video
clips
Changes
Recommendations section
highlighted
Pre-experimental instruction
Rating probe
No rating for five minutes: ask
user to rate the current item
17. Employed algorithm
Vector Space Model Engine
Use the tags associated to a clip to create a vector of each clip
Create a tag vector for the subset of clips rated by the user
Recommends clips with a tag vector similar to the created tag vector
Older ratings are logarithmically discounted, as are older items
18. Experimental procedure
Each participant:
entered demographic details
was shown an instruction on how to use the system
used the system freely for at least 30 minutes
completed the questionnaires
entered an email address for the raffle
Rating items
Users could perpetually rate items and inspect recommendations in
any given order
Rating probe: at least 6 ratings unless ignored
19. Questionnaires
40 statements Choice satisfaction
9 items, e.g. “The videos I chose fitted my
Agree or disagree on a 5-point preference”
scale
General trust in technology
Factor Analysis in two batches 4 items, e.g. “I’m less confident when I use
technology”, reverse-coded
System-specific privacy concern
6 factors 5 items, e.g. “I feel confident that ClipClub
Recommendation set quality respects my privacy”
7 items, e.g. “The recommended videos fitted Intention to rate items
my preference”
5 items, e.g. “I like to give feedback on the
System effectiveness items I’m watching”
6 items, e.g. “The recommender is useless”,
reverse-coded
20. Process data
All clicks were logged
In order to link subjective metrics to observable behavior
Process data measures
Total viewing-time
Number of clicked clips
Number of completed clips
Number of self-initiated ratings
Number of canceled rating requests
22. Path model results
Personalized vs.
.572 (.125)*** Choice .346 (.125)**
random
H2a satisfaction H3a
.696 (.276)* Perceived recom-
H1 mendation quality
Perceived system Intention to
.515 (.135)*** e ectiveness .296 (.123)* provide feedback
H2b H3b
General trust -.268 (.156)1 System-specific -.255 (.113)*
in technology H4 privacy concerns H5
23. Effect of
Personalized vs.
.572 (.125)*** Choice
random
Users notice .696 (.276)*
H2a satisfaction
Perceived recom-
personalization H1 mendation quality
Perceived system
Personalized recommendations .515 (.135)*** e ectiveness
increase perceived H2b
recommendation quality (H1)
Users browse less, but
Users like better watch more
Number of clips watched
recommendations entirely is higher in the
Higher perceived quality personalized condition
increases choice satisfaction Number of clicked clips and
(H2a) and system effectiveness total viewing time are negatively
(H2b) correlated with system
24. Feedback
Choice .346 (.125)**
Better experience satisfaction H3a
increases feedback
Perceived system Intention to
Choice satisfaction and system e ectiveness .296 (.123)* provide feedback
effectiveness increase feedback H3b
intentions (H3a,b)
General trust -.268 (.156)1 System-specific -.255 (.113)*
in technology H4 privacy concerns H5
Privacy decreases Effect of trust in
feedback technology
Users with a higher system- Privacy concerns increase when
specific privacy concern have a users have a lower trust in
lower feedback intention (H5) technology (H4).
25. Intention-behavior gap
Number of canceled rating probes
Significantly lower in the personalized condition
Negatively correlated with intention to provide feedback
Total number of provided ratings
Not significantly correlated with users’ intention to provide feedback
26. To summarize...
Personalized vs.
.572 (.125)*** Choice .346 (.125)**
random
H2a satisfaction H3a
.696 (.276)* Perceived recom-
H1 mendation quality
Perceived system Intention to
.515 (.135)*** e ectiveness .296 (.123)* provide feedback
H2b H3b
General trust -.268 (.156)1 System-specific -.255 (.113)*
in technology H4 privacy concerns H5
28. Remaining questions
True for all recommender systems?
Results should be confirmed in several other systems and with a
higher number and a more diverse range of participants
Other influences?
Incorporate other aspects to get a more detailed understanding of
the mechanisms underlying the user-recommender interaction
Other algorithms?
Test differences between algorithms that only moderately differ in
accuracy
30. Field trails
Full-scale test of the framework
Four different partners, three different countries
Trials are conducted over a longer time-period
Each compares at least three systems (mainly different algorithms)
Questionnaires and process data
Core of evaluation is the same
Algorithm -> perceived recommendation quality -> system
effectiveness
Each partner adds measures of personal interest
31. Want more?
RecSys’10 workshop
User-Centric Evaluation of Recommender
attending
Systems and their Interfaces (UCERSTI)
Barcelona, September 26-30
I am
Line-up:
7 paper presentations !"#$%&'
2 keynotes (Francisco Martin, Pearl Pu)
Panel discussion with 5 prominent researchers
1st internation
al workshop on
User-Centric E
valuation of
Recommender
Systems
and Their Inte
rfaces
Editor's Notes
First I want to thank my co-authors and sponsor
Your typical recommender system works like this:
Right now, researchers seem to focus on the algorithmic performance. They believe that better algorithms lead to a better experience. Is that really true?
It can only be true under two assumptions:
1. users want to get personalized recommendations, and 2. they will provide enough feedback to make this possible
In order to answer these questions, we need to evaluate the user experience, not the algorithm!
What existing evidence do we have?
Increased recommendation accuracy is noticeable, but doesn’t always lead to a better UX
McNee et al.: algorithm with best predictions was rated least helpful
Torres et al.: algorithm with lowest accuracy resulted in highest satisfaction
Ziegler et al.: diversifying recommendation set resulted in lower accuracy but a more positive evaluation
Let’s say we have two systems, one with personalized recommendations, and one without:
Perception tests whether we are able to notice the difference
Evaluation tests whether this increases our satisfaction with the system and, ultimately, our choices
These are measures by questionnaires, but we can also look at process data:
Effective systems may show decreased browsing and overall viewing time
In better systems, users will watch more clips from beginning to end
The more beneficial it seems to be, the more feedback users will provide (Spiekermann et al.; Brodie Karat & Karat; Kobsa & Teltzrow)
Minority = Between 40 and 50% in an overview of privacy surveys
Privacy concerns reduce users’ willingness to disclose personal information (Metzger et al.; Teltzrow & Kobsa)
Most people = 80% of the respondents of a detailed survey
Users’ actual feedback behavior may be different from their intentions (Spiekermann et al.)
So now we look at why users provide preference information
We already know choice satisfaction and perceived system effectiveness, and we hypothesize that a better experience increase the intention to provide feedback
However, privacy concerns may reduce feedback intention, and privacy concerns may be higher for those who don’t trust technology in general
Process data:
Due to the intention-behavior gap actual feedback may only be moderately correlated to feedback intentions
So let’s review the hypotheses (laser-point):
Personalized recommendations should have a perceivably higher quality
This should in turn increase the user experience of the system and the outcome (choices)
A better experience in turn increases their intention to provide feedback
However...
Tip: use two conditions to control the causal relations and to single out the effect
Also: log behavioral data and triangulate this with the constructs
Content and system are in German
To explain the rating feature and its effect on recommendations
Opening recommendations before rating any items showed a similar explanation
Pps were allowed to close this pop-up without rating
After rating, participants were transported to the recommendations
(the length of the vector depends on the impact the tags have)
(in terms of cosine similarity)
Allowing ample opportunity to let their feedback behavior be influenced by their user experience
Unless they ignored the rating-probe
The median number of ratings per user was 15
Tip for UX researchers: you cannot measure UX concepts with a single question. Measurement is far more robust if you construct a scale based on several questions
Exploratory Factor Analysis validates the intended conceptual structure
Finally, test the model with path analysis (mediation on steroids)
The model has a good fit, with a non-significant χ2 of 13.210 (df = 13, p = .4317), a CFI of .996 and an RMSEA between 0 and 0.153 (90% confidence interval)
Let’s review that one more time:
We’ve been developing a framework for this type of research, and validated it in several field trials -->
E.g. Advertisement (MS): Less clips clicked (fewer ads started) but maybe a higher retention (more ads full watched)?
Watch out for our future papers!
Advantages of fitting a model: steps in between reduce variability!