SlideShare a Scribd company logo
1 of 78
Download to read offline
User Studies for APG
How to support system development with user feedback?
Purpose of presentation
1. To share some useful tips and lessons from
user studies we’ve done
2. To explain how to incorporate user studies
into system development flow
3. To encourage more user studies in Social
Computing
Outline
• The APG research roadmap
• Overview of our user studies
• Framework for Value-Driven Development™
• Examples of putting research into practice
• Tips for planning and executing user studies
What is a “user study”?
• A practically oriented research study that gathers real
feedback from real users, based on real scenarios,
testing the usability and functionality of a real system.
• User study is meant to produce information that
helps improve the system!
Why do we need user studies?
1. Break the researcher echo chamber – learn what matters for
people working in the industry
2. Find new use cases for the technologies you’re developing
3. Disseminate our research and systems (get downloads, citations,
real users signing up… ➔ promotion/raise !!!)
4. Maintain relationships to stakeholder organizations and their
key people
5. Get real feedback from real users to improve real systems
➔ the only path to real impact
There is an infinitely gigantic leap from
having “users” to having real users
…meaning: most registered people never use the
system in reality. Real user = active user
Getting your first active user is like achieving
the first moon landing.
…meaning: a small leap for mankind, a giant leap for
your system.
The same problem
is faced by startups
and QCRI projects
(Download Joni’s
dissertation “Startup
Dilemmas” FOR FREE:
http://www.utupub.fi/h
andle/10024/99349)
The APG case
Research Roadmap for Automatic
Persona Generation
Information architecture:
How to choose the correct
information elements and
layouts for a given user, use
case or industry?
Quotes:
How to find representative,
contextually relevant,
and non-distracting
comments describing the
persona?
Evaluation: 1) How to ensure
personas are complete, clear,
consistent and credible?
2) How to measure usefulness
of personas for individuals and
organizations?
Attributes: How to infer
user attributes, such as
psychographics, needs
and wants, political
orientation, and brand
affinities?
Temporal analysis:
How to analyze change
and stability of personas
over time?
Image: How to retrieve,
generate, and choose
persona profile pictures?
Finding better ways to process and choose useful user
information from vast amounts of online data.
”Personas are about giving faces to data.” – Jim Jansen
Applicability: How to create
personas for different
contexts (e.g., e-commerce
personas, political personas,
gamer personas…)?
In our research roadmap, user studies are
relevant in most if not all sections…
Research Roadmap for Automatic
Persona Generation
Information architecture:
How to choose the correct
information elements and
layouts for a given user, use
case or industry?
Quotes:
How to find representative,
contextually relevant,
and non-distracting
comments describing the
persona?
Evaluation: 1) How to ensure
personas are complete, clear,
consistent and credible?
2) How to measure usefulness
of personas for individuals and
organizations?
Attributes: How to infer
user attributes, such as
psychographics, needs
and wants, political
orientation, and brand
affinities?
Temporal analysis:
How to analyze change
and stability of personas
over time?
Image: How to retrieve,
generate, and choose
persona profile pictures?
Finding better ways to process and choose useful user
information from vast amounts of online data.
”Personas are about giving faces to data.” – Jim Jansen
Applicability: How to create
personas for different
contexts (e.g., e-commerce
personas, political personas,
gamer personas…)?
Research Roadmap for Automatic
Persona Generation
Information architecture:
How to choose the correct
information elements and
layouts for a given user, use
case or industry?
Quotes:
How to find representative,
contextually relevant,
and non-distracting
comments describing the
persona?
Evaluation: 1) How to ensure
personas are complete, clear,
consistent and credible?
2) How to measure usefulness
of personas for individuals and
organizations?
Attributes: How to infer
user attributes, such as
psychographics, needs
and wants, political
orientation, and brand
affinities?
Temporal analysis:
How to analyze change
and stability of personas
over time?
Image: How to retrieve,
generate, and choose
persona profile pictures?
Finding better ways to process and choose useful user
information from vast amounts of online data.
”Personas are about giving faces to data.” – Jim Jansen
Applicability: How to create
personas for different
contexts (e.g., e-commerce
personas, political personas,
gamer personas…)?
A list of user studies we have done
Study Publications System
feedbacks
N of
participants
Real users or not
User Study 1 (2017) 4 (+1 in review) ~15 29 Yes
Persona Perception Scale: pilot
(2017)
1 N/A 19 No
QR Workshop (2017) N/A 10 12 Yes
QF Workshop (2018) N/A 9 12 Yes
User Study 2 (2018) 2 (in review) ~15 36 Yes
Persona Perception Scale:
validation (2018)
2 (1 in review
+ 1 in works)
N/A 412 No
Persona Crowd Experiment 1
(Transparent personas) (2018)
1 (in review) 1 (MAJOR) 412 No
Persona Crowd Experiment 2
(Smile Study) (2018)
1 (in review) N/A 2,400 No
User Study 3 (2018) GOAL: 4 TBA (~20) 34 Yes
A list of user studies we have done
Study Publications System
feedbacks
N of
participants
Real users or not
User Study 1 (2017) 4 (+1 in review) ~15 29 Yes
Persona Perception Scale: pilot
(2017)
1 N/A 19 No
QR Workshop (2017) N/A 10 12 Yes
QF Workshop (2018) N/A 9 12 Yes
User Study 2 (2018) 2 (in review) ~15 36 Yes
Persona Perception Scale:
validation (2018)
2 (1 in review
+ 1 in works)
N/A 412 No
Persona Crowd Experiment 1
(Transparent personas) (2018)
1 (in review) 1 (MAJOR) 412 No
Persona Crowd Experiment 2
(Smile Study) (2018)
1 (in review) N/A 2,400 No
User Study 3 (2018) GOAL: 4 TBA (~20) 34 Yes
A list of user studies we have done
Study Publications System
feedbacks
N of
participants
Real users or not
User Study 1 (2017) 4 (+1 in review) ~15 29 Yes
Persona Perception Scale: pilot
(2017)
1 N/A 19 No
QR Workshop (2017) N/A 10 12 Yes
QF Workshop (2018) N/A 9 12 Yes
User Study 2 (2018) 2 (in review) ~15 36 Yes
Persona Perception Scale:
validation (2018)
2 (1 in review
+ 1 in works)
N/A 412 No
Persona Crowd Experiment 1
(Transparent personas) (2018)
1 (in review) 1 (MAJOR) 412 No
Persona Crowd Experiment 2
(Smile Study) (2018)
1 (in review) N/A 2,400 No
User Study 3 (2018) GOAL: 4 TBA (~20) 34 Yes
A list of user studies we have done
Study Publications System
feedbacks
N of
participants
Real users or not
User Study 1 (2017) 4 (+1 in review) ~15 29 Yes
Persona Perception Scale: pilot
(2017)
1 N/A 19 No
QR Workshop (2017) N/A 10 12 Yes
QF Workshop (2018) N/A 9 12 Yes
User Study 2 (2018) 2 (in review) ~15 36 Yes
Persona Perception Scale:
validation (2018)
2 (1 in review
+ 1 in works)
N/A 412 No
Persona Crowd Experiment 1
(Transparent personas) (2018)
1 (in review) 1 (MAJOR) 412 No
Persona Crowd Experiment 2
(Smile Study) (2018)
1 (in review) N/A 2,400 No
User Study 3 (2018) GOAL: 4 TBA (~20) 34 Yes
Observations:
• Number of participants and number of system development
items is not correlated (when including crowd experiments)
• However, when excluding crowd experiments, there is a
correlation (the more people participate, the more ideas you
get)
…both the quality and magnitude of system development items
vary!
As the number of user studies increases,
the number of changes actually grows!
(but the magnitude of changes decreases,
i.e., you start fixing smaller things)
Zeroing in on the
value…
US1
US2
US3
Moment of value
= state and point in
time where the system
provides real value to
real users and is
therefore actively used
by them.
Value-Driven System Development
(VDSD)
Value-Driven
Development
• Implement the
recommended changes
• Conduct another user
study (validate progress
towards value; find new
things to improve)
Filter useful feedback from non-useful:
• How common is the feedback?
• Who gives the feedback?
• Design the user study
• Recruit participants
• Conduct the study
User-driven system development over time
v1 v2 v3
v4
Number of improvements
increasing
Magnitude of
problems decreasing
User study 2 User study 3 User study 4User study 1
time
Challenge: you need to make real
improvements in between the user studies
…otherwise, you keep hearing the same
complaints again and again!
Remember: developing a useful system is a
journey, not a project (there is no “end”).
…that’s why you need years of dedication,
funding and committed people.
Some tips
How to decide the user study focus?
• Behavior = objective information about what the users
looks/clicks at. Considered as the most reliable kind of
data.
• Perceptions = helps understand what are the user’s
impressions and why she or he does things. However,
susceptible to several biases from both researcher and
respondent.
…ideally, you’d measure both!
(mixed method studies are the best user studies)
To-do list for a user study
1. Decide research problem
2. Decide manipulations
3. Create user task
4. Create experimental treatments & flows
5. Create consent form (and get IRB approval if needed)
6. Recruit users
7. Pilot test (devices, software, flows…)
How to design a user study?
• Combine an (a) interesting theoretical problem and (b) opportunity to
get valuable user feedback
• The REAL criteria:
• When possible, use the real system (not mockups, screenshots)
• Use real users (otherwise they won’t get engaged or provide valid feedback)
• Use real or realistic task scenario (to get as close to the users’ reality as
possible)
There’s a lot
to consider…
An example of user study flow (eye-tracking)
1. Welcome the participant
2. Explain what the study is about and how we collect data
3. Get their consent
4. Calibrate the devices
5. Start the recording
6. Make notes
7. When participant finishes, ask some additional questions
8. Thank them (preferably with a gift card ☺)
Tips on running user studies
(based on experience ☺)
• Run 4-6 pilots with different people (include people from outside the
team)
• Be realistic with timing (leave empty time!)
• Always test the machines before, on the spot (Internet….)
• Use two stations
• Have a Plan B (for machines, Internet, people…)
Tips for writing questionnaire introduction
• Be positive (engage people)
• Tell what the study is about (in plain language)
• Tell how long answering will take
• Tell who they can contact if they have questions
Example
What type of data to collect?
• Real engagement data, e.g. eye-tracking, mouse tracking
→ tells what’s going on
• Think aloud data, e.g. notes, voice recordings, transcriptions
→ tells why it is going on
What metrics to analyze?
Metric Type Definition
Success rate Objective How many were able to carry out the
task successfully?
Time to success Objective How long did it take to complete the
task?
Ratio of time spent in AOI Objective How long was the participant
hovering/looking at an area of the
screen?
Frequency of mentions Objective How many times the participant
mentioned a specific term (e.g.,
confusion vocabulary,
positive/negative)?
Explicit ratings Subjective Participants’ judgments on the
system or content (e.g., “How easy
was it to…?)
What to ask?
• Open feedback (“How did you find the system?” “Does the system
solve a practical problem for you? If so, what?”)
• Background information:
• Age
• Gender
• Profession
• Experience with the system / in the domain
• (other things that might be confounding factors: country, native
language…)
When interviewing, ask about the users’
problems, not about the system:
• What is your typical workday like?
• What type of tasks you have in your job?
• What problems do you face in those tasks?
• Interesting…can you give a real example?
“How do I know if participants are paying
attention during studies?”
“Attention checks, or Instructional Manipulation Checks (IMCs), are a
straightforward and simple way to determine who does or doesn't pay
attention to your study instructions (Oppenheimer, Mayvis, &
Davidenko, 2009).
We recommend that you always have at least one attention check in
any given study.”
Source: www.prolific.ac (a great survey platform)
Example:
Rejecting!
This respondent failed the attention test. He or
she was supposed to select “slightly agree” but
selected “strongly agree”. Not likely that he or
she paid attention to the survey -- cannot use
his or her answers.
😢
Challenges with crowd
• Need to communicate super simply (pilot, pilot, pilot!) →
instructions, task description
• Quality control (give example from hate interpretation paper, the list)
• Lack of validity → they’re NOT real users!
• Therefore, you can test UI aspects and general information
processing/usability stuff, but not validate sources of value
Should I use the crowd?
Pros Cons
+Get a large number of participants
easily and cheaply (whereas recruiting
real users is costly and difficult)
-Not real user feedback! (validity
problem)
+Fast to collect data and iterate on
study designs (whereas real user
studies are “hit or miss”)
-Quality concerns (but can also be in real
user studies)
Managing the crowd (Pitkänen & Salminen, 2013)
Pitkänen, L., & Salminen, J. (2013). Managing the Crowd: A Study on
Videography Application. In Proceedings of Applied Business and
Entrepreneurship Association International (ABEAI 2013) Conference,
Honolulu, Hawaii, USA, 14–20 November.
How to analyze the results?
• For quantitative: R/Python/SPSS/Excel
• For qualitative analysis: Word/Excel is fine 90% of the time
• Only need Nvivo/Atlas when multiple layers of complexity
• (e.g., you need to label notes with multiple codes, e.g. “RESEARCH,CHI”,
“RESEARCH,UMAP”; assign priority (e.g., 1-5))
Visualizing eye-tracking data (heatmaps)
Surprising findings from user study
“Find match…no idea what this means…” → and we thought it was
DEAD SIMPLE! *not so* for users.
• (always record think-alouds → transcribe to analyze again)
• Remember: you are not your user. (always test your designs and
features against *real* users)
“To discover real user needs, we’ve been carrying
out several user studies. However, there are many
issues in conducting user studies. The feedback we
get is not always relevant or valid.”
https://www.linkedin.com/pulse/how-identify-useful-user-feedback-three-tips-value-joni-salminen/
The issues in interpreting user feedback
• Some participants might not be truly engaged or interested in
the system and just participate out of duty or because they
were “forced to”.
• Similarly, users may just brainstorm features that really they
would not use but that “sound cool”. (leading to feature
creep)
• Moreover, when compiling the feedback, we find that there
are a lot of requests for new features. Say, the users want 10
new features, but we have time and resources for two and
thus need to prioritize.
How to filter useful user feedback? 3 tips:
1. Who does the feeback come from? → not all people are engaged, motivated, or
knowledgeable to give useful feedback. Therefore, we have to consider if a person is
just “shooting ideas” or if he or she actually wants to provide useful feedback. We
then prioritize the comments from the people whose feedback indicates they are
taking the commenting more seriously.
2. How repetitive is the feedback? → if the request comes from many organizations and
many people within an organization, it is more likely to be a real problem to solve. If
it’s a rare request, the problem is probably also very rare and worthy to focus on.
3. Is the feedback traceable to a real problem the user has? → we want to know if the
request if a nice-to-have or pain killer. We want to solve real problems with the
system, so nice-to-haves must be minimized. Even if many motivated people suggest a
new feature, it could still be a nice-to-have if we cannot logically connect it to a real
problem.
How to filter useful user feedback? 3 tips:
1. Who does the feeback come from? → not all people are engaged, motivated, or
knowledgeable to give useful feedback. Therefore, we have to consider if a person is
just “shooting ideas” or if he or she actually wants to provide useful feedback. We
then prioritize the comments from the people whose feedback indicates they are
taking the commenting more seriously.
2. How repetitive is the feedback? → if the request comes from many organizations and
many people within an organization, it is more likely to be a real problem to solve. If
it’s a rare request, the problem is probably also very rare and worthy to focus on.
3. Is the feedback traceable to a real problem the user has? → we want to know if the
request if a nice-to-have or pain killer. We want to solve real problems with the
system, so nice-to-haves must be minimized. Even if many motivated people suggest a
new feature, it could still be a nice-to-have if we cannot logically connect it to a real
problem.
How to filter useful user feedback? 3 tips:
1. Who does the feeback come from? → not all people are engaged, motivated, or
knowledgeable to give useful feedback. Therefore, we have to consider if a person is
just “shooting ideas” or if he or she actually wants to provide useful feedback. We
then prioritize the comments from the people whose feedback indicates they are
taking the commenting more seriously.
2. How repetitive is the feedback? → if the request comes from many organizations and
many people within an organization, it is more likely to be a real problem to solve. If
it’s a rare request, the problem is probably also very rare and worthy to focus on.
3. Is the feedback traceable to a real problem the user has? → we want to know if the
request if a nice-to-have or pain killer. We want to solve real problems with the
system, so nice-to-haves must be minimized. Even if many motivated people suggest a
new feature, it could still be a nice-to-have if we cannot logically connect it to a real
problem.
Other things to be cautious about
• Social desirability bias (getting *too high* scores)
• Not all adoption problems can be solved with tech. You also need
marketing, time, product champions, etc. (barriers for adoption)
Example of social desirability bias
(QR Workshop 2017)
…yet, 0 active users from QR!
(who would log-in and use the system)
Lessons:
• What people say and do are two different things
• System ratings are poorly correlated with actual behavior
• If you want most reliable data, go for behavioral data
• If you want to ask people, ask about their problems, not about
solutions (when you understand the problem fully, you can create a
solution for it)
How to communicate results to engineer?
• Difficult (sometimes they don’t care ☺)
• Best to make engineer run test sessions him/herself
• Best to have a good relationship and communicate with
them (I love Soon ❤)
How to communicate results to engineer?
• Use quotes, stories, drawings…
• Use numbers (how many people ex
• Convince your engineer the problem really matters for users (buy-in)
• …again, best if they self experience user feedback. (internalization
problem)
Remember: Many engineers
think “user is stupid” when
they don’t understand how
to use the system.
That’s NOT the case!
Example of
communicating
results:
1. Draw on
the spot
2. Take a
screenshot
3. Do a
mock-up
4. Explain to
developer
Done!
Example of communicating results
1. …Joni has an idea
2. …runs to Soon’s office
3. …Soon puts it into the source code
4. Done!
5. (repeat 100 times ☺)
Example
Solves a real problem (usability)
experienced by many users.
Next step: validate the solution in a new user study!
(…and make sure the solution didn’t create a new
problem)
…remember, iteration: user study → system
development
Usability problems are not (that)
important!
…you can solve all the problems,
but if there is no value, your
system fails.
Usability problems are important
(--- when they are blocking access
to value)
Challenges for design guidelines
• How to prioritize one change over another?
• How to imagine the effect of change in user behavior?
• How to distinguish between small UI problem and really
big problems (i.e., the system being “useless”)
Example
of
communicating
results
How to read the table
• Frequency mentioned = How often participants mentioned this issue
• Difficulty to fix = Assessment of how much time it takes
• Priority = Assessment of how much this would make us close to “moment of
value”
• Four types:
• Definition (text)
• UI change
• System change
• New feature
• Three levels of difficulty (correlate quite well with types)
• Low (takes ~5mins from Soon (or maybe a day at max.))
• Medium (takes 1-3 weeks from Soon)
• High (requires 3-6 months of research )
How to read the table
• Frequency mentioned = How often participants mentioned this issue
• Difficulty to fix = Assessment of how much time it takes
• Priority = Assessment of how much this would make us close to
“moment of value”
NOTE: It’s important to be able to accurately
estimate the level of difficulty! (…you need to have
a certain level of understanding about software
development)
…and, you need to know how quick / skilled your
developer is! (again, I  Soon)
Prioritization: Take the quick wins (low hanging
fruits) first.
Then, work systematically to solve the more
challenging problems.
Bottom line: User studies have many benefits
• Build relationship with users / client organizations
• Learn what’s going on in the real world (real people’s problems)
• Solve research problems that matter (and make some good
publications ☺)
• Develop better systems that actually get used
Thanks for listening!
(….and thanks A LOT for participating in pilot studies ☺)
(ps. Full slides at SlideShare: www.slideshare.com/jonis12)
(pps. Visit https://persona.qcri.org for demo!)

More Related Content

What's hot

Point of-interest recommendation for location promotion in location-based soc...
Point of-interest recommendation for location promotion in location-based soc...Point of-interest recommendation for location promotion in location-based soc...
Point of-interest recommendation for location promotion in location-based soc...CloudTechnologies
 
IRJET- Hybrid Book Recommendation System
IRJET- Hybrid Book Recommendation SystemIRJET- Hybrid Book Recommendation System
IRJET- Hybrid Book Recommendation SystemIRJET Journal
 
Social Recommender Systems
Social Recommender SystemsSocial Recommender Systems
Social Recommender Systemsguest77b0cd12
 
Preference Elicitation in Recommender Systems
Preference Elicitation in Recommender SystemsPreference Elicitation in Recommender Systems
Preference Elicitation in Recommender SystemsAnish Shenoy
 
Tutorial on Relationship Mining In Online Social Networks
Tutorial on Relationship Mining In Online Social NetworksTutorial on Relationship Mining In Online Social Networks
Tutorial on Relationship Mining In Online Social Networkspjing2
 
A literature survey on recommendation
A literature survey on recommendationA literature survey on recommendation
A literature survey on recommendationaciijournal
 
Proposal final
Proposal finalProposal final
Proposal finalMido Razaz
 
IRJET- Analysis on Existing Methodologies of User Service Rating Prediction S...
IRJET- Analysis on Existing Methodologies of User Service Rating Prediction S...IRJET- Analysis on Existing Methodologies of User Service Rating Prediction S...
IRJET- Analysis on Existing Methodologies of User Service Rating Prediction S...IRJET Journal
 
Recommendation systems
Recommendation systems  Recommendation systems
Recommendation systems Badr Hirchoua
 
Analysis, design and implementation of a Multi-Criteria Recommender System ba...
Analysis, design and implementation of a Multi-Criteria Recommender System ba...Analysis, design and implementation of a Multi-Criteria Recommender System ba...
Analysis, design and implementation of a Multi-Criteria Recommender System ba...Davide Giannico
 
Active Learning in Collaborative Filtering Recommender Systems : a Survey
Active Learning in Collaborative Filtering Recommender Systems : a SurveyActive Learning in Collaborative Filtering Recommender Systems : a Survey
Active Learning in Collaborative Filtering Recommender Systems : a SurveyUniversity of Bergen
 
Information Retrieval Models for Recommender Systems - PhD slides
Information Retrieval Models for Recommender Systems - PhD slidesInformation Retrieval Models for Recommender Systems - PhD slides
Information Retrieval Models for Recommender Systems - PhD slidesDaniel Valcarce
 
Social Recommender Systems Tutorial - WWW 2011
Social Recommender Systems Tutorial - WWW 2011Social Recommender Systems Tutorial - WWW 2011
Social Recommender Systems Tutorial - WWW 2011idoguy
 
Movie recommendation system using collaborative filtering system
Movie recommendation system using collaborative filtering system Movie recommendation system using collaborative filtering system
Movie recommendation system using collaborative filtering system Mauryasuraj98
 
Guide to Recommender Systems
Guide to Recommender SystemsGuide to Recommender Systems
Guide to Recommender SystemsAmancio Bouza
 
Best Practices in Recommender System Challenges
Best Practices in Recommender System ChallengesBest Practices in Recommender System Challenges
Best Practices in Recommender System ChallengesAlan Said
 
Review on Opinion Targets and Opinion Words Extraction Techniques from Online...
Review on Opinion Targets and Opinion Words Extraction Techniques from Online...Review on Opinion Targets and Opinion Words Extraction Techniques from Online...
Review on Opinion Targets and Opinion Words Extraction Techniques from Online...IRJET Journal
 

What's hot (20)

Point of-interest recommendation for location promotion in location-based soc...
Point of-interest recommendation for location promotion in location-based soc...Point of-interest recommendation for location promotion in location-based soc...
Point of-interest recommendation for location promotion in location-based soc...
 
IRJET- Hybrid Book Recommendation System
IRJET- Hybrid Book Recommendation SystemIRJET- Hybrid Book Recommendation System
IRJET- Hybrid Book Recommendation System
 
Social Recommender Systems
Social Recommender SystemsSocial Recommender Systems
Social Recommender Systems
 
20120140506003
2012014050600320120140506003
20120140506003
 
Preference Elicitation in Recommender Systems
Preference Elicitation in Recommender SystemsPreference Elicitation in Recommender Systems
Preference Elicitation in Recommender Systems
 
Tutorial on Relationship Mining In Online Social Networks
Tutorial on Relationship Mining In Online Social NetworksTutorial on Relationship Mining In Online Social Networks
Tutorial on Relationship Mining In Online Social Networks
 
Recommender system
Recommender systemRecommender system
Recommender system
 
A literature survey on recommendation
A literature survey on recommendationA literature survey on recommendation
A literature survey on recommendation
 
Proposal final
Proposal finalProposal final
Proposal final
 
IRJET- Analysis on Existing Methodologies of User Service Rating Prediction S...
IRJET- Analysis on Existing Methodologies of User Service Rating Prediction S...IRJET- Analysis on Existing Methodologies of User Service Rating Prediction S...
IRJET- Analysis on Existing Methodologies of User Service Rating Prediction S...
 
Recommendation systems
Recommendation systems  Recommendation systems
Recommendation systems
 
Analysis, design and implementation of a Multi-Criteria Recommender System ba...
Analysis, design and implementation of a Multi-Criteria Recommender System ba...Analysis, design and implementation of a Multi-Criteria Recommender System ba...
Analysis, design and implementation of a Multi-Criteria Recommender System ba...
 
Active Learning in Collaborative Filtering Recommender Systems : a Survey
Active Learning in Collaborative Filtering Recommender Systems : a SurveyActive Learning in Collaborative Filtering Recommender Systems : a Survey
Active Learning in Collaborative Filtering Recommender Systems : a Survey
 
Using social network sites
Using social network sites Using social network sites
Using social network sites
 
Information Retrieval Models for Recommender Systems - PhD slides
Information Retrieval Models for Recommender Systems - PhD slidesInformation Retrieval Models for Recommender Systems - PhD slides
Information Retrieval Models for Recommender Systems - PhD slides
 
Social Recommender Systems Tutorial - WWW 2011
Social Recommender Systems Tutorial - WWW 2011Social Recommender Systems Tutorial - WWW 2011
Social Recommender Systems Tutorial - WWW 2011
 
Movie recommendation system using collaborative filtering system
Movie recommendation system using collaborative filtering system Movie recommendation system using collaborative filtering system
Movie recommendation system using collaborative filtering system
 
Guide to Recommender Systems
Guide to Recommender SystemsGuide to Recommender Systems
Guide to Recommender Systems
 
Best Practices in Recommender System Challenges
Best Practices in Recommender System ChallengesBest Practices in Recommender System Challenges
Best Practices in Recommender System Challenges
 
Review on Opinion Targets and Opinion Words Extraction Techniques from Online...
Review on Opinion Targets and Opinion Words Extraction Techniques from Online...Review on Opinion Targets and Opinion Words Extraction Techniques from Online...
Review on Opinion Targets and Opinion Words Extraction Techniques from Online...
 

Similar to User Studies for APG: How to support system development with user feedback?

UX Design Process | Sample Proposal
UX Design Process | Sample Proposal UX Design Process | Sample Proposal
UX Design Process | Sample Proposal Marta Fioni
 
UXPA 2023: Learn how to get over personas by swiping right on user roles
UXPA 2023: Learn how to get over personas by swiping right on user rolesUXPA 2023: Learn how to get over personas by swiping right on user roles
UXPA 2023: Learn how to get over personas by swiping right on user rolesUXPA International
 
Homespun UX: Going Beyond Web Analytics
Homespun UX: Going Beyond Web AnalyticsHomespun UX: Going Beyond Web Analytics
Homespun UX: Going Beyond Web AnalyticsMary Ann Brody
 
Ucd Techniques - Shad MUN 2008
Ucd Techniques - Shad MUN 2008Ucd Techniques - Shad MUN 2008
Ucd Techniques - Shad MUN 2008Patañjali Chary
 
UCD Workshop - Shad MUN 2008
UCD Workshop - Shad MUN 2008UCD Workshop - Shad MUN 2008
UCD Workshop - Shad MUN 2008guest63c15b
 
The Net Promoter Score - What can NPS Tell you about your User Experience
The Net Promoter Score - What can NPS Tell you about your User ExperienceThe Net Promoter Score - What can NPS Tell you about your User Experience
The Net Promoter Score - What can NPS Tell you about your User ExperienceUXPA International
 
Research Operations at Scale (Christian Rohrer at DesignOps Summit 2017)
Research Operations at Scale (Christian Rohrer at DesignOps Summit 2017)Research Operations at Scale (Christian Rohrer at DesignOps Summit 2017)
Research Operations at Scale (Christian Rohrer at DesignOps Summit 2017)Rosenfeld Media
 
180719 테헤란로 런치클럽-실무자 관점에서 본 Spotify의 유저리서치 방식 (Spotify 백원희님)
180719 테헤란로 런치클럽-실무자 관점에서 본 Spotify의 유저리서치 방식 (Spotify 백원희님)180719 테헤란로 런치클럽-실무자 관점에서 본 Spotify의 유저리서치 방식 (Spotify 백원희님)
180719 테헤란로 런치클럽-실무자 관점에서 본 Spotify의 유저리서치 방식 (Spotify 백원희님)StartupAlliance
 
Combining Methods: Web Analytics and User Testing
Combining Methods: Web Analytics and User TestingCombining Methods: Web Analytics and User Testing
Combining Methods: Web Analytics and User TestingUser Intelligence
 
Understanding User Experience Workshop - Interlink Conference 2012
Understanding User Experience Workshop - Interlink Conference 2012Understanding User Experience Workshop - Interlink Conference 2012
Understanding User Experience Workshop - Interlink Conference 2012Lynne Polischuik
 
User Zoom Webinar Monster Aug09 Vf
User Zoom Webinar Monster Aug09 VfUser Zoom Webinar Monster Aug09 Vf
User Zoom Webinar Monster Aug09 VfUserZoom
 
Survey Research In Empirical Software Engineering
Survey Research In Empirical Software EngineeringSurvey Research In Empirical Software Engineering
Survey Research In Empirical Software Engineeringalessio_ferrari
 
Maturing User Research in a Unicorn - UXSEA Summit 2019
Maturing User Research in a Unicorn - UXSEA Summit 2019Maturing User Research in a Unicorn - UXSEA Summit 2019
Maturing User Research in a Unicorn - UXSEA Summit 2019Kuldeep Kulshreshtha
 
Chapter 7 - Evaluation Tekhnique
Chapter 7 - Evaluation TekhniqueChapter 7 - Evaluation Tekhnique
Chapter 7 - Evaluation TekhniqueMuhammad Najib
 
Tutorial on metrics of user engagement -- Applications to Search & E- commerce
Tutorial on metrics of user engagement -- Applications to Search & E- commerceTutorial on metrics of user engagement -- Applications to Search & E- commerce
Tutorial on metrics of user engagement -- Applications to Search & E- commerceMounia Lalmas-Roelleke
 
Telling the Full Story: Adding Qualitative Data To Executive Dashboards
Telling the Full Story: Adding Qualitative Data To Executive DashboardsTelling the Full Story: Adding Qualitative Data To Executive Dashboards
Telling the Full Story: Adding Qualitative Data To Executive DashboardsUserZoom
 
Growth Hacking with Lean UX
Growth Hacking with Lean UXGrowth Hacking with Lean UX
Growth Hacking with Lean UXJohn Rockson
 

Similar to User Studies for APG: How to support system development with user feedback? (20)

UX Design Process | Sample Proposal
UX Design Process | Sample Proposal UX Design Process | Sample Proposal
UX Design Process | Sample Proposal
 
UXPA 2023: Learn how to get over personas by swiping right on user roles
UXPA 2023: Learn how to get over personas by swiping right on user rolesUXPA 2023: Learn how to get over personas by swiping right on user roles
UXPA 2023: Learn how to get over personas by swiping right on user roles
 
Homespun UX: Going Beyond Web Analytics
Homespun UX: Going Beyond Web AnalyticsHomespun UX: Going Beyond Web Analytics
Homespun UX: Going Beyond Web Analytics
 
Lean User Testing Intro
Lean User Testing IntroLean User Testing Intro
Lean User Testing Intro
 
Ucd Techniques - Shad MUN 2008
Ucd Techniques - Shad MUN 2008Ucd Techniques - Shad MUN 2008
Ucd Techniques - Shad MUN 2008
 
UCD Workshop - Shad MUN 2008
UCD Workshop - Shad MUN 2008UCD Workshop - Shad MUN 2008
UCD Workshop - Shad MUN 2008
 
The Net Promoter Score - What can NPS Tell you about your User Experience
The Net Promoter Score - What can NPS Tell you about your User ExperienceThe Net Promoter Score - What can NPS Tell you about your User Experience
The Net Promoter Score - What can NPS Tell you about your User Experience
 
Research Operations at Scale (Christian Rohrer at DesignOps Summit 2017)
Research Operations at Scale (Christian Rohrer at DesignOps Summit 2017)Research Operations at Scale (Christian Rohrer at DesignOps Summit 2017)
Research Operations at Scale (Christian Rohrer at DesignOps Summit 2017)
 
180719 테헤란로 런치클럽-실무자 관점에서 본 Spotify의 유저리서치 방식 (Spotify 백원희님)
180719 테헤란로 런치클럽-실무자 관점에서 본 Spotify의 유저리서치 방식 (Spotify 백원희님)180719 테헤란로 런치클럽-실무자 관점에서 본 Spotify의 유저리서치 방식 (Spotify 백원희님)
180719 테헤란로 런치클럽-실무자 관점에서 본 Spotify의 유저리서치 방식 (Spotify 백원희님)
 
Combining Methods: Web Analytics and User Testing
Combining Methods: Web Analytics and User TestingCombining Methods: Web Analytics and User Testing
Combining Methods: Web Analytics and User Testing
 
When Qual Met UX
When Qual Met UXWhen Qual Met UX
When Qual Met UX
 
Understanding User Experience Workshop - Interlink Conference 2012
Understanding User Experience Workshop - Interlink Conference 2012Understanding User Experience Workshop - Interlink Conference 2012
Understanding User Experience Workshop - Interlink Conference 2012
 
User Zoom Webinar Monster Aug09 Vf
User Zoom Webinar Monster Aug09 VfUser Zoom Webinar Monster Aug09 Vf
User Zoom Webinar Monster Aug09 Vf
 
Survey Research In Empirical Software Engineering
Survey Research In Empirical Software EngineeringSurvey Research In Empirical Software Engineering
Survey Research In Empirical Software Engineering
 
Maturing User Research in a Unicorn - UXSEA Summit 2019
Maturing User Research in a Unicorn - UXSEA Summit 2019Maturing User Research in a Unicorn - UXSEA Summit 2019
Maturing User Research in a Unicorn - UXSEA Summit 2019
 
Chapter 7 - Evaluation Tekhnique
Chapter 7 - Evaluation TekhniqueChapter 7 - Evaluation Tekhnique
Chapter 7 - Evaluation Tekhnique
 
Designing Mobile UX
Designing Mobile UXDesigning Mobile UX
Designing Mobile UX
 
Tutorial on metrics of user engagement -- Applications to Search & E- commerce
Tutorial on metrics of user engagement -- Applications to Search & E- commerceTutorial on metrics of user engagement -- Applications to Search & E- commerce
Tutorial on metrics of user engagement -- Applications to Search & E- commerce
 
Telling the Full Story: Adding Qualitative Data To Executive Dashboards
Telling the Full Story: Adding Qualitative Data To Executive DashboardsTelling the Full Story: Adding Qualitative Data To Executive Dashboards
Telling the Full Story: Adding Qualitative Data To Executive Dashboards
 
Growth Hacking with Lean UX
Growth Hacking with Lean UXGrowth Hacking with Lean UX
Growth Hacking with Lean UX
 

More from Joni Salminen

Automatic Persona Generation: Introduction & Current Challenges
Automatic Persona Generation: Introduction & Current ChallengesAutomatic Persona Generation: Introduction & Current Challenges
Automatic Persona Generation: Introduction & Current ChallengesJoni Salminen
 
Five NLP Challenges in Data-Driven Personas
Five NLP Challenges in Data-Driven PersonasFive NLP Challenges in Data-Driven Personas
Five NLP Challenges in Data-Driven PersonasJoni Salminen
 
Problem of majority voting
Problem of majority votingProblem of majority voting
Problem of majority votingJoni Salminen
 
Persona Analytics: Progress Report and Road Ahead
Persona Analytics: Progress Report and Road AheadPersona Analytics: Progress Report and Road Ahead
Persona Analytics: Progress Report and Road AheadJoni Salminen
 
Enriching social media personas with personality traits
Enriching social media personas with personality traitsEnriching social media personas with personality traits
Enriching social media personas with personality traitsJoni Salminen
 
To Use Branded Keywords or Not? Rationale of Professional Search-engine Marke...
To Use Branded Keywords or Not? Rationale of Professional Search-engine Marke...To Use Branded Keywords or Not? Rationale of Professional Search-engine Marke...
To Use Branded Keywords or Not? Rationale of Professional Search-engine Marke...Joni Salminen
 
Determining Online Brand Reputation with Machine Learning from Social Media M...
Determining Online Brand Reputation with Machine Learning from Social Media M...Determining Online Brand Reputation with Machine Learning from Social Media M...
Determining Online Brand Reputation with Machine Learning from Social Media M...Joni Salminen
 
Is More Better?: Impact of Multiple Photos on Perception of Persona Profiles
Is More Better?: Impact of Multiple Photos on Perception of Persona ProfilesIs More Better?: Impact of Multiple Photos on Perception of Persona Profiles
Is More Better?: Impact of Multiple Photos on Perception of Persona ProfilesJoni Salminen
 
Anatomy of Online Hate: Developing a Taxonomy and Machine Learning Models for...
Anatomy of Online Hate: Developing a Taxonomy and Machine Learning Models for...Anatomy of Online Hate: Developing a Taxonomy and Machine Learning Models for...
Anatomy of Online Hate: Developing a Taxonomy and Machine Learning Models for...Joni Salminen
 
OSS-EBM: Open Source Software Entrepreneurial Business Modelling
OSS-EBM: Open Source Software Entrepreneurial Business ModellingOSS-EBM: Open Source Software Entrepreneurial Business Modelling
OSS-EBM: Open Source Software Entrepreneurial Business ModellingJoni Salminen
 
Gender effect on e-commerce sales of experience gifts: Preliminary empirical ...
Gender effect on e-commerce sales of experience gifts: Preliminary empirical ...Gender effect on e-commerce sales of experience gifts: Preliminary empirical ...
Gender effect on e-commerce sales of experience gifts: Preliminary empirical ...Joni Salminen
 
Tips for Scale Development: Evaluating Automatic Personas
Tips for Scale Development: Evaluating Automatic PersonasTips for Scale Development: Evaluating Automatic Personas
Tips for Scale Development: Evaluating Automatic PersonasJoni Salminen
 
Big Data, Small Personas: Research Agenda for Automatic Persona Generation
Big Data, Small Personas: Research Agenda for Automatic Persona GenerationBig Data, Small Personas: Research Agenda for Automatic Persona Generation
Big Data, Small Personas: Research Agenda for Automatic Persona GenerationJoni Salminen
 
Why do startups avoid difficult problems?
Why do startups avoid difficult problems?Why do startups avoid difficult problems?
Why do startups avoid difficult problems?Joni Salminen
 
Social Espionage: Drawing Benefit from Competitors’ Social Media Presence
Social Espionage: Drawing Benefit from Competitors’ Social Media PresenceSocial Espionage: Drawing Benefit from Competitors’ Social Media Presence
Social Espionage: Drawing Benefit from Competitors’ Social Media PresenceJoni Salminen
 
Strategic Digital Marketing (Digital Marketing '15 @ Oulu University)
Strategic Digital Marketing (Digital Marketing '15 @ Oulu University)Strategic Digital Marketing (Digital Marketing '15 @ Oulu University)
Strategic Digital Marketing (Digital Marketing '15 @ Oulu University)Joni Salminen
 
Social Media Marketing (Digital Marketing '15 @ Oulu University)
Social Media Marketing (Digital Marketing '15 @ Oulu University)Social Media Marketing (Digital Marketing '15 @ Oulu University)
Social Media Marketing (Digital Marketing '15 @ Oulu University)Joni Salminen
 
Search Engine Marketing (Digital Marketing '15 @ Oulu University)
Search Engine Marketing (Digital Marketing '15 @ Oulu University)Search Engine Marketing (Digital Marketing '15 @ Oulu University)
Search Engine Marketing (Digital Marketing '15 @ Oulu University)Joni Salminen
 
Introduction to Digital Marketing (Digital Marketing '15 @ Oulu University)
Introduction to Digital Marketing (Digital Marketing '15 @ Oulu University)Introduction to Digital Marketing (Digital Marketing '15 @ Oulu University)
Introduction to Digital Marketing (Digital Marketing '15 @ Oulu University)Joni Salminen
 
Web Analytics (Digital Marketing '15 @ Oulu University)
Web Analytics (Digital Marketing '15 @ Oulu University)Web Analytics (Digital Marketing '15 @ Oulu University)
Web Analytics (Digital Marketing '15 @ Oulu University)Joni Salminen
 

More from Joni Salminen (20)

Automatic Persona Generation: Introduction & Current Challenges
Automatic Persona Generation: Introduction & Current ChallengesAutomatic Persona Generation: Introduction & Current Challenges
Automatic Persona Generation: Introduction & Current Challenges
 
Five NLP Challenges in Data-Driven Personas
Five NLP Challenges in Data-Driven PersonasFive NLP Challenges in Data-Driven Personas
Five NLP Challenges in Data-Driven Personas
 
Problem of majority voting
Problem of majority votingProblem of majority voting
Problem of majority voting
 
Persona Analytics: Progress Report and Road Ahead
Persona Analytics: Progress Report and Road AheadPersona Analytics: Progress Report and Road Ahead
Persona Analytics: Progress Report and Road Ahead
 
Enriching social media personas with personality traits
Enriching social media personas with personality traitsEnriching social media personas with personality traits
Enriching social media personas with personality traits
 
To Use Branded Keywords or Not? Rationale of Professional Search-engine Marke...
To Use Branded Keywords or Not? Rationale of Professional Search-engine Marke...To Use Branded Keywords or Not? Rationale of Professional Search-engine Marke...
To Use Branded Keywords or Not? Rationale of Professional Search-engine Marke...
 
Determining Online Brand Reputation with Machine Learning from Social Media M...
Determining Online Brand Reputation with Machine Learning from Social Media M...Determining Online Brand Reputation with Machine Learning from Social Media M...
Determining Online Brand Reputation with Machine Learning from Social Media M...
 
Is More Better?: Impact of Multiple Photos on Perception of Persona Profiles
Is More Better?: Impact of Multiple Photos on Perception of Persona ProfilesIs More Better?: Impact of Multiple Photos on Perception of Persona Profiles
Is More Better?: Impact of Multiple Photos on Perception of Persona Profiles
 
Anatomy of Online Hate: Developing a Taxonomy and Machine Learning Models for...
Anatomy of Online Hate: Developing a Taxonomy and Machine Learning Models for...Anatomy of Online Hate: Developing a Taxonomy and Machine Learning Models for...
Anatomy of Online Hate: Developing a Taxonomy and Machine Learning Models for...
 
OSS-EBM: Open Source Software Entrepreneurial Business Modelling
OSS-EBM: Open Source Software Entrepreneurial Business ModellingOSS-EBM: Open Source Software Entrepreneurial Business Modelling
OSS-EBM: Open Source Software Entrepreneurial Business Modelling
 
Gender effect on e-commerce sales of experience gifts: Preliminary empirical ...
Gender effect on e-commerce sales of experience gifts: Preliminary empirical ...Gender effect on e-commerce sales of experience gifts: Preliminary empirical ...
Gender effect on e-commerce sales of experience gifts: Preliminary empirical ...
 
Tips for Scale Development: Evaluating Automatic Personas
Tips for Scale Development: Evaluating Automatic PersonasTips for Scale Development: Evaluating Automatic Personas
Tips for Scale Development: Evaluating Automatic Personas
 
Big Data, Small Personas: Research Agenda for Automatic Persona Generation
Big Data, Small Personas: Research Agenda for Automatic Persona GenerationBig Data, Small Personas: Research Agenda for Automatic Persona Generation
Big Data, Small Personas: Research Agenda for Automatic Persona Generation
 
Why do startups avoid difficult problems?
Why do startups avoid difficult problems?Why do startups avoid difficult problems?
Why do startups avoid difficult problems?
 
Social Espionage: Drawing Benefit from Competitors’ Social Media Presence
Social Espionage: Drawing Benefit from Competitors’ Social Media PresenceSocial Espionage: Drawing Benefit from Competitors’ Social Media Presence
Social Espionage: Drawing Benefit from Competitors’ Social Media Presence
 
Strategic Digital Marketing (Digital Marketing '15 @ Oulu University)
Strategic Digital Marketing (Digital Marketing '15 @ Oulu University)Strategic Digital Marketing (Digital Marketing '15 @ Oulu University)
Strategic Digital Marketing (Digital Marketing '15 @ Oulu University)
 
Social Media Marketing (Digital Marketing '15 @ Oulu University)
Social Media Marketing (Digital Marketing '15 @ Oulu University)Social Media Marketing (Digital Marketing '15 @ Oulu University)
Social Media Marketing (Digital Marketing '15 @ Oulu University)
 
Search Engine Marketing (Digital Marketing '15 @ Oulu University)
Search Engine Marketing (Digital Marketing '15 @ Oulu University)Search Engine Marketing (Digital Marketing '15 @ Oulu University)
Search Engine Marketing (Digital Marketing '15 @ Oulu University)
 
Introduction to Digital Marketing (Digital Marketing '15 @ Oulu University)
Introduction to Digital Marketing (Digital Marketing '15 @ Oulu University)Introduction to Digital Marketing (Digital Marketing '15 @ Oulu University)
Introduction to Digital Marketing (Digital Marketing '15 @ Oulu University)
 
Web Analytics (Digital Marketing '15 @ Oulu University)
Web Analytics (Digital Marketing '15 @ Oulu University)Web Analytics (Digital Marketing '15 @ Oulu University)
Web Analytics (Digital Marketing '15 @ Oulu University)
 

Recently uploaded

Automating Google Workspace (GWS) & more with Apps Script
Automating Google Workspace (GWS) & more with Apps ScriptAutomating Google Workspace (GWS) & more with Apps Script
Automating Google Workspace (GWS) & more with Apps Scriptwesley chun
 
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...Drew Madelung
 
EIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptx
EIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptxEIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptx
EIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptxEarley Information Science
 
Exploring the Future Potential of AI-Enabled Smartphone Processors
Exploring the Future Potential of AI-Enabled Smartphone ProcessorsExploring the Future Potential of AI-Enabled Smartphone Processors
Exploring the Future Potential of AI-Enabled Smartphone Processorsdebabhi2
 
Finology Group – Insurtech Innovation Award 2024
Finology Group – Insurtech Innovation Award 2024Finology Group – Insurtech Innovation Award 2024
Finology Group – Insurtech Innovation Award 2024The Digital Insurer
 
Understanding Discord NSFW Servers A Guide for Responsible Users.pdf
Understanding Discord NSFW Servers A Guide for Responsible Users.pdfUnderstanding Discord NSFW Servers A Guide for Responsible Users.pdf
Understanding Discord NSFW Servers A Guide for Responsible Users.pdfUK Journal
 
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024The Digital Insurer
 
Presentation on how to chat with PDF using ChatGPT code interpreter
Presentation on how to chat with PDF using ChatGPT code interpreterPresentation on how to chat with PDF using ChatGPT code interpreter
Presentation on how to chat with PDF using ChatGPT code interpreternaman860154
 
Histor y of HAM Radio presentation slide
Histor y of HAM Radio presentation slideHistor y of HAM Radio presentation slide
Histor y of HAM Radio presentation slidevu2urc
 
Boost Fertility New Invention Ups Success Rates.pdf
Boost Fertility New Invention Ups Success Rates.pdfBoost Fertility New Invention Ups Success Rates.pdf
Boost Fertility New Invention Ups Success Rates.pdfsudhanshuwaghmare1
 
08448380779 Call Girls In Friends Colony Women Seeking Men
08448380779 Call Girls In Friends Colony Women Seeking Men08448380779 Call Girls In Friends Colony Women Seeking Men
08448380779 Call Girls In Friends Colony Women Seeking MenDelhi Call girls
 
08448380779 Call Girls In Diplomatic Enclave Women Seeking Men
08448380779 Call Girls In Diplomatic Enclave Women Seeking Men08448380779 Call Girls In Diplomatic Enclave Women Seeking Men
08448380779 Call Girls In Diplomatic Enclave Women Seeking MenDelhi Call girls
 
08448380779 Call Girls In Greater Kailash - I Women Seeking Men
08448380779 Call Girls In Greater Kailash - I Women Seeking Men08448380779 Call Girls In Greater Kailash - I Women Seeking Men
08448380779 Call Girls In Greater Kailash - I Women Seeking MenDelhi Call girls
 
Data Cloud, More than a CDP by Matt Robison
Data Cloud, More than a CDP by Matt RobisonData Cloud, More than a CDP by Matt Robison
Data Cloud, More than a CDP by Matt RobisonAnna Loughnan Colquhoun
 
Scaling API-first – The story of a global engineering organization
Scaling API-first – The story of a global engineering organizationScaling API-first – The story of a global engineering organization
Scaling API-first – The story of a global engineering organizationRadu Cotescu
 
Raspberry Pi 5: Challenges and Solutions in Bringing up an OpenGL/Vulkan Driv...
Raspberry Pi 5: Challenges and Solutions in Bringing up an OpenGL/Vulkan Driv...Raspberry Pi 5: Challenges and Solutions in Bringing up an OpenGL/Vulkan Driv...
Raspberry Pi 5: Challenges and Solutions in Bringing up an OpenGL/Vulkan Driv...Igalia
 
The 7 Things I Know About Cyber Security After 25 Years | April 2024
The 7 Things I Know About Cyber Security After 25 Years | April 2024The 7 Things I Know About Cyber Security After 25 Years | April 2024
The 7 Things I Know About Cyber Security After 25 Years | April 2024Rafal Los
 
Factors to Consider When Choosing Accounts Payable Services Providers.pptx
Factors to Consider When Choosing Accounts Payable Services Providers.pptxFactors to Consider When Choosing Accounts Payable Services Providers.pptx
Factors to Consider When Choosing Accounts Payable Services Providers.pptxKatpro Technologies
 
Artificial Intelligence: Facts and Myths
Artificial Intelligence: Facts and MythsArtificial Intelligence: Facts and Myths
Artificial Intelligence: Facts and MythsJoaquim Jorge
 
08448380779 Call Girls In Civil Lines Women Seeking Men
08448380779 Call Girls In Civil Lines Women Seeking Men08448380779 Call Girls In Civil Lines Women Seeking Men
08448380779 Call Girls In Civil Lines Women Seeking MenDelhi Call girls
 

Recently uploaded (20)

Automating Google Workspace (GWS) & more with Apps Script
Automating Google Workspace (GWS) & more with Apps ScriptAutomating Google Workspace (GWS) & more with Apps Script
Automating Google Workspace (GWS) & more with Apps Script
 
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...
 
EIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptx
EIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptxEIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptx
EIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptx
 
Exploring the Future Potential of AI-Enabled Smartphone Processors
Exploring the Future Potential of AI-Enabled Smartphone ProcessorsExploring the Future Potential of AI-Enabled Smartphone Processors
Exploring the Future Potential of AI-Enabled Smartphone Processors
 
Finology Group – Insurtech Innovation Award 2024
Finology Group – Insurtech Innovation Award 2024Finology Group – Insurtech Innovation Award 2024
Finology Group – Insurtech Innovation Award 2024
 
Understanding Discord NSFW Servers A Guide for Responsible Users.pdf
Understanding Discord NSFW Servers A Guide for Responsible Users.pdfUnderstanding Discord NSFW Servers A Guide for Responsible Users.pdf
Understanding Discord NSFW Servers A Guide for Responsible Users.pdf
 
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024
 
Presentation on how to chat with PDF using ChatGPT code interpreter
Presentation on how to chat with PDF using ChatGPT code interpreterPresentation on how to chat with PDF using ChatGPT code interpreter
Presentation on how to chat with PDF using ChatGPT code interpreter
 
Histor y of HAM Radio presentation slide
Histor y of HAM Radio presentation slideHistor y of HAM Radio presentation slide
Histor y of HAM Radio presentation slide
 
Boost Fertility New Invention Ups Success Rates.pdf
Boost Fertility New Invention Ups Success Rates.pdfBoost Fertility New Invention Ups Success Rates.pdf
Boost Fertility New Invention Ups Success Rates.pdf
 
08448380779 Call Girls In Friends Colony Women Seeking Men
08448380779 Call Girls In Friends Colony Women Seeking Men08448380779 Call Girls In Friends Colony Women Seeking Men
08448380779 Call Girls In Friends Colony Women Seeking Men
 
08448380779 Call Girls In Diplomatic Enclave Women Seeking Men
08448380779 Call Girls In Diplomatic Enclave Women Seeking Men08448380779 Call Girls In Diplomatic Enclave Women Seeking Men
08448380779 Call Girls In Diplomatic Enclave Women Seeking Men
 
08448380779 Call Girls In Greater Kailash - I Women Seeking Men
08448380779 Call Girls In Greater Kailash - I Women Seeking Men08448380779 Call Girls In Greater Kailash - I Women Seeking Men
08448380779 Call Girls In Greater Kailash - I Women Seeking Men
 
Data Cloud, More than a CDP by Matt Robison
Data Cloud, More than a CDP by Matt RobisonData Cloud, More than a CDP by Matt Robison
Data Cloud, More than a CDP by Matt Robison
 
Scaling API-first – The story of a global engineering organization
Scaling API-first – The story of a global engineering organizationScaling API-first – The story of a global engineering organization
Scaling API-first – The story of a global engineering organization
 
Raspberry Pi 5: Challenges and Solutions in Bringing up an OpenGL/Vulkan Driv...
Raspberry Pi 5: Challenges and Solutions in Bringing up an OpenGL/Vulkan Driv...Raspberry Pi 5: Challenges and Solutions in Bringing up an OpenGL/Vulkan Driv...
Raspberry Pi 5: Challenges and Solutions in Bringing up an OpenGL/Vulkan Driv...
 
The 7 Things I Know About Cyber Security After 25 Years | April 2024
The 7 Things I Know About Cyber Security After 25 Years | April 2024The 7 Things I Know About Cyber Security After 25 Years | April 2024
The 7 Things I Know About Cyber Security After 25 Years | April 2024
 
Factors to Consider When Choosing Accounts Payable Services Providers.pptx
Factors to Consider When Choosing Accounts Payable Services Providers.pptxFactors to Consider When Choosing Accounts Payable Services Providers.pptx
Factors to Consider When Choosing Accounts Payable Services Providers.pptx
 
Artificial Intelligence: Facts and Myths
Artificial Intelligence: Facts and MythsArtificial Intelligence: Facts and Myths
Artificial Intelligence: Facts and Myths
 
08448380779 Call Girls In Civil Lines Women Seeking Men
08448380779 Call Girls In Civil Lines Women Seeking Men08448380779 Call Girls In Civil Lines Women Seeking Men
08448380779 Call Girls In Civil Lines Women Seeking Men
 

User Studies for APG: How to support system development with user feedback?

  • 1. User Studies for APG How to support system development with user feedback?
  • 2. Purpose of presentation 1. To share some useful tips and lessons from user studies we’ve done 2. To explain how to incorporate user studies into system development flow 3. To encourage more user studies in Social Computing
  • 3. Outline • The APG research roadmap • Overview of our user studies • Framework for Value-Driven Development™ • Examples of putting research into practice • Tips for planning and executing user studies
  • 4. What is a “user study”? • A practically oriented research study that gathers real feedback from real users, based on real scenarios, testing the usability and functionality of a real system. • User study is meant to produce information that helps improve the system!
  • 5. Why do we need user studies? 1. Break the researcher echo chamber – learn what matters for people working in the industry 2. Find new use cases for the technologies you’re developing 3. Disseminate our research and systems (get downloads, citations, real users signing up… ➔ promotion/raise !!!) 4. Maintain relationships to stakeholder organizations and their key people 5. Get real feedback from real users to improve real systems ➔ the only path to real impact
  • 6. There is an infinitely gigantic leap from having “users” to having real users …meaning: most registered people never use the system in reality. Real user = active user
  • 7. Getting your first active user is like achieving the first moon landing. …meaning: a small leap for mankind, a giant leap for your system.
  • 8. The same problem is faced by startups and QCRI projects (Download Joni’s dissertation “Startup Dilemmas” FOR FREE: http://www.utupub.fi/h andle/10024/99349)
  • 10. Research Roadmap for Automatic Persona Generation Information architecture: How to choose the correct information elements and layouts for a given user, use case or industry? Quotes: How to find representative, contextually relevant, and non-distracting comments describing the persona? Evaluation: 1) How to ensure personas are complete, clear, consistent and credible? 2) How to measure usefulness of personas for individuals and organizations? Attributes: How to infer user attributes, such as psychographics, needs and wants, political orientation, and brand affinities? Temporal analysis: How to analyze change and stability of personas over time? Image: How to retrieve, generate, and choose persona profile pictures? Finding better ways to process and choose useful user information from vast amounts of online data. ”Personas are about giving faces to data.” – Jim Jansen Applicability: How to create personas for different contexts (e.g., e-commerce personas, political personas, gamer personas…)?
  • 11. In our research roadmap, user studies are relevant in most if not all sections…
  • 12. Research Roadmap for Automatic Persona Generation Information architecture: How to choose the correct information elements and layouts for a given user, use case or industry? Quotes: How to find representative, contextually relevant, and non-distracting comments describing the persona? Evaluation: 1) How to ensure personas are complete, clear, consistent and credible? 2) How to measure usefulness of personas for individuals and organizations? Attributes: How to infer user attributes, such as psychographics, needs and wants, political orientation, and brand affinities? Temporal analysis: How to analyze change and stability of personas over time? Image: How to retrieve, generate, and choose persona profile pictures? Finding better ways to process and choose useful user information from vast amounts of online data. ”Personas are about giving faces to data.” – Jim Jansen Applicability: How to create personas for different contexts (e.g., e-commerce personas, political personas, gamer personas…)?
  • 13. Research Roadmap for Automatic Persona Generation Information architecture: How to choose the correct information elements and layouts for a given user, use case or industry? Quotes: How to find representative, contextually relevant, and non-distracting comments describing the persona? Evaluation: 1) How to ensure personas are complete, clear, consistent and credible? 2) How to measure usefulness of personas for individuals and organizations? Attributes: How to infer user attributes, such as psychographics, needs and wants, political orientation, and brand affinities? Temporal analysis: How to analyze change and stability of personas over time? Image: How to retrieve, generate, and choose persona profile pictures? Finding better ways to process and choose useful user information from vast amounts of online data. ”Personas are about giving faces to data.” – Jim Jansen Applicability: How to create personas for different contexts (e.g., e-commerce personas, political personas, gamer personas…)?
  • 14. A list of user studies we have done Study Publications System feedbacks N of participants Real users or not User Study 1 (2017) 4 (+1 in review) ~15 29 Yes Persona Perception Scale: pilot (2017) 1 N/A 19 No QR Workshop (2017) N/A 10 12 Yes QF Workshop (2018) N/A 9 12 Yes User Study 2 (2018) 2 (in review) ~15 36 Yes Persona Perception Scale: validation (2018) 2 (1 in review + 1 in works) N/A 412 No Persona Crowd Experiment 1 (Transparent personas) (2018) 1 (in review) 1 (MAJOR) 412 No Persona Crowd Experiment 2 (Smile Study) (2018) 1 (in review) N/A 2,400 No User Study 3 (2018) GOAL: 4 TBA (~20) 34 Yes
  • 15. A list of user studies we have done Study Publications System feedbacks N of participants Real users or not User Study 1 (2017) 4 (+1 in review) ~15 29 Yes Persona Perception Scale: pilot (2017) 1 N/A 19 No QR Workshop (2017) N/A 10 12 Yes QF Workshop (2018) N/A 9 12 Yes User Study 2 (2018) 2 (in review) ~15 36 Yes Persona Perception Scale: validation (2018) 2 (1 in review + 1 in works) N/A 412 No Persona Crowd Experiment 1 (Transparent personas) (2018) 1 (in review) 1 (MAJOR) 412 No Persona Crowd Experiment 2 (Smile Study) (2018) 1 (in review) N/A 2,400 No User Study 3 (2018) GOAL: 4 TBA (~20) 34 Yes
  • 16. A list of user studies we have done Study Publications System feedbacks N of participants Real users or not User Study 1 (2017) 4 (+1 in review) ~15 29 Yes Persona Perception Scale: pilot (2017) 1 N/A 19 No QR Workshop (2017) N/A 10 12 Yes QF Workshop (2018) N/A 9 12 Yes User Study 2 (2018) 2 (in review) ~15 36 Yes Persona Perception Scale: validation (2018) 2 (1 in review + 1 in works) N/A 412 No Persona Crowd Experiment 1 (Transparent personas) (2018) 1 (in review) 1 (MAJOR) 412 No Persona Crowd Experiment 2 (Smile Study) (2018) 1 (in review) N/A 2,400 No User Study 3 (2018) GOAL: 4 TBA (~20) 34 Yes
  • 17. A list of user studies we have done Study Publications System feedbacks N of participants Real users or not User Study 1 (2017) 4 (+1 in review) ~15 29 Yes Persona Perception Scale: pilot (2017) 1 N/A 19 No QR Workshop (2017) N/A 10 12 Yes QF Workshop (2018) N/A 9 12 Yes User Study 2 (2018) 2 (in review) ~15 36 Yes Persona Perception Scale: validation (2018) 2 (1 in review + 1 in works) N/A 412 No Persona Crowd Experiment 1 (Transparent personas) (2018) 1 (in review) 1 (MAJOR) 412 No Persona Crowd Experiment 2 (Smile Study) (2018) 1 (in review) N/A 2,400 No User Study 3 (2018) GOAL: 4 TBA (~20) 34 Yes
  • 18. Observations: • Number of participants and number of system development items is not correlated (when including crowd experiments) • However, when excluding crowd experiments, there is a correlation (the more people participate, the more ideas you get) …both the quality and magnitude of system development items vary!
  • 19. As the number of user studies increases, the number of changes actually grows! (but the magnitude of changes decreases, i.e., you start fixing smaller things)
  • 20. Zeroing in on the value… US1 US2 US3 Moment of value = state and point in time where the system provides real value to real users and is therefore actively used by them.
  • 21. Value-Driven System Development (VDSD) Value-Driven Development • Implement the recommended changes • Conduct another user study (validate progress towards value; find new things to improve) Filter useful feedback from non-useful: • How common is the feedback? • Who gives the feedback? • Design the user study • Recruit participants • Conduct the study
  • 22. User-driven system development over time v1 v2 v3 v4 Number of improvements increasing Magnitude of problems decreasing User study 2 User study 3 User study 4User study 1 time
  • 23. Challenge: you need to make real improvements in between the user studies …otherwise, you keep hearing the same complaints again and again!
  • 24. Remember: developing a useful system is a journey, not a project (there is no “end”). …that’s why you need years of dedication, funding and committed people.
  • 26. How to decide the user study focus? • Behavior = objective information about what the users looks/clicks at. Considered as the most reliable kind of data. • Perceptions = helps understand what are the user’s impressions and why she or he does things. However, susceptible to several biases from both researcher and respondent. …ideally, you’d measure both! (mixed method studies are the best user studies)
  • 27. To-do list for a user study 1. Decide research problem 2. Decide manipulations 3. Create user task 4. Create experimental treatments & flows 5. Create consent form (and get IRB approval if needed) 6. Recruit users 7. Pilot test (devices, software, flows…)
  • 28. How to design a user study? • Combine an (a) interesting theoretical problem and (b) opportunity to get valuable user feedback • The REAL criteria: • When possible, use the real system (not mockups, screenshots) • Use real users (otherwise they won’t get engaged or provide valid feedback) • Use real or realistic task scenario (to get as close to the users’ reality as possible)
  • 29. There’s a lot to consider…
  • 30. An example of user study flow (eye-tracking) 1. Welcome the participant 2. Explain what the study is about and how we collect data 3. Get their consent 4. Calibrate the devices 5. Start the recording 6. Make notes 7. When participant finishes, ask some additional questions 8. Thank them (preferably with a gift card ☺)
  • 31. Tips on running user studies (based on experience ☺) • Run 4-6 pilots with different people (include people from outside the team) • Be realistic with timing (leave empty time!) • Always test the machines before, on the spot (Internet….) • Use two stations • Have a Plan B (for machines, Internet, people…)
  • 32. Tips for writing questionnaire introduction • Be positive (engage people) • Tell what the study is about (in plain language) • Tell how long answering will take • Tell who they can contact if they have questions
  • 34. What type of data to collect? • Real engagement data, e.g. eye-tracking, mouse tracking → tells what’s going on • Think aloud data, e.g. notes, voice recordings, transcriptions → tells why it is going on
  • 35. What metrics to analyze? Metric Type Definition Success rate Objective How many were able to carry out the task successfully? Time to success Objective How long did it take to complete the task? Ratio of time spent in AOI Objective How long was the participant hovering/looking at an area of the screen? Frequency of mentions Objective How many times the participant mentioned a specific term (e.g., confusion vocabulary, positive/negative)? Explicit ratings Subjective Participants’ judgments on the system or content (e.g., “How easy was it to…?)
  • 36. What to ask? • Open feedback (“How did you find the system?” “Does the system solve a practical problem for you? If so, what?”) • Background information: • Age • Gender • Profession • Experience with the system / in the domain • (other things that might be confounding factors: country, native language…)
  • 37. When interviewing, ask about the users’ problems, not about the system: • What is your typical workday like? • What type of tasks you have in your job? • What problems do you face in those tasks? • Interesting…can you give a real example?
  • 38. “How do I know if participants are paying attention during studies?” “Attention checks, or Instructional Manipulation Checks (IMCs), are a straightforward and simple way to determine who does or doesn't pay attention to your study instructions (Oppenheimer, Mayvis, & Davidenko, 2009). We recommend that you always have at least one attention check in any given study.” Source: www.prolific.ac (a great survey platform)
  • 40. Rejecting! This respondent failed the attention test. He or she was supposed to select “slightly agree” but selected “strongly agree”. Not likely that he or she paid attention to the survey -- cannot use his or her answers. 😢
  • 41. Challenges with crowd • Need to communicate super simply (pilot, pilot, pilot!) → instructions, task description • Quality control (give example from hate interpretation paper, the list) • Lack of validity → they’re NOT real users! • Therefore, you can test UI aspects and general information processing/usability stuff, but not validate sources of value
  • 42. Should I use the crowd? Pros Cons +Get a large number of participants easily and cheaply (whereas recruiting real users is costly and difficult) -Not real user feedback! (validity problem) +Fast to collect data and iterate on study designs (whereas real user studies are “hit or miss”) -Quality concerns (but can also be in real user studies)
  • 43. Managing the crowd (Pitkänen & Salminen, 2013) Pitkänen, L., & Salminen, J. (2013). Managing the Crowd: A Study on Videography Application. In Proceedings of Applied Business and Entrepreneurship Association International (ABEAI 2013) Conference, Honolulu, Hawaii, USA, 14–20 November.
  • 44. How to analyze the results? • For quantitative: R/Python/SPSS/Excel • For qualitative analysis: Word/Excel is fine 90% of the time • Only need Nvivo/Atlas when multiple layers of complexity • (e.g., you need to label notes with multiple codes, e.g. “RESEARCH,CHI”, “RESEARCH,UMAP”; assign priority (e.g., 1-5))
  • 46.
  • 47.
  • 48. Surprising findings from user study “Find match…no idea what this means…” → and we thought it was DEAD SIMPLE! *not so* for users. • (always record think-alouds → transcribe to analyze again) • Remember: you are not your user. (always test your designs and features against *real* users)
  • 49. “To discover real user needs, we’ve been carrying out several user studies. However, there are many issues in conducting user studies. The feedback we get is not always relevant or valid.” https://www.linkedin.com/pulse/how-identify-useful-user-feedback-three-tips-value-joni-salminen/
  • 50. The issues in interpreting user feedback • Some participants might not be truly engaged or interested in the system and just participate out of duty or because they were “forced to”. • Similarly, users may just brainstorm features that really they would not use but that “sound cool”. (leading to feature creep) • Moreover, when compiling the feedback, we find that there are a lot of requests for new features. Say, the users want 10 new features, but we have time and resources for two and thus need to prioritize.
  • 51. How to filter useful user feedback? 3 tips: 1. Who does the feeback come from? → not all people are engaged, motivated, or knowledgeable to give useful feedback. Therefore, we have to consider if a person is just “shooting ideas” or if he or she actually wants to provide useful feedback. We then prioritize the comments from the people whose feedback indicates they are taking the commenting more seriously. 2. How repetitive is the feedback? → if the request comes from many organizations and many people within an organization, it is more likely to be a real problem to solve. If it’s a rare request, the problem is probably also very rare and worthy to focus on. 3. Is the feedback traceable to a real problem the user has? → we want to know if the request if a nice-to-have or pain killer. We want to solve real problems with the system, so nice-to-haves must be minimized. Even if many motivated people suggest a new feature, it could still be a nice-to-have if we cannot logically connect it to a real problem.
  • 52. How to filter useful user feedback? 3 tips: 1. Who does the feeback come from? → not all people are engaged, motivated, or knowledgeable to give useful feedback. Therefore, we have to consider if a person is just “shooting ideas” or if he or she actually wants to provide useful feedback. We then prioritize the comments from the people whose feedback indicates they are taking the commenting more seriously. 2. How repetitive is the feedback? → if the request comes from many organizations and many people within an organization, it is more likely to be a real problem to solve. If it’s a rare request, the problem is probably also very rare and worthy to focus on. 3. Is the feedback traceable to a real problem the user has? → we want to know if the request if a nice-to-have or pain killer. We want to solve real problems with the system, so nice-to-haves must be minimized. Even if many motivated people suggest a new feature, it could still be a nice-to-have if we cannot logically connect it to a real problem.
  • 53. How to filter useful user feedback? 3 tips: 1. Who does the feeback come from? → not all people are engaged, motivated, or knowledgeable to give useful feedback. Therefore, we have to consider if a person is just “shooting ideas” or if he or she actually wants to provide useful feedback. We then prioritize the comments from the people whose feedback indicates they are taking the commenting more seriously. 2. How repetitive is the feedback? → if the request comes from many organizations and many people within an organization, it is more likely to be a real problem to solve. If it’s a rare request, the problem is probably also very rare and worthy to focus on. 3. Is the feedback traceable to a real problem the user has? → we want to know if the request if a nice-to-have or pain killer. We want to solve real problems with the system, so nice-to-haves must be minimized. Even if many motivated people suggest a new feature, it could still be a nice-to-have if we cannot logically connect it to a real problem.
  • 54. Other things to be cautious about • Social desirability bias (getting *too high* scores) • Not all adoption problems can be solved with tech. You also need marketing, time, product champions, etc. (barriers for adoption)
  • 55. Example of social desirability bias (QR Workshop 2017)
  • 56. …yet, 0 active users from QR! (who would log-in and use the system)
  • 57. Lessons: • What people say and do are two different things • System ratings are poorly correlated with actual behavior • If you want most reliable data, go for behavioral data • If you want to ask people, ask about their problems, not about solutions (when you understand the problem fully, you can create a solution for it)
  • 58. How to communicate results to engineer? • Difficult (sometimes they don’t care ☺) • Best to make engineer run test sessions him/herself • Best to have a good relationship and communicate with them (I love Soon ❤)
  • 59. How to communicate results to engineer? • Use quotes, stories, drawings… • Use numbers (how many people ex • Convince your engineer the problem really matters for users (buy-in) • …again, best if they self experience user feedback. (internalization problem)
  • 60. Remember: Many engineers think “user is stupid” when they don’t understand how to use the system. That’s NOT the case!
  • 63. 3. Do a mock-up 4. Explain to developer
  • 64. Done!
  • 65. Example of communicating results 1. …Joni has an idea 2. …runs to Soon’s office 3. …Soon puts it into the source code 4. Done! 5. (repeat 100 times ☺)
  • 67. Solves a real problem (usability) experienced by many users.
  • 68. Next step: validate the solution in a new user study! (…and make sure the solution didn’t create a new problem) …remember, iteration: user study → system development
  • 69. Usability problems are not (that) important! …you can solve all the problems, but if there is no value, your system fails.
  • 70. Usability problems are important (--- when they are blocking access to value)
  • 71. Challenges for design guidelines • How to prioritize one change over another? • How to imagine the effect of change in user behavior? • How to distinguish between small UI problem and really big problems (i.e., the system being “useless”)
  • 73. How to read the table • Frequency mentioned = How often participants mentioned this issue • Difficulty to fix = Assessment of how much time it takes • Priority = Assessment of how much this would make us close to “moment of value” • Four types: • Definition (text) • UI change • System change • New feature • Three levels of difficulty (correlate quite well with types) • Low (takes ~5mins from Soon (or maybe a day at max.)) • Medium (takes 1-3 weeks from Soon) • High (requires 3-6 months of research )
  • 74. How to read the table • Frequency mentioned = How often participants mentioned this issue • Difficulty to fix = Assessment of how much time it takes • Priority = Assessment of how much this would make us close to “moment of value”
  • 75. NOTE: It’s important to be able to accurately estimate the level of difficulty! (…you need to have a certain level of understanding about software development) …and, you need to know how quick / skilled your developer is! (again, I  Soon)
  • 76. Prioritization: Take the quick wins (low hanging fruits) first. Then, work systematically to solve the more challenging problems.
  • 77. Bottom line: User studies have many benefits • Build relationship with users / client organizations • Learn what’s going on in the real world (real people’s problems) • Solve research problems that matter (and make some good publications ☺) • Develop better systems that actually get used
  • 78. Thanks for listening! (….and thanks A LOT for participating in pilot studies ☺) (ps. Full slides at SlideShare: www.slideshare.com/jonis12) (pps. Visit https://persona.qcri.org for demo!)