SlideShare a Scribd company logo
1 of 34
Download to read offline
Crowdsourcing the
Semantic Web
Elena Simperl
University of Southampton
Seminar@University of Manchester
4 March 2015
Crowdsourcing Web semantics
• Crowdsourcing is increasingly used
to augment the results of
algorithms solving Semantic Web
problems
• Great challenges
• Which form of crowdsourcing for
which task?
• How to design the crowdsourcing
exercise effectively?
• How to combine human- and
machine-driven approaches?
There is crowdsourcing and crowdsourcing
06-Mar-15 3
Typology of crowdsourcing
Macrotasks Microtasks
Challenges Self-organized
crowds
Crowdfunding
4
Source:Source: (Prpić, Shukla, Kietzmann and McCarthy 2015)
Dimensions of crowdsourcing
• What to crowdsource?
– Tasks you can’t complete in-house or using computers
– A question of time, budget, resources, ethics etc.
• Who is the crowd?
– Crowdsourcing ≠‘turkers’
– Open call, biased through use of platforms and promotion channels
– No traditional means to manage and incentivize
– Crowd has little to no context about the project
• How to crowdsource?
– Macro vs. microtasks
– Complex workflows
– Assessment and aggregation
– Aligning incentives
– Hybrid systems
– Learn to optimize from previous interactions
5
Hybrid systems (or ‚social machines‘)
06-Mar-15 Tutorial@ISWC2013 6
Physical world
(people and devices)
Design and composition Participation and data supply
Model of social interaction
Virtual world
(Network of
social interactions)
Source: Dave Robertson
CROWDSOURCING DATA
CURATION
Crowdsourcing Linked Data quality assessment
M Acosta, A Zaveri, E Simperl, D Kontokostas, S Auer, J Lehmann
ISWC 2013, 260-276
7
Overview
• Compare two forms of crowdsourcing
(challenges and paid microtasks) in
the context of data quality assessment
and repair
• Experiments on DBpedia
– Identify potential errors, classify,
and repair them
– TripleCheckMate challenge vs.
Mechanical Turk
8
What to crowdsource
• Incorrect object
 Example: dbpedia:Dave_Dobbyn dbprop:dateOfBirth “3”.
• Incorrect data type or language tags
 Example: dbpedia:Torishima_Izu_Islands foaf:name “鳥島”@en.
• Incorrect link to “external Web pages”
 Example: dbpedia:John-Two-Hawks dbpedia-owl:wikiPageExternalLink
<http://cedarlakedvd.com/>
Who is the crowd
Challenge
LD Experts
Difficult task
Final prize
Find Verify
Microtasks
Workers
Easy task
Micropayments
TripleCheckMate
[Kontoskostas2013] MTurk
Adapted from [Bernstein2010]
http://mturk.com
How to crowdsource: workflow
11
How to crowdsource: microtask design
• Selection of foaf:name or
rdfs:label to extract human-
readable descriptions
• Values extracted automatically
from Wikipedia infoboxes
• Link to the Wikipedia article via
foaf:isPrimaryTopicOf
• Preview of external pages by
implementing HTML iframe
Incorrect object
Incorrect data type or language tag
Incorrect outlink
Experiments
• Crowdsourcing approaches
• Find stage: Challenge with LD experts
• Verify stage: Microtasks (5 assignments)
• Gold standard
• Two of the authors generated gold standard for all contest triples
indepedently
• Conflicts resolved via mutual agreement
• Validation: precision
Overall results
LD experts Microtask
workers
Number of distinct
participants
50 80
Total time
3 weeks (predefined) 4 days
Total triples evaluated
1,512 1,073
Total cost
~ US$ 400 (predefined) ~ US$ 43
Precision results: Incorrect objects
• Turkers can reduce the error rates of LD experts from the Find stage
• 117 DBpedia triples had predicates related to dates with
incorrect/incomplete values:
”2005 Six Nations Championship” Date 12 .
• 52 DBpedia triples had erroneous values from the source:
”English (programming language)” Influenced by ? .
• Experts classified all these triples as incorrect
• Workers compared values against Wikipedia and successfully classified this
triples as “correct”
Triples compared LD Experts MTurk
(majority voting: n=5)
509 0.7151 0.8977
Precision results: Incorrect data types
0
20
40
60
80
100
120
140
Date English Millimetre Nanometre
Number
Number
with
decimals
Second Volt Year Not
specified /
URI
Numberoftriples
Data types
Experts TP
Experts FP
Crowd TP
Crowd FP
Triples compared LD Experts MTurk
(majority voting: n=5)
341 0.8270 0.4752
Precision results: Incorrect links
• We analyzed the 189 misclassifications by the experts
• The 6% misclassifications by the workers correspond to
pages with a language different from English
50%
39%
11%
Freebase links
Wikipedia images
External links
Triples compared Baseline LD Experts MTurk
(n=5 majority voting)
223 0.2598 0.1525 0.9412
Conclusions
• The effort of LD experts should be invested in tasks that
demand domain-specific skills
– Experts seem less motivated to perform on tasks that
come with an additional overhead (e.g., checking an
external link, going back on Wikipedia, data repair)
• MTurk crowd was exceptionally good at performing object
types and links checks
– But seem not to understand the intricacies of data types
Future work
• Additional experiments with data types
• Workflows for new types of errors, e.g., tasks which experts cannot
answer on the fly
• Training the crowd
• Building the DBpedia ‘social machine’
– Add automatic QA tools
– Close the feedback loop (from crowd to experts)
– Change the parameters of the ‘Find’ step to achieve seamless
integration
– Change crowdsourcing model of ‘Find’ to improve retention
19
CROWDSOURCING IMAGE
ANNOTATION
Improving paid microtasks through gamification and adaptive
furtherance incentives
O Feyisetan, E Simperl, M Van Kleek, N Shadbolt
WWW 2015, to appear
20
Overview
• Make paid microtasks more
cost-effective though
gamification*
*use of game elements and
mechanics in a non-game
context
What to crowdsource: image labeling
22
Who is the crowd: paid microtask platforms
23
[Kaufmann, Schulze, Viet, 2011]
Research hypotheses
• Contributors will perform better if tasks are
more engaging
• Increased accuracy through higher inter-
annotator agreement
• Cost savings through reduced unit costs
• Micro-targeting incentives when players
attempt to quit improves retention
How to crowdsource: microtask design
• Image labeling tasks published
on microtask platform
– Free text labels, varying
numbers of labels per
image, taboo words
– 1st setting: ‘standard’ tasks,
including basic spam
control
– 2nd setting: the same
requirements and rewards,
but contributors were asked
to carry out the task in
WordSmith
25
How to crowdsource: gamification
• Levels – 9 levels from newbie to Wordsmith, function of images
tagged
• Badges – function of number of images tagged
• Bonus points – for new tags
• Treasure points – for multiples of bonus points
• Feedback alerts - related to badges, points, levels
• Leaderboard - hourly scores and level of the top 5 players
• Activities widget – real-time updates on other players
26
How to crowdsource: incentives
• Money – 5 cents extra for the effort
• Power – see how other players tagged the images
• Leaderboard - advance to the ‘Global’ leaderboard seen by everyone
• Badges – receive ’Ultimate’ badge and a shiny new avatar
• Levels – advance straight to the next level
• Access - quicker access to treasure points
27
Assigned at random or targeted based on Bayes inference
(number of tagged images as feature)
Experiments
• Crowdsourcing approaches
– Paid microtasks
– Wordsmith GWAP
• Data set: ESP Game (120K images with curated labels)
• Validation: precision
28
Experiments (2)
• 1st experiment: ‘newcomers’
– Tag one image with two keywords ~ complete
Wordsmith level 1
– 200 images, US$ 0.02 per task, 3 judgements per task
• 2nd experiment: ‘novice’
– Tag 11 images ~ complete Wordsmith level 2
– 2200 images, US$ 0.10 per task, 600 workers
– Furtherance incentives: none, random, targeted (Bayes
inference based on # of images tagged)
29
Results: 1st experiment
Metric CrowdFlower Wordsmith
Total workers 600 423
Total keywords 1,200 41,206
Unique keywords 111 5,708
Avg. agreement 5.72% 37.7%
Mean images/person 1 32
Max images/person 1 200
Results: 2nd experiment
Metric CrowdFlower Wordsmith (no
furtherance
incentives
Total workers 600 514
Total keywords 13,200 35,890
Unique keywords 1,323 4,091
Avg. agreement 6.32% 10.9%
Mean images/person 11 27
Max images/person 1 351
Results: 2nd experiment, incentives
• Incentives led to increased participation
– Power: mean image/person = 132
– People come back (20 times!) and play longer (43 hours,
3 hours without incentives)
32
Conclusions and future work
• Task design matters as much as payment
• Gamification achieves high accuracy for lower costs and
improved engagement
• Workers appreciate social features, but their main
motivation is still task-driven
– Feedback mechanisms, peer learning, training
33
Email: e.simperl@soton.ac.uk
Twitter: @esimperl
34

More Related Content

Similar to Crowdsourcing the Semantic Web

Highlights from the 8th ACM Conference on Recommender Systems (RecSys 2014)
Highlights from the 8th ACM Conference on Recommender Systems (RecSys 2014)Highlights from the 8th ACM Conference on Recommender Systems (RecSys 2014)
Highlights from the 8th ACM Conference on Recommender Systems (RecSys 2014)David Zibriczky
 
Data science in 10 steps
Data science in 10 stepsData science in 10 steps
Data science in 10 stepsQuantUniversity
 
Evaluation and User Study in HCI
Evaluation and User Study in HCIEvaluation and User Study in HCI
Evaluation and User Study in HCIByungkyu (Jay) Kang
 
Best Practices in Recommender System Challenges
Best Practices in Recommender System ChallengesBest Practices in Recommender System Challenges
Best Practices in Recommender System ChallengesAlan Said
 
Human computation, crowdsourcing and social: An industrial perspective
Human computation, crowdsourcing and social: An industrial perspectiveHuman computation, crowdsourcing and social: An industrial perspective
Human computation, crowdsourcing and social: An industrial perspectiveoralonso
 
Comparison GWAP Mechanical Turk
Comparison GWAP Mechanical TurkComparison GWAP Mechanical Turk
Comparison GWAP Mechanical TurkElena Simperl
 
Large scale computing
Large scale computing Large scale computing
Large scale computing Bhupesh Bansal
 
Presentation Selan dos Santos 4Eyes Lab
Presentation Selan dos Santos 4Eyes LabPresentation Selan dos Santos 4Eyes Lab
Presentation Selan dos Santos 4Eyes Labselan_rds
 
Bazley understanding online audiences vsg conf march 2016 for uploading
Bazley understanding online audiences vsg conf march 2016 for uploadingBazley understanding online audiences vsg conf march 2016 for uploading
Bazley understanding online audiences vsg conf march 2016 for uploadingMartin Bazley
 
Rokach-GomaxSlides.pptx
Rokach-GomaxSlides.pptxRokach-GomaxSlides.pptx
Rokach-GomaxSlides.pptxJadna Almeida
 
Rokach-GomaxSlides (1).pptx
Rokach-GomaxSlides (1).pptxRokach-GomaxSlides (1).pptx
Rokach-GomaxSlides (1).pptxJadna Almeida
 
Lecture_1_Intro.pdf
Lecture_1_Intro.pdfLecture_1_Intro.pdf
Lecture_1_Intro.pdfpaijitk
 
How to do science in a large IT company (ICPC World Finals 2021, Moscow)
How to do science in a large IT company (ICPC World Finals 2021, Moscow)How to do science in a large IT company (ICPC World Finals 2021, Moscow)
How to do science in a large IT company (ICPC World Finals 2021, Moscow)Alexander Borzunov
 
Staying Competitive in Data Analytics: Analyze Boulder 20140903
Staying Competitive in Data Analytics: Analyze Boulder 20140903Staying Competitive in Data Analytics: Analyze Boulder 20140903
Staying Competitive in Data Analytics: Analyze Boulder 20140903Richard Hackathorn
 
Where Search Meets Machine Learning: Presented by Diana Hu & Joaquin Delgado,...
Where Search Meets Machine Learning: Presented by Diana Hu & Joaquin Delgado,...Where Search Meets Machine Learning: Presented by Diana Hu & Joaquin Delgado,...
Where Search Meets Machine Learning: Presented by Diana Hu & Joaquin Delgado,...Lucidworks
 
Human Computation for Big Data
Human Computation for Big DataHuman Computation for Big Data
Human Computation for Big DataeXascale Infolab
 
Lucene/Solr Revolution 2015: Where Search Meets Machine Learning
Lucene/Solr Revolution 2015: Where Search Meets Machine LearningLucene/Solr Revolution 2015: Where Search Meets Machine Learning
Lucene/Solr Revolution 2015: Where Search Meets Machine LearningJoaquin Delgado PhD.
 
Lucene/Solr Revolution 2015: Where Search Meets Machine Learning
Lucene/Solr Revolution 2015: Where Search Meets Machine LearningLucene/Solr Revolution 2015: Where Search Meets Machine Learning
Lucene/Solr Revolution 2015: Where Search Meets Machine LearningS. Diana Hu
 

Similar to Crowdsourcing the Semantic Web (20)

Highlights from the 8th ACM Conference on Recommender Systems (RecSys 2014)
Highlights from the 8th ACM Conference on Recommender Systems (RecSys 2014)Highlights from the 8th ACM Conference on Recommender Systems (RecSys 2014)
Highlights from the 8th ACM Conference on Recommender Systems (RecSys 2014)
 
ARlab RESEARCH | Social search
ARlab RESEARCH | Social searchARlab RESEARCH | Social search
ARlab RESEARCH | Social search
 
Data science in 10 steps
Data science in 10 stepsData science in 10 steps
Data science in 10 steps
 
Evaluation and User Study in HCI
Evaluation and User Study in HCIEvaluation and User Study in HCI
Evaluation and User Study in HCI
 
Best Practices in Recommender System Challenges
Best Practices in Recommender System ChallengesBest Practices in Recommender System Challenges
Best Practices in Recommender System Challenges
 
Human computation, crowdsourcing and social: An industrial perspective
Human computation, crowdsourcing and social: An industrial perspectiveHuman computation, crowdsourcing and social: An industrial perspective
Human computation, crowdsourcing and social: An industrial perspective
 
Comparison GWAP Mechanical Turk
Comparison GWAP Mechanical TurkComparison GWAP Mechanical Turk
Comparison GWAP Mechanical Turk
 
Social Network Analysis Applications and Approach
Social Network Analysis Applications and ApproachSocial Network Analysis Applications and Approach
Social Network Analysis Applications and Approach
 
Large scale computing
Large scale computing Large scale computing
Large scale computing
 
Presentation Selan dos Santos 4Eyes Lab
Presentation Selan dos Santos 4Eyes LabPresentation Selan dos Santos 4Eyes Lab
Presentation Selan dos Santos 4Eyes Lab
 
Bazley understanding online audiences vsg conf march 2016 for uploading
Bazley understanding online audiences vsg conf march 2016 for uploadingBazley understanding online audiences vsg conf march 2016 for uploading
Bazley understanding online audiences vsg conf march 2016 for uploading
 
Rokach-GomaxSlides.pptx
Rokach-GomaxSlides.pptxRokach-GomaxSlides.pptx
Rokach-GomaxSlides.pptx
 
Rokach-GomaxSlides (1).pptx
Rokach-GomaxSlides (1).pptxRokach-GomaxSlides (1).pptx
Rokach-GomaxSlides (1).pptx
 
Lecture_1_Intro.pdf
Lecture_1_Intro.pdfLecture_1_Intro.pdf
Lecture_1_Intro.pdf
 
How to do science in a large IT company (ICPC World Finals 2021, Moscow)
How to do science in a large IT company (ICPC World Finals 2021, Moscow)How to do science in a large IT company (ICPC World Finals 2021, Moscow)
How to do science in a large IT company (ICPC World Finals 2021, Moscow)
 
Staying Competitive in Data Analytics: Analyze Boulder 20140903
Staying Competitive in Data Analytics: Analyze Boulder 20140903Staying Competitive in Data Analytics: Analyze Boulder 20140903
Staying Competitive in Data Analytics: Analyze Boulder 20140903
 
Where Search Meets Machine Learning: Presented by Diana Hu & Joaquin Delgado,...
Where Search Meets Machine Learning: Presented by Diana Hu & Joaquin Delgado,...Where Search Meets Machine Learning: Presented by Diana Hu & Joaquin Delgado,...
Where Search Meets Machine Learning: Presented by Diana Hu & Joaquin Delgado,...
 
Human Computation for Big Data
Human Computation for Big DataHuman Computation for Big Data
Human Computation for Big Data
 
Lucene/Solr Revolution 2015: Where Search Meets Machine Learning
Lucene/Solr Revolution 2015: Where Search Meets Machine LearningLucene/Solr Revolution 2015: Where Search Meets Machine Learning
Lucene/Solr Revolution 2015: Where Search Meets Machine Learning
 
Lucene/Solr Revolution 2015: Where Search Meets Machine Learning
Lucene/Solr Revolution 2015: Where Search Meets Machine LearningLucene/Solr Revolution 2015: Where Search Meets Machine Learning
Lucene/Solr Revolution 2015: Where Search Meets Machine Learning
 

More from Elena Simperl

This talk was not generated with ChatGPT: how AI is changing science
This talk was not generated with ChatGPT: how AI is changing scienceThis talk was not generated with ChatGPT: how AI is changing science
This talk was not generated with ChatGPT: how AI is changing scienceElena Simperl
 
Knowledge graph use cases in natural language generation
Knowledge graph use cases in natural language generationKnowledge graph use cases in natural language generation
Knowledge graph use cases in natural language generationElena Simperl
 
Knowledge engineering: from people to machines and back
Knowledge engineering: from people to machines and backKnowledge engineering: from people to machines and back
Knowledge engineering: from people to machines and backElena Simperl
 
The web of data: how are we doing so far
The web of data: how are we doing so farThe web of data: how are we doing so far
The web of data: how are we doing so farElena Simperl
 
What Wikidata teaches us about knowledge engineering
What Wikidata teaches us about knowledge engineeringWhat Wikidata teaches us about knowledge engineering
What Wikidata teaches us about knowledge engineeringElena Simperl
 
Open government data portals: from publishing to use and impact
Open government data portals: from publishing to use and impactOpen government data portals: from publishing to use and impact
Open government data portals: from publishing to use and impactElena Simperl
 
Ten myths about knowledge graphs.pdf
Ten myths about knowledge graphs.pdfTen myths about knowledge graphs.pdf
Ten myths about knowledge graphs.pdfElena Simperl
 
What Wikidata teaches us about knowledge engineering
What Wikidata teaches us about knowledge engineeringWhat Wikidata teaches us about knowledge engineering
What Wikidata teaches us about knowledge engineeringElena Simperl
 
Data commons and their role in fighting misinformation.pdf
Data commons and their role in fighting misinformation.pdfData commons and their role in fighting misinformation.pdf
Data commons and their role in fighting misinformation.pdfElena Simperl
 
Are our knowledge graphs trustworthy?
Are our knowledge graphs trustworthy?Are our knowledge graphs trustworthy?
Are our knowledge graphs trustworthy?Elena Simperl
 
The web of data: how are we doing so far?
The web of data: how are we doing so far?The web of data: how are we doing so far?
The web of data: how are we doing so far?Elena Simperl
 
Crowdsourcing and citizen engagement for people-centric smart cities
Crowdsourcing and citizen engagement for people-centric smart citiesCrowdsourcing and citizen engagement for people-centric smart cities
Crowdsourcing and citizen engagement for people-centric smart citiesElena Simperl
 
Pie chart or pizza: identifying chart types and their virality on Twitter
Pie chart or pizza: identifying chart types and their virality on TwitterPie chart or pizza: identifying chart types and their virality on Twitter
Pie chart or pizza: identifying chart types and their virality on TwitterElena Simperl
 
High-value datasets: from publication to impact
High-value datasets: from publication to impactHigh-value datasets: from publication to impact
High-value datasets: from publication to impactElena Simperl
 
The story of Data Stories
The story of Data StoriesThe story of Data Stories
The story of Data StoriesElena Simperl
 
The human face of AI: how collective and augmented intelligence can help sol...
The human face of AI:  how collective and augmented intelligence can help sol...The human face of AI:  how collective and augmented intelligence can help sol...
The human face of AI: how collective and augmented intelligence can help sol...Elena Simperl
 
Qrowd and the city: designing people-centric smart cities
Qrowd and the city: designing people-centric smart citiesQrowd and the city: designing people-centric smart cities
Qrowd and the city: designing people-centric smart citiesElena Simperl
 
One does not simply crowdsource the Semantic Web: 10 years with people, URIs,...
One does not simply crowdsource the Semantic Web: 10 years with people, URIs,...One does not simply crowdsource the Semantic Web: 10 years with people, URIs,...
One does not simply crowdsource the Semantic Web: 10 years with people, URIs,...Elena Simperl
 
Inclusive cities: a crowdsourcing approach
Inclusive cities: a crowdsourcing approachInclusive cities: a crowdsourcing approach
Inclusive cities: a crowdsourcing approachElena Simperl
 

More from Elena Simperl (20)

This talk was not generated with ChatGPT: how AI is changing science
This talk was not generated with ChatGPT: how AI is changing scienceThis talk was not generated with ChatGPT: how AI is changing science
This talk was not generated with ChatGPT: how AI is changing science
 
Knowledge graph use cases in natural language generation
Knowledge graph use cases in natural language generationKnowledge graph use cases in natural language generation
Knowledge graph use cases in natural language generation
 
Knowledge engineering: from people to machines and back
Knowledge engineering: from people to machines and backKnowledge engineering: from people to machines and back
Knowledge engineering: from people to machines and back
 
The web of data: how are we doing so far
The web of data: how are we doing so farThe web of data: how are we doing so far
The web of data: how are we doing so far
 
What Wikidata teaches us about knowledge engineering
What Wikidata teaches us about knowledge engineeringWhat Wikidata teaches us about knowledge engineering
What Wikidata teaches us about knowledge engineering
 
Open government data portals: from publishing to use and impact
Open government data portals: from publishing to use and impactOpen government data portals: from publishing to use and impact
Open government data portals: from publishing to use and impact
 
Ten myths about knowledge graphs.pdf
Ten myths about knowledge graphs.pdfTen myths about knowledge graphs.pdf
Ten myths about knowledge graphs.pdf
 
What Wikidata teaches us about knowledge engineering
What Wikidata teaches us about knowledge engineeringWhat Wikidata teaches us about knowledge engineering
What Wikidata teaches us about knowledge engineering
 
Data commons and their role in fighting misinformation.pdf
Data commons and their role in fighting misinformation.pdfData commons and their role in fighting misinformation.pdf
Data commons and their role in fighting misinformation.pdf
 
Are our knowledge graphs trustworthy?
Are our knowledge graphs trustworthy?Are our knowledge graphs trustworthy?
Are our knowledge graphs trustworthy?
 
The web of data: how are we doing so far?
The web of data: how are we doing so far?The web of data: how are we doing so far?
The web of data: how are we doing so far?
 
Crowdsourcing and citizen engagement for people-centric smart cities
Crowdsourcing and citizen engagement for people-centric smart citiesCrowdsourcing and citizen engagement for people-centric smart cities
Crowdsourcing and citizen engagement for people-centric smart cities
 
Pie chart or pizza: identifying chart types and their virality on Twitter
Pie chart or pizza: identifying chart types and their virality on TwitterPie chart or pizza: identifying chart types and their virality on Twitter
Pie chart or pizza: identifying chart types and their virality on Twitter
 
High-value datasets: from publication to impact
High-value datasets: from publication to impactHigh-value datasets: from publication to impact
High-value datasets: from publication to impact
 
The story of Data Stories
The story of Data StoriesThe story of Data Stories
The story of Data Stories
 
The human face of AI: how collective and augmented intelligence can help sol...
The human face of AI:  how collective and augmented intelligence can help sol...The human face of AI:  how collective and augmented intelligence can help sol...
The human face of AI: how collective and augmented intelligence can help sol...
 
Qrowd and the city: designing people-centric smart cities
Qrowd and the city: designing people-centric smart citiesQrowd and the city: designing people-centric smart cities
Qrowd and the city: designing people-centric smart cities
 
One does not simply crowdsource the Semantic Web: 10 years with people, URIs,...
One does not simply crowdsource the Semantic Web: 10 years with people, URIs,...One does not simply crowdsource the Semantic Web: 10 years with people, URIs,...
One does not simply crowdsource the Semantic Web: 10 years with people, URIs,...
 
Qrowd and the city
Qrowd and the cityQrowd and the city
Qrowd and the city
 
Inclusive cities: a crowdsourcing approach
Inclusive cities: a crowdsourcing approachInclusive cities: a crowdsourcing approach
Inclusive cities: a crowdsourcing approach
 

Recently uploaded

Python Notes for mca i year students osmania university.docx
Python Notes for mca i year students osmania university.docxPython Notes for mca i year students osmania university.docx
Python Notes for mca i year students osmania university.docxRamakrishna Reddy Bijjam
 
The basics of sentences session 3pptx.pptx
The basics of sentences session 3pptx.pptxThe basics of sentences session 3pptx.pptx
The basics of sentences session 3pptx.pptxheathfieldcps1
 
Basic Civil Engineering first year Notes- Chapter 4 Building.pptx
Basic Civil Engineering first year Notes- Chapter 4 Building.pptxBasic Civil Engineering first year Notes- Chapter 4 Building.pptx
Basic Civil Engineering first year Notes- Chapter 4 Building.pptxDenish Jangid
 
Food Chain and Food Web (Ecosystem) EVS, B. Pharmacy 1st Year, Sem-II
Food Chain and Food Web (Ecosystem) EVS, B. Pharmacy 1st Year, Sem-IIFood Chain and Food Web (Ecosystem) EVS, B. Pharmacy 1st Year, Sem-II
Food Chain and Food Web (Ecosystem) EVS, B. Pharmacy 1st Year, Sem-IIShubhangi Sonawane
 
Measures of Dispersion and Variability: Range, QD, AD and SD
Measures of Dispersion and Variability: Range, QD, AD and SDMeasures of Dispersion and Variability: Range, QD, AD and SD
Measures of Dispersion and Variability: Range, QD, AD and SDThiyagu K
 
TỔNG ÔN TẬP THI VÀO LỚP 10 MÔN TIẾNG ANH NĂM HỌC 2023 - 2024 CÓ ĐÁP ÁN (NGỮ Â...
TỔNG ÔN TẬP THI VÀO LỚP 10 MÔN TIẾNG ANH NĂM HỌC 2023 - 2024 CÓ ĐÁP ÁN (NGỮ Â...TỔNG ÔN TẬP THI VÀO LỚP 10 MÔN TIẾNG ANH NĂM HỌC 2023 - 2024 CÓ ĐÁP ÁN (NGỮ Â...
TỔNG ÔN TẬP THI VÀO LỚP 10 MÔN TIẾNG ANH NĂM HỌC 2023 - 2024 CÓ ĐÁP ÁN (NGỮ Â...Nguyen Thanh Tu Collection
 
Beyond the EU: DORA and NIS 2 Directive's Global Impact
Beyond the EU: DORA and NIS 2 Directive's Global ImpactBeyond the EU: DORA and NIS 2 Directive's Global Impact
Beyond the EU: DORA and NIS 2 Directive's Global ImpactPECB
 
1029-Danh muc Sach Giao Khoa khoi 6.pdf
1029-Danh muc Sach Giao Khoa khoi  6.pdf1029-Danh muc Sach Giao Khoa khoi  6.pdf
1029-Danh muc Sach Giao Khoa khoi 6.pdfQucHHunhnh
 
Key note speaker Neum_Admir Softic_ENG.pdf
Key note speaker Neum_Admir Softic_ENG.pdfKey note speaker Neum_Admir Softic_ENG.pdf
Key note speaker Neum_Admir Softic_ENG.pdfAdmir Softic
 
Holdier Curriculum Vitae (April 2024).pdf
Holdier Curriculum Vitae (April 2024).pdfHoldier Curriculum Vitae (April 2024).pdf
Holdier Curriculum Vitae (April 2024).pdfagholdier
 
Mixin Classes in Odoo 17 How to Extend Models Using Mixin Classes
Mixin Classes in Odoo 17  How to Extend Models Using Mixin ClassesMixin Classes in Odoo 17  How to Extend Models Using Mixin Classes
Mixin Classes in Odoo 17 How to Extend Models Using Mixin ClassesCeline George
 
Z Score,T Score, Percential Rank and Box Plot Graph
Z Score,T Score, Percential Rank and Box Plot GraphZ Score,T Score, Percential Rank and Box Plot Graph
Z Score,T Score, Percential Rank and Box Plot GraphThiyagu K
 
2024-NATIONAL-LEARNING-CAMP-AND-OTHER.pptx
2024-NATIONAL-LEARNING-CAMP-AND-OTHER.pptx2024-NATIONAL-LEARNING-CAMP-AND-OTHER.pptx
2024-NATIONAL-LEARNING-CAMP-AND-OTHER.pptxMaritesTamaniVerdade
 
Presentation by Andreas Schleicher Tackling the School Absenteeism Crisis 30 ...
Presentation by Andreas Schleicher Tackling the School Absenteeism Crisis 30 ...Presentation by Andreas Schleicher Tackling the School Absenteeism Crisis 30 ...
Presentation by Andreas Schleicher Tackling the School Absenteeism Crisis 30 ...EduSkills OECD
 
Energy Resources. ( B. Pharmacy, 1st Year, Sem-II) Natural Resources
Energy Resources. ( B. Pharmacy, 1st Year, Sem-II) Natural ResourcesEnergy Resources. ( B. Pharmacy, 1st Year, Sem-II) Natural Resources
Energy Resources. ( B. Pharmacy, 1st Year, Sem-II) Natural ResourcesShubhangi Sonawane
 
Ecological Succession. ( ECOSYSTEM, B. Pharmacy, 1st Year, Sem-II, Environmen...
Ecological Succession. ( ECOSYSTEM, B. Pharmacy, 1st Year, Sem-II, Environmen...Ecological Succession. ( ECOSYSTEM, B. Pharmacy, 1st Year, Sem-II, Environmen...
Ecological Succession. ( ECOSYSTEM, B. Pharmacy, 1st Year, Sem-II, Environmen...Shubhangi Sonawane
 
microwave assisted reaction. General introduction
microwave assisted reaction. General introductionmicrowave assisted reaction. General introduction
microwave assisted reaction. General introductionMaksud Ahmed
 
PROCESS RECORDING FORMAT.docx
PROCESS      RECORDING        FORMAT.docxPROCESS      RECORDING        FORMAT.docx
PROCESS RECORDING FORMAT.docxPoojaSen20
 

Recently uploaded (20)

Python Notes for mca i year students osmania university.docx
Python Notes for mca i year students osmania university.docxPython Notes for mca i year students osmania university.docx
Python Notes for mca i year students osmania university.docx
 
The basics of sentences session 3pptx.pptx
The basics of sentences session 3pptx.pptxThe basics of sentences session 3pptx.pptx
The basics of sentences session 3pptx.pptx
 
Basic Civil Engineering first year Notes- Chapter 4 Building.pptx
Basic Civil Engineering first year Notes- Chapter 4 Building.pptxBasic Civil Engineering first year Notes- Chapter 4 Building.pptx
Basic Civil Engineering first year Notes- Chapter 4 Building.pptx
 
Food Chain and Food Web (Ecosystem) EVS, B. Pharmacy 1st Year, Sem-II
Food Chain and Food Web (Ecosystem) EVS, B. Pharmacy 1st Year, Sem-IIFood Chain and Food Web (Ecosystem) EVS, B. Pharmacy 1st Year, Sem-II
Food Chain and Food Web (Ecosystem) EVS, B. Pharmacy 1st Year, Sem-II
 
Measures of Dispersion and Variability: Range, QD, AD and SD
Measures of Dispersion and Variability: Range, QD, AD and SDMeasures of Dispersion and Variability: Range, QD, AD and SD
Measures of Dispersion and Variability: Range, QD, AD and SD
 
TỔNG ÔN TẬP THI VÀO LỚP 10 MÔN TIẾNG ANH NĂM HỌC 2023 - 2024 CÓ ĐÁP ÁN (NGỮ Â...
TỔNG ÔN TẬP THI VÀO LỚP 10 MÔN TIẾNG ANH NĂM HỌC 2023 - 2024 CÓ ĐÁP ÁN (NGỮ Â...TỔNG ÔN TẬP THI VÀO LỚP 10 MÔN TIẾNG ANH NĂM HỌC 2023 - 2024 CÓ ĐÁP ÁN (NGỮ Â...
TỔNG ÔN TẬP THI VÀO LỚP 10 MÔN TIẾNG ANH NĂM HỌC 2023 - 2024 CÓ ĐÁP ÁN (NGỮ Â...
 
Beyond the EU: DORA and NIS 2 Directive's Global Impact
Beyond the EU: DORA and NIS 2 Directive's Global ImpactBeyond the EU: DORA and NIS 2 Directive's Global Impact
Beyond the EU: DORA and NIS 2 Directive's Global Impact
 
INDIA QUIZ 2024 RLAC DELHI UNIVERSITY.pptx
INDIA QUIZ 2024 RLAC DELHI UNIVERSITY.pptxINDIA QUIZ 2024 RLAC DELHI UNIVERSITY.pptx
INDIA QUIZ 2024 RLAC DELHI UNIVERSITY.pptx
 
1029-Danh muc Sach Giao Khoa khoi 6.pdf
1029-Danh muc Sach Giao Khoa khoi  6.pdf1029-Danh muc Sach Giao Khoa khoi  6.pdf
1029-Danh muc Sach Giao Khoa khoi 6.pdf
 
Asian American Pacific Islander Month DDSD 2024.pptx
Asian American Pacific Islander Month DDSD 2024.pptxAsian American Pacific Islander Month DDSD 2024.pptx
Asian American Pacific Islander Month DDSD 2024.pptx
 
Key note speaker Neum_Admir Softic_ENG.pdf
Key note speaker Neum_Admir Softic_ENG.pdfKey note speaker Neum_Admir Softic_ENG.pdf
Key note speaker Neum_Admir Softic_ENG.pdf
 
Holdier Curriculum Vitae (April 2024).pdf
Holdier Curriculum Vitae (April 2024).pdfHoldier Curriculum Vitae (April 2024).pdf
Holdier Curriculum Vitae (April 2024).pdf
 
Mixin Classes in Odoo 17 How to Extend Models Using Mixin Classes
Mixin Classes in Odoo 17  How to Extend Models Using Mixin ClassesMixin Classes in Odoo 17  How to Extend Models Using Mixin Classes
Mixin Classes in Odoo 17 How to Extend Models Using Mixin Classes
 
Z Score,T Score, Percential Rank and Box Plot Graph
Z Score,T Score, Percential Rank and Box Plot GraphZ Score,T Score, Percential Rank and Box Plot Graph
Z Score,T Score, Percential Rank and Box Plot Graph
 
2024-NATIONAL-LEARNING-CAMP-AND-OTHER.pptx
2024-NATIONAL-LEARNING-CAMP-AND-OTHER.pptx2024-NATIONAL-LEARNING-CAMP-AND-OTHER.pptx
2024-NATIONAL-LEARNING-CAMP-AND-OTHER.pptx
 
Presentation by Andreas Schleicher Tackling the School Absenteeism Crisis 30 ...
Presentation by Andreas Schleicher Tackling the School Absenteeism Crisis 30 ...Presentation by Andreas Schleicher Tackling the School Absenteeism Crisis 30 ...
Presentation by Andreas Schleicher Tackling the School Absenteeism Crisis 30 ...
 
Energy Resources. ( B. Pharmacy, 1st Year, Sem-II) Natural Resources
Energy Resources. ( B. Pharmacy, 1st Year, Sem-II) Natural ResourcesEnergy Resources. ( B. Pharmacy, 1st Year, Sem-II) Natural Resources
Energy Resources. ( B. Pharmacy, 1st Year, Sem-II) Natural Resources
 
Ecological Succession. ( ECOSYSTEM, B. Pharmacy, 1st Year, Sem-II, Environmen...
Ecological Succession. ( ECOSYSTEM, B. Pharmacy, 1st Year, Sem-II, Environmen...Ecological Succession. ( ECOSYSTEM, B. Pharmacy, 1st Year, Sem-II, Environmen...
Ecological Succession. ( ECOSYSTEM, B. Pharmacy, 1st Year, Sem-II, Environmen...
 
microwave assisted reaction. General introduction
microwave assisted reaction. General introductionmicrowave assisted reaction. General introduction
microwave assisted reaction. General introduction
 
PROCESS RECORDING FORMAT.docx
PROCESS      RECORDING        FORMAT.docxPROCESS      RECORDING        FORMAT.docx
PROCESS RECORDING FORMAT.docx
 

Crowdsourcing the Semantic Web

  • 1. Crowdsourcing the Semantic Web Elena Simperl University of Southampton Seminar@University of Manchester 4 March 2015
  • 2. Crowdsourcing Web semantics • Crowdsourcing is increasingly used to augment the results of algorithms solving Semantic Web problems • Great challenges • Which form of crowdsourcing for which task? • How to design the crowdsourcing exercise effectively? • How to combine human- and machine-driven approaches?
  • 3. There is crowdsourcing and crowdsourcing 06-Mar-15 3
  • 4. Typology of crowdsourcing Macrotasks Microtasks Challenges Self-organized crowds Crowdfunding 4 Source:Source: (Prpić, Shukla, Kietzmann and McCarthy 2015)
  • 5. Dimensions of crowdsourcing • What to crowdsource? – Tasks you can’t complete in-house or using computers – A question of time, budget, resources, ethics etc. • Who is the crowd? – Crowdsourcing ≠‘turkers’ – Open call, biased through use of platforms and promotion channels – No traditional means to manage and incentivize – Crowd has little to no context about the project • How to crowdsource? – Macro vs. microtasks – Complex workflows – Assessment and aggregation – Aligning incentives – Hybrid systems – Learn to optimize from previous interactions 5
  • 6. Hybrid systems (or ‚social machines‘) 06-Mar-15 Tutorial@ISWC2013 6 Physical world (people and devices) Design and composition Participation and data supply Model of social interaction Virtual world (Network of social interactions) Source: Dave Robertson
  • 7. CROWDSOURCING DATA CURATION Crowdsourcing Linked Data quality assessment M Acosta, A Zaveri, E Simperl, D Kontokostas, S Auer, J Lehmann ISWC 2013, 260-276 7
  • 8. Overview • Compare two forms of crowdsourcing (challenges and paid microtasks) in the context of data quality assessment and repair • Experiments on DBpedia – Identify potential errors, classify, and repair them – TripleCheckMate challenge vs. Mechanical Turk 8
  • 9. What to crowdsource • Incorrect object  Example: dbpedia:Dave_Dobbyn dbprop:dateOfBirth “3”. • Incorrect data type or language tags  Example: dbpedia:Torishima_Izu_Islands foaf:name “鳥島”@en. • Incorrect link to “external Web pages”  Example: dbpedia:John-Two-Hawks dbpedia-owl:wikiPageExternalLink <http://cedarlakedvd.com/>
  • 10. Who is the crowd Challenge LD Experts Difficult task Final prize Find Verify Microtasks Workers Easy task Micropayments TripleCheckMate [Kontoskostas2013] MTurk Adapted from [Bernstein2010] http://mturk.com
  • 11. How to crowdsource: workflow 11
  • 12. How to crowdsource: microtask design • Selection of foaf:name or rdfs:label to extract human- readable descriptions • Values extracted automatically from Wikipedia infoboxes • Link to the Wikipedia article via foaf:isPrimaryTopicOf • Preview of external pages by implementing HTML iframe Incorrect object Incorrect data type or language tag Incorrect outlink
  • 13. Experiments • Crowdsourcing approaches • Find stage: Challenge with LD experts • Verify stage: Microtasks (5 assignments) • Gold standard • Two of the authors generated gold standard for all contest triples indepedently • Conflicts resolved via mutual agreement • Validation: precision
  • 14. Overall results LD experts Microtask workers Number of distinct participants 50 80 Total time 3 weeks (predefined) 4 days Total triples evaluated 1,512 1,073 Total cost ~ US$ 400 (predefined) ~ US$ 43
  • 15. Precision results: Incorrect objects • Turkers can reduce the error rates of LD experts from the Find stage • 117 DBpedia triples had predicates related to dates with incorrect/incomplete values: ”2005 Six Nations Championship” Date 12 . • 52 DBpedia triples had erroneous values from the source: ”English (programming language)” Influenced by ? . • Experts classified all these triples as incorrect • Workers compared values against Wikipedia and successfully classified this triples as “correct” Triples compared LD Experts MTurk (majority voting: n=5) 509 0.7151 0.8977
  • 16. Precision results: Incorrect data types 0 20 40 60 80 100 120 140 Date English Millimetre Nanometre Number Number with decimals Second Volt Year Not specified / URI Numberoftriples Data types Experts TP Experts FP Crowd TP Crowd FP Triples compared LD Experts MTurk (majority voting: n=5) 341 0.8270 0.4752
  • 17. Precision results: Incorrect links • We analyzed the 189 misclassifications by the experts • The 6% misclassifications by the workers correspond to pages with a language different from English 50% 39% 11% Freebase links Wikipedia images External links Triples compared Baseline LD Experts MTurk (n=5 majority voting) 223 0.2598 0.1525 0.9412
  • 18. Conclusions • The effort of LD experts should be invested in tasks that demand domain-specific skills – Experts seem less motivated to perform on tasks that come with an additional overhead (e.g., checking an external link, going back on Wikipedia, data repair) • MTurk crowd was exceptionally good at performing object types and links checks – But seem not to understand the intricacies of data types
  • 19. Future work • Additional experiments with data types • Workflows for new types of errors, e.g., tasks which experts cannot answer on the fly • Training the crowd • Building the DBpedia ‘social machine’ – Add automatic QA tools – Close the feedback loop (from crowd to experts) – Change the parameters of the ‘Find’ step to achieve seamless integration – Change crowdsourcing model of ‘Find’ to improve retention 19
  • 20. CROWDSOURCING IMAGE ANNOTATION Improving paid microtasks through gamification and adaptive furtherance incentives O Feyisetan, E Simperl, M Van Kleek, N Shadbolt WWW 2015, to appear 20
  • 21. Overview • Make paid microtasks more cost-effective though gamification* *use of game elements and mechanics in a non-game context
  • 22. What to crowdsource: image labeling 22
  • 23. Who is the crowd: paid microtask platforms 23 [Kaufmann, Schulze, Viet, 2011]
  • 24. Research hypotheses • Contributors will perform better if tasks are more engaging • Increased accuracy through higher inter- annotator agreement • Cost savings through reduced unit costs • Micro-targeting incentives when players attempt to quit improves retention
  • 25. How to crowdsource: microtask design • Image labeling tasks published on microtask platform – Free text labels, varying numbers of labels per image, taboo words – 1st setting: ‘standard’ tasks, including basic spam control – 2nd setting: the same requirements and rewards, but contributors were asked to carry out the task in WordSmith 25
  • 26. How to crowdsource: gamification • Levels – 9 levels from newbie to Wordsmith, function of images tagged • Badges – function of number of images tagged • Bonus points – for new tags • Treasure points – for multiples of bonus points • Feedback alerts - related to badges, points, levels • Leaderboard - hourly scores and level of the top 5 players • Activities widget – real-time updates on other players 26
  • 27. How to crowdsource: incentives • Money – 5 cents extra for the effort • Power – see how other players tagged the images • Leaderboard - advance to the ‘Global’ leaderboard seen by everyone • Badges – receive ’Ultimate’ badge and a shiny new avatar • Levels – advance straight to the next level • Access - quicker access to treasure points 27 Assigned at random or targeted based on Bayes inference (number of tagged images as feature)
  • 28. Experiments • Crowdsourcing approaches – Paid microtasks – Wordsmith GWAP • Data set: ESP Game (120K images with curated labels) • Validation: precision 28
  • 29. Experiments (2) • 1st experiment: ‘newcomers’ – Tag one image with two keywords ~ complete Wordsmith level 1 – 200 images, US$ 0.02 per task, 3 judgements per task • 2nd experiment: ‘novice’ – Tag 11 images ~ complete Wordsmith level 2 – 2200 images, US$ 0.10 per task, 600 workers – Furtherance incentives: none, random, targeted (Bayes inference based on # of images tagged) 29
  • 30. Results: 1st experiment Metric CrowdFlower Wordsmith Total workers 600 423 Total keywords 1,200 41,206 Unique keywords 111 5,708 Avg. agreement 5.72% 37.7% Mean images/person 1 32 Max images/person 1 200
  • 31. Results: 2nd experiment Metric CrowdFlower Wordsmith (no furtherance incentives Total workers 600 514 Total keywords 13,200 35,890 Unique keywords 1,323 4,091 Avg. agreement 6.32% 10.9% Mean images/person 11 27 Max images/person 1 351
  • 32. Results: 2nd experiment, incentives • Incentives led to increased participation – Power: mean image/person = 132 – People come back (20 times!) and play longer (43 hours, 3 hours without incentives) 32
  • 33. Conclusions and future work • Task design matters as much as payment • Gamification achieves high accuracy for lower costs and improved engagement • Workers appreciate social features, but their main motivation is still task-driven – Feedback mechanisms, peer learning, training 33