Pests of mustard_Identification_Management_Dr.UPR.pdf
#kbdata: Exploring potential impact of technology limitations on DH research
1. #kbdata: Exploring potential impact of
technology limitations on DH research
Myriam C. Traub, Jacco van Ossenbruggen
Centrum Wiskunde & Informatica, Amsterdam
2. Translate the established tradition of source
criticism to the digital world and create a new
tradition of tool criticism to systematically
identify and explain technology-induced bias.
http://event.cwi.nl/toolcriticism/ #toolcrit
2
3. Context
✤ SealincMedia project, original
goals:
✤ crowdsourcing enrichment
✤ measure effect on scholarly
tasks
✤ Who are the scholars?
✤ What are their tasks?
3
4. Interviews
✤ Aim:
✤ Find out what types of
research tasks scholars
perform on digital archives
✤ Which quantitative / distant
reading tasks are not
(sufficiently) supported
✤ Scholars with experience in
performing historical research
on digital archives
4
(seeTPDL 2015 paper for details)
5. 5
I mostly use digital archives for
exploration of a topic, selecting
material for close reading (T1, T2) or
external processing (T4).
OCR quality in digital archives /
libraries is partly very bad.
I cannot quantify its impact on my
research tasks.
I would not trust quantitative
analyses (T3a, T3b) based on this data
sufficiently to use it in publications.
6. Categorisation of research tasks
T1 find the first mention of a concept
T2 find a subset with relevant documents
T3 investigate quantitative results over time
T3.a compare quantitative results for two terms
T3.b compare quantitative results from two corpora
T4 tasks using external tools on archive data
7. Literature
✤ OCR quality is addressed from
the perspective of the collection
owner/OCR software developer
✤ Usability studies for digital
libraries
✤ Robustness of search engines
towards OCR errors
✤ Error removal in post-
processing either systematically
or intellectually
7
8. We care
about average
performance on
representative subsets
for generic cases.
I care about
actual performance
on my non-
representative subset
for my specific
query.
8
Two different perspectives of quality evaluation
9. Use case
✤ Aims:
✤ To study the impact on
research tasks in detail
✤ Identify starting points for
workarounds and/or further
research
✤ Tasks T1 - T3
9
10. T1: Finding the
first mention
✤ Key requirement: recall
✤ 100% recall is unrealistic
✤ Aim: Find out how a scholar
can assess the reliability of
results
10
14. 01
OCR confidence
values useful?
✤ Available for all items in the
collection: page, word,
character
✤ Only for highest ranked
words / characters, other
candidates missing
✤ This information would be
required to estimate recall.
14
15. Confusion table
✤ Applied frequent OCR
confusions to query
✤ 23 alternative spellings, but
none of them yielded an
earlier mention
✤ Problem: long tail
Amstcrdam 16-01-1743
Amstordam 01-08-1772
Amsttrdam 04-08-1705
Amslerdam 12-12-1673
Amslcrdam 20-06-1797
Amslordam 29-06-1813
Amsltrdam 13-04-1810
Amscerdam 17-10-1753
Amsccrdam 16-02-1816
Amscordam 01-11-1813
Amsctrdam 16-06-1823
Amfterdam already found
Amftcrdam 17-08-1644
Amftordam 31-01-1749
Amfttrdam 26-11-1675
Amflerdam 03-03-1629
Amflcrdam 01-03-1663
Amflordam 05-03-1723
Amfltrdam 01-09-1672
Amfcerdam 22-04-1700
Amfccrdam 27-11-1742
Amfcordam -
Amfctrdam 09-10-1880
correct confused
s f
n u
e c
n a
t l
t c
h b
l i
e o
e t
full table available online:
http://dx.doi.org/10.6084/m9.figshare.1448810
19. Confusion Matrix OCR Confidence
Values
Alternative
Confidence
Values
available: sample only full corpus not available
T1 find all queries for x,
impractical
estimated precision, not
helpful
improve recall
T2 as above estimated precision,
requires improved UI
improve recall
T3 pattern summarized over
set of alternative queries
estimates of corrected
precision
estimates of
corrected recall
T3.a warn for different
susceptibility to errors
as above, warn for
different distribution of
confidence values
as above
T3.b as above as above as above
19
20. Conclusions
Problems
✤ Scholars see OCR
quality as a serious
problem, but cannot
assess its impact
✤ OCR technology is
unlikely to be perfect
✤ OCR errors are
reported in terms of
averages measured
over representative
samples
✤ Impact on a specific
research task cannot
be assessed based on
average error metrics
Start of solutions
✤ Impact of OCR is
different for different
research tasks, so
these tasks need to
made be explicit
✤ OCR errors often
assumed to be
random but are often
partly systematic
✤ Tool pipelines and
their limitations need
to be transparent &
better documented
21. No silver bullet
✤ we propose novel strategies that solve
part of the problem:
✤ critical attitude
(awareness and better support)
✤ transparency
(provenance, open source,
documentation, …)
✤ alternative quality metrics
(taking research context into account)
21
23. User logs
✤ 5 months on 8 servers
✤ March - July 2015
✤ 100 M requests
✤ 4 M queries
✤ 1 M unique queries
(dominated by named entities)
✤ 2.7 M unique documents
viewed
27. Candidate metric to
measure search bias
✤ Retrievability
(IR, Azzopardi, CIKM 2008)
✤ measures how often documents
are retrieved for a given set Q
✤ compares popular documents
against non-popular
✤ Inequality expressed with Gini
coefficient and Lorenz curve
✤ Inequality correlated with user
interest is fine…
28. Experimental setup
✤ Repeat original experiment with synthesised queries
✤ Run experiment with real queries from log
✤ note the ratio: 1M queries vs 100M documents
✤ To do: test known item search for different quality
OCR, different media, different titles, …
38. Conclusions
✤ Only small fragment of newspaper corpus is viewed or even
retrieved in top #10, 100, 100
✤ No clear evidence retrieval bias is correlated with OCR errors.
Why?
✤ there is no relation
✤ we look for patterns at a too generic level
✤ back to the specificity of the use cases?
✤ Other forms of bias that are measurable/quantifiable?