This document describes the Search and Hyperlinking task that was developed between 2012-2014. It involved developing technologies for video search and hyperlinking based on user needs expressed as queries. Users tested the system using a BBC video collection and provided relevance judgments which were used to evaluate performance using metrics like MAP, P@5/10. Results showed that ASR transcripts, prosodic features and metadata improved search accuracy while concept detection approaches worked best for hyperlinking. Lessons learned included how device and features affected user behavior and system performance.
4. Users
Main
group
User
Target
Media
Professionals
Broadcast
Researchers
&
Educators
Journalists
Research
Academic
researchers
&
students
InvesDgate
Academic
educators
Educate
Public
users
CiMzens
Entertainment,
Infotainment
Professionals
Reuse
Media
Archivists
Annotate
7. 1998 2002 2008
2010 2013 2015
DATA
BIG DATA?
not
representaMve
representaMve
8. Search
&
Hyperlinking
task
• User
oriented:
aim
to
explore
the
needs
of
real
users
expressed
as
queries.
– How:
UK
ciMzens
and
crowd
sourcing
for
retrieval
assessment
• Temporal
aspect:
seek
to
direct
users
to
the
relevant
parts
of
retrieved
video
(“jump-‐in
point”).
– How:
segmentaMon,
segment
overlap,
transcripts.
prosodic,
visual
(low-‐level,
high-‐level;
keyframes)
• MulDmodal:
want
to
invesMgate
technologies
for
addressing
variety
in
user
needs
and
expectaMons
– varied
visual
and
audio
contribuMons,
intenMonal
gap
between
query
and
mulMmodal
descriptors
in
content
9. ME
Search
&
Hyperlinking
task
in
development:
2012
–
2014
Search
Hyperlinking
2012
2013
2014
2012
2013
2014
Dataset
BlipTv
BBC
BlipTv
BBC
Features
released:
!
Transcripts
2
ASR
3
ASR
2
ASR
3
ASR
!
Prosodic
features
no
yes
no
yes
!
Visual
clues
for
queries
yes
no
no
!
Concept
detecDon
yes
yes
Type
of
the
task
Known-‐item
Ad-‐hoc
Ad-‐hoc
Query/Anchors
creaDon
PC
iPad
PC
iPad
Number
of
queries/anchors
30/30
4/50
50/30
30/30
11/
98/30
Relevance
assessment
MTurk
users
(BBC)
MTurk
MTurk
Numbers
of
assessed
cases
30
50
9
900
3
517
9
975
13
141
EvaluaDon
metrics
MRR,
MASP,
MASDWP
MAP(-‐bin/
tol),
P@5/10
MAP
MAP(-‐bin/tol),
P@5/10
10. Dataset:
Video
collecMon
• BBC
copyright
cleared
broadcast
material:
– Videos:
• Development
set:
6
weeks
between
01.04.2008
and
11.05.2008
(1335
hours/2323
videos)
•
Test
set:
11
weeks
between
12.05.2008
and
31.07.2008
(2686
hours,
3528
videos)
– Manually
transcribed
subMtles
– Metadata
• AddiDonal
data:
– ASR:
LIMSI/Vocapia,
LIUM,
NST-‐Sheffield
– Shot
boundaries,
keyframes
– Output
of
visual
concept
detectors
by
University
of
Leuven,
and
University
of
Oxford
11. Dataset:
Query
• 28
Users
-‐
Policemen,
Hair
dresser,
Bouncer,
Sales
manger,
Student,
Self-‐employed
• Two
hour
session
on
iPads:
– Search
the
archive
(document
level)
– Define
clips
(segment
level)
– Define
anchors
(anchor
level)
Statement
of
InformaMon
Need
Search
Refine
Relevant
Clips
Define
Anchors
12. User
study
@
BBC:
1.)
Statement
of
InformaMon
Need
13. User
study
@
BBC:
2.)
Search
Relevant
clips
Goto
1.)
Goto
3.)
16. Data
cleaning:
Usable
InformaMon
Need
• DescripMon
clearly
specifies
what
is
relevant
• A
query
with
a
suitable
Mtle
exists
• Sufficient
relevant
segments
exist
(try
query)
17. Data
cleaning:
Process
• For
each
informaMon
need
in
batch
1. check
if
usable
2. If
in
doubt
use
search
to
search
for
relevant
data
3. reword
&
spellcheck
descripMon
4. select
the
first
suitable
query
5. Save
18. Data
cleaning:
Usable
Anchor
• Longer
than
5
seconds
• DesMnaMon
descripMon
clearly
idenMfies
the
material
the
user
wants
to
see
when
he
would
acMvate
the
anchor
described
by
label
• It
is
likely
that
there
are
some
relevant
items
in
the
collecMon
19. Data
cleaning:
Process
• For
each
informaMon
need
in
assigned
batch
– Go
through
anchors
• check
if
usable
• reword
&
spellcheck
descripMon
• Assess
whether
it
is
like
to
find
links
in
the
collecMon
(possibly
using
search)
– Save
22. Ground
truth
creaMon
• Queries/Anchors:
user
studies
at
BBC:
-‐
28
users
with
following
profile:
" Age:
18-‐30
years
old
" Use
of
search
engines
and
services
on
iPads
on
the
daily
basis
• Relevance
assessment:
via
crowdsourcing
on
Amazon
MTurk
plaporm:
– Top
10
results
from
58
search
and
62
hyperlinking
submissions
– 1
judgment
per
query
or
anchor
that
was
accepted/rejected
based
on
an
automated
algorithm,
special
cases
of
users
typos
checked
manually
– Number
of
evaluated
HITs:
9
900
for
search,
and
13
141
for
hyperlinking
23. •
P@5/10/20
•
MAP
based:
•
EvaluaMon
metrics
MAP:
taking
into
account
any
overlapping
segment:
•
MAP-‐bin:
relevant
segments
are
binned
for
relevance:
•
MAP-‐tol:
only
start
Mmes
of
the
segments
are
considered:
31. Lessons
learned
1.
iPad
vs
PC
=
different
user
behaviour
and
expectaMon
from
the
system.
2.
Prosodic
features
broaden
the
scope
of
the
search
sub-‐task.
3.
Use
of
shot
segmentaMon
based
units
achieves
the
worst
scores
for
both
sub-‐tasks.
4.
Use
of
metadata
improves
results
for
both
sub-‐tasks.
32. The
Search
and
Hyperlinking
task
was
supported
by
We
are
grateful
to
Jana
Eggink
and
Andy
O'Dwyer
from
the
BBC
for
preparing
the
collecMon
and
hosMng
the
user
trials.
...
and
of
course
Martha
for
advise
&
crowdsourcing
access.
33. JRS
at
Search
and
Hyperlinking
of
Television
Content
Task
Werner
Bailer,
Harald
SMegler
MediaEval
Workshop,
Barcelona,
Oct.
2014
34. Linking
sub-‐task
• Matching
terms
from
textual
resources
• Reranking
based
on
visual
similarity
(VLAT)
• Using
visual
concepts
(only/in
addiMon)
• Results
– Differences
between
different
text
resources
– Context
helped
only
in
few
of
the
cases
– Visual
reranking
provides
small
improvement
– Visual
concepts
did
not
provide
improvements
34
35. Zsombor
Paróczi,
Bálint
Fodor,
Gábor
Szűcs
SoluDon
with
concept
enrichment
• Concept
enrichment:
the
set
of
words
is
extended
with
their
synonyms
or
other
conceptually
connected
words.
• Top
10
vs
top
50
conceptually
connected
words
for
each
word
• Conclusion:
the
results
show
that
concept
enrichment
with
less
words
give
beuer
precision
because
at
the
opposite
case
the
noise
is
greater.
36. Television
Linked
To
The
Web
H.A.
Le1,
Q.M.
Bui1,
B.
Huet1,
B.
Cervenková2,
J.
Bouchner2,
E.
Apostolidis3,
F.
Markatopoulou3,
A.
Pournaras3,
V.
Mezaris3,
D.
Stein4,
S.
Eickeler4,
and
M.
Stadtschnitzer4
1 - Eurecom, Sophia Antipolis, France.
2 - University of Economics, Prague, Czech Republic.
3 - Information Technologies Institute, CERTH, Thessaloniki, Greece.
4 - Fraunhofer IAIS, Sankt Augustin, Germany.
16-‐17
Oct
2014
www.linkedtv.eu
37. Reasons
to
visit
the
LinkedTV
poster
LinkedTV
@
MediaEval
2014
Search
and
Hyperlinking
Task
38. Reasons
to
visit
the
LinkedTV
poster
LinkedTV
@
MediaEval
2014
Search
and
Hyperlinking
Task
39. Reasons
to
visit
the
LinkedTV
poster
LinkedTV
@
MediaEval
2014
Search
and
Hyperlinking
Task