SlideShare a Scribd company logo
1
Introduction
Tools
Design principles
Open questions
Listening Test Interface Design
Brecht De Man
Correspondence: <b.deman@qmul.ac.uk>
Centre for Digital Music
School of Electronic Engineering and Computer Science
Queen Mary, University of London
July 12, 2016
@BrechtDeMan
1
Introduction
Tools
Design principles
Open questions
In the news
Bug in fMRI software calls 15 years of research into question
(Wired UK)
Software faults raise questions about the validity of brain
studies (Ars Technica)
Source: iStock
“In theory, we should find 5% false positives (for a
significance threshold of 5%), but instead we found that
the most common software packages for fMRI analysis
can result in false-positive rates of up to
70%. These results question the validity of some
40,000 fMRI studies and may have a large
impact on the interpretation of neuroimaging results.”
[1] A. Eklund, T. E. Nichols, and H. Knutsson, “Cluster failure: Why fMRI inferences for spatial extent have
inflated false-positive rates,” Proceedings of the National Academy of Sciences, 2016.
@BrechtDeMan Listening Test Interface Design July 12, 2016 1 / 35
1
Introduction
Tools
Design principles
Open questions
In the news
Bug in fMRI software calls 15 years of research into question
(Wired UK)
Software faults raise questions about the validity of brain
studies (Ars Technica)
“In theory, we should find 5% false positives (for a
significance threshold of 5%), but instead we found that
the most common software packages for fMRI analysis
can result in false-positive rates of up to
70%. These results question the validity of some
40,000 fMRI studies and may have a large
impact on the interpretation of neuroimaging results.”
[1] A. Eklund, T. E. Nichols, and H. Knutsson, “Cluster failure: Why fMRI inferences for spatial extent have
inflated false-positive rates,” Proceedings of the National Academy of Sciences, 2016.
[2] Bennett et al., “Neural Correlates of Interspecies Perspective Taking in the Post-Mortem Atlantic Salmon: An
Argument For Proper Multiple Comparisons Correction,” Journal of Serendipitous and Unexpected Results, 2010.
@BrechtDeMan Listening Test Interface Design July 12, 2016 1 / 35
1
Introduction
Tools
Design principles
Open questions
Applications
I Audio codecs
I Electroacoustic design (loudspeakers, headphones, ...)
I Human perception
I Standards
I ...
@BrechtDeMan Listening Test Interface Design July 12, 2016 2 / 35
1
Introduction
Tools
Design principles
Open questions
Perceptual evaluation
I Interface design
I Subject selection
I Subject training
I Stimulus selection
I Listening environment
I Playback system
I Subject exclusion
I Research questions
I ...
@BrechtDeMan Listening Test Interface Design July 12, 2016 3 / 35
1
Introduction
Tools
Design principles
Open questions
Goals
I Accurate, reproducible, representative
I High discrimination
I Low time and e↵ort
@BrechtDeMan Listening Test Interface Design July 12, 2016 4 / 35
1
Introduction
Tools
Design principles
Open questions
Basic reading
[3] Søren Bech and Nick Zacharov, “Perceptual Audio Evaluation -
Theory, Method and Application,” Wiley, 2006.
@BrechtDeMan Listening Test Interface Design July 12, 2016 5 / 35
1
Introduction
Tools
Design principles
Open questions
Table of Contents
1 Introduction
2 Tools
3 Design principles
4 Open questions
@BrechtDeMan Listening Test Interface Design July 12, 2016
1
Introduction
Tools
Design principles
Open questions
Table of Contents
1 Introduction
2 Tools
3 Design principles
4 Open questions
@BrechtDeMan Listening Test Interface Design July 12, 2016
1
Introduction
Tools
Design principles
Open questions
Available listening test tools
Name Supported interfaces Language
APE multi-stimulus multi-axis, pairwise MATLAB
BeaqleJS MUSHRA + ABX JavaScript
HULTI-GEN many Max
MUSHRAM MUSHRA MATLAB
Scale semantic di↵erential, n-AFC MATLAB
WhisPER many (7) MATLAB
@BrechtDeMan Listening Test Interface Design July 12, 2016 6 / 35
1
Introduction
Tools
Design principles
Open questions
APE
I MATLAB-based
I Pairwise (AB) mode
I Multiple stimulus mode
• All stimuli on one axis
• Multiple axes for multiple attributes
I Made available to public
• Others can spot bugs and make improvements
• code.soundsoftware.ac.uk/projects/APE
• AES 136th
Convention [4]
@BrechtDeMan Listening Test Interface Design July 12, 2016 7 / 35
1
Introduction
Tools
Design principles
Open questions
Web Audio Evaluation Tool
I Browser-based: OS-independent, no 3rd party software
I Wide range of highly customisable interfaces
I Remote or local tests
I Easy quick test setup: no programming required
I High compatibility: configuration and results in XML
I Immediate diagnostics and result in browser
I Intuitive, flexible, works out of the box
@BrechtDeMan Listening Test Interface Design July 12, 2016 8 / 35
1
Introduction
Tools
Design principles
Open questions
Web Audio Evaluation Tool
Developers
I Nicholas Jillings Web Audio
I Brecht De Man perception of music production
I David Mo↵at realism of sound synthesis
Contributors
I Giulio Moro Hammond organs
I Adib Mehrabi vocal imitation of musical sounds
I Alex Wilson audio quality in music
I Alejandro Osses Vecchi psychoacoustics
@BrechtDeMan Listening Test Interface Design July 12, 2016 9 / 35
1
Introduction
Tools
Design principles
Open questions
Table of Contents
1 Introduction
2 Tools
3 Design principles
4 Open questions
@BrechtDeMan Listening Test Interface Design July 12, 2016
1
Introduction
Tools
Design principles
Open questions
Double blind
No blind:
@BrechtDeMan Listening Test Interface Design July 12, 2016 10 / 35
1
Introduction
Tools
Design principles
Open questions
Double blind
Single blind:
@BrechtDeMan Listening Test Interface Design July 12, 2016 10 / 35
1
Introduction
Tools
Design principles
Open questions
Double blind
Double blind:
@BrechtDeMan Listening Test Interface Design July 12, 2016 10 / 35
1
Introduction
Tools
Design principles
Open questions
Double blind
[5] F. E. Toole and S. Olive, “Hearing is believing vs. believing is hearing: Blind vs. sighted listening tests, and
other interesting things,” in Audio Engineering Society Convention 97, November 1994.
@BrechtDeMan Listening Test Interface Design July 12, 2016 10 / 35
1
Introduction
Tools
Design principles
Open questions
Individual tests
@BrechtDeMan Listening Test Interface Design July 12, 2016 11 / 35
1
Introduction
Tools
Design principles
Open questions
Randomisation
Remove bias due to order of
I (Presentation and labelling of) stimuli
I Parts of the experiment
I Tests
I ...
@BrechtDeMan Listening Test Interface Design July 12, 2016 12 / 35
1
Introduction
Tools
Design principles
Open questions
Free switching
“Increases ability to perceive more delicate di↵erences”
[6] J. Berg and F. Rumsey, “Spatial attribute identification and scaling by repertory grid technique and other
methods,” 16th AES Conference: Spatial Sound Reproduction, March 1999.
@BrechtDeMan Listening Test Interface Design July 12, 2016 13 / 35
1
Introduction
Tools
Design principles
Open questions
Time-aligned stimuli
I Switch to corresponding position in next stimulus
I More balanced over full length of stimuli
@BrechtDeMan Listening Test Interface Design July 12, 2016 14 / 35
1
Introduction
Tools
Design principles
Open questions
Reduce visual distractions
I “Visual information processing [induces] Inattentional
Deafness”
I “Requiring use of scroll bar: significant negative e↵ect on
critical listening reaction times”
[7] Joshua Mycroft, Joshua D. Reiss and Tony Stockman, “The Influence of Graphical User Interface Design on
Critical Listening Skills,” Sound and Music Computing (SMC), July 2013.
@BrechtDeMan Listening Test Interface Design July 12, 2016 15 / 35
1
Introduction
Tools
Design principles
Open questions
Automated, computerised tests
I Double blind
I Individual tests
I Randomisation
I Free switching
I Time-aligned stimuli
I Reduce visual distractions
@BrechtDeMan Listening Test Interface Design July 12, 2016 16 / 35
1
Introduction
Tools
Design principles
Open questions
Automated, computerised tests
[8] Helmut Haas, “The influence of a single echo on the audibility of speech,” JAES Vol. 20 No. 2, pp. 146-159,
1972.
@BrechtDeMan Listening Test Interface Design July 12, 2016 16 / 35
1
Introduction
Tools
Design principles
Open questions
Automated, computerised tests
Dr Sean Olive (HARMAN)
I Pencil and paper
I HyperCard (late 1980s)
@BrechtDeMan Listening Test Interface Design July 12, 2016 16 / 35
1
Introduction
Tools
Design principles
Open questions
Automated, computerised tests
Dr Sean Olive (HARMAN)
I Pencil and paper
I HyperCard (late 1980s)
I Pocket PC / PDA (early 2000s)
@BrechtDeMan Listening Test Interface Design July 12, 2016 16 / 35
1
Introduction
Tools
Design principles
Open questions
Automated, computerised tests
Dr Sean Olive (HARMAN)
I Pencil and paper
I HyperCard (late 1980s)
I Pocket PC / PDA (early 2000s)
I Apple iPad
@BrechtDeMan Listening Test Interface Design July 12, 2016 16 / 35
1
Introduction
Tools
Design principles
Open questions
Pilot test and dummy analysis
Source: Muppet Show Wikia
I Feedback on interface and stimuli
I Know what type of data to expect
@BrechtDeMan Listening Test Interface Design July 12, 2016 17 / 35
1
Introduction
Tools
Design principles
Open questions
Language
Native tongue: “Minimise (risk of) semantic constraints on
subjects”
[9] J. Berg and F. Rumsey, “Correlation between emotive, descriptive and naturalness attributes in subjective data
relating to spatial sound reproduction,” AES Convention 109, September 2000.
@BrechtDeMan Listening Test Interface Design July 12, 2016 18 / 35
1
Introduction
Tools
Design principles
Open questions
Safety
I Don’t allow accidental closing of window
I Intermediate session saves
@BrechtDeMan Listening Test Interface Design July 12, 2016 19 / 35
1
Introduction
Tools
Design principles
Open questions
Reward
I Immediately display all / own results
I Candy
@BrechtDeMan Listening Test Interface Design July 12, 2016 20 / 35
1
Introduction
Tools
Design principles
Open questions
Diagnostics
I Spot errors early
I Exclude subjects
I Insights to better design tests
I Statistics for experiment report
@BrechtDeMan Listening Test Interface Design July 12, 2016 21 / 35
1
Introduction
Tools
Design principles
Open questions
Surveys
I Immediately before/after test/page
• Avoid chasing subjects afterwards
• Some questions might relate to the test
@BrechtDeMan Listening Test Interface Design July 12, 2016 22 / 35
1
Introduction
Tools
Design principles
Open questions
Comments
I Learn more than simple, single-attribute ratings
I Understand how subjects used the scale
I “People get frustrated when they don’t have comment boxes”
(Sean Olive, AES140 Workshop)
I Separate comment boxes (one for each stimulus) instead of
one
• Proportion of stimuli commented on: 96.5% instead of 82.1%
• Comments 47% longer
[4] Brecht De Man and Joshua D. Reiss, “APE: Audio perceptual evaluation toolbox for MATLAB,” 136th
Convention of the Audio Engineering Society, April 2014.
@BrechtDeMan Listening Test Interface Design July 12, 2016 23 / 35
1
Introduction
Tools
Design principles
Open questions
Axis ticks
[10] Jouni Paulus, Christian Uhle and J¨urgen Herre, “Perceived Level of Late Reverberation in Speech and Music,”
AES Convention 130, 2011.
@BrechtDeMan Listening Test Interface Design July 12, 2016 24 / 35
1
Introduction
Tools
Design principles
Open questions
Table of Contents
1 Introduction
2 Tools
3 Design principles
4 Open questions
@BrechtDeMan Listening Test Interface Design July 12, 2016
1
Introduction
Tools
Design principles
Open questions
Calibration
I Audiometric test
I Test of playback system
@BrechtDeMan Listening Test Interface Design July 12, 2016 25 / 35
1
Introduction
Tools
Design principles
Open questions
Multiple stimulus or pairwise?
“Multiple comparison methods produce more reliable and
discriminating judgments of sound quality compared to single
stimulus or paired comparison methods.”
[11] S. Olive and T. Welti, “The relationship between perception and measurement of headphone sound quality,” in
Audio Engineering Society Convention 133, October 2012.
“Multiple stimulus tests are substantially less time-consuming.”
[12] B. De Man and J. D. Reiss, “A pairwise and multiple stimuli approach to perceptual evaluation of microphone
types,” in Audio Engineering Society Convention 134, May 2013.
Multiple stimulus only viable for limited number of stimuli to be
compared against each other.
@BrechtDeMan Listening Test Interface Design July 12, 2016 26 / 35
1
Introduction
Tools
Design principles
Open questions
Listening level
Options
I Everyone at same level
I Choose most comfortable/familiar level
• Set once at start
• Change throughout test
@BrechtDeMan Listening Test Interface Design July 12, 2016 27 / 35
1
Introduction
Tools
Design principles
Open questions
To MUSHRA or not to MUSHRA?
[13] Method for the subjective assessment of intermediate quality level of coding systems. Recommendation ITU-R
BS.1534-1, 2003.
[14] E. Vincent, M. G. Jafari, and M. D. Plumbley, “Preliminary guidelines for subjective evalutation of audio
source separation algorithms,” in UK ICA Research Network Workshop, 2006.
@BrechtDeMan Listening Test Interface Design July 12, 2016 28 / 35
1
Introduction
Tools
Design principles
Open questions
To MUSHRA or not to MUSHRA?
I “[MUSHRA aims] to establish the detectability, magnitude of
impairment or basic audio quality of the codec relative to a
reference. [It does] not attempt to measure the listeners’
preference for the audio codec when compared to the hidden
lossless reference.”
[15] Sean E. Olive, “Some new evidence that teenagers and college students may prefer accurate sound
reproduction,” in Audio Engineering Society Convention 132, April 2012.
I What if there is no reference to compare to?
@BrechtDeMan Listening Test Interface Design July 12, 2016 28 / 35
1
Introduction
Tools
Design principles
Open questions
To MUSHRA or not to MUSHRA?
“More accessible, particularly to technically experienced subjects”
[16] Christoph V¨olker and Rainer Huber, “Adaptations for the MUlti Stimulus test with Hidden Reference and
Anchor (MUSHRA) for elder and technical unexperienced participants,” DAGA, 2015.
@BrechtDeMan Listening Test Interface Design July 12, 2016 28 / 35
1
Introduction
Tools
Design principles
Open questions
To MUSHRA or not to MUSHRA?
[4] Brecht De Man and Joshua D. Reiss, “APE: Audio perceptual evaluation toolbox for MATLAB,” 136th
Convention of the Audio Engineering Society, April 2014.
@BrechtDeMan Listening Test Interface Design July 12, 2016 28 / 35
1
Introduction
Tools
Design principles
Open questions
To MUSHRA or not to MUSHRA?
[17] Nicholas Jillings, Brecht De Man, David Mo↵at and Joshua D. Reiss, “Web Audio Evaluation Tool: A
browser-based listening test environment,” 12th Sound and Music Computing Conference, July 2015.
@BrechtDeMan Listening Test Interface Design July 12, 2016 28 / 35
1
Introduction
Tools
Design principles
Open questions
To MUSHRA or not to MUSHRA?
[18] Jan Berg, “OPAQUE - A tool for the elicitation and grading of audio quality attributes,” AES Convention 118,
2005.
@BrechtDeMan Listening Test Interface Design July 12, 2016 28 / 35
1
Introduction
Tools
Design principles
Open questions
Constraints on use of scale
Examples
I “Put at least one slider at 100%”
I “The anchor needs to be below 20%”
I ...
Options
I Instructions
I Enforced (not possible to proceed)
I Remove subject from test
I Normalise ratings after the fact
@BrechtDeMan Listening Test Interface Design July 12, 2016 29 / 35
1
Introduction
Tools
Design principles
Open questions
Remote tests
I Pro: Low e↵ort, scales easily, di↵erent
locations/cultures/languages within reach
I Con: Loss of control ... or is there?
[19] M. Cartwright, B. Pardo, G. Mysore and M. Ho↵man “Fast and Easy Crowdsourced Perceptual Audio
Evaluation,” ICASSP, 2016.
[20] M. Schoe✏er, F.-R. St¨oter, H. Bayerlein, B. Edler and J. Herre, “An Experiment about Estimating the Number
of Instruments in Polyphonic Music: A Comparison Between Internet and Laboratory Results,” ISMIR, 2013.
[21] F.-R. St¨oter, M. Schoe✏er, B. Edler and J. Herre, “Human Ability of Counting the Number of Instruments in
Polyphonic Music,” Meetings on Acoustics Vol. 19, 2013.
@BrechtDeMan Listening Test Interface Design July 12, 2016 30 / 35
1
Introduction
Tools
Design principles
Open questions
Online tests
Why online?
I Cross-platform
I Centralised results collection (over web or local server)
I Multiple machines at once
I Leveraging other web technologies
@BrechtDeMan Listening Test Interface Design July 12, 2016 31 / 35
1
Introduction
Tools
Design principles
Open questions
Conclusion
Try it!
I Conduct an experiment
I Contribute
(github.com/BrechtDeMan/WebAudioEvaluationTool)
I Any problems? Feedback?
• Report issue
• b.deman@qmul.ac.uk
@BrechtDeMan Listening Test Interface Design July 12, 2016 32 / 35
WIMP
References
Q&A
AES 2nd Workshop on Intelligent Music Production
I Tuesday 13 September 2016
I Charterhouse Square, Queen Mary University of London
I Short (<2 pages) paper submission deadline: 1 August 2016
I Confirmed speakers:
• Bryan Pardo (Northwestern University)
• Fran¸cois Pachet (Sony CSL Music Paris)
I More info and registration: aes-uk.org/wimp
I Volunteers needed
@BrechtDeMan Listening Test Interface Design July 12, 2016 33 / 35
WIMP
References
Q&A
References
A. Eklund, T. E. Nichols, and H. Knutsson, “Cluster failure: Why fMRI inferences for spatial extent have inflated
false-positive rates,” Proceedings of the National Academy of Sciences, 2016.
B. et al., “Neural correlates of interspecies perspective taking in the post-mortem atlantic salmon: An argument
for proper multiple comparisons correction,” Journal of Serendipitous and Unexpected Results, 2010.
S. Bech and N. Zacharov, Perceptual Audio Evaluation - Theory, Method and Application.
John Wiley & Sons, 2007.
B. De Man and J. D. Reiss, “APE: Audio Perceptual Evaluation toolbox for MATLAB,” in Audio Engineering
Society Convention 136, April 2014.
F. E. Toole and S. Olive, “Hearing is believing vs. believing is hearing: Blind vs. sighted listening tests, and other
interesting things,” in Audio Engineering Society Convention 97, November 1994.
J. Berg and F. Rumsey, “Spatial attribute identification and scaling by repertory grid technique and other
methods,” in Audio Engineering Society Conference: 16th International Conference: Spatial Sound Reproduction,
March 1999.
J. Mycroft, J. D. Reiss, and T. Stockman, “The influence of graphical user interface design on critical listening
skills,” Sound and Music Computing (SMC), Stockholm, 2013.
H. Haas, “The influence of a single echo on the audibility of speech,” J. Audio Eng. Soc, vol. 20, no. 2,
pp. 146–159, 1972.
J. Berg and F. Rumsey, “Correlation between emotive, descriptive and naturalness attributes in subjective data
relating to spatial sound reproduction,” in Audio Engineering Society Convention 109, September 2000.
J. Paulus, C. Uhle, and J. Herre, “Perceived level of late reverberation in speech and music,” in Audio
Engineering Society Convention 130, May 2011.
@BrechtDeMan Listening Test Interface Design July 12, 2016 34 / 35
WIMP
References
Q&A
References (cont.)
S. Olive and T. Welti, “The relationship between perception and measurement of headphone sound quality,” in
Audio Engineering Society Convention 133, October 2012.
B. De Man and J. D. Reiss, “A pairwise and multiple stimuli approach to perceptual evaluation of microphone
types,” in Audio Engineering Society Convention 134, May 2013.
Method for the subjective assessment of intermediate quality level of coding systems.
Recommendation ITU-R BS.1534-1, 2003.
E. Vincent, M. G. Jafari, and M. D. Plumbley, “Preliminary guidelines for subjective evalutation of audio source
separation algorithms,” in UK ICA Research Network Workshop, 2006.
S. E. Olive, “Some new evidence that teenagers and college students may prefer accurate sound reproduction,” in
Audio Engineering Society Convention 132, April 2012.
C. V¨olker and R. Huber, “Adaptions for the MUlti Stimulus test with Hidden Reference and Anchor (MUSHRA)
for elder and technical unexperienced participants,” in DAGA 2015 N¨urnberg, 2015.
N. Jillings, D. Mo↵at, B. De Man, and J. D. Reiss, “Web Audio Evaluation Tool: A browser-based listening test
environment,” in 12th Sound and Music Computing Conference, July 2015.
J. Berg, “OPAQUE – A tool for the elicitation and grading of audio quality attributes,” in 118th Convention of
the Audio Engineering Society, May 2005.
M. Cartwright, “Fast and easy crowdsourced perceptual audio evaluation,” in IEEE International Conference on
Acoustics, Speech and Signal Processing, 2016.
M. Schoe✏er, F.-R. St¨oter, H. Bayerlein, B. Edler, and J. Herre, “An experiment about estimating the number of
instruments in polyphonic music: A comparison between internet and laboratory results,” in 14th International
Society for Music Information Retrieval Conference (ISMIR 2013), pp. 389–394, 2013.
@BrechtDeMan Listening Test Interface Design July 12, 2016 34 / 35
WIMP
References
Q&A
References (cont.)
F.-R. St¨oter, M. Schoe✏er, B. Edler, and J. Herre, “Human ability of counting the number of instruments in
polyphonic music,” in Proceedings of Meetings on Acoustics, vol. 19, p. 035034, Acoustical Society of America,
2013.
@BrechtDeMan Listening Test Interface Design July 12, 2016 34 / 35
WIMP
References
Q&A
Sound sources
I Fredy V: “In The Meantime”
I ‘counting.wav’ by Corsica S (Freesound)
@BrechtDeMan Listening Test Interface Design July 12, 2016 34 / 35
1
WIMP
References
Q&A
Q&A
b.deman@qmul.ac.uk
www.brechtdeman.com
@BrechtDeMan
github.com/BrechtDeMan/WebAudioEvaluationTool
linkedin.com/in/BrechtDeMan
aes.org/aes/BrechtDeMan
@BrechtDeMan Listening Test Interface Design July 12, 2016 35 / 35

More Related Content

Similar to C4DM Seminar 2016-07-12: Brecht De Man

QALD-7 Question Answering over Linked Data Challenge
QALD-7 Question Answering over Linked Data ChallengeQALD-7 Question Answering over Linked Data Challenge
QALD-7 Question Answering over Linked Data Challenge
Holistic Benchmarking of Big Linked Data
 
Qald 7 at ESWC2017
Qald 7 at ESWC2017Qald 7 at ESWC2017
Qald 7 at ESWC2017
Giulio Napolitano
 
Immersive Recommendation Workshop, NYC Media Lab'17
Immersive Recommendation Workshop, NYC Media Lab'17Immersive Recommendation Workshop, NYC Media Lab'17
Immersive Recommendation Workshop, NYC Media Lab'17
Longqi Yang
 
Machine Support for Interacting with Scientific Publications Improving Inform...
Machine Support for Interacting with Scientific Publications Improving Inform...Machine Support for Interacting with Scientific Publications Improving Inform...
Machine Support for Interacting with Scientific Publications Improving Inform...
Christoph Lange
 
The culture of researchData
The culture of researchDataThe culture of researchData
The culture of researchData
petermurrayrust
 
Developing and sharing tools for bioelectromagnetic research
Developing and sharing tools for bioelectromagnetic researchDeveloping and sharing tools for bioelectromagnetic research
Developing and sharing tools for bioelectromagnetic research
Robert Oostenveld
 
The culture of researchData
The culture of researchData The culture of researchData
The culture of researchData
TheContentMine
 
The Culture of Research Data, by Peter Murray-Rust
The Culture of Research Data, by Peter Murray-RustThe Culture of Research Data, by Peter Murray-Rust
The Culture of Research Data, by Peter Murray-Rust
LEARN Project
 
Machine Learning Pitfalls
Machine Learning Pitfalls Machine Learning Pitfalls
Machine Learning Pitfalls
Dan Elton
 
The Dark Art: Is Music Recommendation Science a Science
The Dark Art: Is Music Recommendation Science a ScienceThe Dark Art: Is Music Recommendation Science a Science
The Dark Art: Is Music Recommendation Science a Science
mpapish
 
Mediated participatory design for contextually aware in vehicle experiences
Mediated participatory design for contextually aware in vehicle experiencesMediated participatory design for contextually aware in vehicle experiences
Mediated participatory design for contextually aware in vehicle experiences
Stavros Tasoudis
 
Using Bibliometrics to Keep Up with the Joneses
Using Bibliometrics to Keep Up with the JonesesUsing Bibliometrics to Keep Up with the Joneses
Using Bibliometrics to Keep Up with the Joneses
Christina Pikas
 
Noteworks SI 622 Presentation Slides
Noteworks SI 622 Presentation SlidesNoteworks SI 622 Presentation Slides
Noteworks SI 622 Presentation Slideskebuhc
 
Sound is not speech
Sound is not speechSound is not speech
Sound is not speech
Audio Analytic
 
SoundSoftware: Software Sustainability for audio and Music Researchers
SoundSoftware: Software Sustainability for audio and Music Researchers SoundSoftware: Software Sustainability for audio and Music Researchers
SoundSoftware: Software Sustainability for audio and Music Researchers
SoundSoftware ac.uk
 
Session 1 Lecture 2 PACT A Framework for Designing Interactive Systems
Session 1 Lecture 2 PACT A Framework for Designing Interactive SystemsSession 1 Lecture 2 PACT A Framework for Designing Interactive Systems
Session 1 Lecture 2 PACT A Framework for Designing Interactive Systems
Khalid Md Saifuddin
 
Potentials and limitations of ‘Automated Sentiment Analysis
Potentials and limitations of ‘Automated Sentiment AnalysisPotentials and limitations of ‘Automated Sentiment Analysis
Potentials and limitations of ‘Automated Sentiment Analysis
Karthik Sharma
 
Engaging with DARPA - Stefanie Thompkins
Engaging with DARPA - Stefanie ThompkinsEngaging with DARPA - Stefanie Thompkins
Engaging with DARPA - Stefanie Thompkins
Michigan State University Research
 
Berlin 6 Open Access Conference: Mark Liberman
Berlin 6 Open Access Conference: Mark LibermanBerlin 6 Open Access Conference: Mark Liberman
Berlin 6 Open Access Conference: Mark Liberman
Cornelius Puschmann
 

Similar to C4DM Seminar 2016-07-12: Brecht De Man (20)

QALD-7 Question Answering over Linked Data Challenge
QALD-7 Question Answering over Linked Data ChallengeQALD-7 Question Answering over Linked Data Challenge
QALD-7 Question Answering over Linked Data Challenge
 
Qald 7 at ESWC2017
Qald 7 at ESWC2017Qald 7 at ESWC2017
Qald 7 at ESWC2017
 
Immersive Recommendation Workshop, NYC Media Lab'17
Immersive Recommendation Workshop, NYC Media Lab'17Immersive Recommendation Workshop, NYC Media Lab'17
Immersive Recommendation Workshop, NYC Media Lab'17
 
Machine Support for Interacting with Scientific Publications Improving Inform...
Machine Support for Interacting with Scientific Publications Improving Inform...Machine Support for Interacting with Scientific Publications Improving Inform...
Machine Support for Interacting with Scientific Publications Improving Inform...
 
The culture of researchData
The culture of researchDataThe culture of researchData
The culture of researchData
 
Developing and sharing tools for bioelectromagnetic research
Developing and sharing tools for bioelectromagnetic researchDeveloping and sharing tools for bioelectromagnetic research
Developing and sharing tools for bioelectromagnetic research
 
The culture of researchData
The culture of researchData The culture of researchData
The culture of researchData
 
The Culture of Research Data, by Peter Murray-Rust
The Culture of Research Data, by Peter Murray-RustThe Culture of Research Data, by Peter Murray-Rust
The Culture of Research Data, by Peter Murray-Rust
 
Machine Learning Pitfalls
Machine Learning Pitfalls Machine Learning Pitfalls
Machine Learning Pitfalls
 
The Dark Art: Is Music Recommendation Science a Science
The Dark Art: Is Music Recommendation Science a ScienceThe Dark Art: Is Music Recommendation Science a Science
The Dark Art: Is Music Recommendation Science a Science
 
Mediated participatory design for contextually aware in vehicle experiences
Mediated participatory design for contextually aware in vehicle experiencesMediated participatory design for contextually aware in vehicle experiences
Mediated participatory design for contextually aware in vehicle experiences
 
Using Bibliometrics to Keep Up with the Joneses
Using Bibliometrics to Keep Up with the JonesesUsing Bibliometrics to Keep Up with the Joneses
Using Bibliometrics to Keep Up with the Joneses
 
Noteworks SI 622 Presentation Slides
Noteworks SI 622 Presentation SlidesNoteworks SI 622 Presentation Slides
Noteworks SI 622 Presentation Slides
 
Sound is not speech
Sound is not speechSound is not speech
Sound is not speech
 
SoundSoftware: Software Sustainability for audio and Music Researchers
SoundSoftware: Software Sustainability for audio and Music Researchers SoundSoftware: Software Sustainability for audio and Music Researchers
SoundSoftware: Software Sustainability for audio and Music Researchers
 
Session 1 Lecture 2 PACT A Framework for Designing Interactive Systems
Session 1 Lecture 2 PACT A Framework for Designing Interactive SystemsSession 1 Lecture 2 PACT A Framework for Designing Interactive Systems
Session 1 Lecture 2 PACT A Framework for Designing Interactive Systems
 
Potentials and limitations of ‘Automated Sentiment Analysis
Potentials and limitations of ‘Automated Sentiment AnalysisPotentials and limitations of ‘Automated Sentiment Analysis
Potentials and limitations of ‘Automated Sentiment Analysis
 
Engaging with DARPA - Stefanie Thompkins
Engaging with DARPA - Stefanie ThompkinsEngaging with DARPA - Stefanie Thompkins
Engaging with DARPA - Stefanie Thompkins
 
Berlin 6 Open Access Conference: Mark Liberman
Berlin 6 Open Access Conference: Mark LibermanBerlin 6 Open Access Conference: Mark Liberman
Berlin 6 Open Access Conference: Mark Liberman
 
Thompson_Aaron_MP
Thompson_Aaron_MPThompson_Aaron_MP
Thompson_Aaron_MP
 

More from sebastianewert

British library radio collections
British library radio collectionsBritish library radio collections
British library radio collectionssebastianewert
 
Large-Scale Capture of Producer-Defined Musical Semantics - Ryan Stables (Sem...
Large-Scale Capture of Producer-Defined Musical Semantics - Ryan Stables (Sem...Large-Scale Capture of Producer-Defined Musical Semantics - Ryan Stables (Sem...
Large-Scale Capture of Producer-Defined Musical Semantics - Ryan Stables (Sem...
sebastianewert
 
Exploring the British Library's audio collections - Richard Ranft (Semantic M...
Exploring the British Library's audio collections - Richard Ranft (Semantic M...Exploring the British Library's audio collections - Richard Ranft (Semantic M...
Exploring the British Library's audio collections - Richard Ranft (Semantic M...
sebastianewert
 
Semantic Linking of Information, Content and Metadata for Early Music (SLICKM...
Semantic Linking of Information, Content and Metadata for Early Music (SLICKM...Semantic Linking of Information, Content and Metadata for Early Music (SLICKM...
Semantic Linking of Information, Content and Metadata for Early Music (SLICKM...
sebastianewert
 
News collections at the British Library - Luke McKernan (Semantic Media @ The...
News collections at the British Library - Luke McKernan (Semantic Media @ The...News collections at the British Library - Luke McKernan (Semantic Media @ The...
News collections at the British Library - Luke McKernan (Semantic Media @ The...
sebastianewert
 
Mood-based Classification of TV Programmes - Jana Eggink, Sam Davies, Denise...
Mood-based Classification of TV Programmes - Jana Eggink, Sam Davies, Denise...Mood-based Classification of TV Programmes - Jana Eggink, Sam Davies, Denise...
Mood-based Classification of TV Programmes - Jana Eggink, Sam Davies, Denise...
sebastianewert
 
BBC R&D Datasets - Jana Eggink (Semantic Media @ BBC, Feb 2013)
BBC R&D Datasets - Jana Eggink (Semantic Media @ BBC, Feb 2013)BBC R&D Datasets - Jana Eggink (Semantic Media @ BBC, Feb 2013)
BBC R&D Datasets - Jana Eggink (Semantic Media @ BBC, Feb 2013)
sebastianewert
 
Semantic Media Project Introduction - Mark Sandler (Barbican Arts Centre, Oct...
Semantic Media Project Introduction - Mark Sandler (Barbican Arts Centre, Oct...Semantic Media Project Introduction - Mark Sandler (Barbican Arts Centre, Oct...
Semantic Media Project Introduction - Mark Sandler (Barbican Arts Centre, Oct...
sebastianewert
 

More from sebastianewert (8)

British library radio collections
British library radio collectionsBritish library radio collections
British library radio collections
 
Large-Scale Capture of Producer-Defined Musical Semantics - Ryan Stables (Sem...
Large-Scale Capture of Producer-Defined Musical Semantics - Ryan Stables (Sem...Large-Scale Capture of Producer-Defined Musical Semantics - Ryan Stables (Sem...
Large-Scale Capture of Producer-Defined Musical Semantics - Ryan Stables (Sem...
 
Exploring the British Library's audio collections - Richard Ranft (Semantic M...
Exploring the British Library's audio collections - Richard Ranft (Semantic M...Exploring the British Library's audio collections - Richard Ranft (Semantic M...
Exploring the British Library's audio collections - Richard Ranft (Semantic M...
 
Semantic Linking of Information, Content and Metadata for Early Music (SLICKM...
Semantic Linking of Information, Content and Metadata for Early Music (SLICKM...Semantic Linking of Information, Content and Metadata for Early Music (SLICKM...
Semantic Linking of Information, Content and Metadata for Early Music (SLICKM...
 
News collections at the British Library - Luke McKernan (Semantic Media @ The...
News collections at the British Library - Luke McKernan (Semantic Media @ The...News collections at the British Library - Luke McKernan (Semantic Media @ The...
News collections at the British Library - Luke McKernan (Semantic Media @ The...
 
Mood-based Classification of TV Programmes - Jana Eggink, Sam Davies, Denise...
Mood-based Classification of TV Programmes - Jana Eggink, Sam Davies, Denise...Mood-based Classification of TV Programmes - Jana Eggink, Sam Davies, Denise...
Mood-based Classification of TV Programmes - Jana Eggink, Sam Davies, Denise...
 
BBC R&D Datasets - Jana Eggink (Semantic Media @ BBC, Feb 2013)
BBC R&D Datasets - Jana Eggink (Semantic Media @ BBC, Feb 2013)BBC R&D Datasets - Jana Eggink (Semantic Media @ BBC, Feb 2013)
BBC R&D Datasets - Jana Eggink (Semantic Media @ BBC, Feb 2013)
 
Semantic Media Project Introduction - Mark Sandler (Barbican Arts Centre, Oct...
Semantic Media Project Introduction - Mark Sandler (Barbican Arts Centre, Oct...Semantic Media Project Introduction - Mark Sandler (Barbican Arts Centre, Oct...
Semantic Media Project Introduction - Mark Sandler (Barbican Arts Centre, Oct...
 

Recently uploaded

DERIVATION OF MODIFIED BERNOULLI EQUATION WITH VISCOUS EFFECTS AND TERMINAL V...
DERIVATION OF MODIFIED BERNOULLI EQUATION WITH VISCOUS EFFECTS AND TERMINAL V...DERIVATION OF MODIFIED BERNOULLI EQUATION WITH VISCOUS EFFECTS AND TERMINAL V...
DERIVATION OF MODIFIED BERNOULLI EQUATION WITH VISCOUS EFFECTS AND TERMINAL V...
Wasswaderrick3
 
原版制作(carleton毕业证书)卡尔顿大学毕业证硕士文凭原版一模一样
原版制作(carleton毕业证书)卡尔顿大学毕业证硕士文凭原版一模一样原版制作(carleton毕业证书)卡尔顿大学毕业证硕士文凭原版一模一样
原版制作(carleton毕业证书)卡尔顿大学毕业证硕士文凭原版一模一样
yqqaatn0
 
Nutraceutical market, scope and growth: Herbal drug technology
Nutraceutical market, scope and growth: Herbal drug technologyNutraceutical market, scope and growth: Herbal drug technology
Nutraceutical market, scope and growth: Herbal drug technology
Lokesh Patil
 
NuGOweek 2024 Ghent - programme - final version
NuGOweek 2024 Ghent - programme - final versionNuGOweek 2024 Ghent - programme - final version
NuGOweek 2024 Ghent - programme - final version
pablovgd
 
BLOOD AND BLOOD COMPONENT- introduction to blood physiology
BLOOD AND BLOOD COMPONENT- introduction to blood physiologyBLOOD AND BLOOD COMPONENT- introduction to blood physiology
BLOOD AND BLOOD COMPONENT- introduction to blood physiology
NoelManyise1
 
Lateral Ventricles.pdf very easy good diagrams comprehensive
Lateral Ventricles.pdf very easy good diagrams comprehensiveLateral Ventricles.pdf very easy good diagrams comprehensive
Lateral Ventricles.pdf very easy good diagrams comprehensive
silvermistyshot
 
Comparing Evolved Extractive Text Summary Scores of Bidirectional Encoder Rep...
Comparing Evolved Extractive Text Summary Scores of Bidirectional Encoder Rep...Comparing Evolved Extractive Text Summary Scores of Bidirectional Encoder Rep...
Comparing Evolved Extractive Text Summary Scores of Bidirectional Encoder Rep...
University of Maribor
 
DMARDs Pharmacolgy Pharm D 5th Semester.pdf
DMARDs Pharmacolgy Pharm D 5th Semester.pdfDMARDs Pharmacolgy Pharm D 5th Semester.pdf
DMARDs Pharmacolgy Pharm D 5th Semester.pdf
fafyfskhan251kmf
 
Leaf Initiation, Growth and Differentiation.pdf
Leaf Initiation, Growth and Differentiation.pdfLeaf Initiation, Growth and Differentiation.pdf
Leaf Initiation, Growth and Differentiation.pdf
RenuJangid3
 
Deep Software Variability and Frictionless Reproducibility
Deep Software Variability and Frictionless ReproducibilityDeep Software Variability and Frictionless Reproducibility
Deep Software Variability and Frictionless Reproducibility
University of Rennes, INSA Rennes, Inria/IRISA, CNRS
 
extra-chromosomal-inheritance[1].pptx.pdfpdf
extra-chromosomal-inheritance[1].pptx.pdfpdfextra-chromosomal-inheritance[1].pptx.pdfpdf
extra-chromosomal-inheritance[1].pptx.pdfpdf
DiyaBiswas10
 
GBSN- Microbiology (Lab 3) Gram Staining
GBSN- Microbiology (Lab 3) Gram StainingGBSN- Microbiology (Lab 3) Gram Staining
GBSN- Microbiology (Lab 3) Gram Staining
Areesha Ahmad
 
(May 29th, 2024) Advancements in Intravital Microscopy- Insights for Preclini...
(May 29th, 2024) Advancements in Intravital Microscopy- Insights for Preclini...(May 29th, 2024) Advancements in Intravital Microscopy- Insights for Preclini...
(May 29th, 2024) Advancements in Intravital Microscopy- Insights for Preclini...
Scintica Instrumentation
 
如何办理(uvic毕业证书)维多利亚大学毕业证本科学位证书原版一模一样
如何办理(uvic毕业证书)维多利亚大学毕业证本科学位证书原版一模一样如何办理(uvic毕业证书)维多利亚大学毕业证本科学位证书原版一模一样
如何办理(uvic毕业证书)维多利亚大学毕业证本科学位证书原版一模一样
yqqaatn0
 
Unveiling the Energy Potential of Marshmallow Deposits.pdf
Unveiling the Energy Potential of Marshmallow Deposits.pdfUnveiling the Energy Potential of Marshmallow Deposits.pdf
Unveiling the Energy Potential of Marshmallow Deposits.pdf
Erdal Coalmaker
 
Hemostasis_importance& clinical significance.pptx
Hemostasis_importance& clinical significance.pptxHemostasis_importance& clinical significance.pptx
Hemostasis_importance& clinical significance.pptx
muralinath2
 
What is greenhouse gasses and how many gasses are there to affect the Earth.
What is greenhouse gasses and how many gasses are there to affect the Earth.What is greenhouse gasses and how many gasses are there to affect the Earth.
What is greenhouse gasses and how many gasses are there to affect the Earth.
moosaasad1975
 
Hemoglobin metabolism_pathophysiology.pptx
Hemoglobin metabolism_pathophysiology.pptxHemoglobin metabolism_pathophysiology.pptx
Hemoglobin metabolism_pathophysiology.pptx
muralinath2
 
Observation of Io’s Resurfacing via Plume Deposition Using Ground-based Adapt...
Observation of Io’s Resurfacing via Plume Deposition Using Ground-based Adapt...Observation of Io’s Resurfacing via Plume Deposition Using Ground-based Adapt...
Observation of Io’s Resurfacing via Plume Deposition Using Ground-based Adapt...
Sérgio Sacani
 
erythropoiesis-I_mechanism& clinical significance.pptx
erythropoiesis-I_mechanism& clinical significance.pptxerythropoiesis-I_mechanism& clinical significance.pptx
erythropoiesis-I_mechanism& clinical significance.pptx
muralinath2
 

Recently uploaded (20)

DERIVATION OF MODIFIED BERNOULLI EQUATION WITH VISCOUS EFFECTS AND TERMINAL V...
DERIVATION OF MODIFIED BERNOULLI EQUATION WITH VISCOUS EFFECTS AND TERMINAL V...DERIVATION OF MODIFIED BERNOULLI EQUATION WITH VISCOUS EFFECTS AND TERMINAL V...
DERIVATION OF MODIFIED BERNOULLI EQUATION WITH VISCOUS EFFECTS AND TERMINAL V...
 
原版制作(carleton毕业证书)卡尔顿大学毕业证硕士文凭原版一模一样
原版制作(carleton毕业证书)卡尔顿大学毕业证硕士文凭原版一模一样原版制作(carleton毕业证书)卡尔顿大学毕业证硕士文凭原版一模一样
原版制作(carleton毕业证书)卡尔顿大学毕业证硕士文凭原版一模一样
 
Nutraceutical market, scope and growth: Herbal drug technology
Nutraceutical market, scope and growth: Herbal drug technologyNutraceutical market, scope and growth: Herbal drug technology
Nutraceutical market, scope and growth: Herbal drug technology
 
NuGOweek 2024 Ghent - programme - final version
NuGOweek 2024 Ghent - programme - final versionNuGOweek 2024 Ghent - programme - final version
NuGOweek 2024 Ghent - programme - final version
 
BLOOD AND BLOOD COMPONENT- introduction to blood physiology
BLOOD AND BLOOD COMPONENT- introduction to blood physiologyBLOOD AND BLOOD COMPONENT- introduction to blood physiology
BLOOD AND BLOOD COMPONENT- introduction to blood physiology
 
Lateral Ventricles.pdf very easy good diagrams comprehensive
Lateral Ventricles.pdf very easy good diagrams comprehensiveLateral Ventricles.pdf very easy good diagrams comprehensive
Lateral Ventricles.pdf very easy good diagrams comprehensive
 
Comparing Evolved Extractive Text Summary Scores of Bidirectional Encoder Rep...
Comparing Evolved Extractive Text Summary Scores of Bidirectional Encoder Rep...Comparing Evolved Extractive Text Summary Scores of Bidirectional Encoder Rep...
Comparing Evolved Extractive Text Summary Scores of Bidirectional Encoder Rep...
 
DMARDs Pharmacolgy Pharm D 5th Semester.pdf
DMARDs Pharmacolgy Pharm D 5th Semester.pdfDMARDs Pharmacolgy Pharm D 5th Semester.pdf
DMARDs Pharmacolgy Pharm D 5th Semester.pdf
 
Leaf Initiation, Growth and Differentiation.pdf
Leaf Initiation, Growth and Differentiation.pdfLeaf Initiation, Growth and Differentiation.pdf
Leaf Initiation, Growth and Differentiation.pdf
 
Deep Software Variability and Frictionless Reproducibility
Deep Software Variability and Frictionless ReproducibilityDeep Software Variability and Frictionless Reproducibility
Deep Software Variability and Frictionless Reproducibility
 
extra-chromosomal-inheritance[1].pptx.pdfpdf
extra-chromosomal-inheritance[1].pptx.pdfpdfextra-chromosomal-inheritance[1].pptx.pdfpdf
extra-chromosomal-inheritance[1].pptx.pdfpdf
 
GBSN- Microbiology (Lab 3) Gram Staining
GBSN- Microbiology (Lab 3) Gram StainingGBSN- Microbiology (Lab 3) Gram Staining
GBSN- Microbiology (Lab 3) Gram Staining
 
(May 29th, 2024) Advancements in Intravital Microscopy- Insights for Preclini...
(May 29th, 2024) Advancements in Intravital Microscopy- Insights for Preclini...(May 29th, 2024) Advancements in Intravital Microscopy- Insights for Preclini...
(May 29th, 2024) Advancements in Intravital Microscopy- Insights for Preclini...
 
如何办理(uvic毕业证书)维多利亚大学毕业证本科学位证书原版一模一样
如何办理(uvic毕业证书)维多利亚大学毕业证本科学位证书原版一模一样如何办理(uvic毕业证书)维多利亚大学毕业证本科学位证书原版一模一样
如何办理(uvic毕业证书)维多利亚大学毕业证本科学位证书原版一模一样
 
Unveiling the Energy Potential of Marshmallow Deposits.pdf
Unveiling the Energy Potential of Marshmallow Deposits.pdfUnveiling the Energy Potential of Marshmallow Deposits.pdf
Unveiling the Energy Potential of Marshmallow Deposits.pdf
 
Hemostasis_importance& clinical significance.pptx
Hemostasis_importance& clinical significance.pptxHemostasis_importance& clinical significance.pptx
Hemostasis_importance& clinical significance.pptx
 
What is greenhouse gasses and how many gasses are there to affect the Earth.
What is greenhouse gasses and how many gasses are there to affect the Earth.What is greenhouse gasses and how many gasses are there to affect the Earth.
What is greenhouse gasses and how many gasses are there to affect the Earth.
 
Hemoglobin metabolism_pathophysiology.pptx
Hemoglobin metabolism_pathophysiology.pptxHemoglobin metabolism_pathophysiology.pptx
Hemoglobin metabolism_pathophysiology.pptx
 
Observation of Io’s Resurfacing via Plume Deposition Using Ground-based Adapt...
Observation of Io’s Resurfacing via Plume Deposition Using Ground-based Adapt...Observation of Io’s Resurfacing via Plume Deposition Using Ground-based Adapt...
Observation of Io’s Resurfacing via Plume Deposition Using Ground-based Adapt...
 
erythropoiesis-I_mechanism& clinical significance.pptx
erythropoiesis-I_mechanism& clinical significance.pptxerythropoiesis-I_mechanism& clinical significance.pptx
erythropoiesis-I_mechanism& clinical significance.pptx
 

C4DM Seminar 2016-07-12: Brecht De Man

  • 1. 1 Introduction Tools Design principles Open questions Listening Test Interface Design Brecht De Man Correspondence: <b.deman@qmul.ac.uk> Centre for Digital Music School of Electronic Engineering and Computer Science Queen Mary, University of London July 12, 2016 @BrechtDeMan
  • 2. 1 Introduction Tools Design principles Open questions In the news Bug in fMRI software calls 15 years of research into question (Wired UK) Software faults raise questions about the validity of brain studies (Ars Technica) Source: iStock “In theory, we should find 5% false positives (for a significance threshold of 5%), but instead we found that the most common software packages for fMRI analysis can result in false-positive rates of up to 70%. These results question the validity of some 40,000 fMRI studies and may have a large impact on the interpretation of neuroimaging results.” [1] A. Eklund, T. E. Nichols, and H. Knutsson, “Cluster failure: Why fMRI inferences for spatial extent have inflated false-positive rates,” Proceedings of the National Academy of Sciences, 2016. @BrechtDeMan Listening Test Interface Design July 12, 2016 1 / 35
  • 3. 1 Introduction Tools Design principles Open questions In the news Bug in fMRI software calls 15 years of research into question (Wired UK) Software faults raise questions about the validity of brain studies (Ars Technica) “In theory, we should find 5% false positives (for a significance threshold of 5%), but instead we found that the most common software packages for fMRI analysis can result in false-positive rates of up to 70%. These results question the validity of some 40,000 fMRI studies and may have a large impact on the interpretation of neuroimaging results.” [1] A. Eklund, T. E. Nichols, and H. Knutsson, “Cluster failure: Why fMRI inferences for spatial extent have inflated false-positive rates,” Proceedings of the National Academy of Sciences, 2016. [2] Bennett et al., “Neural Correlates of Interspecies Perspective Taking in the Post-Mortem Atlantic Salmon: An Argument For Proper Multiple Comparisons Correction,” Journal of Serendipitous and Unexpected Results, 2010. @BrechtDeMan Listening Test Interface Design July 12, 2016 1 / 35
  • 4. 1 Introduction Tools Design principles Open questions Applications I Audio codecs I Electroacoustic design (loudspeakers, headphones, ...) I Human perception I Standards I ... @BrechtDeMan Listening Test Interface Design July 12, 2016 2 / 35
  • 5. 1 Introduction Tools Design principles Open questions Perceptual evaluation I Interface design I Subject selection I Subject training I Stimulus selection I Listening environment I Playback system I Subject exclusion I Research questions I ... @BrechtDeMan Listening Test Interface Design July 12, 2016 3 / 35
  • 6. 1 Introduction Tools Design principles Open questions Goals I Accurate, reproducible, representative I High discrimination I Low time and e↵ort @BrechtDeMan Listening Test Interface Design July 12, 2016 4 / 35
  • 7. 1 Introduction Tools Design principles Open questions Basic reading [3] Søren Bech and Nick Zacharov, “Perceptual Audio Evaluation - Theory, Method and Application,” Wiley, 2006. @BrechtDeMan Listening Test Interface Design July 12, 2016 5 / 35
  • 8. 1 Introduction Tools Design principles Open questions Table of Contents 1 Introduction 2 Tools 3 Design principles 4 Open questions @BrechtDeMan Listening Test Interface Design July 12, 2016
  • 9. 1 Introduction Tools Design principles Open questions Table of Contents 1 Introduction 2 Tools 3 Design principles 4 Open questions @BrechtDeMan Listening Test Interface Design July 12, 2016
  • 10. 1 Introduction Tools Design principles Open questions Available listening test tools Name Supported interfaces Language APE multi-stimulus multi-axis, pairwise MATLAB BeaqleJS MUSHRA + ABX JavaScript HULTI-GEN many Max MUSHRAM MUSHRA MATLAB Scale semantic di↵erential, n-AFC MATLAB WhisPER many (7) MATLAB @BrechtDeMan Listening Test Interface Design July 12, 2016 6 / 35
  • 11. 1 Introduction Tools Design principles Open questions APE I MATLAB-based I Pairwise (AB) mode I Multiple stimulus mode • All stimuli on one axis • Multiple axes for multiple attributes I Made available to public • Others can spot bugs and make improvements • code.soundsoftware.ac.uk/projects/APE • AES 136th Convention [4] @BrechtDeMan Listening Test Interface Design July 12, 2016 7 / 35
  • 12. 1 Introduction Tools Design principles Open questions Web Audio Evaluation Tool I Browser-based: OS-independent, no 3rd party software I Wide range of highly customisable interfaces I Remote or local tests I Easy quick test setup: no programming required I High compatibility: configuration and results in XML I Immediate diagnostics and result in browser I Intuitive, flexible, works out of the box @BrechtDeMan Listening Test Interface Design July 12, 2016 8 / 35
  • 13. 1 Introduction Tools Design principles Open questions Web Audio Evaluation Tool Developers I Nicholas Jillings Web Audio I Brecht De Man perception of music production I David Mo↵at realism of sound synthesis Contributors I Giulio Moro Hammond organs I Adib Mehrabi vocal imitation of musical sounds I Alex Wilson audio quality in music I Alejandro Osses Vecchi psychoacoustics @BrechtDeMan Listening Test Interface Design July 12, 2016 9 / 35
  • 14. 1 Introduction Tools Design principles Open questions Table of Contents 1 Introduction 2 Tools 3 Design principles 4 Open questions @BrechtDeMan Listening Test Interface Design July 12, 2016
  • 15. 1 Introduction Tools Design principles Open questions Double blind No blind: @BrechtDeMan Listening Test Interface Design July 12, 2016 10 / 35
  • 16. 1 Introduction Tools Design principles Open questions Double blind Single blind: @BrechtDeMan Listening Test Interface Design July 12, 2016 10 / 35
  • 17. 1 Introduction Tools Design principles Open questions Double blind Double blind: @BrechtDeMan Listening Test Interface Design July 12, 2016 10 / 35
  • 18. 1 Introduction Tools Design principles Open questions Double blind [5] F. E. Toole and S. Olive, “Hearing is believing vs. believing is hearing: Blind vs. sighted listening tests, and other interesting things,” in Audio Engineering Society Convention 97, November 1994. @BrechtDeMan Listening Test Interface Design July 12, 2016 10 / 35
  • 19. 1 Introduction Tools Design principles Open questions Individual tests @BrechtDeMan Listening Test Interface Design July 12, 2016 11 / 35
  • 20. 1 Introduction Tools Design principles Open questions Randomisation Remove bias due to order of I (Presentation and labelling of) stimuli I Parts of the experiment I Tests I ... @BrechtDeMan Listening Test Interface Design July 12, 2016 12 / 35
  • 21. 1 Introduction Tools Design principles Open questions Free switching “Increases ability to perceive more delicate di↵erences” [6] J. Berg and F. Rumsey, “Spatial attribute identification and scaling by repertory grid technique and other methods,” 16th AES Conference: Spatial Sound Reproduction, March 1999. @BrechtDeMan Listening Test Interface Design July 12, 2016 13 / 35
  • 22. 1 Introduction Tools Design principles Open questions Time-aligned stimuli I Switch to corresponding position in next stimulus I More balanced over full length of stimuli @BrechtDeMan Listening Test Interface Design July 12, 2016 14 / 35
  • 23. 1 Introduction Tools Design principles Open questions Reduce visual distractions I “Visual information processing [induces] Inattentional Deafness” I “Requiring use of scroll bar: significant negative e↵ect on critical listening reaction times” [7] Joshua Mycroft, Joshua D. Reiss and Tony Stockman, “The Influence of Graphical User Interface Design on Critical Listening Skills,” Sound and Music Computing (SMC), July 2013. @BrechtDeMan Listening Test Interface Design July 12, 2016 15 / 35
  • 24. 1 Introduction Tools Design principles Open questions Automated, computerised tests I Double blind I Individual tests I Randomisation I Free switching I Time-aligned stimuli I Reduce visual distractions @BrechtDeMan Listening Test Interface Design July 12, 2016 16 / 35
  • 25. 1 Introduction Tools Design principles Open questions Automated, computerised tests [8] Helmut Haas, “The influence of a single echo on the audibility of speech,” JAES Vol. 20 No. 2, pp. 146-159, 1972. @BrechtDeMan Listening Test Interface Design July 12, 2016 16 / 35
  • 26. 1 Introduction Tools Design principles Open questions Automated, computerised tests Dr Sean Olive (HARMAN) I Pencil and paper I HyperCard (late 1980s) @BrechtDeMan Listening Test Interface Design July 12, 2016 16 / 35
  • 27. 1 Introduction Tools Design principles Open questions Automated, computerised tests Dr Sean Olive (HARMAN) I Pencil and paper I HyperCard (late 1980s) I Pocket PC / PDA (early 2000s) @BrechtDeMan Listening Test Interface Design July 12, 2016 16 / 35
  • 28. 1 Introduction Tools Design principles Open questions Automated, computerised tests Dr Sean Olive (HARMAN) I Pencil and paper I HyperCard (late 1980s) I Pocket PC / PDA (early 2000s) I Apple iPad @BrechtDeMan Listening Test Interface Design July 12, 2016 16 / 35
  • 29. 1 Introduction Tools Design principles Open questions Pilot test and dummy analysis Source: Muppet Show Wikia I Feedback on interface and stimuli I Know what type of data to expect @BrechtDeMan Listening Test Interface Design July 12, 2016 17 / 35
  • 30. 1 Introduction Tools Design principles Open questions Language Native tongue: “Minimise (risk of) semantic constraints on subjects” [9] J. Berg and F. Rumsey, “Correlation between emotive, descriptive and naturalness attributes in subjective data relating to spatial sound reproduction,” AES Convention 109, September 2000. @BrechtDeMan Listening Test Interface Design July 12, 2016 18 / 35
  • 31. 1 Introduction Tools Design principles Open questions Safety I Don’t allow accidental closing of window I Intermediate session saves @BrechtDeMan Listening Test Interface Design July 12, 2016 19 / 35
  • 32. 1 Introduction Tools Design principles Open questions Reward I Immediately display all / own results I Candy @BrechtDeMan Listening Test Interface Design July 12, 2016 20 / 35
  • 33. 1 Introduction Tools Design principles Open questions Diagnostics I Spot errors early I Exclude subjects I Insights to better design tests I Statistics for experiment report @BrechtDeMan Listening Test Interface Design July 12, 2016 21 / 35
  • 34. 1 Introduction Tools Design principles Open questions Surveys I Immediately before/after test/page • Avoid chasing subjects afterwards • Some questions might relate to the test @BrechtDeMan Listening Test Interface Design July 12, 2016 22 / 35
  • 35. 1 Introduction Tools Design principles Open questions Comments I Learn more than simple, single-attribute ratings I Understand how subjects used the scale I “People get frustrated when they don’t have comment boxes” (Sean Olive, AES140 Workshop) I Separate comment boxes (one for each stimulus) instead of one • Proportion of stimuli commented on: 96.5% instead of 82.1% • Comments 47% longer [4] Brecht De Man and Joshua D. Reiss, “APE: Audio perceptual evaluation toolbox for MATLAB,” 136th Convention of the Audio Engineering Society, April 2014. @BrechtDeMan Listening Test Interface Design July 12, 2016 23 / 35
  • 36. 1 Introduction Tools Design principles Open questions Axis ticks [10] Jouni Paulus, Christian Uhle and J¨urgen Herre, “Perceived Level of Late Reverberation in Speech and Music,” AES Convention 130, 2011. @BrechtDeMan Listening Test Interface Design July 12, 2016 24 / 35
  • 37. 1 Introduction Tools Design principles Open questions Table of Contents 1 Introduction 2 Tools 3 Design principles 4 Open questions @BrechtDeMan Listening Test Interface Design July 12, 2016
  • 38. 1 Introduction Tools Design principles Open questions Calibration I Audiometric test I Test of playback system @BrechtDeMan Listening Test Interface Design July 12, 2016 25 / 35
  • 39. 1 Introduction Tools Design principles Open questions Multiple stimulus or pairwise? “Multiple comparison methods produce more reliable and discriminating judgments of sound quality compared to single stimulus or paired comparison methods.” [11] S. Olive and T. Welti, “The relationship between perception and measurement of headphone sound quality,” in Audio Engineering Society Convention 133, October 2012. “Multiple stimulus tests are substantially less time-consuming.” [12] B. De Man and J. D. Reiss, “A pairwise and multiple stimuli approach to perceptual evaluation of microphone types,” in Audio Engineering Society Convention 134, May 2013. Multiple stimulus only viable for limited number of stimuli to be compared against each other. @BrechtDeMan Listening Test Interface Design July 12, 2016 26 / 35
  • 40. 1 Introduction Tools Design principles Open questions Listening level Options I Everyone at same level I Choose most comfortable/familiar level • Set once at start • Change throughout test @BrechtDeMan Listening Test Interface Design July 12, 2016 27 / 35
  • 41. 1 Introduction Tools Design principles Open questions To MUSHRA or not to MUSHRA? [13] Method for the subjective assessment of intermediate quality level of coding systems. Recommendation ITU-R BS.1534-1, 2003. [14] E. Vincent, M. G. Jafari, and M. D. Plumbley, “Preliminary guidelines for subjective evalutation of audio source separation algorithms,” in UK ICA Research Network Workshop, 2006. @BrechtDeMan Listening Test Interface Design July 12, 2016 28 / 35
  • 42. 1 Introduction Tools Design principles Open questions To MUSHRA or not to MUSHRA? I “[MUSHRA aims] to establish the detectability, magnitude of impairment or basic audio quality of the codec relative to a reference. [It does] not attempt to measure the listeners’ preference for the audio codec when compared to the hidden lossless reference.” [15] Sean E. Olive, “Some new evidence that teenagers and college students may prefer accurate sound reproduction,” in Audio Engineering Society Convention 132, April 2012. I What if there is no reference to compare to? @BrechtDeMan Listening Test Interface Design July 12, 2016 28 / 35
  • 43. 1 Introduction Tools Design principles Open questions To MUSHRA or not to MUSHRA? “More accessible, particularly to technically experienced subjects” [16] Christoph V¨olker and Rainer Huber, “Adaptations for the MUlti Stimulus test with Hidden Reference and Anchor (MUSHRA) for elder and technical unexperienced participants,” DAGA, 2015. @BrechtDeMan Listening Test Interface Design July 12, 2016 28 / 35
  • 44. 1 Introduction Tools Design principles Open questions To MUSHRA or not to MUSHRA? [4] Brecht De Man and Joshua D. Reiss, “APE: Audio perceptual evaluation toolbox for MATLAB,” 136th Convention of the Audio Engineering Society, April 2014. @BrechtDeMan Listening Test Interface Design July 12, 2016 28 / 35
  • 45. 1 Introduction Tools Design principles Open questions To MUSHRA or not to MUSHRA? [17] Nicholas Jillings, Brecht De Man, David Mo↵at and Joshua D. Reiss, “Web Audio Evaluation Tool: A browser-based listening test environment,” 12th Sound and Music Computing Conference, July 2015. @BrechtDeMan Listening Test Interface Design July 12, 2016 28 / 35
  • 46. 1 Introduction Tools Design principles Open questions To MUSHRA or not to MUSHRA? [18] Jan Berg, “OPAQUE - A tool for the elicitation and grading of audio quality attributes,” AES Convention 118, 2005. @BrechtDeMan Listening Test Interface Design July 12, 2016 28 / 35
  • 47. 1 Introduction Tools Design principles Open questions Constraints on use of scale Examples I “Put at least one slider at 100%” I “The anchor needs to be below 20%” I ... Options I Instructions I Enforced (not possible to proceed) I Remove subject from test I Normalise ratings after the fact @BrechtDeMan Listening Test Interface Design July 12, 2016 29 / 35
  • 48. 1 Introduction Tools Design principles Open questions Remote tests I Pro: Low e↵ort, scales easily, di↵erent locations/cultures/languages within reach I Con: Loss of control ... or is there? [19] M. Cartwright, B. Pardo, G. Mysore and M. Ho↵man “Fast and Easy Crowdsourced Perceptual Audio Evaluation,” ICASSP, 2016. [20] M. Schoe✏er, F.-R. St¨oter, H. Bayerlein, B. Edler and J. Herre, “An Experiment about Estimating the Number of Instruments in Polyphonic Music: A Comparison Between Internet and Laboratory Results,” ISMIR, 2013. [21] F.-R. St¨oter, M. Schoe✏er, B. Edler and J. Herre, “Human Ability of Counting the Number of Instruments in Polyphonic Music,” Meetings on Acoustics Vol. 19, 2013. @BrechtDeMan Listening Test Interface Design July 12, 2016 30 / 35
  • 49. 1 Introduction Tools Design principles Open questions Online tests Why online? I Cross-platform I Centralised results collection (over web or local server) I Multiple machines at once I Leveraging other web technologies @BrechtDeMan Listening Test Interface Design July 12, 2016 31 / 35
  • 50. 1 Introduction Tools Design principles Open questions Conclusion Try it! I Conduct an experiment I Contribute (github.com/BrechtDeMan/WebAudioEvaluationTool) I Any problems? Feedback? • Report issue • b.deman@qmul.ac.uk @BrechtDeMan Listening Test Interface Design July 12, 2016 32 / 35
  • 51. WIMP References Q&A AES 2nd Workshop on Intelligent Music Production I Tuesday 13 September 2016 I Charterhouse Square, Queen Mary University of London I Short (<2 pages) paper submission deadline: 1 August 2016 I Confirmed speakers: • Bryan Pardo (Northwestern University) • Fran¸cois Pachet (Sony CSL Music Paris) I More info and registration: aes-uk.org/wimp I Volunteers needed @BrechtDeMan Listening Test Interface Design July 12, 2016 33 / 35
  • 52. WIMP References Q&A References A. Eklund, T. E. Nichols, and H. Knutsson, “Cluster failure: Why fMRI inferences for spatial extent have inflated false-positive rates,” Proceedings of the National Academy of Sciences, 2016. B. et al., “Neural correlates of interspecies perspective taking in the post-mortem atlantic salmon: An argument for proper multiple comparisons correction,” Journal of Serendipitous and Unexpected Results, 2010. S. Bech and N. Zacharov, Perceptual Audio Evaluation - Theory, Method and Application. John Wiley & Sons, 2007. B. De Man and J. D. Reiss, “APE: Audio Perceptual Evaluation toolbox for MATLAB,” in Audio Engineering Society Convention 136, April 2014. F. E. Toole and S. Olive, “Hearing is believing vs. believing is hearing: Blind vs. sighted listening tests, and other interesting things,” in Audio Engineering Society Convention 97, November 1994. J. Berg and F. Rumsey, “Spatial attribute identification and scaling by repertory grid technique and other methods,” in Audio Engineering Society Conference: 16th International Conference: Spatial Sound Reproduction, March 1999. J. Mycroft, J. D. Reiss, and T. Stockman, “The influence of graphical user interface design on critical listening skills,” Sound and Music Computing (SMC), Stockholm, 2013. H. Haas, “The influence of a single echo on the audibility of speech,” J. Audio Eng. Soc, vol. 20, no. 2, pp. 146–159, 1972. J. Berg and F. Rumsey, “Correlation between emotive, descriptive and naturalness attributes in subjective data relating to spatial sound reproduction,” in Audio Engineering Society Convention 109, September 2000. J. Paulus, C. Uhle, and J. Herre, “Perceived level of late reverberation in speech and music,” in Audio Engineering Society Convention 130, May 2011. @BrechtDeMan Listening Test Interface Design July 12, 2016 34 / 35
  • 53. WIMP References Q&A References (cont.) S. Olive and T. Welti, “The relationship between perception and measurement of headphone sound quality,” in Audio Engineering Society Convention 133, October 2012. B. De Man and J. D. Reiss, “A pairwise and multiple stimuli approach to perceptual evaluation of microphone types,” in Audio Engineering Society Convention 134, May 2013. Method for the subjective assessment of intermediate quality level of coding systems. Recommendation ITU-R BS.1534-1, 2003. E. Vincent, M. G. Jafari, and M. D. Plumbley, “Preliminary guidelines for subjective evalutation of audio source separation algorithms,” in UK ICA Research Network Workshop, 2006. S. E. Olive, “Some new evidence that teenagers and college students may prefer accurate sound reproduction,” in Audio Engineering Society Convention 132, April 2012. C. V¨olker and R. Huber, “Adaptions for the MUlti Stimulus test with Hidden Reference and Anchor (MUSHRA) for elder and technical unexperienced participants,” in DAGA 2015 N¨urnberg, 2015. N. Jillings, D. Mo↵at, B. De Man, and J. D. Reiss, “Web Audio Evaluation Tool: A browser-based listening test environment,” in 12th Sound and Music Computing Conference, July 2015. J. Berg, “OPAQUE – A tool for the elicitation and grading of audio quality attributes,” in 118th Convention of the Audio Engineering Society, May 2005. M. Cartwright, “Fast and easy crowdsourced perceptual audio evaluation,” in IEEE International Conference on Acoustics, Speech and Signal Processing, 2016. M. Schoe✏er, F.-R. St¨oter, H. Bayerlein, B. Edler, and J. Herre, “An experiment about estimating the number of instruments in polyphonic music: A comparison between internet and laboratory results,” in 14th International Society for Music Information Retrieval Conference (ISMIR 2013), pp. 389–394, 2013. @BrechtDeMan Listening Test Interface Design July 12, 2016 34 / 35
  • 54. WIMP References Q&A References (cont.) F.-R. St¨oter, M. Schoe✏er, B. Edler, and J. Herre, “Human ability of counting the number of instruments in polyphonic music,” in Proceedings of Meetings on Acoustics, vol. 19, p. 035034, Acoustical Society of America, 2013. @BrechtDeMan Listening Test Interface Design July 12, 2016 34 / 35
  • 55. WIMP References Q&A Sound sources I Fredy V: “In The Meantime” I ‘counting.wav’ by Corsica S (Freesound) @BrechtDeMan Listening Test Interface Design July 12, 2016 34 / 35