Automating Business Process via MuleSoft Composer | Bangalore MuleSoft Meetup...
Introduction of my research histroy: From instrument recognition to support of amateurs' music creation
1. Introduction of my research history:
From instrument recognition
to support of amateurs' music
creation
Tetsuro Kitahara
Nihon University, Japan
kitahara@chs.nihon-u.ac.jp
http://www.kthrlab.jp/
twitter: @tetsurokitahara
2. Introduction of myself
● Name: Tetsuro Kitahara
● Living in: Tokyo, Japan
● Age: 37
● Position: Associate professor in Nihon University
● Favorites: Music, Drinking alchol, etc.
4. Research when I was a student
● Instrument recognition for polyphonic music
● Content-based Music Information Retrieval
5. Instrument recognition
● Instrumentation is an important factor in MIR
● Not many attempts at polyphonic instrument
recognition at that time
● Typical framework:
Note detection -> feature extract. for each note
1:1:000
C4
B3
D4
E4
F4
G4
2:1:000 1:1:000
C4
B3
D4
E4
F4
G4
2:1:000
VN
VN
VN
PF
PF
(VN:Violin, PF: Piano)
For each note…
Pitch: C4
Start: 1:1:000
End: 1:3:000
Piano: 99.0%
Violin: 0.6%
……
A posteriori probabilitiesHarmonic structure
X1 = 0.124
X2 = 0.635
……
Feature vector
6. Instrument recognition
● Instrogram: Subsymbolic time-freq. representation
of instrument existence
● p(ωi; t, f ) : the probabilty that the sound of instru-
ment ωi with F0 of f exists at time t
Piano
Flute
Time [s]
7. Formulation of Instrogram
Non-specific IEP Conditional IEP
H
M
M
S E
…
S E
S E
HMM for each semitone
t
f
Instrument existence probability (IEP)
…
PreFEst [Goto, 2000]
(Estimate the weight for
tone model for each semitone)
= w110×
+ w660×
Observed spectrum
Tone model
for 110Hz
Tone model
for 660Hz
p(ωi; t, f ) = p(X; t, f ) p(ωi | X; t, f )
Demo VIdeo
8. Research when I was a PostDoc
● Chord voicing based on Bayesian network (skip)
● BayesianBand
● OrpheusBB
● CrestMuseXML and CrestMuse Toolkit (skip)
9. BayesianBand
A jam session system based on mutual prediction
of the user and system
C ? ? ?
Chord progression:
Predict
Determine
in real time
Musically
match
Pleasant
ChordMelody
Concept Melody
Each keystroke
Infer the next chord
Next chord
When changing
the measure
Get the latest result
Generate MIDI data
C
Dm
Em
F
G
Am
Bm(-5)
Likelihood
Chord Chord Chord
Melody
Note
Melody
Note
Melody
Note
Implementation
Demo VIdeo
10. OrpheusBB
Human-in-the-loop
Initial
input
Generate
music
Listen to
the music
Edit the
melody
Regenerate
the backing
Edit the
backing
Regenerate
the melody
Finish
Model
(Collaboration with Univ. of Tokyo)
Demo
Allows users to edit outputs of the system
Automatically re-generate the remaining part
according to part edited by users
12. Research in the current university
● Four-part harmonization using Bayes nets (skip)
● Humming-based composition support (skip)
● Melody editing using melodic outline
● Smart loop sequencer
13. Melody editing using melodic outline
● Melody is represented as a continous curve
● User can edit the melody by redrawing this curve
16. High-order coeffs. of
original melody
Low-order coeffs. of
edited outline
Fourier transform
Inverse Fourier
transform
Do Mi Fa So Ti Ra So Do Re Mi
Hidden Markov model
How to generate a melody from the edited outline
Demo
17. Smart loop sequencer
● Automatically selects music loops from collection
● Uses the degree of excitement as an input
The degree of
excitement
18. Tuple of loop IDs for 5 parts at measure n ("0" for no loop)
Estimate the most likely
[s1, ..., sN] for given x
Observed signal
Fomulating with HMM
19. Simplify the formulation by independently considering
qn, i × s'n, i
Whether a loop is placed
at measure n in part i
If so, which loop is placed
there
P(xn | qn)
The more loops are inserted,
the higher xn is emitted
5 1
the higher deg. of excitement
is annotated in the loop,
the higher xn is emitted
P(xn | s'n)
20. High deg. of excitement
Drums
Drums
Sequence
Sequence
Low deg. of excitement
How to estimate the degree of excitement for each music loop
Demo
21. Discussions
● Generate music data
from users’ inputs
● Users’ inputs are
usually imcomplete
● Typically based on
probabilistic models
These works have two aspects:
Automatic music generation Human-computer interaction
● Allow users to input
their intent easily
● More abstract than
specific music data
● Details are hidden
● Tradeoff btwn details
and intuitiveness
22. Conclusion
● My research
– MIR-related subjects (-2007)
● Musical instrument recognition
● MIR based on instrumentation similarity
– Music generation (2007-)
● BayesianBand, Melodic outline, Smart loop sequencer, …
● 2 aspects in my recent works
– Automatic music generation
– human-computer interaction
● Research plan during this stay
– Improvement of melody generation model, ...