SlideShare a Scribd company logo
1 of 31
Download to read offline
Introduction to Audio Content Analysis
module 5.0: data, data splits, and augmentation
alexander lerch
overview intro data augmentation summary
introduction
overview
corresponding textbook section
chapter 5
lecture content
• data requirements
• data splits for train and test
• N-Fold cross-validation
• data augmentation
learning objectives
• understand the importance of data in machine learning
• define task-specific data requirements
• discuss possibilities of data augmentation
• implement N-Fold cross-validation in Python
module 5.0: data, data splits, and augmentation 1 / 9
overview intro data augmentation summary
introduction
overview
corresponding textbook section
chapter 5
lecture content
• data requirements
• data splits for train and test
• N-Fold cross-validation
• data augmentation
learning objectives
• understand the importance of data in machine learning
• define task-specific data requirements
• discuss possibilities of data augmentation
• implement N-Fold cross-validation in Python
module 5.0: data, data splits, and augmentation 1 / 9
overview intro data augmentation summary
machine learning
data-driven
derive classification parameters from data, e.g.,
⇒ learn feature distributions/separation metrics per class
typical steps
1 define training set: annotated results
2 normalize training set
3 train classifier
4 evaluate classifier with test (or validation) set
5 (adjust classifier settings, return to 4.)
module 5.0: data, data splits, and augmentation 2 / 9
overview intro data augmentation summary
machine learning
data-driven
derive classification parameters from data, e.g.,
⇒ learn feature distributions/separation metrics per class
typical steps
1 define training set: annotated results
2 normalize training set
3 train classifier
4 evaluate classifier with test (or validation) set
5 (adjust classifier settings, return to 4.)
module 5.0: data, data splits, and augmentation 2 / 9
overview intro data augmentation summary
machine learning
data-driven
derive classification parameters from data, e.g.,
⇒ learn feature distributions/separation metrics per class
typical steps
1 define training set: annotated results
2 normalize training set
3 train classifier
4 evaluate classifier with test (or validation) set
5 (adjust classifier settings, return to 4.)
module 5.0: data, data splits, and augmentation 2 / 9
overview intro data augmentation summary
machine learning
data-driven
derive classification parameters from data, e.g.,
⇒ learn feature distributions/separation metrics per class
typical steps
1 define training set: annotated results
2 normalize training set
3 train classifier
4 evaluate classifier with test (or validation) set
5 (adjust classifier settings, return to 4.)
module 5.0: data, data splits, and augmentation 2 / 9
overview intro data augmentation summary
machine learning
data-driven
derive classification parameters from data, e.g.,
⇒ learn feature distributions/separation metrics per class
typical steps
1 define training set: annotated results
2 normalize training set
3 train classifier
4 evaluate classifier with test (or validation) set
5 (adjust classifier settings, return to 4.)
module 5.0: data, data splits, and augmentation 2 / 9
overview intro data augmentation summary
data
requirements
what are important properties of our data
module 5.0: data, data splits, and augmentation 3 / 9
overview intro data augmentation summary
data
requirements
what are important properties of our data
representative
• represent all necessary factors of input data (e.g., range of genres, audio qualities,
musical complexity, etc.)
• unbiased representation of class balance/label distribution
clean, non-noisy
• potential issues with subjective tasks
sufficient
• complex tasks/systems require lots of data
module 5.0: data, data splits, and augmentation 3 / 9
overview intro data augmentation summary
data
data split
a bigger data set is commonly split in subsets
• training data (≈ 70 − 80%)
▶ used to build the machine learning model
• validation data (≈ 10 − 15%)
▶ used to tweak model parameters
• testing data (≈ 10 − 15%)
▶ used to evaluate the model
▶ needs to be unseen!
no overlap between subsets!
• also make sure that similar content (from one
recording, album, artist, ...) is grouped into
one subset only
module 5.0: data, data splits, and augmentation 4 / 9
matlab
source:
plotDataSplit.m
overview intro data augmentation summary
data
data split
a bigger data set is commonly split in subsets
• training data (≈ 70 − 80%)
▶ used to build the machine learning model
• validation data (≈ 10 − 15%)
▶ used to tweak model parameters
• testing data (≈ 10 − 15%)
▶ used to evaluate the model
▶ needs to be unseen!
no overlap between subsets!
• also make sure that similar content (from one
recording, album, artist, ...) is grouped into
one subset only
module 5.0: data, data splits, and augmentation 4 / 9
matlab
source:
plotDataSplit.m
overview intro data augmentation summary
data
N-Fold cross validation 1/2
trying to utilize ALL data as both training and testing data
special case: Leave One Out CV
tends to be time-consuming
1 split training set into N parts (randomly, but preferably identical number per class)
2 select one part as test set
3 train the classifier with all observations from remaining N − 1 parts
4 compute the classification rate for the test set
5 repeat until all N parts have been tested
6 overall result: average classification rate
module 5.0: data, data splits, and augmentation 5 / 9
overview intro data augmentation summary
data
N-Fold cross validation 1/2
trying to utilize ALL data as both training and testing data
special case: Leave One Out CV
tends to be time-consuming
1 split training set into N parts (randomly, but preferably identical number per class)
2 select one part as test set
3 train the classifier with all observations from remaining N − 1 parts
4 compute the classification rate for the test set
5 repeat until all N parts have been tested
6 overall result: average classification rate
module 5.0: data, data splits, and augmentation 5 / 9
overview intro data augmentation summary
data
N-Fold cross validation 1/2
trying to utilize ALL data as both training and testing data
special case: Leave One Out CV
tends to be time-consuming
1 split training set into N parts (randomly, but preferably identical number per class)
2 select one part as test set
3 train the classifier with all observations from remaining N − 1 parts
4 compute the classification rate for the test set
5 repeat until all N parts have been tested
6 overall result: average classification rate
module 5.0: data, data splits, and augmentation 5 / 9
overview intro data augmentation summary
data
N-Fold cross validation 1/2
trying to utilize ALL data as both training and testing data
special case: Leave One Out CV
tends to be time-consuming
1 split training set into N parts (randomly, but preferably identical number per class)
2 select one part as test set
3 train the classifier with all observations from remaining N − 1 parts
4 compute the classification rate for the test set
5 repeat until all N parts have been tested
6 overall result: average classification rate
module 5.0: data, data splits, and augmentation 5 / 9
overview intro data augmentation summary
data
N-Fold cross validation 1/2
trying to utilize ALL data as both training and testing data
special case: Leave One Out CV
tends to be time-consuming
1 split training set into N parts (randomly, but preferably identical number per class)
2 select one part as test set
3 train the classifier with all observations from remaining N − 1 parts
4 compute the classification rate for the test set
5 repeat until all N parts have been tested
6 overall result: average classification rate
module 5.0: data, data splits, and augmentation 5 / 9
overview intro data augmentation summary
data
N-Fold cross validation 1/2
trying to utilize ALL data as both training and testing data
special case: Leave One Out CV
tends to be time-consuming
1 split training set into N parts (randomly, but preferably identical number per class)
2 select one part as test set
3 train the classifier with all observations from remaining N − 1 parts
4 compute the classification rate for the test set
5 repeat until all N parts have been tested
6 overall result: average classification rate
module 5.0: data, data splits, and augmentation 5 / 9
overview intro data augmentation summary
data
N-Fold cross validation 1/2
trying to utilize ALL data as both training and testing data
special case: Leave One Out CV
tends to be time-consuming
1 split training set into N parts (randomly, but preferably identical number per class)
2 select one part as test set
3 train the classifier with all observations from remaining N − 1 parts
4 compute the classification rate for the test set
5 repeat until all N parts have been tested
6 overall result: average classification rate
module 5.0: data, data splits, and augmentation 5 / 9
overview intro data augmentation summary
data
N-Fold cross validation 2/2
1.
2.
3.
4.
Data
Split 1 Split 2 Split 3 Split 4
Test Train Train Train
Train Test Train Train
Train Train Test Train
Train Train Train Test
module 5.0: data, data splits, and augmentation 6 / 9
overview intro data augmentation summary
classification
interaction of data, features, and classifier
training set
• training set too small, feature number too large
⇒ overfitting
• training set too noisy
⇒ underfitting
• training set not representative
⇒ bad classification performance
classifier
• classifier too complex
⇒ overfitting
• poor classifier
⇒ bad classification performance
features
• poor features
⇒ bad classification performance
module 5.0: data, data splits, and augmentation 7 / 9
overview intro data augmentation summary
classification
interaction of data, features, and classifier
training set
• training set too small, feature number too large
⇒ overfitting
• training set too noisy
⇒ underfitting
• training set not representative
⇒ bad classification performance
classifier
• classifier too complex
⇒ overfitting
• poor classifier
⇒ bad classification performance
features
• poor features
⇒ bad classification performance
module 5.0: data, data splits, and augmentation 7 / 9
overview intro data augmentation summary
classification
interaction of data, features, and classifier
training set
• training set too small, feature number too large
⇒ overfitting
• training set too noisy
⇒ underfitting
• training set not representative
⇒ bad classification performance
classifier
• classifier too complex
⇒ overfitting
• poor classifier
⇒ bad classification performance
features
• poor features
⇒ bad classification performance
module 5.0: data, data splits, and augmentation 7 / 9
overview intro data augmentation summary
data
augmentation
if annotated data is insufficient, we can ’cheat’ by increasing the amount of
training data
⇒ data augmentation: apply irrelevant transforms to audio data
• data segmentation
▶ treat audio snippets as separate observations
• quality degradation
▶ add noise and distortion, limit bandwidth, etc.
• audio effects
▶ apply reverb, etc.
• changing pitch/tempo
• combine data
▶ mix different audio inputs together (if labels can be “mixed”)
• mask out parts of the signal
module 5.0: data, data splits, and augmentation 8 / 9
overview intro data augmentation summary
data
augmentation
if annotated data is insufficient, we can ’cheat’ by increasing the amount of
training data
⇒ data augmentation: apply irrelevant transforms to audio data
• data segmentation
▶ treat audio snippets as separate observations
• quality degradation
▶ add noise and distortion, limit bandwidth, etc.
• audio effects
▶ apply reverb, etc.
• changing pitch/tempo
• combine data
▶ mix different audio inputs together (if labels can be “mixed”)
• mask out parts of the signal
module 5.0: data, data splits, and augmentation 8 / 9
overview intro data augmentation summary
data
augmentation
if annotated data is insufficient, we can ’cheat’ by increasing the amount of
training data
⇒ data augmentation: apply irrelevant transforms to audio data
• data segmentation
▶ treat audio snippets as separate observations
• quality degradation
▶ add noise and distortion, limit bandwidth, etc.
• audio effects
▶ apply reverb, etc.
• changing pitch/tempo
• combine data
▶ mix different audio inputs together (if labels can be “mixed”)
• mask out parts of the signal
module 5.0: data, data splits, and augmentation 8 / 9
overview intro data augmentation summary
data
augmentation
if annotated data is insufficient, we can ’cheat’ by increasing the amount of
training data
⇒ data augmentation: apply irrelevant transforms to audio data
• data segmentation
▶ treat audio snippets as separate observations
• quality degradation
▶ add noise and distortion, limit bandwidth, etc.
• audio effects
▶ apply reverb, etc.
• changing pitch/tempo
• combine data
▶ mix different audio inputs together (if labels can be “mixed”)
• mask out parts of the signal
module 5.0: data, data splits, and augmentation 8 / 9
overview intro data augmentation summary
data
augmentation
if annotated data is insufficient, we can ’cheat’ by increasing the amount of
training data
⇒ data augmentation: apply irrelevant transforms to audio data
• data segmentation
▶ treat audio snippets as separate observations
• quality degradation
▶ add noise and distortion, limit bandwidth, etc.
• audio effects
▶ apply reverb, etc.
• changing pitch/tempo
• combine data
▶ mix different audio inputs together (if labels can be “mixed”)
• mask out parts of the signal
module 5.0: data, data splits, and augmentation 8 / 9
overview intro data augmentation summary
data
augmentation
if annotated data is insufficient, we can ’cheat’ by increasing the amount of
training data
⇒ data augmentation: apply irrelevant transforms to audio data
• data segmentation
▶ treat audio snippets as separate observations
• quality degradation
▶ add noise and distortion, limit bandwidth, etc.
• audio effects
▶ apply reverb, etc.
• changing pitch/tempo
• combine data
▶ mix different audio inputs together (if labels can be “mixed”)
• mask out parts of the signal
module 5.0: data, data splits, and augmentation 8 / 9
overview intro data augmentation summary
data
augmentation
if annotated data is insufficient, we can ’cheat’ by increasing the amount of
training data
⇒ data augmentation: apply irrelevant transforms to audio data
• data segmentation
▶ treat audio snippets as separate observations
• quality degradation
▶ add noise and distortion, limit bandwidth, etc.
• audio effects
▶ apply reverb, etc.
• changing pitch/tempo
• combine data
▶ mix different audio inputs together (if labels can be “mixed”)
• mask out parts of the signal
module 5.0: data, data splits, and augmentation 8 / 9
overview intro data augmentation summary
summary
lecture content
data
• representative
• clean, non-noisy
• sufficient
data split
• train
• validation
• test
cross validation
• multiple runs with varying data splits
• maximum data utilization
module 5.0: data, data splits, and augmentation 9 / 9

More Related Content

Similar to Audio Content Analysis: Data Splits, Augmentation & Cross-Validation

6 Evaluating Predictive Performance and ensemble.pptx
6 Evaluating Predictive Performance and ensemble.pptx6 Evaluating Predictive Performance and ensemble.pptx
6 Evaluating Predictive Performance and ensemble.pptxmohammedalherwi1
 
Cross validation.pptx
Cross validation.pptxCross validation.pptx
Cross validation.pptxYouKnowwho28
 
06-00-ACA-Evaluation.pdf
06-00-ACA-Evaluation.pdf06-00-ACA-Evaluation.pdf
06-00-ACA-Evaluation.pdfAlexanderLerch4
 
Data mining chapter04and5-best
Data mining chapter04and5-bestData mining chapter04and5-best
Data mining chapter04and5-bestABDUmomo
 
shubhampresentation-180430060134.pptx
shubhampresentation-180430060134.pptxshubhampresentation-180430060134.pptx
shubhampresentation-180430060134.pptxABINASHPADHY6
 
Presentation on supervised learning
Presentation on supervised learningPresentation on supervised learning
Presentation on supervised learningTonmoy Bhagawati
 
Barga Data Science lecture 10
Barga Data Science lecture 10Barga Data Science lecture 10
Barga Data Science lecture 10Roger Barga
 
K-Folds Cross Validation Method
K-Folds Cross Validation MethodK-Folds Cross Validation Method
K-Folds Cross Validation MethodSHUBHAM GUPTA
 
Chinese Named Entity Recognition with Graph-based Semi-supervised Learning Model
Chinese Named Entity Recognition with Graph-based Semi-supervised Learning ModelChinese Named Entity Recognition with Graph-based Semi-supervised Learning Model
Chinese Named Entity Recognition with Graph-based Semi-supervised Learning ModelLifeng (Aaron) Han
 
Bridging the Gap: Machine Learning for Ubiquitous Computing -- Evaluation
Bridging the Gap: Machine Learning for Ubiquitous Computing -- EvaluationBridging the Gap: Machine Learning for Ubiquitous Computing -- Evaluation
Bridging the Gap: Machine Learning for Ubiquitous Computing -- EvaluationThomas Ploetz
 
Chap_05_Data_Collection_and_Analysis.ppt
Chap_05_Data_Collection_and_Analysis.pptChap_05_Data_Collection_and_Analysis.ppt
Chap_05_Data_Collection_and_Analysis.pptRosaHildaFlix
 
Building largescalepredictionsystemv1
Building largescalepredictionsystemv1Building largescalepredictionsystemv1
Building largescalepredictionsystemv1arthi v
 
Customer Churn Analytics using Microsoft R Open
Customer Churn Analytics using Microsoft R OpenCustomer Churn Analytics using Microsoft R Open
Customer Churn Analytics using Microsoft R OpenPoo Kuan Hoong
 
Statistical Learning and Model Selection (1).pptx
Statistical Learning and Model Selection (1).pptxStatistical Learning and Model Selection (1).pptx
Statistical Learning and Model Selection (1).pptxrajalakshmi5921
 
Weka tutorial
Weka tutorialWeka tutorial
Weka tutorialSajib Sen
 
Cross-validation Tutorial: What, how and which?
Cross-validation Tutorial: What, how and which?Cross-validation Tutorial: What, how and which?
Cross-validation Tutorial: What, how and which?Pradeep Redddy Raamana
 
LETS PUBLISH WITH MORE RELIABLE & PRESENTABLE MODELLING.pptx
LETS PUBLISH WITH MORE RELIABLE & PRESENTABLE MODELLING.pptxLETS PUBLISH WITH MORE RELIABLE & PRESENTABLE MODELLING.pptx
LETS PUBLISH WITH MORE RELIABLE & PRESENTABLE MODELLING.pptxshamsul2010
 
Cmpe 255 cross validation
Cmpe 255 cross validationCmpe 255 cross validation
Cmpe 255 cross validationAbraham Kong
 

Similar to Audio Content Analysis: Data Splits, Augmentation & Cross-Validation (20)

6 Evaluating Predictive Performance and ensemble.pptx
6 Evaluating Predictive Performance and ensemble.pptx6 Evaluating Predictive Performance and ensemble.pptx
6 Evaluating Predictive Performance and ensemble.pptx
 
Cross validation.pptx
Cross validation.pptxCross validation.pptx
Cross validation.pptx
 
06-00-ACA-Evaluation.pdf
06-00-ACA-Evaluation.pdf06-00-ACA-Evaluation.pdf
06-00-ACA-Evaluation.pdf
 
Data mining chapter04and5-best
Data mining chapter04and5-bestData mining chapter04and5-best
Data mining chapter04and5-best
 
shubhampresentation-180430060134.pptx
shubhampresentation-180430060134.pptxshubhampresentation-180430060134.pptx
shubhampresentation-180430060134.pptx
 
Presentation on supervised learning
Presentation on supervised learningPresentation on supervised learning
Presentation on supervised learning
 
Barga Data Science lecture 10
Barga Data Science lecture 10Barga Data Science lecture 10
Barga Data Science lecture 10
 
K-Folds Cross Validation Method
K-Folds Cross Validation MethodK-Folds Cross Validation Method
K-Folds Cross Validation Method
 
Chinese Named Entity Recognition with Graph-based Semi-supervised Learning Model
Chinese Named Entity Recognition with Graph-based Semi-supervised Learning ModelChinese Named Entity Recognition with Graph-based Semi-supervised Learning Model
Chinese Named Entity Recognition with Graph-based Semi-supervised Learning Model
 
Bridging the Gap: Machine Learning for Ubiquitous Computing -- Evaluation
Bridging the Gap: Machine Learning for Ubiquitous Computing -- EvaluationBridging the Gap: Machine Learning for Ubiquitous Computing -- Evaluation
Bridging the Gap: Machine Learning for Ubiquitous Computing -- Evaluation
 
evaluation and credibility-Part 2
evaluation and credibility-Part 2evaluation and credibility-Part 2
evaluation and credibility-Part 2
 
Chap_05_Data_Collection_and_Analysis.ppt
Chap_05_Data_Collection_and_Analysis.pptChap_05_Data_Collection_and_Analysis.ppt
Chap_05_Data_Collection_and_Analysis.ppt
 
Building largescalepredictionsystemv1
Building largescalepredictionsystemv1Building largescalepredictionsystemv1
Building largescalepredictionsystemv1
 
Customer Churn Analytics using Microsoft R Open
Customer Churn Analytics using Microsoft R OpenCustomer Churn Analytics using Microsoft R Open
Customer Churn Analytics using Microsoft R Open
 
Statistical Learning and Model Selection (1).pptx
Statistical Learning and Model Selection (1).pptxStatistical Learning and Model Selection (1).pptx
Statistical Learning and Model Selection (1).pptx
 
Weka tutorial
Weka tutorialWeka tutorial
Weka tutorial
 
Cross-validation Tutorial: What, how and which?
Cross-validation Tutorial: What, how and which?Cross-validation Tutorial: What, how and which?
Cross-validation Tutorial: What, how and which?
 
LETS PUBLISH WITH MORE RELIABLE & PRESENTABLE MODELLING.pptx
LETS PUBLISH WITH MORE RELIABLE & PRESENTABLE MODELLING.pptxLETS PUBLISH WITH MORE RELIABLE & PRESENTABLE MODELLING.pptx
LETS PUBLISH WITH MORE RELIABLE & PRESENTABLE MODELLING.pptx
 
Dbm630 lecture08
Dbm630 lecture08Dbm630 lecture08
Dbm630 lecture08
 
Cmpe 255 cross validation
Cmpe 255 cross validationCmpe 255 cross validation
Cmpe 255 cross validation
 

More from AlexanderLerch4

0A-02-ACA-Fundamentals-Convolution.pdf
0A-02-ACA-Fundamentals-Convolution.pdf0A-02-ACA-Fundamentals-Convolution.pdf
0A-02-ACA-Fundamentals-Convolution.pdfAlexanderLerch4
 
09-03-ACA-Temporal-Onsets.pdf
09-03-ACA-Temporal-Onsets.pdf09-03-ACA-Temporal-Onsets.pdf
09-03-ACA-Temporal-Onsets.pdfAlexanderLerch4
 
09-07-ACA-Temporal-Structure.pdf
09-07-ACA-Temporal-Structure.pdf09-07-ACA-Temporal-Structure.pdf
09-07-ACA-Temporal-Structure.pdfAlexanderLerch4
 
09-01-ACA-Temporal-Intro.pdf
09-01-ACA-Temporal-Intro.pdf09-01-ACA-Temporal-Intro.pdf
09-01-ACA-Temporal-Intro.pdfAlexanderLerch4
 
07-06-ACA-Tonal-Chord.pdf
07-06-ACA-Tonal-Chord.pdf07-06-ACA-Tonal-Chord.pdf
07-06-ACA-Tonal-Chord.pdfAlexanderLerch4
 
15-00-ACA-Music-Performance-Analysis.pdf
15-00-ACA-Music-Performance-Analysis.pdf15-00-ACA-Music-Performance-Analysis.pdf
15-00-ACA-Music-Performance-Analysis.pdfAlexanderLerch4
 
12-01-ACA-Similarity.pdf
12-01-ACA-Similarity.pdf12-01-ACA-Similarity.pdf
12-01-ACA-Similarity.pdfAlexanderLerch4
 
09-04-ACA-Temporal-BeatHisto.pdf
09-04-ACA-Temporal-BeatHisto.pdf09-04-ACA-Temporal-BeatHisto.pdf
09-04-ACA-Temporal-BeatHisto.pdfAlexanderLerch4
 
09-05-ACA-Temporal-Tempo.pdf
09-05-ACA-Temporal-Tempo.pdf09-05-ACA-Temporal-Tempo.pdf
09-05-ACA-Temporal-Tempo.pdfAlexanderLerch4
 
11-00-ACA-Fingerprinting.pdf
11-00-ACA-Fingerprinting.pdf11-00-ACA-Fingerprinting.pdf
11-00-ACA-Fingerprinting.pdfAlexanderLerch4
 
03-05-ACA-Input-Features.pdf
03-05-ACA-Input-Features.pdf03-05-ACA-Input-Features.pdf
03-05-ACA-Input-Features.pdfAlexanderLerch4
 
07-03-03-ACA-Tonal-Monof0.pdf
07-03-03-ACA-Tonal-Monof0.pdf07-03-03-ACA-Tonal-Monof0.pdf
07-03-03-ACA-Tonal-Monof0.pdfAlexanderLerch4
 
07-03-05-ACA-Tonal-F0Eval.pdf
07-03-05-ACA-Tonal-F0Eval.pdf07-03-05-ACA-Tonal-F0Eval.pdf
07-03-05-ACA-Tonal-F0Eval.pdfAlexanderLerch4
 
03-03-01-ACA-Input-TF-Fourier.pdf
03-03-01-ACA-Input-TF-Fourier.pdf03-03-01-ACA-Input-TF-Fourier.pdf
03-03-01-ACA-Input-TF-Fourier.pdfAlexanderLerch4
 
03-06-ACA-Input-FeatureLearning.pdf
03-06-ACA-Input-FeatureLearning.pdf03-06-ACA-Input-FeatureLearning.pdf
03-06-ACA-Input-FeatureLearning.pdfAlexanderLerch4
 
07-03-04-ACA-Tonal-Polyf0.pdf
07-03-04-ACA-Tonal-Polyf0.pdf07-03-04-ACA-Tonal-Polyf0.pdf
07-03-04-ACA-Tonal-Polyf0.pdfAlexanderLerch4
 

More from AlexanderLerch4 (20)

0A-02-ACA-Fundamentals-Convolution.pdf
0A-02-ACA-Fundamentals-Convolution.pdf0A-02-ACA-Fundamentals-Convolution.pdf
0A-02-ACA-Fundamentals-Convolution.pdf
 
12-02-ACA-Genre.pdf
12-02-ACA-Genre.pdf12-02-ACA-Genre.pdf
12-02-ACA-Genre.pdf
 
10-02-ACA-Alignment.pdf
10-02-ACA-Alignment.pdf10-02-ACA-Alignment.pdf
10-02-ACA-Alignment.pdf
 
09-03-ACA-Temporal-Onsets.pdf
09-03-ACA-Temporal-Onsets.pdf09-03-ACA-Temporal-Onsets.pdf
09-03-ACA-Temporal-Onsets.pdf
 
13-00-ACA-Mood.pdf
13-00-ACA-Mood.pdf13-00-ACA-Mood.pdf
13-00-ACA-Mood.pdf
 
09-07-ACA-Temporal-Structure.pdf
09-07-ACA-Temporal-Structure.pdf09-07-ACA-Temporal-Structure.pdf
09-07-ACA-Temporal-Structure.pdf
 
09-01-ACA-Temporal-Intro.pdf
09-01-ACA-Temporal-Intro.pdf09-01-ACA-Temporal-Intro.pdf
09-01-ACA-Temporal-Intro.pdf
 
07-06-ACA-Tonal-Chord.pdf
07-06-ACA-Tonal-Chord.pdf07-06-ACA-Tonal-Chord.pdf
07-06-ACA-Tonal-Chord.pdf
 
15-00-ACA-Music-Performance-Analysis.pdf
15-00-ACA-Music-Performance-Analysis.pdf15-00-ACA-Music-Performance-Analysis.pdf
15-00-ACA-Music-Performance-Analysis.pdf
 
12-01-ACA-Similarity.pdf
12-01-ACA-Similarity.pdf12-01-ACA-Similarity.pdf
12-01-ACA-Similarity.pdf
 
09-04-ACA-Temporal-BeatHisto.pdf
09-04-ACA-Temporal-BeatHisto.pdf09-04-ACA-Temporal-BeatHisto.pdf
09-04-ACA-Temporal-BeatHisto.pdf
 
09-05-ACA-Temporal-Tempo.pdf
09-05-ACA-Temporal-Tempo.pdf09-05-ACA-Temporal-Tempo.pdf
09-05-ACA-Temporal-Tempo.pdf
 
11-00-ACA-Fingerprinting.pdf
11-00-ACA-Fingerprinting.pdf11-00-ACA-Fingerprinting.pdf
11-00-ACA-Fingerprinting.pdf
 
08-00-ACA-Intensity.pdf
08-00-ACA-Intensity.pdf08-00-ACA-Intensity.pdf
08-00-ACA-Intensity.pdf
 
03-05-ACA-Input-Features.pdf
03-05-ACA-Input-Features.pdf03-05-ACA-Input-Features.pdf
03-05-ACA-Input-Features.pdf
 
07-03-03-ACA-Tonal-Monof0.pdf
07-03-03-ACA-Tonal-Monof0.pdf07-03-03-ACA-Tonal-Monof0.pdf
07-03-03-ACA-Tonal-Monof0.pdf
 
07-03-05-ACA-Tonal-F0Eval.pdf
07-03-05-ACA-Tonal-F0Eval.pdf07-03-05-ACA-Tonal-F0Eval.pdf
07-03-05-ACA-Tonal-F0Eval.pdf
 
03-03-01-ACA-Input-TF-Fourier.pdf
03-03-01-ACA-Input-TF-Fourier.pdf03-03-01-ACA-Input-TF-Fourier.pdf
03-03-01-ACA-Input-TF-Fourier.pdf
 
03-06-ACA-Input-FeatureLearning.pdf
03-06-ACA-Input-FeatureLearning.pdf03-06-ACA-Input-FeatureLearning.pdf
03-06-ACA-Input-FeatureLearning.pdf
 
07-03-04-ACA-Tonal-Polyf0.pdf
07-03-04-ACA-Tonal-Polyf0.pdf07-03-04-ACA-Tonal-Polyf0.pdf
07-03-04-ACA-Tonal-Polyf0.pdf
 

Recently uploaded

Understanding the Laravel MVC Architecture
Understanding the Laravel MVC ArchitectureUnderstanding the Laravel MVC Architecture
Understanding the Laravel MVC ArchitecturePixlogix Infotech
 
Scanning the Internet for External Cloud Exposures via SSL Certs
Scanning the Internet for External Cloud Exposures via SSL CertsScanning the Internet for External Cloud Exposures via SSL Certs
Scanning the Internet for External Cloud Exposures via SSL CertsRizwan Syed
 
SQL Database Design For Developers at php[tek] 2024
SQL Database Design For Developers at php[tek] 2024SQL Database Design For Developers at php[tek] 2024
SQL Database Design For Developers at php[tek] 2024Scott Keck-Warren
 
Making_way_through_DLL_hollowing_inspite_of_CFG_by_Debjeet Banerjee.pptx
Making_way_through_DLL_hollowing_inspite_of_CFG_by_Debjeet Banerjee.pptxMaking_way_through_DLL_hollowing_inspite_of_CFG_by_Debjeet Banerjee.pptx
Making_way_through_DLL_hollowing_inspite_of_CFG_by_Debjeet Banerjee.pptxnull - The Open Security Community
 
Maximizing Board Effectiveness 2024 Webinar.pptx
Maximizing Board Effectiveness 2024 Webinar.pptxMaximizing Board Effectiveness 2024 Webinar.pptx
Maximizing Board Effectiveness 2024 Webinar.pptxOnBoard
 
Presentation on how to chat with PDF using ChatGPT code interpreter
Presentation on how to chat with PDF using ChatGPT code interpreterPresentation on how to chat with PDF using ChatGPT code interpreter
Presentation on how to chat with PDF using ChatGPT code interpreternaman860154
 
Key Features Of Token Development (1).pptx
Key  Features Of Token  Development (1).pptxKey  Features Of Token  Development (1).pptx
Key Features Of Token Development (1).pptxLBM Solutions
 
Designing IA for AI - Information Architecture Conference 2024
Designing IA for AI - Information Architecture Conference 2024Designing IA for AI - Information Architecture Conference 2024
Designing IA for AI - Information Architecture Conference 2024Enterprise Knowledge
 
My Hashitalk Indonesia April 2024 Presentation
My Hashitalk Indonesia April 2024 PresentationMy Hashitalk Indonesia April 2024 Presentation
My Hashitalk Indonesia April 2024 PresentationRidwan Fadjar
 
"LLMs for Python Engineers: Advanced Data Analysis and Semantic Kernel",Oleks...
"LLMs for Python Engineers: Advanced Data Analysis and Semantic Kernel",Oleks..."LLMs for Python Engineers: Advanced Data Analysis and Semantic Kernel",Oleks...
"LLMs for Python Engineers: Advanced Data Analysis and Semantic Kernel",Oleks...Fwdays
 
Integration and Automation in Practice: CI/CD in Mule Integration and Automat...
Integration and Automation in Practice: CI/CD in Mule Integration and Automat...Integration and Automation in Practice: CI/CD in Mule Integration and Automat...
Integration and Automation in Practice: CI/CD in Mule Integration and Automat...Patryk Bandurski
 
Benefits Of Flutter Compared To Other Frameworks
Benefits Of Flutter Compared To Other FrameworksBenefits Of Flutter Compared To Other Frameworks
Benefits Of Flutter Compared To Other FrameworksSoftradix Technologies
 
08448380779 Call Girls In Friends Colony Women Seeking Men
08448380779 Call Girls In Friends Colony Women Seeking Men08448380779 Call Girls In Friends Colony Women Seeking Men
08448380779 Call Girls In Friends Colony Women Seeking MenDelhi Call girls
 
Enhancing Worker Digital Experience: A Hands-on Workshop for Partners
Enhancing Worker Digital Experience: A Hands-on Workshop for PartnersEnhancing Worker Digital Experience: A Hands-on Workshop for Partners
Enhancing Worker Digital Experience: A Hands-on Workshop for PartnersThousandEyes
 
How to convert PDF to text with Nanonets
How to convert PDF to text with NanonetsHow to convert PDF to text with Nanonets
How to convert PDF to text with Nanonetsnaman860154
 
Breaking the Kubernetes Kill Chain: Host Path Mount
Breaking the Kubernetes Kill Chain: Host Path MountBreaking the Kubernetes Kill Chain: Host Path Mount
Breaking the Kubernetes Kill Chain: Host Path MountPuma Security, LLC
 
Advanced Test Driven-Development @ php[tek] 2024
Advanced Test Driven-Development @ php[tek] 2024Advanced Test Driven-Development @ php[tek] 2024
Advanced Test Driven-Development @ php[tek] 2024Scott Keck-Warren
 

Recently uploaded (20)

Understanding the Laravel MVC Architecture
Understanding the Laravel MVC ArchitectureUnderstanding the Laravel MVC Architecture
Understanding the Laravel MVC Architecture
 
Scanning the Internet for External Cloud Exposures via SSL Certs
Scanning the Internet for External Cloud Exposures via SSL CertsScanning the Internet for External Cloud Exposures via SSL Certs
Scanning the Internet for External Cloud Exposures via SSL Certs
 
SQL Database Design For Developers at php[tek] 2024
SQL Database Design For Developers at php[tek] 2024SQL Database Design For Developers at php[tek] 2024
SQL Database Design For Developers at php[tek] 2024
 
Making_way_through_DLL_hollowing_inspite_of_CFG_by_Debjeet Banerjee.pptx
Making_way_through_DLL_hollowing_inspite_of_CFG_by_Debjeet Banerjee.pptxMaking_way_through_DLL_hollowing_inspite_of_CFG_by_Debjeet Banerjee.pptx
Making_way_through_DLL_hollowing_inspite_of_CFG_by_Debjeet Banerjee.pptx
 
Maximizing Board Effectiveness 2024 Webinar.pptx
Maximizing Board Effectiveness 2024 Webinar.pptxMaximizing Board Effectiveness 2024 Webinar.pptx
Maximizing Board Effectiveness 2024 Webinar.pptx
 
Presentation on how to chat with PDF using ChatGPT code interpreter
Presentation on how to chat with PDF using ChatGPT code interpreterPresentation on how to chat with PDF using ChatGPT code interpreter
Presentation on how to chat with PDF using ChatGPT code interpreter
 
Key Features Of Token Development (1).pptx
Key  Features Of Token  Development (1).pptxKey  Features Of Token  Development (1).pptx
Key Features Of Token Development (1).pptx
 
Designing IA for AI - Information Architecture Conference 2024
Designing IA for AI - Information Architecture Conference 2024Designing IA for AI - Information Architecture Conference 2024
Designing IA for AI - Information Architecture Conference 2024
 
My Hashitalk Indonesia April 2024 Presentation
My Hashitalk Indonesia April 2024 PresentationMy Hashitalk Indonesia April 2024 Presentation
My Hashitalk Indonesia April 2024 Presentation
 
E-Vehicle_Hacking_by_Parul Sharma_null_owasp.pptx
E-Vehicle_Hacking_by_Parul Sharma_null_owasp.pptxE-Vehicle_Hacking_by_Parul Sharma_null_owasp.pptx
E-Vehicle_Hacking_by_Parul Sharma_null_owasp.pptx
 
"LLMs for Python Engineers: Advanced Data Analysis and Semantic Kernel",Oleks...
"LLMs for Python Engineers: Advanced Data Analysis and Semantic Kernel",Oleks..."LLMs for Python Engineers: Advanced Data Analysis and Semantic Kernel",Oleks...
"LLMs for Python Engineers: Advanced Data Analysis and Semantic Kernel",Oleks...
 
DMCC Future of Trade Web3 - Special Edition
DMCC Future of Trade Web3 - Special EditionDMCC Future of Trade Web3 - Special Edition
DMCC Future of Trade Web3 - Special Edition
 
Integration and Automation in Practice: CI/CD in Mule Integration and Automat...
Integration and Automation in Practice: CI/CD in Mule Integration and Automat...Integration and Automation in Practice: CI/CD in Mule Integration and Automat...
Integration and Automation in Practice: CI/CD in Mule Integration and Automat...
 
Benefits Of Flutter Compared To Other Frameworks
Benefits Of Flutter Compared To Other FrameworksBenefits Of Flutter Compared To Other Frameworks
Benefits Of Flutter Compared To Other Frameworks
 
08448380779 Call Girls In Friends Colony Women Seeking Men
08448380779 Call Girls In Friends Colony Women Seeking Men08448380779 Call Girls In Friends Colony Women Seeking Men
08448380779 Call Girls In Friends Colony Women Seeking Men
 
Enhancing Worker Digital Experience: A Hands-on Workshop for Partners
Enhancing Worker Digital Experience: A Hands-on Workshop for PartnersEnhancing Worker Digital Experience: A Hands-on Workshop for Partners
Enhancing Worker Digital Experience: A Hands-on Workshop for Partners
 
How to convert PDF to text with Nanonets
How to convert PDF to text with NanonetsHow to convert PDF to text with Nanonets
How to convert PDF to text with Nanonets
 
The transition to renewables in India.pdf
The transition to renewables in India.pdfThe transition to renewables in India.pdf
The transition to renewables in India.pdf
 
Breaking the Kubernetes Kill Chain: Host Path Mount
Breaking the Kubernetes Kill Chain: Host Path MountBreaking the Kubernetes Kill Chain: Host Path Mount
Breaking the Kubernetes Kill Chain: Host Path Mount
 
Advanced Test Driven-Development @ php[tek] 2024
Advanced Test Driven-Development @ php[tek] 2024Advanced Test Driven-Development @ php[tek] 2024
Advanced Test Driven-Development @ php[tek] 2024
 

Audio Content Analysis: Data Splits, Augmentation & Cross-Validation

  • 1. Introduction to Audio Content Analysis module 5.0: data, data splits, and augmentation alexander lerch
  • 2. overview intro data augmentation summary introduction overview corresponding textbook section chapter 5 lecture content • data requirements • data splits for train and test • N-Fold cross-validation • data augmentation learning objectives • understand the importance of data in machine learning • define task-specific data requirements • discuss possibilities of data augmentation • implement N-Fold cross-validation in Python module 5.0: data, data splits, and augmentation 1 / 9
  • 3. overview intro data augmentation summary introduction overview corresponding textbook section chapter 5 lecture content • data requirements • data splits for train and test • N-Fold cross-validation • data augmentation learning objectives • understand the importance of data in machine learning • define task-specific data requirements • discuss possibilities of data augmentation • implement N-Fold cross-validation in Python module 5.0: data, data splits, and augmentation 1 / 9
  • 4. overview intro data augmentation summary machine learning data-driven derive classification parameters from data, e.g., ⇒ learn feature distributions/separation metrics per class typical steps 1 define training set: annotated results 2 normalize training set 3 train classifier 4 evaluate classifier with test (or validation) set 5 (adjust classifier settings, return to 4.) module 5.0: data, data splits, and augmentation 2 / 9
  • 5. overview intro data augmentation summary machine learning data-driven derive classification parameters from data, e.g., ⇒ learn feature distributions/separation metrics per class typical steps 1 define training set: annotated results 2 normalize training set 3 train classifier 4 evaluate classifier with test (or validation) set 5 (adjust classifier settings, return to 4.) module 5.0: data, data splits, and augmentation 2 / 9
  • 6. overview intro data augmentation summary machine learning data-driven derive classification parameters from data, e.g., ⇒ learn feature distributions/separation metrics per class typical steps 1 define training set: annotated results 2 normalize training set 3 train classifier 4 evaluate classifier with test (or validation) set 5 (adjust classifier settings, return to 4.) module 5.0: data, data splits, and augmentation 2 / 9
  • 7. overview intro data augmentation summary machine learning data-driven derive classification parameters from data, e.g., ⇒ learn feature distributions/separation metrics per class typical steps 1 define training set: annotated results 2 normalize training set 3 train classifier 4 evaluate classifier with test (or validation) set 5 (adjust classifier settings, return to 4.) module 5.0: data, data splits, and augmentation 2 / 9
  • 8. overview intro data augmentation summary machine learning data-driven derive classification parameters from data, e.g., ⇒ learn feature distributions/separation metrics per class typical steps 1 define training set: annotated results 2 normalize training set 3 train classifier 4 evaluate classifier with test (or validation) set 5 (adjust classifier settings, return to 4.) module 5.0: data, data splits, and augmentation 2 / 9
  • 9. overview intro data augmentation summary data requirements what are important properties of our data module 5.0: data, data splits, and augmentation 3 / 9
  • 10. overview intro data augmentation summary data requirements what are important properties of our data representative • represent all necessary factors of input data (e.g., range of genres, audio qualities, musical complexity, etc.) • unbiased representation of class balance/label distribution clean, non-noisy • potential issues with subjective tasks sufficient • complex tasks/systems require lots of data module 5.0: data, data splits, and augmentation 3 / 9
  • 11. overview intro data augmentation summary data data split a bigger data set is commonly split in subsets • training data (≈ 70 − 80%) ▶ used to build the machine learning model • validation data (≈ 10 − 15%) ▶ used to tweak model parameters • testing data (≈ 10 − 15%) ▶ used to evaluate the model ▶ needs to be unseen! no overlap between subsets! • also make sure that similar content (from one recording, album, artist, ...) is grouped into one subset only module 5.0: data, data splits, and augmentation 4 / 9 matlab source: plotDataSplit.m
  • 12. overview intro data augmentation summary data data split a bigger data set is commonly split in subsets • training data (≈ 70 − 80%) ▶ used to build the machine learning model • validation data (≈ 10 − 15%) ▶ used to tweak model parameters • testing data (≈ 10 − 15%) ▶ used to evaluate the model ▶ needs to be unseen! no overlap between subsets! • also make sure that similar content (from one recording, album, artist, ...) is grouped into one subset only module 5.0: data, data splits, and augmentation 4 / 9 matlab source: plotDataSplit.m
  • 13. overview intro data augmentation summary data N-Fold cross validation 1/2 trying to utilize ALL data as both training and testing data special case: Leave One Out CV tends to be time-consuming 1 split training set into N parts (randomly, but preferably identical number per class) 2 select one part as test set 3 train the classifier with all observations from remaining N − 1 parts 4 compute the classification rate for the test set 5 repeat until all N parts have been tested 6 overall result: average classification rate module 5.0: data, data splits, and augmentation 5 / 9
  • 14. overview intro data augmentation summary data N-Fold cross validation 1/2 trying to utilize ALL data as both training and testing data special case: Leave One Out CV tends to be time-consuming 1 split training set into N parts (randomly, but preferably identical number per class) 2 select one part as test set 3 train the classifier with all observations from remaining N − 1 parts 4 compute the classification rate for the test set 5 repeat until all N parts have been tested 6 overall result: average classification rate module 5.0: data, data splits, and augmentation 5 / 9
  • 15. overview intro data augmentation summary data N-Fold cross validation 1/2 trying to utilize ALL data as both training and testing data special case: Leave One Out CV tends to be time-consuming 1 split training set into N parts (randomly, but preferably identical number per class) 2 select one part as test set 3 train the classifier with all observations from remaining N − 1 parts 4 compute the classification rate for the test set 5 repeat until all N parts have been tested 6 overall result: average classification rate module 5.0: data, data splits, and augmentation 5 / 9
  • 16. overview intro data augmentation summary data N-Fold cross validation 1/2 trying to utilize ALL data as both training and testing data special case: Leave One Out CV tends to be time-consuming 1 split training set into N parts (randomly, but preferably identical number per class) 2 select one part as test set 3 train the classifier with all observations from remaining N − 1 parts 4 compute the classification rate for the test set 5 repeat until all N parts have been tested 6 overall result: average classification rate module 5.0: data, data splits, and augmentation 5 / 9
  • 17. overview intro data augmentation summary data N-Fold cross validation 1/2 trying to utilize ALL data as both training and testing data special case: Leave One Out CV tends to be time-consuming 1 split training set into N parts (randomly, but preferably identical number per class) 2 select one part as test set 3 train the classifier with all observations from remaining N − 1 parts 4 compute the classification rate for the test set 5 repeat until all N parts have been tested 6 overall result: average classification rate module 5.0: data, data splits, and augmentation 5 / 9
  • 18. overview intro data augmentation summary data N-Fold cross validation 1/2 trying to utilize ALL data as both training and testing data special case: Leave One Out CV tends to be time-consuming 1 split training set into N parts (randomly, but preferably identical number per class) 2 select one part as test set 3 train the classifier with all observations from remaining N − 1 parts 4 compute the classification rate for the test set 5 repeat until all N parts have been tested 6 overall result: average classification rate module 5.0: data, data splits, and augmentation 5 / 9
  • 19. overview intro data augmentation summary data N-Fold cross validation 1/2 trying to utilize ALL data as both training and testing data special case: Leave One Out CV tends to be time-consuming 1 split training set into N parts (randomly, but preferably identical number per class) 2 select one part as test set 3 train the classifier with all observations from remaining N − 1 parts 4 compute the classification rate for the test set 5 repeat until all N parts have been tested 6 overall result: average classification rate module 5.0: data, data splits, and augmentation 5 / 9
  • 20. overview intro data augmentation summary data N-Fold cross validation 2/2 1. 2. 3. 4. Data Split 1 Split 2 Split 3 Split 4 Test Train Train Train Train Test Train Train Train Train Test Train Train Train Train Test module 5.0: data, data splits, and augmentation 6 / 9
  • 21. overview intro data augmentation summary classification interaction of data, features, and classifier training set • training set too small, feature number too large ⇒ overfitting • training set too noisy ⇒ underfitting • training set not representative ⇒ bad classification performance classifier • classifier too complex ⇒ overfitting • poor classifier ⇒ bad classification performance features • poor features ⇒ bad classification performance module 5.0: data, data splits, and augmentation 7 / 9
  • 22. overview intro data augmentation summary classification interaction of data, features, and classifier training set • training set too small, feature number too large ⇒ overfitting • training set too noisy ⇒ underfitting • training set not representative ⇒ bad classification performance classifier • classifier too complex ⇒ overfitting • poor classifier ⇒ bad classification performance features • poor features ⇒ bad classification performance module 5.0: data, data splits, and augmentation 7 / 9
  • 23. overview intro data augmentation summary classification interaction of data, features, and classifier training set • training set too small, feature number too large ⇒ overfitting • training set too noisy ⇒ underfitting • training set not representative ⇒ bad classification performance classifier • classifier too complex ⇒ overfitting • poor classifier ⇒ bad classification performance features • poor features ⇒ bad classification performance module 5.0: data, data splits, and augmentation 7 / 9
  • 24. overview intro data augmentation summary data augmentation if annotated data is insufficient, we can ’cheat’ by increasing the amount of training data ⇒ data augmentation: apply irrelevant transforms to audio data • data segmentation ▶ treat audio snippets as separate observations • quality degradation ▶ add noise and distortion, limit bandwidth, etc. • audio effects ▶ apply reverb, etc. • changing pitch/tempo • combine data ▶ mix different audio inputs together (if labels can be “mixed”) • mask out parts of the signal module 5.0: data, data splits, and augmentation 8 / 9
  • 25. overview intro data augmentation summary data augmentation if annotated data is insufficient, we can ’cheat’ by increasing the amount of training data ⇒ data augmentation: apply irrelevant transforms to audio data • data segmentation ▶ treat audio snippets as separate observations • quality degradation ▶ add noise and distortion, limit bandwidth, etc. • audio effects ▶ apply reverb, etc. • changing pitch/tempo • combine data ▶ mix different audio inputs together (if labels can be “mixed”) • mask out parts of the signal module 5.0: data, data splits, and augmentation 8 / 9
  • 26. overview intro data augmentation summary data augmentation if annotated data is insufficient, we can ’cheat’ by increasing the amount of training data ⇒ data augmentation: apply irrelevant transforms to audio data • data segmentation ▶ treat audio snippets as separate observations • quality degradation ▶ add noise and distortion, limit bandwidth, etc. • audio effects ▶ apply reverb, etc. • changing pitch/tempo • combine data ▶ mix different audio inputs together (if labels can be “mixed”) • mask out parts of the signal module 5.0: data, data splits, and augmentation 8 / 9
  • 27. overview intro data augmentation summary data augmentation if annotated data is insufficient, we can ’cheat’ by increasing the amount of training data ⇒ data augmentation: apply irrelevant transforms to audio data • data segmentation ▶ treat audio snippets as separate observations • quality degradation ▶ add noise and distortion, limit bandwidth, etc. • audio effects ▶ apply reverb, etc. • changing pitch/tempo • combine data ▶ mix different audio inputs together (if labels can be “mixed”) • mask out parts of the signal module 5.0: data, data splits, and augmentation 8 / 9
  • 28. overview intro data augmentation summary data augmentation if annotated data is insufficient, we can ’cheat’ by increasing the amount of training data ⇒ data augmentation: apply irrelevant transforms to audio data • data segmentation ▶ treat audio snippets as separate observations • quality degradation ▶ add noise and distortion, limit bandwidth, etc. • audio effects ▶ apply reverb, etc. • changing pitch/tempo • combine data ▶ mix different audio inputs together (if labels can be “mixed”) • mask out parts of the signal module 5.0: data, data splits, and augmentation 8 / 9
  • 29. overview intro data augmentation summary data augmentation if annotated data is insufficient, we can ’cheat’ by increasing the amount of training data ⇒ data augmentation: apply irrelevant transforms to audio data • data segmentation ▶ treat audio snippets as separate observations • quality degradation ▶ add noise and distortion, limit bandwidth, etc. • audio effects ▶ apply reverb, etc. • changing pitch/tempo • combine data ▶ mix different audio inputs together (if labels can be “mixed”) • mask out parts of the signal module 5.0: data, data splits, and augmentation 8 / 9
  • 30. overview intro data augmentation summary data augmentation if annotated data is insufficient, we can ’cheat’ by increasing the amount of training data ⇒ data augmentation: apply irrelevant transforms to audio data • data segmentation ▶ treat audio snippets as separate observations • quality degradation ▶ add noise and distortion, limit bandwidth, etc. • audio effects ▶ apply reverb, etc. • changing pitch/tempo • combine data ▶ mix different audio inputs together (if labels can be “mixed”) • mask out parts of the signal module 5.0: data, data splits, and augmentation 8 / 9
  • 31. overview intro data augmentation summary summary lecture content data • representative • clean, non-noisy • sufficient data split • train • validation • test cross validation • multiple runs with varying data splits • maximum data utilization module 5.0: data, data splits, and augmentation 9 / 9