SlideShare a Scribd company logo
1 of 19
Download to read offline
Multi-modal NLP Systems
In Healthcare
November, 2021
Jekaterina Novikova
Background: Main Approaches
1. Linguistic and Acoustic Feature extraction and generation
a. Extracting pre-defined and/or hand-crafted features from speech and text
b. Generating automatic representations of speech and language
c. Comparing hand-crafted and automatic representations
d. Hybrid approach - combination of the above
2. Multi-modality
a. Predictive power of linguistic and acoustic information on cognitive impairment
b. Vulnerability and predictive power of different linguistic modalities (syntactic vs lexical)
3. Model development
a. Cross-language detection of cognitive impairment
b. Semi-supervised models in the absence of labels
c. Model with removed age bias
4. Challenges in Model Development: QA and validation
a. Effect of ASR errors on the features and on model predictive power
b. Effect of heterogeneous data on model predictive power
c. Automatic noise removal
Part 1.
Linguistic and Acoustic
Feature Extraction and Generation
● Extracting acoustic features (pitch, energy, pauses) and representations (MFCC) from speech
● Generating transcripts using ASR
● Extracting linguistic features from the transcripts (syntactic / semantic / lexical)
● ML models based on hand-crafted features
the boy is handing the girl &uh cookies and she's
telling him to be quiet i guess.
1a. Extracting hand-crafted features from speech
Motivation: using existing knowledge, generate new features and improve models performance.
Example: Several publications have shown that healthy (HC) and cognitively impaired (AD+MCI) subjects
pause before different kinds of words. Pausing can signify word finding difficulty. We investigate the context in
which pauses occur, not just the pauses themselves.
Our method:
1. Identify which words around the pause contain
the most distinguishing information
2. Extract features from tokens within that range
shown to be the most distinguishing.
3. Improve transcript level classification performance.
1a. Generating new hand-crafted features
Pause
Extraction
Feature
Extraction
... ...
Eyre, B., Balagopalan, A., &
Novikova, J.. Fantastic Features
and Where to Find Them:
Detecting Cognitive Impairment
with a Subsequence
Classification Guided Approach.
In Proceedings of the Sixth
Workshop on Noisy User-generated
Text (W-NUT 2020) at EMNLP
2020
Motivation: it is difficult to get access to large amounts of data that are required
for achieving state-of-the-art performance for ML models. Transfer learning approach
lets achieve higher accuracy due to very deep neural models pre-trained on huge
Amounts of data.
Automatic representations: are extracted from language using pre-trained large
neural language models, such as BERT.
Our method: fine-tune pre-trained models for the tasks of interest, which in our case
is the task of detecting cognitive impairment from language.
Results: ADReSS dataset
1b. Generating automatic representations of speech and language
Features (automatic
representations)
extracted
Balagopalan, A., Eyre, B., Rudzicz, F., & Novikova, J. To BERT or Not To BERT: Comparing Speech and Language-based Approaches for Alzheimer’s Disease
Detection. In Proceedings of INTERSPEECH 2020
1c. Comparing hand-crafted and automatic representations
Balagopalan, A., Novikova, J. Comparing Acoustic-based Approaches for Alzheimer’s Disease Detection. In Proceedings of INTERSPEECH 2021
Motivation: it is difficult to get access to large amounts of audio data
that are required for achieving state-of-the-art performance for ML
models. Transfer learning approach lets achieve higher accuracy on
acoustic data classification.
Automatic representations: are extracted from audio of human
speech using transfer learning methods like wav2vec2.
Our method: compare audio classification approaches and decide
which one is the most promising for the AD detection task.
Results: ADReSS dataset
Motivation: some information that is known to be important for the task of our interest is not encoded in large
pre-trained neural language models.
Our method:
1. Identify linguistic features not encoded well via probing tasks
2. Combine representations from final BERT layer and these features, and finetune
Results: Experiments done with the AD detection task on the DementiaBank dataset. FS1 here are the features
identified as under-represented in a BERT model.
1d. Hybrid approach to feature extraction
Balagopalan, A., Novikova, J. Augmenting BERT Carefully with Underrepresented Linguistic Features. In: NeurIPS Workshop on Machine Learning for Health
ML4H, 2020
Part 2.
Multi-modality
Motivation: linguistic features can be of different modalities, and each modality
has a specific influence and importance in cognitive impairment classification.
Our method:
1. Divide linguistic features into non-overlapping subsets according to their
modalities
2. Let neural networks learn low-dimensional representations that agree with
each other
3. These representations are passed into a classifier network
Results: we illustrate the effectiveness of modality division when our developed ML
model (Control vs AD classification) is seeking to find a consensus among
modalities. Presented modalities are: acoustic, semantic and syntactic.
2a. Multi-modality. Predictive power of linguistic and acoustic
information on cognitive impairment
Zhu, Z., Novikova, J., & Rudzicz, F.. Detecting cognitive impairments by agreeing on interpretations of linguistic features. In Proceedings of the 2019
Conference of the North American Chapter of the Association for Computational Linguistics NAACL 2019
Motivation: Understanding the vulnerability of linguistic features extracted from
noisy text is important.
Our method:
1. Analysing vulnerability of lexical and syntactic features to various levels of
text alterations such as deletion, insertion and substitution
2. Feature significance and the impact of alterations on feature predictive
power
3. Computing coefficient of importance for lexical and syntactic features
separately for several text classification tasks
Results:
● Values of lexical features are easily affected by even slight changes in text.
Syntactic features, however, are more robust to such modifications.
● But lower changes of syntactic features result in stronger effects on
classification performance.
2a. Multi-modality. Vulnerability and predictive power of
different linguistic modalities
J. Novikova, A. Balagopalan, K. Shkaruta and F. Rudzicz. Lexical Features Are More Vulnerable, Syntactic Features Have More Predictive Power. In: The 5th
Workshop on Noisy User-generated Text at EMNLP 2019, Hong Kong, 2019
Part 3.
Model Development
Motivation: Most developments are made in
resource-rich languages (especially English).
Multi-language clinical speech datasets are small.
Our method:
Develop cross-language model:
1. We use Optimal Transport (OT) domain
adaptation systems to adapt French and
Mandarin to English.
2. Utilize out-of-domain, single-speaker, healthy
speech data,
3. Train aphasia detection models on English data,
test on French and Mandarin
Results: Such a model improves aphasia detection over
unilingual baselines and direct feature transfer.
3a. Models. Cross-language detection of cognitive
impairment
Aparna Balagopalan, Jekaterina Novikova, Matthew B A Mcdermott, Bret Nestor, Tristan Naumann, Marzyeh Ghassemi ; Proceedings of the Machine
Learning for Health NeurIPS Workshop, PMLR 116:202-219, 2020
Motivation: acquiring sufficient labeled data can be expensive or difficult,
especially from people with cognitive impairment.
Our method:
Develop Transductive Consensus Networks (TCNs), suitable for
semi-supervised learning:
1. ePhysicians try to produce indistinguishable representations for each
modality
2. Discriminators recognize modal-specific information retained in
representations
3. Classifier trains the networks to make a correct decision
Results: TCNs outperform or align with the best benchmark algorithms given
20 to 200 labeled samples on the Bank Marketing and the DementiaBank
datasets (Controls vs AD classification).
3b. Models. Semi-supervised models in the absence of
labels
Z. Zhu, J. Novikova, and F. Rudzicz. Semi-supervised classification by reaching consensus among modalities. In: NeurIPS Workshop on Interpretability and
Robustness in Audio, Speech, and Language IRASL, Montreal, 2018
Model F1-macro
QDA 0.5243
RF 0.6184
GP 0.6775
MLP 0.7528
CN 0.7998*
Motivation: DNN classifiers are able to estimate age from linguistic features, and
could bias on them to detect dementia.
Our method:
We put forward four fair representation learning models that learn low-dimensional
representations of data samples containing as little age information as possible.
Results: Our best models compromise as little as 2.56% accuracy (on the
DementiaBank dataset) and 1.54% accuracy (on the FamousPeople dataset).
Moreover, they have better fairness scores than statistical adjustment methods
3c. Models. Removing age bias from the model
Zhu, Z., Novikova, J., & Rudzicz, F. (2018). Isolating effects of age with fair representation learning when assessing dementia. arXiv preprint arXiv:1807.07217.
Part 4.
Challenges in Model Development:
QA and Validation
Motivation: Errors in ASR may affect predictive performance of the ML
models.
Our method:
We introduce three types of artificial errors to the manual transcripts of the
DemBank and Healthy Aging datasets:
Deletions. Words are missed in the transcript.
Insertions. New word is introduced.
Substitutions. Word is replaced with another one.
Results: Simulated deletion errors have a strong effect on classification
performance when detecting cognitive impairment from speech and
language.
4a. QA and Validation. Effect of ASR errors
Balagopalan, A., Shkaruta, K., and Novikova, J. Impact of ASR on Alzheimer's Disease Detection: All Errors are Equal, but Deletions are More Equal than Others.
In: The 6th Workshop on Noisy User-generated Text at EMNLP 2020.
Motivation:
Datasets we have are small enough. We need more data to train more accurate models. Previous work shows
that same-task data of healthy participants helps improve AD detection on a single-task dataset of pathological
speech
Our method:
Adding in a large amount of healthy data from different tasks.
Results:
● Increase of up to 9% in F1 scores.
● Effect is especially pronounced when
data come from healthy subjects of
age > 60.
4b. QA and Validation. Effect of heterogeneous data
A.Balagopalan, J.Novikova, F.Rudzicz and M.Ghassemi. The Effect of Heterogeneous Data for Alzheimer's Disease Detection from Speech. In: NeurIPS Workshop
on Machine Learning for Health ML4H, Montreal, 2018
Thank you!

More Related Content

What's hot

LiDeng-BerlinOct2015-ASR-GenDisc-4by3.pptx
LiDeng-BerlinOct2015-ASR-GenDisc-4by3.pptxLiDeng-BerlinOct2015-ASR-GenDisc-4by3.pptx
LiDeng-BerlinOct2015-ASR-GenDisc-4by3.pptx
VishnuRajuV
 
Amharic WSD using WordNet
Amharic WSD using WordNetAmharic WSD using WordNet
Amharic WSD using WordNet
Seid Hassen
 
A word sense disambiguation technique for sinhala
A word sense disambiguation technique  for sinhalaA word sense disambiguation technique  for sinhala
A word sense disambiguation technique for sinhala
Vijayindu Gamage
 
An Improved Approach to Word Sense Disambiguation
An Improved Approach to Word Sense DisambiguationAn Improved Approach to Word Sense Disambiguation
An Improved Approach to Word Sense Disambiguation
Surabhi Verma
 
Natural Language Processing: Definition and Application
Natural Language Processing: Definition and ApplicationNatural Language Processing: Definition and Application
Natural Language Processing: Definition and Application
Stephen Shellman
 

What's hot (18)

Frontiers of Natural Language Processing
Frontiers of Natural Language ProcessingFrontiers of Natural Language Processing
Frontiers of Natural Language Processing
 
NLP
NLPNLP
NLP
 
2106 ACM DIS
2106 ACM DIS2106 ACM DIS
2106 ACM DIS
 
Natural language processing
Natural language processingNatural language processing
Natural language processing
 
LiDeng-BerlinOct2015-ASR-GenDisc-4by3.pptx
LiDeng-BerlinOct2015-ASR-GenDisc-4by3.pptxLiDeng-BerlinOct2015-ASR-GenDisc-4by3.pptx
LiDeng-BerlinOct2015-ASR-GenDisc-4by3.pptx
 
Suggestion Generation for Specific Erroneous Part in a Sentence using Deep Le...
Suggestion Generation for Specific Erroneous Part in a Sentence using Deep Le...Suggestion Generation for Specific Erroneous Part in a Sentence using Deep Le...
Suggestion Generation for Specific Erroneous Part in a Sentence using Deep Le...
 
Artificial Intelligence Notes Unit 4
Artificial Intelligence Notes Unit 4Artificial Intelligence Notes Unit 4
Artificial Intelligence Notes Unit 4
 
Amharic WSD using WordNet
Amharic WSD using WordNetAmharic WSD using WordNet
Amharic WSD using WordNet
 
Word sense dissambiguation
Word sense dissambiguationWord sense dissambiguation
Word sense dissambiguation
 
Word Segmentation and Lexical Normalization for Unsegmented Languages
Word Segmentation and Lexical Normalization for Unsegmented LanguagesWord Segmentation and Lexical Normalization for Unsegmented Languages
Word Segmentation and Lexical Normalization for Unsegmented Languages
 
Applsci 09-02758
Applsci 09-02758Applsci 09-02758
Applsci 09-02758
 
A word sense disambiguation technique for sinhala
A word sense disambiguation technique  for sinhalaA word sense disambiguation technique  for sinhala
A word sense disambiguation technique for sinhala
 
Natural language processing PPT presentation
Natural language processing PPT presentationNatural language processing PPT presentation
Natural language processing PPT presentation
 
Natural language processing
Natural language processingNatural language processing
Natural language processing
 
An Improved Approach to Word Sense Disambiguation
An Improved Approach to Word Sense DisambiguationAn Improved Approach to Word Sense Disambiguation
An Improved Approach to Word Sense Disambiguation
 
Natural Language Processing: Definition and Application
Natural Language Processing: Definition and ApplicationNatural Language Processing: Definition and Application
Natural Language Processing: Definition and Application
 
NLP and its Use in Education
NLP and its Use in EducationNLP and its Use in Education
NLP and its Use in Education
 
Introduction to natural language processing
Introduction to natural language processingIntroduction to natural language processing
Introduction to natural language processing
 

Similar to Multi-modal NLP Systems in Healthcare

Chunker Based Sentiment Analysis and Tense Classification for Nepali Text
Chunker Based Sentiment Analysis and Tense Classification for Nepali TextChunker Based Sentiment Analysis and Tense Classification for Nepali Text
Chunker Based Sentiment Analysis and Tense Classification for Nepali Text
kevig
 
Chunker Based Sentiment Analysis and Tense Classification for Nepali Text
Chunker Based Sentiment Analysis and Tense Classification for Nepali TextChunker Based Sentiment Analysis and Tense Classification for Nepali Text
Chunker Based Sentiment Analysis and Tense Classification for Nepali Text
kevig
 
Intelligent Systems - Predictive Analytics Project
Intelligent Systems - Predictive Analytics ProjectIntelligent Systems - Predictive Analytics Project
Intelligent Systems - Predictive Analytics Project
Shreya Chakrabarti
 
Paper id 23201490
Paper id 23201490Paper id 23201490
Paper id 23201490
IJRAT
 

Similar to Multi-modal NLP Systems in Healthcare (20)

CMPE258 Short story.pptx
CMPE258 Short story.pptxCMPE258 Short story.pptx
CMPE258 Short story.pptx
 
MULTILINGUAL SPEECH IDENTIFICATION USING ARTIFICIAL NEURAL NETWORK
 MULTILINGUAL SPEECH IDENTIFICATION USING ARTIFICIAL NEURAL NETWORK MULTILINGUAL SPEECH IDENTIFICATION USING ARTIFICIAL NEURAL NETWORK
MULTILINGUAL SPEECH IDENTIFICATION USING ARTIFICIAL NEURAL NETWORK
 
The Power of Natural Language Processing (NLP) | Enterprise Wired
The Power of Natural Language Processing (NLP) | Enterprise WiredThe Power of Natural Language Processing (NLP) | Enterprise Wired
The Power of Natural Language Processing (NLP) | Enterprise Wired
 
Cognitive plausibility in learning algorithms
Cognitive plausibility in learning algorithmsCognitive plausibility in learning algorithms
Cognitive plausibility in learning algorithms
 
A Review Paper on Speech Based Emotion Detection Using Deep Learning
A Review Paper on Speech Based Emotion Detection Using Deep LearningA Review Paper on Speech Based Emotion Detection Using Deep Learning
A Review Paper on Speech Based Emotion Detection Using Deep Learning
 
NLP Meetup 2023
NLP Meetup 2023NLP Meetup 2023
NLP Meetup 2023
 
NLP.pptx
NLP.pptxNLP.pptx
NLP.pptx
 
1808.10245v1 (1).pdf
1808.10245v1 (1).pdf1808.10245v1 (1).pdf
1808.10245v1 (1).pdf
 
Chunker Based Sentiment Analysis and Tense Classification for Nepali Text
Chunker Based Sentiment Analysis and Tense Classification for Nepali TextChunker Based Sentiment Analysis and Tense Classification for Nepali Text
Chunker Based Sentiment Analysis and Tense Classification for Nepali Text
 
Chunker Based Sentiment Analysis and Tense Classification for Nepali Text
Chunker Based Sentiment Analysis and Tense Classification for Nepali TextChunker Based Sentiment Analysis and Tense Classification for Nepali Text
Chunker Based Sentiment Analysis and Tense Classification for Nepali Text
 
Intelligent Systems - Predictive Analytics Project
Intelligent Systems - Predictive Analytics ProjectIntelligent Systems - Predictive Analytics Project
Intelligent Systems - Predictive Analytics Project
 
A prior case study of natural language processing on different domain
A prior case study of natural language processing  on different domain A prior case study of natural language processing  on different domain
A prior case study of natural language processing on different domain
 
Paper id 23201490
Paper id 23201490Paper id 23201490
Paper id 23201490
 
srinu.pptx
srinu.pptxsrinu.pptx
srinu.pptx
 
Sentiment analysis by deep learning approaches
Sentiment analysis by deep learning approachesSentiment analysis by deep learning approaches
Sentiment analysis by deep learning approaches
 
NLPinAAC
NLPinAACNLPinAAC
NLPinAAC
 
ppt for IASNLP.pptx
ppt for IASNLP.pptxppt for IASNLP.pptx
ppt for IASNLP.pptx
 
IRJET- Sign Language Recognition using Machine Learning Algorithm
IRJET- Sign Language Recognition using Machine Learning AlgorithmIRJET- Sign Language Recognition using Machine Learning Algorithm
IRJET- Sign Language Recognition using Machine Learning Algorithm
 
Spoken language identification using i-vectors, x-vectors, PLDA and logistic ...
Spoken language identification using i-vectors, x-vectors, PLDA and logistic ...Spoken language identification using i-vectors, x-vectors, PLDA and logistic ...
Spoken language identification using i-vectors, x-vectors, PLDA and logistic ...
 
Natural Language Processing Through Different Classes of Machine Learning
Natural Language Processing Through Different Classes of Machine LearningNatural Language Processing Through Different Classes of Machine Learning
Natural Language Processing Through Different Classes of Machine Learning
 

Recently uploaded

Recently uploaded (14)

BIG DEVELOPMENTS IN LESOTHO(DAMS & MINES
BIG DEVELOPMENTS IN LESOTHO(DAMS & MINESBIG DEVELOPMENTS IN LESOTHO(DAMS & MINES
BIG DEVELOPMENTS IN LESOTHO(DAMS & MINES
 
TSM unit 5 Toxicokinetics seminar by Ansari Aashif Raza.pptx
TSM unit 5 Toxicokinetics seminar by  Ansari Aashif Raza.pptxTSM unit 5 Toxicokinetics seminar by  Ansari Aashif Raza.pptx
TSM unit 5 Toxicokinetics seminar by Ansari Aashif Raza.pptx
 
"I hear you": Moving beyond empathy in UXR
"I hear you": Moving beyond empathy in UXR"I hear you": Moving beyond empathy in UXR
"I hear you": Moving beyond empathy in UXR
 
2024-05-15-Surat Meetup-Hyperautomation.pptx
2024-05-15-Surat Meetup-Hyperautomation.pptx2024-05-15-Surat Meetup-Hyperautomation.pptx
2024-05-15-Surat Meetup-Hyperautomation.pptx
 
Microsoft Fabric Analytics Engineer (DP-600) Exam Dumps 2024.pdf
Microsoft Fabric Analytics Engineer (DP-600) Exam Dumps 2024.pdfMicrosoft Fabric Analytics Engineer (DP-600) Exam Dumps 2024.pdf
Microsoft Fabric Analytics Engineer (DP-600) Exam Dumps 2024.pdf
 
STM valmiusseminaari 26-04-2024 PUUMALAINEN Ajankohtaista kansainvälisestä yh...
STM valmiusseminaari 26-04-2024 PUUMALAINEN Ajankohtaista kansainvälisestä yh...STM valmiusseminaari 26-04-2024 PUUMALAINEN Ajankohtaista kansainvälisestä yh...
STM valmiusseminaari 26-04-2024 PUUMALAINEN Ajankohtaista kansainvälisestä yh...
 
2024 mega trends for the digital workplace - FINAL.pdf
2024 mega trends for the digital workplace - FINAL.pdf2024 mega trends for the digital workplace - FINAL.pdf
2024 mega trends for the digital workplace - FINAL.pdf
 
SaaStr Workshop Wednesday with CEO of Guru
SaaStr Workshop Wednesday with CEO of GuruSaaStr Workshop Wednesday with CEO of Guru
SaaStr Workshop Wednesday with CEO of Guru
 
Using AI to boost productivity for developers
Using AI to boost productivity for developersUsing AI to boost productivity for developers
Using AI to boost productivity for developers
 
BIG DEVELOPMENTS IN LESOTHO(DAMS & MINES
BIG DEVELOPMENTS IN LESOTHO(DAMS & MINESBIG DEVELOPMENTS IN LESOTHO(DAMS & MINES
BIG DEVELOPMENTS IN LESOTHO(DAMS & MINES
 
Modernizing The Transport System:Dhaka Metro Rail
Modernizing The Transport System:Dhaka Metro RailModernizing The Transport System:Dhaka Metro Rail
Modernizing The Transport System:Dhaka Metro Rail
 
Abortion Pills Fahaheel ௹+918133066128💬@ Safe and Effective Mifepristion and ...
Abortion Pills Fahaheel ௹+918133066128💬@ Safe and Effective Mifepristion and ...Abortion Pills Fahaheel ௹+918133066128💬@ Safe and Effective Mifepristion and ...
Abortion Pills Fahaheel ௹+918133066128💬@ Safe and Effective Mifepristion and ...
 
The Concession of Asaba International Airport: Balancing Politics and Policy ...
The Concession of Asaba International Airport: Balancing Politics and Policy ...The Concession of Asaba International Airport: Balancing Politics and Policy ...
The Concession of Asaba International Airport: Balancing Politics and Policy ...
 
Databricks Machine Learning Associate Exam Dumps 2024.pdf
Databricks Machine Learning Associate Exam Dumps 2024.pdfDatabricks Machine Learning Associate Exam Dumps 2024.pdf
Databricks Machine Learning Associate Exam Dumps 2024.pdf
 

Multi-modal NLP Systems in Healthcare

  • 1. Multi-modal NLP Systems In Healthcare November, 2021 Jekaterina Novikova
  • 2. Background: Main Approaches 1. Linguistic and Acoustic Feature extraction and generation a. Extracting pre-defined and/or hand-crafted features from speech and text b. Generating automatic representations of speech and language c. Comparing hand-crafted and automatic representations d. Hybrid approach - combination of the above 2. Multi-modality a. Predictive power of linguistic and acoustic information on cognitive impairment b. Vulnerability and predictive power of different linguistic modalities (syntactic vs lexical) 3. Model development a. Cross-language detection of cognitive impairment b. Semi-supervised models in the absence of labels c. Model with removed age bias 4. Challenges in Model Development: QA and validation a. Effect of ASR errors on the features and on model predictive power b. Effect of heterogeneous data on model predictive power c. Automatic noise removal
  • 3. Part 1. Linguistic and Acoustic Feature Extraction and Generation
  • 4. ● Extracting acoustic features (pitch, energy, pauses) and representations (MFCC) from speech ● Generating transcripts using ASR ● Extracting linguistic features from the transcripts (syntactic / semantic / lexical) ● ML models based on hand-crafted features the boy is handing the girl &uh cookies and she's telling him to be quiet i guess. 1a. Extracting hand-crafted features from speech
  • 5. Motivation: using existing knowledge, generate new features and improve models performance. Example: Several publications have shown that healthy (HC) and cognitively impaired (AD+MCI) subjects pause before different kinds of words. Pausing can signify word finding difficulty. We investigate the context in which pauses occur, not just the pauses themselves. Our method: 1. Identify which words around the pause contain the most distinguishing information 2. Extract features from tokens within that range shown to be the most distinguishing. 3. Improve transcript level classification performance. 1a. Generating new hand-crafted features Pause Extraction Feature Extraction ... ... Eyre, B., Balagopalan, A., & Novikova, J.. Fantastic Features and Where to Find Them: Detecting Cognitive Impairment with a Subsequence Classification Guided Approach. In Proceedings of the Sixth Workshop on Noisy User-generated Text (W-NUT 2020) at EMNLP 2020
  • 6. Motivation: it is difficult to get access to large amounts of data that are required for achieving state-of-the-art performance for ML models. Transfer learning approach lets achieve higher accuracy due to very deep neural models pre-trained on huge Amounts of data. Automatic representations: are extracted from language using pre-trained large neural language models, such as BERT. Our method: fine-tune pre-trained models for the tasks of interest, which in our case is the task of detecting cognitive impairment from language. Results: ADReSS dataset 1b. Generating automatic representations of speech and language Features (automatic representations) extracted Balagopalan, A., Eyre, B., Rudzicz, F., & Novikova, J. To BERT or Not To BERT: Comparing Speech and Language-based Approaches for Alzheimer’s Disease Detection. In Proceedings of INTERSPEECH 2020
  • 7. 1c. Comparing hand-crafted and automatic representations Balagopalan, A., Novikova, J. Comparing Acoustic-based Approaches for Alzheimer’s Disease Detection. In Proceedings of INTERSPEECH 2021 Motivation: it is difficult to get access to large amounts of audio data that are required for achieving state-of-the-art performance for ML models. Transfer learning approach lets achieve higher accuracy on acoustic data classification. Automatic representations: are extracted from audio of human speech using transfer learning methods like wav2vec2. Our method: compare audio classification approaches and decide which one is the most promising for the AD detection task. Results: ADReSS dataset
  • 8. Motivation: some information that is known to be important for the task of our interest is not encoded in large pre-trained neural language models. Our method: 1. Identify linguistic features not encoded well via probing tasks 2. Combine representations from final BERT layer and these features, and finetune Results: Experiments done with the AD detection task on the DementiaBank dataset. FS1 here are the features identified as under-represented in a BERT model. 1d. Hybrid approach to feature extraction Balagopalan, A., Novikova, J. Augmenting BERT Carefully with Underrepresented Linguistic Features. In: NeurIPS Workshop on Machine Learning for Health ML4H, 2020
  • 10. Motivation: linguistic features can be of different modalities, and each modality has a specific influence and importance in cognitive impairment classification. Our method: 1. Divide linguistic features into non-overlapping subsets according to their modalities 2. Let neural networks learn low-dimensional representations that agree with each other 3. These representations are passed into a classifier network Results: we illustrate the effectiveness of modality division when our developed ML model (Control vs AD classification) is seeking to find a consensus among modalities. Presented modalities are: acoustic, semantic and syntactic. 2a. Multi-modality. Predictive power of linguistic and acoustic information on cognitive impairment Zhu, Z., Novikova, J., & Rudzicz, F.. Detecting cognitive impairments by agreeing on interpretations of linguistic features. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics NAACL 2019
  • 11. Motivation: Understanding the vulnerability of linguistic features extracted from noisy text is important. Our method: 1. Analysing vulnerability of lexical and syntactic features to various levels of text alterations such as deletion, insertion and substitution 2. Feature significance and the impact of alterations on feature predictive power 3. Computing coefficient of importance for lexical and syntactic features separately for several text classification tasks Results: ● Values of lexical features are easily affected by even slight changes in text. Syntactic features, however, are more robust to such modifications. ● But lower changes of syntactic features result in stronger effects on classification performance. 2a. Multi-modality. Vulnerability and predictive power of different linguistic modalities J. Novikova, A. Balagopalan, K. Shkaruta and F. Rudzicz. Lexical Features Are More Vulnerable, Syntactic Features Have More Predictive Power. In: The 5th Workshop on Noisy User-generated Text at EMNLP 2019, Hong Kong, 2019
  • 13. Motivation: Most developments are made in resource-rich languages (especially English). Multi-language clinical speech datasets are small. Our method: Develop cross-language model: 1. We use Optimal Transport (OT) domain adaptation systems to adapt French and Mandarin to English. 2. Utilize out-of-domain, single-speaker, healthy speech data, 3. Train aphasia detection models on English data, test on French and Mandarin Results: Such a model improves aphasia detection over unilingual baselines and direct feature transfer. 3a. Models. Cross-language detection of cognitive impairment Aparna Balagopalan, Jekaterina Novikova, Matthew B A Mcdermott, Bret Nestor, Tristan Naumann, Marzyeh Ghassemi ; Proceedings of the Machine Learning for Health NeurIPS Workshop, PMLR 116:202-219, 2020
  • 14. Motivation: acquiring sufficient labeled data can be expensive or difficult, especially from people with cognitive impairment. Our method: Develop Transductive Consensus Networks (TCNs), suitable for semi-supervised learning: 1. ePhysicians try to produce indistinguishable representations for each modality 2. Discriminators recognize modal-specific information retained in representations 3. Classifier trains the networks to make a correct decision Results: TCNs outperform or align with the best benchmark algorithms given 20 to 200 labeled samples on the Bank Marketing and the DementiaBank datasets (Controls vs AD classification). 3b. Models. Semi-supervised models in the absence of labels Z. Zhu, J. Novikova, and F. Rudzicz. Semi-supervised classification by reaching consensus among modalities. In: NeurIPS Workshop on Interpretability and Robustness in Audio, Speech, and Language IRASL, Montreal, 2018 Model F1-macro QDA 0.5243 RF 0.6184 GP 0.6775 MLP 0.7528 CN 0.7998*
  • 15. Motivation: DNN classifiers are able to estimate age from linguistic features, and could bias on them to detect dementia. Our method: We put forward four fair representation learning models that learn low-dimensional representations of data samples containing as little age information as possible. Results: Our best models compromise as little as 2.56% accuracy (on the DementiaBank dataset) and 1.54% accuracy (on the FamousPeople dataset). Moreover, they have better fairness scores than statistical adjustment methods 3c. Models. Removing age bias from the model Zhu, Z., Novikova, J., & Rudzicz, F. (2018). Isolating effects of age with fair representation learning when assessing dementia. arXiv preprint arXiv:1807.07217.
  • 16. Part 4. Challenges in Model Development: QA and Validation
  • 17. Motivation: Errors in ASR may affect predictive performance of the ML models. Our method: We introduce three types of artificial errors to the manual transcripts of the DemBank and Healthy Aging datasets: Deletions. Words are missed in the transcript. Insertions. New word is introduced. Substitutions. Word is replaced with another one. Results: Simulated deletion errors have a strong effect on classification performance when detecting cognitive impairment from speech and language. 4a. QA and Validation. Effect of ASR errors Balagopalan, A., Shkaruta, K., and Novikova, J. Impact of ASR on Alzheimer's Disease Detection: All Errors are Equal, but Deletions are More Equal than Others. In: The 6th Workshop on Noisy User-generated Text at EMNLP 2020.
  • 18. Motivation: Datasets we have are small enough. We need more data to train more accurate models. Previous work shows that same-task data of healthy participants helps improve AD detection on a single-task dataset of pathological speech Our method: Adding in a large amount of healthy data from different tasks. Results: ● Increase of up to 9% in F1 scores. ● Effect is especially pronounced when data come from healthy subjects of age > 60. 4b. QA and Validation. Effect of heterogeneous data A.Balagopalan, J.Novikova, F.Rudzicz and M.Ghassemi. The Effect of Heterogeneous Data for Alzheimer's Disease Detection from Speech. In: NeurIPS Workshop on Machine Learning for Health ML4H, Montreal, 2018