SlideShare a Scribd company logo
1 of 9
Download to read offline
Analysis of Hierarchical Multi-Content Text
Classi
fi
cation Model on B-SHARP Dataset

for Early Detection of Alzheimer’s Disease
Asia-Paci
fi
c Chapter of the Association for Computational Linguistic
s

Presented by Jinho D. Cho
i

October 28, 2020
♠
Renxuan A. Li, ♦
Ihab Hajjar, ♦
Felicia Goldstein, ♠
Jinho D. Choi
♠
Department of Computer Science, ♦
Department of Neurolog
y

Emory University, Atlanta GA, USA
jinho.choi@emory.edu
Detection of Alzheimer’s Disease
2
Cerebrospinal Fluid Analysis (CFA)
Positron Emission Tomography (PET)
Mild Cognition Impairment (MCI)
3
Mild Cognitiv
e

Impairment
Mil
d

Dementia
Moderat
e

Dementia
Sever
e

Dementia
Impairment does not Interfere
with activities or daily living
First work to detec
t

MCI using NLP
Impairment starts Interfering
with activities or daily living
B-SHARP Dataset
4
Brain, Stress, Hypertension, and Aging Research Program
Collect 1-2 minute recordings for 3 tasks from MCI patients and Control subjects.
Task 1
:

Daily Activity
Task 2
:

Room Environment
Task 3
:

Picture Description
B-SHARP Dataset
5
1st-visit 2nd-visit 3rd-visit Recordings MoCA BNT
Control 185 100 50 385 26.2 (±2.6) 14.2 (±1.2)
MCI 141 68 28 265 21.5 (±3.5) 13.4 (±1.5)
Total 326 168 78 650 24.2 (±3.8) 13.9 (±1.4)
Subject make multiple visits to take more recordings.
The term between the previous and each visit is 1 year.
Tokens Sentences Nouns Verbs Conjuncts Complex Discourse
Q1
Control 186.6 (±60.4) 10.4 (±4.5) 28.1 (±9.6) 30.4 (±11.5) 8.5 (±4.5) 2.3 (±1.7) 8.1 (±5.4)
MCI 175.6 (±54.5) 9.8 (±4.1) 23.7 (±8.3) 29.3 (±10.4) 8.5 (±4.2) 2.0 (±1.6) 9.2 (±6.0)
Q2
Control 191.5 (±11.8) 11.7 (±4.7) 41.1 (±13.3) 24.3 (±11.2) 6.6 (±4.5) 3.6 (±2.7) 7.1 (±4.8)
MCI 178.6 (±11.7) 11.6 (±4.7) 36.7 (±12.1) 23.2 (±10.6) 6.4 (±4.4) 2.9 (±2.3) 8.4 (±5.3)
Q3
Control 193.4 (±63.4) 12.6 (±5.4) 39.5 (±13.5) 28.4 (±10.1) 8.0 (±4.8) 3.3 (±2.1) 6.1 (±5.5)
MCI 187.8 (±63.4) 12.7 (±5.1) 36.2 (±13.2) 27.7 (±10.9) 7.2 (±4.2) 2.6 (±2.0) 7.3 (±5.5)
All
Control 578.1 (±149.8) 34.5 (±10.7) 110.5 (±27.9) 84.2 (±25.4) 23.5 (±10.1) 9.3 (±4.5) 21.4 (±13.0)
MCI 548.7 (±140.6) 34.0 (±10.5) 98.1 (±26.1) 81.2 (±24.1) 22.5 (±9.7) 7.7 (±4.2) 25.3 (±15.0)
p 0.0110 0.5541 < 0.0001 0.1277 0.2046 < 0.0001 0.0006
Table 1: Average counts and their standard deviations of linguistic features per transcript in the B-SHARP dataset.
Hierarchical Multi-Content Classi
fi
cation
6
w11 w12 ⋯ w1n
[CLS1] w21 w22 ⋯ w2n
[CLS2] w31 w32 ⋯ w3n
[CLS3]
c1 e11 e12 ⋯ e1n c2 e21 e22 ⋯ e2n c3 e31 e32 ⋯ e3n
c1 c2 c3
MLP1 MLP2 MLP3
o2
o1
o3
⊕ ⊕
MLPe oe
Transformer1 (T1) Transformer2 (T2) Transformer3 (T3)
Task 1 Task 2 Task 3
Experiments
7
5-folds Cross Validation
Transformer Encoder
s

BERT (Devlin et al., 2019
)

RoBERTa (Liu et al., 2020
)

ALBERT (Lan et al., 2019)
CV0 CV1 CV2 CV3 CV4 ALL
Control 77 77 77 77 77 385
MCI 53 53 53 53 53 265
Control 37 37 37 37 37 185
MCI 27 28 28 29 29 141
Recordings
Subjects
Subjects in each set are mutually exclusive to the other sets.
Evaluation
8
BERT RoBERTa ALBERT
Q1 Q2 Q3 Q1 Q2 Q3 Q1 Q2 Q3
ACC 67.6 (±0.4) 69.0 (±1.2) 67.7 (±0.7) 69.0 (±1.5) 69.9 (±0.2) 65.2 (±0.3) 67.6 (±1.5) 69.5 (±0.3) 66.6 (±1.3)
SEN 48.9 (±1.8) 57.1 (±2.5) 41.5 (±3.6) 44.3 (±4.5) 55.3 (±1.2) 37.1 (±3.7) 45.9 (±1.9) 52.2 (±0.6) 37.4 (±3.3)
SPE 80.4 (±1.2) 77.3 (±2.8) 85.2 (±3.0) 85.8 (±2.1) 79.7 (±0.7) 84.5 (±3.0) 82.6 (±3.7) 81.4 (±0.3) 86.8 (±3.3)
Table 3: Model performance on the individual tasks. ACC: accuracy, SEN: sensitivity, SPE: specificity.
CNN BERTe RoBERTae ALBERTe Be + Re Ae + Re Be + Ae + Re
ACC 69.5 (±0.2) 69.9 (±1.1) 71.6 (±1.5) 69.7 (±2.9) 72.2 (±0.7) 71.5 (±1.9) 74.1 (±0.3)
SEN 49.2 (±0.8) 57.6 (±3.4) 48.5 (±6.1) 46.2 (±8.3) 56.5 (±2.5) 51.7 (±1.3) 60.9 (±5.2)
SPE 83.5 (±0.9) 77.4 (±4.8) 87.5 (±1.8) 85.4 (±0.5) 83.1 (±0.9) 86.7 (±3.4) 84.0 (±2.4)
Table 4: Performance of ensemble models. Berte/RoBERTae/ALBERTe use transcript embeddings from all 3 tasks
trained by the BERT/RoBERTa/ALBERT models in Table 3, respectively. Be+Re uses transcript embeddings from
both Berte and RoBERTae (so the total of 6 embeddings), Ae+Re uses transcript embeddings from both ALBERTe
and RoBERTae (6 embeddings), and Be+Ae+Re uses transcript embeddings from all three models (9 embeddings).
BERT RoBERTa ALBERT
Q1 Q2 Q3 Q1 Q2 Q3 Q1 Q2 Q3
ACC 67.6 (±0.4) 69.0 (±1.2) 67.7 (±0.7) 69.0 (±1.5) 69.9 (±0.2) 65.2 (±0.3) 67.6 (±1.5) 69.5 (±0.3) 66.6 (±1.3)
SEN 48.9 (±1.8) 57.1 (±2.5) 41.5 (±3.6) 44.3 (±4.5) 55.3 (±1.2) 37.1 (±3.7) 45.9 (±1.9) 52.2 (±0.6) 37.4 (±3.3)
SPE 80.4 (±1.2) 77.3 (±2.8) 85.2 (±3.0) 85.8 (±2.1) 79.7 (±0.7) 84.5 (±3.0) 82.6 (±3.7) 81.4 (±0.3) 86.8 (±3.3)
Table 3: Model performance on the individual tasks. ACC: accuracy, SEN: sensitivity, SPE: specificity.
CNN BERTe RoBERTae ALBERTe Be + Re Ae + Re Be + Ae + Re
ACC 69.5 (±0.2) 69.9 (±1.1) 71.6 (±1.5) 69.7 (±2.9) 72.2 (±0.7) 71.5 (±1.9) 74.1 (±0.3)
SEN 49.2 (±0.8) 57.6 (±3.4) 48.5 (±6.1) 46.2 (±8.3) 56.5 (±2.5) 51.7 (±1.3) 60.9 (±5.2)
SPE 83.5 (±0.9) 77.4 (±4.8) 87.5 (±1.8) 85.4 (±0.5) 83.1 (±0.9) 86.7 (±3.4) 84.0 (±2.4)
Table 4: Performance of ensemble models. Berte/RoBERTae/ALBERTe use transcript embeddings from all 3 tasks
trained by the BERT/RoBERTa/ALBERT models in Table 3, respectively. Be+Re uses transcript embeddings from
both Berte and RoBERTae (so the total of 6 embeddings), Ae+Re uses transcript embeddings from both ALBERTe
and RoBERTae (6 embeddings), and Be+Ae+Re uses transcript embeddings from all three models (9 embeddings).
Performance on the Individual Tasks
Performance of the Ensemble Models
Conclusion
9
Introduced the new dataset, B-SHARP
,

for the detection of Mild Cognitive Impairment (MCI)
Presented Hierarchical Multi-Content Classi
fi
cation Mode
l

to jointly learn multiple documents from different tasks
Achieved the state-of-the-art results with an ensemble mode
l

using three types of transformer encoders
Please visit our lab webpag
e

http://nlp.cs.emory.edu

More Related Content

Similar to Analysis of Hierarchical Multi-Content Text Classification Model on B-SHARP Dataset for Early Detection of Alzheimer’s Disease

Manifold learning for bankruptcy prediction
Manifold learning for bankruptcy predictionManifold learning for bankruptcy prediction
Manifold learning for bankruptcy prediction
Armando Vieira
 
Myocardial Viability - the STICH Trial NEJM May 2011
Myocardial Viability - the STICH Trial NEJM May 2011Myocardial Viability - the STICH Trial NEJM May 2011
Myocardial Viability - the STICH Trial NEJM May 2011
callroom
 
Weight & thickness
Weight & thicknessWeight & thickness
Weight & thickness
poonmaram
 

Similar to Analysis of Hierarchical Multi-Content Text Classification Model on B-SHARP Dataset for Early Detection of Alzheimer’s Disease (20)

Manifold learning for bankruptcy prediction
Manifold learning for bankruptcy predictionManifold learning for bankruptcy prediction
Manifold learning for bankruptcy prediction
 
Nutritional value of soybean meal: Influence of origin and opportunities for ...
Nutritional value of soybean meal: Influence of origin and opportunities for ...Nutritional value of soybean meal: Influence of origin and opportunities for ...
Nutritional value of soybean meal: Influence of origin and opportunities for ...
 
Stated choice design variables - do they play a role on valuation estimates
Stated choice design variables - do they play a role on valuation estimatesStated choice design variables - do they play a role on valuation estimates
Stated choice design variables - do they play a role on valuation estimates
 
Myocardial Viability - the STICH Trial NEJM May 2011
Myocardial Viability - the STICH Trial NEJM May 2011Myocardial Viability - the STICH Trial NEJM May 2011
Myocardial Viability - the STICH Trial NEJM May 2011
 
Reducing Waste in Expandable Collections
Reducing Waste in Expandable CollectionsReducing Waste in Expandable Collections
Reducing Waste in Expandable Collections
 
ESN Presentation
ESN PresentationESN Presentation
ESN Presentation
 
CHI'16 Journal "A Mouse With Two Optical Sensors That Eliminates Coordinate D...
CHI'16 Journal "A Mouse With Two Optical Sensors That Eliminates Coordinate D...CHI'16 Journal "A Mouse With Two Optical Sensors That Eliminates Coordinate D...
CHI'16 Journal "A Mouse With Two Optical Sensors That Eliminates Coordinate D...
 
CT liver segmentation using artificial bee colony optimization
CT liver segmentation using artificial bee colony optimizationCT liver segmentation using artificial bee colony optimization
CT liver segmentation using artificial bee colony optimization
 
Tables nhanes
Tables nhanesTables nhanes
Tables nhanes
 
Manual de alimentación de rumiantes Errata version 2
Manual de alimentación de rumiantes Errata version 2Manual de alimentación de rumiantes Errata version 2
Manual de alimentación de rumiantes Errata version 2
 
Diabetic Retinopathy Detection
Diabetic Retinopathy DetectionDiabetic Retinopathy Detection
Diabetic Retinopathy Detection
 
Whole-body and localized bioelectrical impedance vector analysis (BIVA) in pr...
Whole-body and localized bioelectrical impedance vector analysis (BIVA) in pr...Whole-body and localized bioelectrical impedance vector analysis (BIVA) in pr...
Whole-body and localized bioelectrical impedance vector analysis (BIVA) in pr...
 
chap08_01.ppt
chap08_01.pptchap08_01.ppt
chap08_01.ppt
 
Image Classification
Image ClassificationImage Classification
Image Classification
 
COMPUTING THE GROWTH RATE OF STEM CELLS USING DIGITAL IMAGE PROCESSING
COMPUTING THE GROWTH RATE OF STEM CELLS USING DIGITAL IMAGE PROCESSING COMPUTING THE GROWTH RATE OF STEM CELLS USING DIGITAL IMAGE PROCESSING
COMPUTING THE GROWTH RATE OF STEM CELLS USING DIGITAL IMAGE PROCESSING
 
STUDIES ON INTEGRATED BIO-HYDROGEN PRODUCTION PROCESS-EXPERIMENTAL AND MODELING
STUDIES ON INTEGRATED BIO-HYDROGEN PRODUCTION PROCESS-EXPERIMENTAL AND MODELINGSTUDIES ON INTEGRATED BIO-HYDROGEN PRODUCTION PROCESS-EXPERIMENTAL AND MODELING
STUDIES ON INTEGRATED BIO-HYDROGEN PRODUCTION PROCESS-EXPERIMENTAL AND MODELING
 
Motion Artifact in Ambulatory EEG
Motion Artifact in Ambulatory EEGMotion Artifact in Ambulatory EEG
Motion Artifact in Ambulatory EEG
 
Weight & thickness
Weight & thicknessWeight & thickness
Weight & thickness
 
GEODETIC Least Square with fx 991-es plus
GEODETIC Least Square with fx 991-es plusGEODETIC Least Square with fx 991-es plus
GEODETIC Least Square with fx 991-es plus
 
Primer designgeneprediction
Primer designgenepredictionPrimer designgeneprediction
Primer designgeneprediction
 

More from Jinho Choi

More from Jinho Choi (20)

Adaptation of Multilingual Transformer Encoder for Robust Enhanced Universal ...
Adaptation of Multilingual Transformer Encoder for Robust Enhanced Universal ...Adaptation of Multilingual Transformer Encoder for Robust Enhanced Universal ...
Adaptation of Multilingual Transformer Encoder for Robust Enhanced Universal ...
 
Competence-Level Prediction and Resume & Job Description Matching Using Conte...
Competence-Level Prediction and Resume & Job Description Matching Using Conte...Competence-Level Prediction and Resume & Job Description Matching Using Conte...
Competence-Level Prediction and Resume & Job Description Matching Using Conte...
 
Transformers to Learn Hierarchical Contexts in Multiparty Dialogue for Span-b...
Transformers to Learn Hierarchical Contexts in Multiparty Dialogue for Span-b...Transformers to Learn Hierarchical Contexts in Multiparty Dialogue for Span-b...
Transformers to Learn Hierarchical Contexts in Multiparty Dialogue for Span-b...
 
The Myth of Higher-Order Inference in Coreference Resolution
The Myth of Higher-Order Inference in Coreference ResolutionThe Myth of Higher-Order Inference in Coreference Resolution
The Myth of Higher-Order Inference in Coreference Resolution
 
Noise Pollution in Hospital Readmission Prediction: Long Document Classificat...
Noise Pollution in Hospital Readmission Prediction: Long Document Classificat...Noise Pollution in Hospital Readmission Prediction: Long Document Classificat...
Noise Pollution in Hospital Readmission Prediction: Long Document Classificat...
 
Abstract Meaning Representation
Abstract Meaning RepresentationAbstract Meaning Representation
Abstract Meaning Representation
 
Semantic Role Labeling
Semantic Role LabelingSemantic Role Labeling
Semantic Role Labeling
 
CKY Parsing
CKY ParsingCKY Parsing
CKY Parsing
 
CS329 - WordNet Similarities
CS329 - WordNet SimilaritiesCS329 - WordNet Similarities
CS329 - WordNet Similarities
 
CS329 - Lexical Relations
CS329 - Lexical RelationsCS329 - Lexical Relations
CS329 - Lexical Relations
 
Automatic Knowledge Base Expansion for Dialogue Management
Automatic Knowledge Base Expansion for Dialogue ManagementAutomatic Knowledge Base Expansion for Dialogue Management
Automatic Knowledge Base Expansion for Dialogue Management
 
Attention is All You Need for AMR Parsing
Attention is All You Need for AMR ParsingAttention is All You Need for AMR Parsing
Attention is All You Need for AMR Parsing
 
Graph-to-Text Generation and its Applications to Dialogue
Graph-to-Text Generation and its Applications to DialogueGraph-to-Text Generation and its Applications to Dialogue
Graph-to-Text Generation and its Applications to Dialogue
 
Real-time Coreference Resolution for Dialogue Understanding
Real-time Coreference Resolution for Dialogue UnderstandingReal-time Coreference Resolution for Dialogue Understanding
Real-time Coreference Resolution for Dialogue Understanding
 
Topological Sort
Topological SortTopological Sort
Topological Sort
 
Tries - Put
Tries - PutTries - Put
Tries - Put
 
Multi-modal Embedding Learning for Early Detection of Alzheimer's Disease
Multi-modal Embedding Learning for Early Detection of Alzheimer's DiseaseMulti-modal Embedding Learning for Early Detection of Alzheimer's Disease
Multi-modal Embedding Learning for Early Detection of Alzheimer's Disease
 
Building Widely-Interpretable Semantic Networks for Dialogue Contexts
Building Widely-Interpretable Semantic Networks for Dialogue ContextsBuilding Widely-Interpretable Semantic Networks for Dialogue Contexts
Building Widely-Interpretable Semantic Networks for Dialogue Contexts
 
How to make Emora talk about Sports Intelligently
How to make Emora talk about Sports IntelligentlyHow to make Emora talk about Sports Intelligently
How to make Emora talk about Sports Intelligently
 
Text-to-SQL with Data-Driven Templates
Text-to-SQL with Data-Driven TemplatesText-to-SQL with Data-Driven Templates
Text-to-SQL with Data-Driven Templates
 

Recently uploaded

Cloud Frontiers: A Deep Dive into Serverless Spatial Data and FME
Cloud Frontiers:  A Deep Dive into Serverless Spatial Data and FMECloud Frontiers:  A Deep Dive into Serverless Spatial Data and FME
Cloud Frontiers: A Deep Dive into Serverless Spatial Data and FME
Safe Software
 
Why Teams call analytics are critical to your entire business
Why Teams call analytics are critical to your entire businessWhy Teams call analytics are critical to your entire business
Why Teams call analytics are critical to your entire business
panagenda
 

Recently uploaded (20)

Powerful Google developer tools for immediate impact! (2023-24 C)
Powerful Google developer tools for immediate impact! (2023-24 C)Powerful Google developer tools for immediate impact! (2023-24 C)
Powerful Google developer tools for immediate impact! (2023-24 C)
 
A Year of the Servo Reboot: Where Are We Now?
A Year of the Servo Reboot: Where Are We Now?A Year of the Servo Reboot: Where Are We Now?
A Year of the Servo Reboot: Where Are We Now?
 
Artificial Intelligence Chap.5 : Uncertainty
Artificial Intelligence Chap.5 : UncertaintyArtificial Intelligence Chap.5 : Uncertainty
Artificial Intelligence Chap.5 : Uncertainty
 
Web Form Automation for Bonterra Impact Management (fka Social Solutions Apri...
Web Form Automation for Bonterra Impact Management (fka Social Solutions Apri...Web Form Automation for Bonterra Impact Management (fka Social Solutions Apri...
Web Form Automation for Bonterra Impact Management (fka Social Solutions Apri...
 
Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...
Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...
Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...
 
Apidays New York 2024 - The Good, the Bad and the Governed by David O'Neill, ...
Apidays New York 2024 - The Good, the Bad and the Governed by David O'Neill, ...Apidays New York 2024 - The Good, the Bad and the Governed by David O'Neill, ...
Apidays New York 2024 - The Good, the Bad and the Governed by David O'Neill, ...
 
Strategies for Landing an Oracle DBA Job as a Fresher
Strategies for Landing an Oracle DBA Job as a FresherStrategies for Landing an Oracle DBA Job as a Fresher
Strategies for Landing an Oracle DBA Job as a Fresher
 
Apidays New York 2024 - Scaling API-first by Ian Reasor and Radu Cotescu, Adobe
Apidays New York 2024 - Scaling API-first by Ian Reasor and Radu Cotescu, AdobeApidays New York 2024 - Scaling API-first by Ian Reasor and Radu Cotescu, Adobe
Apidays New York 2024 - Scaling API-first by Ian Reasor and Radu Cotescu, Adobe
 
GenAI Risks & Security Meetup 01052024.pdf
GenAI Risks & Security Meetup 01052024.pdfGenAI Risks & Security Meetup 01052024.pdf
GenAI Risks & Security Meetup 01052024.pdf
 
TrustArc Webinar - Unlock the Power of AI-Driven Data Discovery
TrustArc Webinar - Unlock the Power of AI-Driven Data DiscoveryTrustArc Webinar - Unlock the Power of AI-Driven Data Discovery
TrustArc Webinar - Unlock the Power of AI-Driven Data Discovery
 
Cloud Frontiers: A Deep Dive into Serverless Spatial Data and FME
Cloud Frontiers:  A Deep Dive into Serverless Spatial Data and FMECloud Frontiers:  A Deep Dive into Serverless Spatial Data and FME
Cloud Frontiers: A Deep Dive into Serverless Spatial Data and FME
 
Connector Corner: Accelerate revenue generation using UiPath API-centric busi...
Connector Corner: Accelerate revenue generation using UiPath API-centric busi...Connector Corner: Accelerate revenue generation using UiPath API-centric busi...
Connector Corner: Accelerate revenue generation using UiPath API-centric busi...
 
presentation ICT roal in 21st century education
presentation ICT roal in 21st century educationpresentation ICT roal in 21st century education
presentation ICT roal in 21st century education
 
MINDCTI Revenue Release Quarter One 2024
MINDCTI Revenue Release Quarter One 2024MINDCTI Revenue Release Quarter One 2024
MINDCTI Revenue Release Quarter One 2024
 
Apidays New York 2024 - Accelerating FinTech Innovation by Vasa Krishnan, Fin...
Apidays New York 2024 - Accelerating FinTech Innovation by Vasa Krishnan, Fin...Apidays New York 2024 - Accelerating FinTech Innovation by Vasa Krishnan, Fin...
Apidays New York 2024 - Accelerating FinTech Innovation by Vasa Krishnan, Fin...
 
ProductAnonymous-April2024-WinProductDiscovery-MelissaKlemke
ProductAnonymous-April2024-WinProductDiscovery-MelissaKlemkeProductAnonymous-April2024-WinProductDiscovery-MelissaKlemke
ProductAnonymous-April2024-WinProductDiscovery-MelissaKlemke
 
Exploring the Future Potential of AI-Enabled Smartphone Processors
Exploring the Future Potential of AI-Enabled Smartphone ProcessorsExploring the Future Potential of AI-Enabled Smartphone Processors
Exploring the Future Potential of AI-Enabled Smartphone Processors
 
Why Teams call analytics are critical to your entire business
Why Teams call analytics are critical to your entire businessWhy Teams call analytics are critical to your entire business
Why Teams call analytics are critical to your entire business
 
AXA XL - Insurer Innovation Award Americas 2024
AXA XL - Insurer Innovation Award Americas 2024AXA XL - Insurer Innovation Award Americas 2024
AXA XL - Insurer Innovation Award Americas 2024
 
Axa Assurance Maroc - Insurer Innovation Award 2024
Axa Assurance Maroc - Insurer Innovation Award 2024Axa Assurance Maroc - Insurer Innovation Award 2024
Axa Assurance Maroc - Insurer Innovation Award 2024
 

Analysis of Hierarchical Multi-Content Text Classification Model on B-SHARP Dataset for Early Detection of Alzheimer’s Disease

  • 1. Analysis of Hierarchical Multi-Content Text Classi fi cation Model on B-SHARP Dataset
 for Early Detection of Alzheimer’s Disease Asia-Paci fi c Chapter of the Association for Computational Linguistic s Presented by Jinho D. Cho i October 28, 2020 ♠ Renxuan A. Li, ♦ Ihab Hajjar, ♦ Felicia Goldstein, ♠ Jinho D. Choi ♠ Department of Computer Science, ♦ Department of Neurolog y Emory University, Atlanta GA, USA jinho.choi@emory.edu
  • 2. Detection of Alzheimer’s Disease 2 Cerebrospinal Fluid Analysis (CFA) Positron Emission Tomography (PET)
  • 3. Mild Cognition Impairment (MCI) 3 Mild Cognitiv e Impairment Mil d Dementia Moderat e Dementia Sever e Dementia Impairment does not Interfere with activities or daily living First work to detec t MCI using NLP Impairment starts Interfering with activities or daily living
  • 4. B-SHARP Dataset 4 Brain, Stress, Hypertension, and Aging Research Program Collect 1-2 minute recordings for 3 tasks from MCI patients and Control subjects. Task 1 : Daily Activity Task 2 : Room Environment Task 3 : Picture Description
  • 5. B-SHARP Dataset 5 1st-visit 2nd-visit 3rd-visit Recordings MoCA BNT Control 185 100 50 385 26.2 (±2.6) 14.2 (±1.2) MCI 141 68 28 265 21.5 (±3.5) 13.4 (±1.5) Total 326 168 78 650 24.2 (±3.8) 13.9 (±1.4) Subject make multiple visits to take more recordings. The term between the previous and each visit is 1 year. Tokens Sentences Nouns Verbs Conjuncts Complex Discourse Q1 Control 186.6 (±60.4) 10.4 (±4.5) 28.1 (±9.6) 30.4 (±11.5) 8.5 (±4.5) 2.3 (±1.7) 8.1 (±5.4) MCI 175.6 (±54.5) 9.8 (±4.1) 23.7 (±8.3) 29.3 (±10.4) 8.5 (±4.2) 2.0 (±1.6) 9.2 (±6.0) Q2 Control 191.5 (±11.8) 11.7 (±4.7) 41.1 (±13.3) 24.3 (±11.2) 6.6 (±4.5) 3.6 (±2.7) 7.1 (±4.8) MCI 178.6 (±11.7) 11.6 (±4.7) 36.7 (±12.1) 23.2 (±10.6) 6.4 (±4.4) 2.9 (±2.3) 8.4 (±5.3) Q3 Control 193.4 (±63.4) 12.6 (±5.4) 39.5 (±13.5) 28.4 (±10.1) 8.0 (±4.8) 3.3 (±2.1) 6.1 (±5.5) MCI 187.8 (±63.4) 12.7 (±5.1) 36.2 (±13.2) 27.7 (±10.9) 7.2 (±4.2) 2.6 (±2.0) 7.3 (±5.5) All Control 578.1 (±149.8) 34.5 (±10.7) 110.5 (±27.9) 84.2 (±25.4) 23.5 (±10.1) 9.3 (±4.5) 21.4 (±13.0) MCI 548.7 (±140.6) 34.0 (±10.5) 98.1 (±26.1) 81.2 (±24.1) 22.5 (±9.7) 7.7 (±4.2) 25.3 (±15.0) p 0.0110 0.5541 < 0.0001 0.1277 0.2046 < 0.0001 0.0006 Table 1: Average counts and their standard deviations of linguistic features per transcript in the B-SHARP dataset.
  • 6. Hierarchical Multi-Content Classi fi cation 6 w11 w12 ⋯ w1n [CLS1] w21 w22 ⋯ w2n [CLS2] w31 w32 ⋯ w3n [CLS3] c1 e11 e12 ⋯ e1n c2 e21 e22 ⋯ e2n c3 e31 e32 ⋯ e3n c1 c2 c3 MLP1 MLP2 MLP3 o2 o1 o3 ⊕ ⊕ MLPe oe Transformer1 (T1) Transformer2 (T2) Transformer3 (T3) Task 1 Task 2 Task 3
  • 7. Experiments 7 5-folds Cross Validation Transformer Encoder s BERT (Devlin et al., 2019 ) RoBERTa (Liu et al., 2020 ) ALBERT (Lan et al., 2019) CV0 CV1 CV2 CV3 CV4 ALL Control 77 77 77 77 77 385 MCI 53 53 53 53 53 265 Control 37 37 37 37 37 185 MCI 27 28 28 29 29 141 Recordings Subjects Subjects in each set are mutually exclusive to the other sets.
  • 8. Evaluation 8 BERT RoBERTa ALBERT Q1 Q2 Q3 Q1 Q2 Q3 Q1 Q2 Q3 ACC 67.6 (±0.4) 69.0 (±1.2) 67.7 (±0.7) 69.0 (±1.5) 69.9 (±0.2) 65.2 (±0.3) 67.6 (±1.5) 69.5 (±0.3) 66.6 (±1.3) SEN 48.9 (±1.8) 57.1 (±2.5) 41.5 (±3.6) 44.3 (±4.5) 55.3 (±1.2) 37.1 (±3.7) 45.9 (±1.9) 52.2 (±0.6) 37.4 (±3.3) SPE 80.4 (±1.2) 77.3 (±2.8) 85.2 (±3.0) 85.8 (±2.1) 79.7 (±0.7) 84.5 (±3.0) 82.6 (±3.7) 81.4 (±0.3) 86.8 (±3.3) Table 3: Model performance on the individual tasks. ACC: accuracy, SEN: sensitivity, SPE: specificity. CNN BERTe RoBERTae ALBERTe Be + Re Ae + Re Be + Ae + Re ACC 69.5 (±0.2) 69.9 (±1.1) 71.6 (±1.5) 69.7 (±2.9) 72.2 (±0.7) 71.5 (±1.9) 74.1 (±0.3) SEN 49.2 (±0.8) 57.6 (±3.4) 48.5 (±6.1) 46.2 (±8.3) 56.5 (±2.5) 51.7 (±1.3) 60.9 (±5.2) SPE 83.5 (±0.9) 77.4 (±4.8) 87.5 (±1.8) 85.4 (±0.5) 83.1 (±0.9) 86.7 (±3.4) 84.0 (±2.4) Table 4: Performance of ensemble models. Berte/RoBERTae/ALBERTe use transcript embeddings from all 3 tasks trained by the BERT/RoBERTa/ALBERT models in Table 3, respectively. Be+Re uses transcript embeddings from both Berte and RoBERTae (so the total of 6 embeddings), Ae+Re uses transcript embeddings from both ALBERTe and RoBERTae (6 embeddings), and Be+Ae+Re uses transcript embeddings from all three models (9 embeddings). BERT RoBERTa ALBERT Q1 Q2 Q3 Q1 Q2 Q3 Q1 Q2 Q3 ACC 67.6 (±0.4) 69.0 (±1.2) 67.7 (±0.7) 69.0 (±1.5) 69.9 (±0.2) 65.2 (±0.3) 67.6 (±1.5) 69.5 (±0.3) 66.6 (±1.3) SEN 48.9 (±1.8) 57.1 (±2.5) 41.5 (±3.6) 44.3 (±4.5) 55.3 (±1.2) 37.1 (±3.7) 45.9 (±1.9) 52.2 (±0.6) 37.4 (±3.3) SPE 80.4 (±1.2) 77.3 (±2.8) 85.2 (±3.0) 85.8 (±2.1) 79.7 (±0.7) 84.5 (±3.0) 82.6 (±3.7) 81.4 (±0.3) 86.8 (±3.3) Table 3: Model performance on the individual tasks. ACC: accuracy, SEN: sensitivity, SPE: specificity. CNN BERTe RoBERTae ALBERTe Be + Re Ae + Re Be + Ae + Re ACC 69.5 (±0.2) 69.9 (±1.1) 71.6 (±1.5) 69.7 (±2.9) 72.2 (±0.7) 71.5 (±1.9) 74.1 (±0.3) SEN 49.2 (±0.8) 57.6 (±3.4) 48.5 (±6.1) 46.2 (±8.3) 56.5 (±2.5) 51.7 (±1.3) 60.9 (±5.2) SPE 83.5 (±0.9) 77.4 (±4.8) 87.5 (±1.8) 85.4 (±0.5) 83.1 (±0.9) 86.7 (±3.4) 84.0 (±2.4) Table 4: Performance of ensemble models. Berte/RoBERTae/ALBERTe use transcript embeddings from all 3 tasks trained by the BERT/RoBERTa/ALBERT models in Table 3, respectively. Be+Re uses transcript embeddings from both Berte and RoBERTae (so the total of 6 embeddings), Ae+Re uses transcript embeddings from both ALBERTe and RoBERTae (6 embeddings), and Be+Ae+Re uses transcript embeddings from all three models (9 embeddings). Performance on the Individual Tasks Performance of the Ensemble Models
  • 9. Conclusion 9 Introduced the new dataset, B-SHARP , for the detection of Mild Cognitive Impairment (MCI) Presented Hierarchical Multi-Content Classi fi cation Mode l to jointly learn multiple documents from different tasks Achieved the state-of-the-art results with an ensemble mode l using three types of transformer encoders Please visit our lab webpag e http://nlp.cs.emory.edu