MRI-Based Alzheimer’s Disease Classification Using Deep Learning: A Novel Small-Data Approach
Raja Haseeb
(raja@rit.kaist.ac.kr)
Advisor: Prof. Jong-Hwan Kim
May 26, 2021
M.S. Dissertation
School of Electrical Engineering, KAIST
RIT
Robot Intelligence Technology Laboratory�Challenge for Knowledge Creation and Innovative Technology
1.1 Research Background
1.2 Research Motivation
1.3 Research Outline
2.1 Overall Approach
2.2 Data Augmentation
2.3 Attention Mechanism
2.4 Contrastive Learning
2.5 Classification Network
3.1 Dataset
3.2 Data Augmentation
3.3 Comparison of Various Architectures
3.4 Proposed Architecture
3.5 Results
3.6 Comparison with Existing Methods
Contents
1. Introduction
3
1.1 Research Background
4 / 59
1.1 Research Background
5 / 59
1.1 Research Background
6 / 59
1.1 Research Background
GRU-based (Lee et al., 2019), Non-linear SVM (Rallabandi et al., 2020),
Multi-modal deep learning (Goto et al., 2020), Transfer learning (Khan et al., 2020),
LSTM-based (Hong et al., 2019)
7 / 59
1.1 Research Background
Multimodal and Multiscale Deep Neural Networks (Lu et al., 2018)
Images." Scientific reports 8.1 (2018): 1-13.
8 / 59
1.1 Research Background
Multimodal Deep Learning (Lee et al., 2019)
9 / 59
1.1 Research Background
Architecture design (Basheera et al., 2019)
independent component analysis based segmented gray matter of T2 weighted magnetic resonance imaging with clinical valuation.“
Alzheimer's & Dementia: Translational Research & Clinical Interventions 5 (2019): 974-986.
10 / 59
1.1 Research Background
interface
structural MRI analysis." Informatics in Medicine Unlocked 18 (2020): 100305.
11 / 59
1.1 Research Background
Schematic diagram of proposed approach (Rallabandi et al., 2020)
structural MRI analysis." Informatics in Medicine Unlocked 18 (2020): 100305.
12 / 59
1.1 Research Background
13 / 59
1.1 Research Background
Schematic diagram of proposed approach (Ahmed et al., 2019)
14 / 59
1.1 Research Background
IEEE Access 7 (2019): 72726-72735.
15 / 59
1.2 Research Motivation
16 / 59
1.2 Research Motivation
17 / 59
1.2 Research Motivation
18 / 59
1.2 Research Motivation
19 / 59
1.3 Research Outline
20 / 59
1.3 Research Outline
21 / 59
1.3 Research Outline
Input Data
Data
Augmentation
Contrastive Learning
Classification Network
Results
Traditional
+
PGGAN-Based
Pretraining with SimCLR to learn useful
representations
ResNet-18
+
CBAM
AD
CN
MCI
22 / 59
1.3 Research Outline
23 / 59
2. Proposed framework
24
2.1 Overall Approach
Generated data
Training data
+
AD
CN
MCI
Contrastive
Learning
Classification phase
Pre-training phase
(ResNet-18 + CBAM)
ResNet-18 architecture
ResNet-18 + CBAM
Input
3x3 conv, 64
3x3 conv, 64
3x3 conv, 64
3x3 conv, 64
3x3 conv, 64
3x3 conv, 64
3x3 conv, 64
3x3 conv, 64
3x3 conv, 64
3x3 conv, 64
3x3 conv, 64
3x3 conv, 64
3x3 conv, 64
3x3 conv, 64
3x3 conv, 64
3x3 conv, 64
3x3 conv, 64
FC
Softmax
AvgPool
25 / 59
2.2 Data Augmentation
26 / 59
2.2 Data Augmentation
27 / 59
2.2 Data Augmentation
Generator
Generated data
Training data
Discriminator
Real
or
Fake
28 / 59
2.2 Data Augnmentation
29 / 59
2.2 Data Augnmentation
PGGAN architecture
30 / 59
2.2 Data Augnmentation
Generator and Discriminator architecture
31 / 59
2.2 Data Augnmentation
32 / 59
2.3 Attention Mechanism
.
33 / 59
2.3 Attention Mechanism
.
Channel
Attention
Module
Spatial
Attention
Module
Input Feature
Refined Feature
Convolutional Block Attention Module
34 / 59
2.3 Attention Mechanism
.
Input Feature F
AvgPool
MaxPool
MLP
Channel Attention
Mc
35 / 59
2.3 Attention Mechanism
.
[MaxPool, AvgPool]
Spatial Attention
Ms
Conv
layer
Channel-refined feature F’
36 / 59
2.4 Contrastive Learning
.
37 / 59
2.4 Contrastive Learning
.
An illustration of SimCLR by Google AI Blog
xi
xj
hi
zj
zi
hj
38 / 59
2.4 Contrastive Learning
.
39 / 59
2.4 Contrastive Learning
.
40 / 59
2.5 Classification Network
ResNet-18 architecture
Input
3x3 conv, 64
3x3 conv, 64
3x3 conv, 64
3x3 conv, 64
3x3 conv, 64
3x3 conv, 64
3x3 conv, 64
3x3 conv, 64
3x3 conv, 64
3x3 conv, 64
3x3 conv, 64
3x3 conv, 64
3x3 conv, 64
3x3 conv, 64
3x3 conv, 64
3x3 conv, 64
3x3 conv, 64
FC
Softmax
AvgPool
conv
Previous conv blocks
Mc
Ms
Next
conv blocks
F
F’
F’’
ResBlock + CBAM
41 / 59
2.5 Classification Network
42 / 59
3. Experiments and results
43
3.1 Dataset
Demographic representation of training data
44 / 59
3.2 Data Augmentation
45 / 59
3.2 Data Augmentation
which captures the similarity of
generated images to real ones.
46 / 59
3.2 Data Augmentation
Training data set after augmentation
47 / 59
3.3 Comparison of various architectures
Custom CNN architecture
Input
5x5 conv, 32
5x5 conv, 64
3x3 conv, 128
3x3 conv, 256
3x3 conv, 512
FC1
Softmax
AvgPool
FC2
FC3
FC4
FC5
48 / 59
3.3 Comparison of various architectures
Comparison of various architectures for AD vs. CN classification task
49 / 59
3.4 Proposed Architecture
50 / 59
3.4 Proposed Architecture
51 / 59
3.5 Results
AD vs. CN classification results of proposed framework
52 / 59
3.5 Results
AD vs. CN vs. MCI classification results of proposed framework
53 / 59
3.6 Comparison with existing methods
Study | Total Subjects | Performance | Approach | Data Leakage |
Aderghal et al., 2017 Cheng and Liu, 2017 Korolev et al., 2017 Valliani and Soni, 2017 Senanayake et al., 2018 Li et al., 2017 Basaia et al., 2019 Hon and Khan, 2017 Hosseini et al., 2018 Lin et al., 2018 Taqi et al., 2018 Vu et al., 2017 Vu et al., 2018 Wang et al., 2019 Basheera et al., 2019 Proposed | 815 (T1 MRI) 193 (T1 MRI + PET) 231 (T1 MRI) 417 (T1 MRI) 515 (T1 MRI) 427 (T1 MRI) 646 (T1 MRI) 416 (T1 MRI) 140 (T1 MRI) 417 (T1 MRI) 400 (T2 MRI) 317 (T1 MRI) 400 (T1 MRI) 400 (T1 MRI) 242 (T2 MRI) 164 (T2 MRI) | ACC=0.84 ACC=0.85 ACC=0.80 ACC=0.81 ACC=0.76 ACC=0.88 ACC=0.99 ACC=0.96 ACC=0.99 ACC=0.89 ACC=1.00 ACC=0.85 ACC=0.86 ACC=0.99 ACC=1.00 ACC=0.83 | ROI-based 3D subject-level 3D subject-level 2D slice-level 3D subject-level 3D patch-level 3D subject-level 2D slice-level 3D subject-level ROI-based 2D slice-level 3D subject-level 3D subject-level 3D subject-level 2D slice-level 2D slice-level | None None None None None None Unclear (b) Unclear (a, c) Unclear (a) Unclear (b) Unclear (b) Unclear (a) Clear (a, c) Clear (b) Clear (b) None |
Comparison of AD vs. CN classification task
54 / 59
3.6 Comparison with existing methods
55 / 59
3.6 Comparison with existing methods
Study | Total Subjects | Performance | Approach | Data Leakage |
Valliani and Soni, 2017 Hosseini et al., 2018 Farooq et al., 2017 Vu et al., 2018 Wang et al., 2019 Basheera et al., 2019 Proposed | 660 (T1 MRI) 210 (T1 MRI) 355 (T1 MRI) 615 (T1 MRI) 624 (T1 MRI) 349 (T2 MRI) 246 (T2 MRI) | ACC=0.57 ACC=0.97 ACC=0.99 ACC=0.80 ACC=0.97 ACC-0.86 ACC=0.65 | 2D slice-level 3D subject-level 2D slice-level 3D subject-level 3D subject-level 2D slice-level 2D slice-level | None Unclear (a) Clear (a, c) Clear (a, c) Clear (b) Clear (b) None |
Comparison of AD vs. CN vs. MCI classification task
56 / 59
4. Conclusion and Future Work
57
4. Conclusion
58 / 59
4. Conclusion
59 / 59
4. Future Work
60 / 59
4. Future Work
61 / 59
Thank you!
62