SlideShare a Scribd company logo
Contents
1. Class Activation Map (B. Zhou et al., 2015.)
2. Guided Backpropagation(J. T. Springenberg et al., 2015.)
3. Grad-CAM
2
Class Activation Map (B. Zhou et al., 2015.)
• Global Average Pooling (M. Lin et al., 2014.)
• class에 해당하는 weight들로 마지막 conv layer에 weighted sum
• 해당 class가 활성화되는 부분을 확인!
3
Guided Backpropagation (J. T. Springenberg et al., 2015.)
• 기존 backpropagation은 forward의 activation 방식을 따름
• deconvnet (M. D. Zeiler et al., 2014.) 은 gradient의 값으로 activation을 적용
• Guided Backpropagation은 backpropagation + deconvnet
4
Grad-CAM
5
Grad-CAM
- CAM은 GAP(Global Average Pooling)을 사용해야만 하고, 재학습이 필수적임!
→ gradient를 계산해서 그것으로 weighted sum
6
Grad-CAM
- CAM은 GAP(Global Average Pooling)을 사용해야만 하고, 재학습이 필수적임!
→ layer가 깊을수록 뚜렷한 효과가 나타남
7
Grad-CAM
- CAM은 GAP(Global Average Pooling)을 사용해야만 하고, 재학습이 필수적임!
→ gradient에 ReLU를 적용해서 positive값만 사용
8
Grad-CAM
- CAM은 GAP(Global Average Pooling)을 사용해야만 하고, 재학습이 필수적임!
→ 다양한 task에서 사용 가능
9
Captioning Visual Question Answering (VQA)
Grad-CAM
- Guided Backpropagation과 결합하여 사용 가능 (Guided Grad-CAM)
10
Grad-CAM
- Guided Backpropagation과 결합하여 사용 가능 (Guided Grad-CAM)
→ 마찬가지로 다양한 task에 적용 가능
11
Captioning Visual Question Answering (VQA)
Grad-CAM
- Guided Backpropagation과 결합하여 사용 가능 (Guided Grad-CAM)
→ 제대로 예측하지 못하는 부분을 알아내어 문제를 해결할 수 있음!
12
Reference
- R. R. Selvaraju, M. Cogswell, A. Das, R. Vedantam, D. Parikh and D. Batra, “Grad-CAM: Visual Explanations from Deep Networks via Gradient-based Localization”, in
ICCV, 2017.
- R. R. Selvaraju, A. Das, R. Vedantam, M. Cogswell, D. Parikh and D. Batra, “Grad-CAM: Why did you say that? Visual Explanations from Deep Networks via Gradient-
based Localization”, in CORR, abs/1610.02391, 2016.
- B. Zhou, A. Khosla, A. Lapedriza, A. Oliva and A. Torralba, “Learning Deep Features for Discriminative Localization”, in CVPR, 2016.
- M. Lin, Q. Chen, and S. Yan. “Network in network”, in ICLR, 2014.
- J. T. Springenberg, A. Dosovitskiy, T. Brox and M. Riedmiller, “Striving for Simplicity: The All Convolutional Net”, in ICLR, 2015.
- M. D. Zeiler and R. Fergus, “Visualizing and understanding convolutional networks”, in ECCV, 2014.
13

More Related Content

More from Sungchul Kim

Deeplabv1, v2, v3, v3+
Deeplabv1, v2, v3, v3+Deeplabv1, v2, v3, v3+
Deeplabv1, v2, v3, v3+
Sungchul Kim
 
Going Deeper with Convolutions
Going Deeper with ConvolutionsGoing Deeper with Convolutions
Going Deeper with Convolutions
Sungchul Kim
 
Generalized Intersection over Union: A Metric and A Loss for Bounding Box Reg...
Generalized Intersection over Union: A Metric and A Loss for Bounding Box Reg...Generalized Intersection over Union: A Metric and A Loss for Bounding Box Reg...
Generalized Intersection over Union: A Metric and A Loss for Bounding Box Reg...
Sungchul Kim
 
Panoptic Segmentation
Panoptic SegmentationPanoptic Segmentation
Panoptic Segmentation
Sungchul Kim
 
On the Variance of the Adaptive Learning Rate and Beyond
On the Variance of the Adaptive Learning Rate and BeyondOn the Variance of the Adaptive Learning Rate and Beyond
On the Variance of the Adaptive Learning Rate and Beyond
Sungchul Kim
 
A Benchmark for Interpretability Methods in Deep Neural Networks
A Benchmark for Interpretability Methods in Deep Neural NetworksA Benchmark for Interpretability Methods in Deep Neural Networks
A Benchmark for Interpretability Methods in Deep Neural Networks
Sungchul Kim
 
KDGAN: Knowledge Distillation with Generative Adversarial Networks
KDGAN: Knowledge Distillation with Generative Adversarial NetworksKDGAN: Knowledge Distillation with Generative Adversarial Networks
KDGAN: Knowledge Distillation with Generative Adversarial Networks
Sungchul Kim
 
Designing Network Design Spaces
Designing Network Design SpacesDesigning Network Design Spaces
Designing Network Design Spaces
Sungchul Kim
 
Search to Distill: Pearls are Everywhere but not the Eyes
Search to Distill: Pearls are Everywhere but not the EyesSearch to Distill: Pearls are Everywhere but not the Eyes
Search to Distill: Pearls are Everywhere but not the Eyes
Sungchul Kim
 
Supervised Constrastive Learning
Supervised Constrastive LearningSupervised Constrastive Learning
Supervised Constrastive Learning
Sungchul Kim
 
Bootstrap Your Own Latent: A New Approach to Self-Supervised Learning
Bootstrap Your Own Latent: A New Approach to Self-Supervised LearningBootstrap Your Own Latent: A New Approach to Self-Supervised Learning
Bootstrap Your Own Latent: A New Approach to Self-Supervised Learning
Sungchul Kim
 
FickleNet: Weakly and Semi-supervised Semantic Image Segmentation using Stoch...
FickleNet: Weakly and Semi-supervised Semantic Image Segmentation using Stoch...FickleNet: Weakly and Semi-supervised Semantic Image Segmentation using Stoch...
FickleNet: Weakly and Semi-supervised Semantic Image Segmentation using Stoch...
Sungchul Kim
 
Regularizing Class-wise Predictions via Self-knowledge Distillation
Regularizing Class-wise Predictions via Self-knowledge DistillationRegularizing Class-wise Predictions via Self-knowledge Distillation
Regularizing Class-wise Predictions via Self-knowledge Distillation
Sungchul Kim
 
FixMatch: Simplifying Semi-Supervised Learning with Consistency and Confidence
FixMatch: Simplifying Semi-Supervised Learning with Consistency and ConfidenceFixMatch: Simplifying Semi-Supervised Learning with Consistency and Confidence
FixMatch: Simplifying Semi-Supervised Learning with Consistency and Confidence
Sungchul Kim
 
Unsupervised Data Augmentation for Consistency Training
Unsupervised Data Augmentation for Consistency TrainingUnsupervised Data Augmentation for Consistency Training
Unsupervised Data Augmentation for Consistency Training
Sungchul Kim
 

More from Sungchul Kim (15)

Deeplabv1, v2, v3, v3+
Deeplabv1, v2, v3, v3+Deeplabv1, v2, v3, v3+
Deeplabv1, v2, v3, v3+
 
Going Deeper with Convolutions
Going Deeper with ConvolutionsGoing Deeper with Convolutions
Going Deeper with Convolutions
 
Generalized Intersection over Union: A Metric and A Loss for Bounding Box Reg...
Generalized Intersection over Union: A Metric and A Loss for Bounding Box Reg...Generalized Intersection over Union: A Metric and A Loss for Bounding Box Reg...
Generalized Intersection over Union: A Metric and A Loss for Bounding Box Reg...
 
Panoptic Segmentation
Panoptic SegmentationPanoptic Segmentation
Panoptic Segmentation
 
On the Variance of the Adaptive Learning Rate and Beyond
On the Variance of the Adaptive Learning Rate and BeyondOn the Variance of the Adaptive Learning Rate and Beyond
On the Variance of the Adaptive Learning Rate and Beyond
 
A Benchmark for Interpretability Methods in Deep Neural Networks
A Benchmark for Interpretability Methods in Deep Neural NetworksA Benchmark for Interpretability Methods in Deep Neural Networks
A Benchmark for Interpretability Methods in Deep Neural Networks
 
KDGAN: Knowledge Distillation with Generative Adversarial Networks
KDGAN: Knowledge Distillation with Generative Adversarial NetworksKDGAN: Knowledge Distillation with Generative Adversarial Networks
KDGAN: Knowledge Distillation with Generative Adversarial Networks
 
Designing Network Design Spaces
Designing Network Design SpacesDesigning Network Design Spaces
Designing Network Design Spaces
 
Search to Distill: Pearls are Everywhere but not the Eyes
Search to Distill: Pearls are Everywhere but not the EyesSearch to Distill: Pearls are Everywhere but not the Eyes
Search to Distill: Pearls are Everywhere but not the Eyes
 
Supervised Constrastive Learning
Supervised Constrastive LearningSupervised Constrastive Learning
Supervised Constrastive Learning
 
Bootstrap Your Own Latent: A New Approach to Self-Supervised Learning
Bootstrap Your Own Latent: A New Approach to Self-Supervised LearningBootstrap Your Own Latent: A New Approach to Self-Supervised Learning
Bootstrap Your Own Latent: A New Approach to Self-Supervised Learning
 
FickleNet: Weakly and Semi-supervised Semantic Image Segmentation using Stoch...
FickleNet: Weakly and Semi-supervised Semantic Image Segmentation using Stoch...FickleNet: Weakly and Semi-supervised Semantic Image Segmentation using Stoch...
FickleNet: Weakly and Semi-supervised Semantic Image Segmentation using Stoch...
 
Regularizing Class-wise Predictions via Self-knowledge Distillation
Regularizing Class-wise Predictions via Self-knowledge DistillationRegularizing Class-wise Predictions via Self-knowledge Distillation
Regularizing Class-wise Predictions via Self-knowledge Distillation
 
FixMatch: Simplifying Semi-Supervised Learning with Consistency and Confidence
FixMatch: Simplifying Semi-Supervised Learning with Consistency and ConfidenceFixMatch: Simplifying Semi-Supervised Learning with Consistency and Confidence
FixMatch: Simplifying Semi-Supervised Learning with Consistency and Confidence
 
Unsupervised Data Augmentation for Consistency Training
Unsupervised Data Augmentation for Consistency TrainingUnsupervised Data Augmentation for Consistency Training
Unsupervised Data Augmentation for Consistency Training
 

Grad-CAM: Visual Explanations from Deep Networks via Gradient-based Localization

  • 1.
  • 2. Contents 1. Class Activation Map (B. Zhou et al., 2015.) 2. Guided Backpropagation(J. T. Springenberg et al., 2015.) 3. Grad-CAM 2
  • 3. Class Activation Map (B. Zhou et al., 2015.) • Global Average Pooling (M. Lin et al., 2014.) • class에 해당하는 weight들로 마지막 conv layer에 weighted sum • 해당 class가 활성화되는 부분을 확인! 3
  • 4. Guided Backpropagation (J. T. Springenberg et al., 2015.) • 기존 backpropagation은 forward의 activation 방식을 따름 • deconvnet (M. D. Zeiler et al., 2014.) 은 gradient의 값으로 activation을 적용 • Guided Backpropagation은 backpropagation + deconvnet 4
  • 6. Grad-CAM - CAM은 GAP(Global Average Pooling)을 사용해야만 하고, 재학습이 필수적임! → gradient를 계산해서 그것으로 weighted sum 6
  • 7. Grad-CAM - CAM은 GAP(Global Average Pooling)을 사용해야만 하고, 재학습이 필수적임! → layer가 깊을수록 뚜렷한 효과가 나타남 7
  • 8. Grad-CAM - CAM은 GAP(Global Average Pooling)을 사용해야만 하고, 재학습이 필수적임! → gradient에 ReLU를 적용해서 positive값만 사용 8
  • 9. Grad-CAM - CAM은 GAP(Global Average Pooling)을 사용해야만 하고, 재학습이 필수적임! → 다양한 task에서 사용 가능 9 Captioning Visual Question Answering (VQA)
  • 10. Grad-CAM - Guided Backpropagation과 결합하여 사용 가능 (Guided Grad-CAM) 10
  • 11. Grad-CAM - Guided Backpropagation과 결합하여 사용 가능 (Guided Grad-CAM) → 마찬가지로 다양한 task에 적용 가능 11 Captioning Visual Question Answering (VQA)
  • 12. Grad-CAM - Guided Backpropagation과 결합하여 사용 가능 (Guided Grad-CAM) → 제대로 예측하지 못하는 부분을 알아내어 문제를 해결할 수 있음! 12
  • 13. Reference - R. R. Selvaraju, M. Cogswell, A. Das, R. Vedantam, D. Parikh and D. Batra, “Grad-CAM: Visual Explanations from Deep Networks via Gradient-based Localization”, in ICCV, 2017. - R. R. Selvaraju, A. Das, R. Vedantam, M. Cogswell, D. Parikh and D. Batra, “Grad-CAM: Why did you say that? Visual Explanations from Deep Networks via Gradient- based Localization”, in CORR, abs/1610.02391, 2016. - B. Zhou, A. Khosla, A. Lapedriza, A. Oliva and A. Torralba, “Learning Deep Features for Discriminative Localization”, in CVPR, 2016. - M. Lin, Q. Chen, and S. Yan. “Network in network”, in ICLR, 2014. - J. T. Springenberg, A. Dosovitskiy, T. Brox and M. Riedmiller, “Striving for Simplicity: The All Convolutional Net”, in ICLR, 2015. - M. D. Zeiler and R. Fergus, “Visualizing and understanding convolutional networks”, in ECCV, 2014. 13