|
1 | 1 | # GRADCAM-Tensorflow2-Visual-Explainable-AI
|
2 |
| -Grad-CAM: Visual Explanations from Deep Networks via Gradient-based Localization |
| 2 | +### Grad-CAM: Visual Explanations from Deep Networks via Gradient-based Localization |
| 3 | + |
| 4 | +* Install Grad CAM : `!pip install tf-explain` |
| 5 | +* src : https://github.com/sicara/tf-explain |
| 6 | +* paper : Grad-CAM: Visual Explanations from Deep Networks via Gradient-based Localization |
| 7 | +* Reference : https://arxiv.org/abs/1610.02391 |
| 8 | +* Abstract : We propose a technique for producing "visual explanations" for decisions from a large class |
| 9 | + of CNN-based models, making them more transparent. Our approach - Gradient-weighted Class Activation Mapping |
| 10 | + (Grad-CAM), uses the gradients of any target concept, flowing into the final convolutional layer to produce |
| 11 | + a coarse localization map highlighting important regions in the image for predicting the concept. Grad-CAM |
| 12 | + is applicable to a wide variety of CNN model-families: |
| 13 | + (1) CNNs with fully-connected layers, |
| 14 | + (2) CNNs used for structured outputs, |
| 15 | + (3) CNNs used in tasks with multimodal inputs or reinforcement learning, |
| 16 | + without any architectural changes or re-training. We combine Grad-CAM with fine-grained visualizations to create |
| 17 | + a high-resolution class-discriminative visualization and apply it to off-the-shelf image classification, captioning, |
| 18 | + and visual question answering (VQA) models, including ResNet-based architectures. In the context of image classification |
| 19 | + models, our visualizations (a) lend insights into their failure modes, |
| 20 | + (b) are robust to adversarial images, (c) outperform previous methods on localization, (d) are more faithful to the |
| 21 | + underlying model and (e) help achieve generalization by identifying dataset bias. For captioning and VQA, we show that even |
| 22 | + non-attention based models can localize inputs. We devise a way to identify important neurons through Grad-CAM and combine it |
| 23 | + with neuron names to provide textual explanations for model decisions. Finally, we design and conduct human studies to measure |
| 24 | + if Grad-CAM helps users establish appropriate trust in predictions from models and show that Grad-CAM helps untrained users |
| 25 | + successfully discern a 'stronger' nodel from a 'weaker' one even when both make identical predictions. |
| 26 | + |
| 27 | +* Note : you can pass `model` object as any tensorflow keras model. |
0 commit comments