Skip to content
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.

Commit 1cf00d8

Browse files
authoredApr 8, 2021
Update README.md
1 parent 184fd6d commit 1cf00d8

File tree

1 file changed

+26
-1
lines changed

1 file changed

+26
-1
lines changed
 

‎README.md

Lines changed: 26 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,2 +1,27 @@
11
# GRADCAM-Tensorflow2-Visual-Explainable-AI
2-
Grad-CAM: Visual Explanations from Deep Networks via Gradient-based Localization
2+
### Grad-CAM: Visual Explanations from Deep Networks via Gradient-based Localization
3+
4+
* Install Grad CAM : `!pip install tf-explain`
5+
* src : https://github.com/sicara/tf-explain
6+
* paper : Grad-CAM: Visual Explanations from Deep Networks via Gradient-based Localization
7+
* Reference : https://arxiv.org/abs/1610.02391
8+
* Abstract : We propose a technique for producing "visual explanations" for decisions from a large class
9+
of CNN-based models, making them more transparent. Our approach - Gradient-weighted Class Activation Mapping
10+
(Grad-CAM), uses the gradients of any target concept, flowing into the final convolutional layer to produce
11+
a coarse localization map highlighting important regions in the image for predicting the concept. Grad-CAM
12+
is applicable to a wide variety of CNN model-families:
13+
(1) CNNs with fully-connected layers,
14+
(2) CNNs used for structured outputs,
15+
(3) CNNs used in tasks with multimodal inputs or reinforcement learning,
16+
without any architectural changes or re-training. We combine Grad-CAM with fine-grained visualizations to create
17+
a high-resolution class-discriminative visualization and apply it to off-the-shelf image classification, captioning,
18+
and visual question answering (VQA) models, including ResNet-based architectures. In the context of image classification
19+
models, our visualizations (a) lend insights into their failure modes,
20+
(b) are robust to adversarial images, (c) outperform previous methods on localization, (d) are more faithful to the
21+
underlying model and (e) help achieve generalization by identifying dataset bias. For captioning and VQA, we show that even
22+
non-attention based models can localize inputs. We devise a way to identify important neurons through Grad-CAM and combine it
23+
with neuron names to provide textual explanations for model decisions. Finally, we design and conduct human studies to measure
24+
if Grad-CAM helps users establish appropriate trust in predictions from models and show that Grad-CAM helps untrained users
25+
successfully discern a 'stronger' nodel from a 'weaker' one even when both make identical predictions.
26+
27+
* Note : you can pass `model` object as any tensorflow keras model.

0 commit comments

Comments
 (0)
Please sign in to comment.