Slide 1

Slide 1 text

𝐢!-LRP : Visual Explanation Generation based on Layer-Wise Relevance Propagation for ResNet Felix Doublet, Seitaro Otsuki, Tsumugi Iida, Komei Sugiura Keio University

Slide 2

Slide 2 text

- 2 - Summary : We adapted LRP to handle ResNet models β€’ A new Layer-wise Relevance Propagation (LRP) rule to handle residual connections of ResNet β€’ π‘ͺπŸ‘-LRP that selects the most noteworthy area based on the generated relevance regions. We introduce : LRP [Bach+, PLoS15] GradCAM [Selvaraju+, ICCV17] Ours

Slide 3

Slide 3 text

- 3 - Background : The need for an eXplainable AI Deep Neural Networks (DNNs) = black box nature : β€’ Lack of trust by end-users (medical analysis, autonomous vehicles) β€’ Legal need (RGPD, AI Act) to explain decisions Anastasia Grivia : https://impact.universityofgalway.ie/articles/to- black-box-or-not-to-black-box/ β‡’ Explainable AI (XAI) Layer-Wise Relevance Propagation [Bach+, PLoS15] : β€’ Strong theoretical background β€’ Transparency in Decision-Making β€’ Interpretable Relevance scores β‡’ White Box Analysis

Slide 4

Slide 4 text

- 4 - Background : Generating visual explanations for ResNet with LRP Input Image n Layer-Wise Relevance Propagation (LRP) fails to generate insightful explanations for ResNet models LRP Output LRP FAIL !

Slide 5

Slide 5 text

- 5 - n Layer-Wise Relevance Propagation (LRP) fails to generate insightful explanations for ResNet models LRP Output Residual connections ResNet[He+, CVPR16] No designed rule βœ• Background : Generating visual explanations for ResNet with LRP ???

Slide 6

Slide 6 text

- 6 - n Layer-Wise Relevance Propagation (LRP) fails to generate insightful explanations for ResNet models LRP Output Residual connections ResNet[He+, CVPR16] Tailored rule βœ“ Our method Background : Generating visual explanations for ResNet with LRP

Slide 7

Slide 7 text

- 7 - GradCAM [Selvaraju+, ICCV17] Related work : Back-Propagation methods GradCAM [Selvaraju+, ICCV17] Gradients flowing to produce a coarse localization map. Input x Gradient [Shrikumar+, ICLR18] Multiplying the input image by the gradient of the output. LRP [Bach+, PLoS15] Propagating the prediction backward through the network layers.

Slide 8

Slide 8 text

- 8 - Related work : Layer-Wise Relevance Propagation - an established theoretical framework LRP [Bach+, PloS15] n Layer-wise Relevance Propagation [Bach+, PLoS15] n Established theoretical framework n Transparent computational processes

Slide 9

Slide 9 text

- 9 - : the relevance score of neuron π‘˜ at layer 𝑙 + 1 : activation of neuron π‘˜ : weight of the connection between neuron 𝑖 and neuron π‘˜ Related work : Layer-Wise Relevance Propagation - an established theoretical framework LRP [Bach+, PloS15]

Slide 10

Slide 10 text

- 10 - : the relevance score of neuron π‘˜ at layer 𝑙 + 1 : activation of neuron π‘˜ : weight of the connection between neuron 𝑖 and neuron π‘˜ Related work : Layer-Wise Relevance Propagation - an established theoretical framework LRP [Bach+, PloS15] A rule must be defined for each type of layer

Slide 11

Slide 11 text

- 11 - Proposed Method : The need for a new rule for the ResNet models ResNet50 architecture :

Slide 12

Slide 12 text

- 12 - Proposed Method : The need for a new rule for the ResNet models ResNet50 architecture :

Slide 13

Slide 13 text

- 13 - One bottleneck block: ResNet50 architecture : Proposed Method : The need for a new rule for the ResNet models Need for a tailored backpropagation rule

Slide 14

Slide 14 text

- 14 - IDEA : Considering each bottleneck block as a single dense layer Proposed Method : Relevance backpropagation for Bottleneck layers

Slide 15

Slide 15 text

- 15 - : Relevance of the layer L IDEA : Considering each bottleneck block as a single dense layer Proposed Method : Relevance backpropagation for Bottleneck layers

Slide 16

Slide 16 text

- 16 - : Input of the Bottleneck block : Output of the Bottleneck block 𝐢 : Output channels π‘ˆ : Height 𝑉 : Width : Avoiding a zero division IDEA : Considering each bottleneck block as a single dense layer : Relevance of the layer L Proposed Method : Relevance backpropagation for Bottleneck layers

Slide 17

Slide 17 text

- 17 - Choice Contour Component (𝐢") : Removing unnecessary information Relevance map (LRP output) Proposed Method : Choice Contour Component : π‘ͺπŸ‘

Slide 18

Slide 18 text

- 18 - Choice Contour Component (𝐢") : Removing unnecessary information Relevance map (LRP output) Proposed Method : Choice Contour Component : π‘ͺπŸ‘ 1st Contour (findContours) 1st Component (connectedComponents) C1C [Iida+, SIG-AM23] TRUE Two binary masks are created

Slide 19

Slide 19 text

- 19 - Choice Contour Component (𝐢") : Removing unnecessary information Relevance map (LRP output) Proposed Method : Choice Contour Component : π‘ͺπŸ‘ 1st Contour (findContours) 1st Component (connectedComponents) C1C [Iida+, SIG-AM23] AND TRUE

Slide 20

Slide 20 text

- 20 - Choice Contour Component (𝐢") : Removing unnecessary information Relevance map (LRP output) Proposed Method : Choice Contour Component : π‘ͺπŸ‘ 1st Contour (findContours) 1st Component (connectedComponents) C1C [Iida+, SIG-AM23] AND Final Relevance map Only the pixels from BOTH the 1st contour and the 1st component are kept TRUE

Slide 21

Slide 21 text

- 21 - Experimental setup : The CUB-200-2011 dataset n The CUB-200-2011 dataset [Wah+, 11] : n 11,788 images of birds n 200 classes n Split : 5,794:200:5,794 (train:val:test) Laysan Albatros Least Auklet

Slide 22

Slide 22 text

Method Acc. ↑ Insertion↑ Deletion↓ ID score↑ RISE [Petsuik+, RMVC18] 0.815Β±0.001 0.371Β±0.015 0.043Β±0.004 0.328Β±0.004 GradCAM [Selvaraju+, ICCV19] 0.815Β±0.001 0.466Β±0.019 0.156Β±0.008 0.310Β±0.020 LRP [Bach+, PLoS15] 0.815Β±0.001 0.063Β±0.007 0.051Β±0.006 0.011Β±0.001 ABN [Fukui+, CVPR19] 0.642Β±0.009 0.282Β±0.052 0.075Β±0.011 0.207Β±0.054 Ours 0.815Β±0.001 0.685Β±0.015 0.017Β±0.001 0.668Β±0.015 - 22 - Quantitative results : Our method outperformed the baseline methods

Slide 23

Slide 23 text

- 23 - +𝟎. πŸπŸπŸ— Quantitative results : Our method outperformed the baseline methods

Slide 24

Slide 24 text

- 24 - βˆ’πŸŽ. πŸŽπŸπŸ” Quantitative results : Our method outperformed the baseline methods

Slide 25

Slide 25 text

- 25 - +𝟎. πŸ‘πŸ‘πŸŽ Quantitative results : Our method outperformed the baseline methods

Slide 26

Slide 26 text

- 26 - Qualitative results : Our method provided the most insightful explanations Original Ours ABN [Fukui+, CVPR19] LRP [Bach+, PloS15] GradCAM [Selvaraju+, ICCV17] RISE [Petsuik+, BMVC18]

Slide 27

Slide 27 text

- 27 - Ours Qualitative results : Our method provided the most insightful explanations Β§ Only one wing is attended Β§ The attended area is vague Original ABN [Fukui+, CVPR19] LRP [Bach+, PloS15] GradCAM [Selvaraju+, ICCV17] RISE [Petsuik+, BMVC18]

Slide 28

Slide 28 text

- 28 - Ours Qualitative results : Our method provided the most insightful explanations Β§ The bird is correctly attended Β§ The attended area is vague Original ABN [Fukui+, CVPR19] LRP [Bach+, PloS15] GradCAM [Selvaraju+, ICCV17] RISE [Petsuik+, BMVC18]

Slide 29

Slide 29 text

- 29 - Ours Qualitative results : Our method provided the most insightful explanations Β§ Almost no pixel is attended Β§ Not an insightful explanation Original ABN [Fukui+, CVPR19] LRP [Bach+, PloS15] GradCAM [Selvaraju+, ICCV17] RISE [Petsuik+, BMVC18]

Slide 30

Slide 30 text

- 30 - Ours Qualitative results : Our method provided the most insightful explanations Β§ The bird is correctly attended Β§ The background is also attended Original ABN [Fukui+, CVPR19] LRP [Bach+, PloS15] GradCAM [Selvaraju+, ICCV17] RISE [Petsuik+, BMVC18]

Slide 31

Slide 31 text

- 31 - Ours Qualitative results : Our method provided the most insightful explanations Β§ The bird is correctly attended Β§ The shape of the bird’s body is precisely attended Original ABN [Fukui+, CVPR19] LRP [Bach+, PloS15] GradCAM [Selvaraju+, ICCV17] RISE [Petsuik+, BMVC18]

Slide 32

Slide 32 text

- 32 - Ablation study : π‘ͺπŸ‘ effectively generated high quality explanations +𝟎. πŸ’πŸ‘πŸ”

Slide 33

Slide 33 text

- 33 - Ablation study : π‘ͺπŸ‘ effectively generated high quality explanations +𝟎. πŸŽπŸ”πŸ–

Slide 34

Slide 34 text

- 34 - Conclusion : We successfully generated high quality explanations Contributions: β€’ A method for calculating LRP in models with residual connections β€’ 𝐢!-LRP, which improves the quality of explanations π‘ͺπŸ‘-LRP LRP [Bach+, PloS15] output 𝐢!-LRP output

Slide 35

Slide 35 text

- 35 - Appendix : Failure case example – All the different methods failed to generate an insightful explanation Original Ours ABN [Fukui 19] LRP [Bach 15] GradCAM [Selvaraju 18] RISE [Petsuik 18] β‡’ None of the different methods correctly attend the bird

Slide 36

Slide 36 text

- 36 - Appendix : Error analysis - The method focusing on an insufficient part of the image Error Type IA OA WA #Error 63 21 16 Insufficiently Attended (IA) Over-Attended (OA) Wrongly-Attended (WA) β€’ IA : The area of attention is too small. β€’ OA : The area of relevance is excessively large β€’ WA : The relevance is given to pixels that do not directly contribute to the classification.