machine learning and ai for the sciences – towards

20
1/20 Müller, Montavon, Samek | ICML Workshop XAI Machine Learning and AI for the sciences – Towards Understanding Part 2: Theory and Extensions Klaus-Robert Müller, Grégoire Montavon , Wojciech Samek ICML 2021 Workshop on XAI

Upload: others

Post on 16-Jul-2022

0 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Machine Learning and AI for the sciences – Towards

1/20Müller, Montavon, Samek | ICML Workshop XAI

Machine Learning and AI for the sciences – Towards Understanding

Part 2: Theory and Extensions

Klaus-Robert Müller, Grégoire Montavon, Wojciech Samek

ICML 2021 Workshop on XAI

Page 2: Machine Learning and AI for the sciences – Towards

2/20Müller, Montavon, Samek | ICML Workshop XAI

Why A Theory of XAI?

Better understand explanation methods, their strength and weaknesses, the underlying assumptions they make and the way they connect.1

2 A good theoretical understanding can be a starting point for new developments (e.g. higher-order explanations).

Page 3: Machine Learning and AI for the sciences – Towards

3/20Müller, Montavon, Samek | ICML Workshop XAI

Example: Can LRP Be Justified Theoretically?

Page 4: Machine Learning and AI for the sciences – Towards

4/20Müller, Montavon, Samek | ICML Workshop XAI

Deep Taylor Decomposition (DTD)

Montavon et al. (2017) Explaining nonlinear classification decisions with deep Taylor decomposition

Montavon et al. Explaining nonlinear classification decisions with deep Taylor decomposition, PatRec 2017

Page 5: Machine Learning and AI for the sciences – Towards

5/20Müller, Montavon, Samek | ICML Workshop XAI

Deep Taylor Decomposition (DTD)

Taylor expansion:

Montavon et al. (2017) Explaining nonlinear classification decisions with deep Taylor decomposition

Montavon et al. Explaining nonlinear classification decisions with deep Taylor decomposition, PatRec 2017

Page 6: Machine Learning and AI for the sciences – Towards

6/20Müller, Montavon, Samek | ICML Workshop XAI

Deep Taylor Decomposition (DTD)

Montavon et al. (2017) Explaining nonlinear classification decisions with deep Taylor decomposition

Taylor expansion:

Montavon et al. Explaining nonlinear classification decisions with deep Taylor decomposition, PatRec 2017

LRP

Page 7: Machine Learning and AI for the sciences – Towards

7/20Müller, Montavon, Samek | ICML Workshop XAI

Two Views on Propagation

DTD view LRP view

Montavon et al. (2017) Explaining nonlinear classification decisions with deep Taylor decomposition

Montavon et al. Explaining nonlinear classification decisions with deep Taylor decomposition, PatRec 2017

Page 8: Machine Learning and AI for the sciences – Towards

8/20Müller, Montavon, Samek | ICML Workshop XAI

Two Views on Propagation

DTD view LRP view

Montavon et al. (2017) Explaining nonlinear classification decisions with deep Taylor decomposition

Montavon et al. Explaining nonlinear classification decisions with deep Taylor decomposition, PatRec 2017

Page 9: Machine Learning and AI for the sciences – Towards

9/20Müller, Montavon, Samek | ICML Workshop XAI

Understand Methods Strengths and Weaknesses

Question: What makes the image below predicted by VGG-16 to be a viaduct?

XAI

LRP-0 LRP-0/γ/ϵ

Montavon et al. Gradient-Based vs. Propagation-Based Explanations: An Axiomatic Comparison, in Explainable AI, Springer LNCS 2019

Page 10: Machine Learning and AI for the sciences – Towards

10/20Müller, Montavon, Samek | ICML Workshop XAI

Understand Methods Strengths and Weaknesses

Question: What makes the image below predicted by VGG-16 to be a viaduct?

XAI

LRP-0 LRP-0/γ/ϵ

general statement?Montavon et al. Gradient-Based vs. Propagation-Based Explanations: An Axiomatic Comparison, in Explainable AI, Springer LNCS 2019

Page 11: Machine Learning and AI for the sciences – Towards

11/20Müller, Montavon, Samek | ICML Workshop XAI

Understand Methods Strengths and Weaknesses

Montavon et al. Gradient-Based vs. Propagation-Based Explanations: An Axiomatic Comparison, in Explainable AI, Springer LNCS 2019

Page 12: Machine Learning and AI for the sciences – Towards

12/20Müller, Montavon, Samek | ICML Workshop XAI

Understand Methods Strengths and Weaknesses

Structurally similar to gradient propagation but with some modifications. → LRP propagation can be seen as smoothing/biasing gradient propagation.

Montavon et al. Gradient-Based vs. Propagation-Based Explanations: An Axiomatic Comparison, in Explainable AI, Springer LNCS 2019

Page 13: Machine Learning and AI for the sciences – Towards

13/20Müller, Montavon, Samek | ICML Workshop XAI

Explaining Deep Similarity Models

Similarity prediction is typically implemented with a product operation in feature space. With this product structure, the similiarity score becomes (locally) bilinear with its input.

Eberle et al. Building and Interpreting Deep Similarity Models, IEEE TPAMI 2020

Page 14: Machine Learning and AI for the sciences – Towards

14/20Müller, Montavon, Samek | ICML Workshop XAI

Second-Order Explanations

Idea: Use the DTD framework to propagate second-order terms of the Taylor expansions at each layer instead of the first-order terms.

Eberle et al. Building and Interpreting Deep Similarity Models, IEEE TPAMI 2020

Page 15: Machine Learning and AI for the sciences – Towards

15/20Müller, Montavon, Samek | ICML Workshop XAI

Second-Order DTD Cast into an LRP Algorithm

Second-order DTD can be cast into multiple LRP computations, followed by an outer product computation. The algorithm is called BiLRP.

Eberle et al. Building and Interpreting Deep Similarity Models, IEEE TPAMI 2020

Page 16: Machine Learning and AI for the sciences – Towards

16/20Müller, Montavon, Samek | ICML Workshop XAI

Examples of Second-Order BiLRP Explanations

BiLRP explanations finds relevant pairs of pixels/patches rather than individual pixels/patches.

Eberle et al. Building and Interpreting Deep Similarity Models, IEEE TPAMI 2020

Page 17: Machine Learning and AI for the sciences – Towards

17/20Müller, Montavon, Samek | ICML Workshop XAI

Explaining GNN Predictions

Graph classification can be achieved using Graph Neural Networks (GNNs).In such architecture, the input graph is applied repeatedly at each layer, and the prediction becomes (locally) polynomial with the input graph.

Schnake et al. Higher-Order Explanations of Graph Neural Networks via Relevant Walks, CoRR 2020

Page 18: Machine Learning and AI for the sciences – Towards

18/20Müller, Montavon, Samek | ICML Workshop XAI

Higher-Order Explanations with DTD

Idea: Use the DTD framework to propagate second-order terms of the Taylor expansions at each layer instead of the first-order terms.

Schnake et al. Higher-Order Explanations of Graph Neural Networks via Relevant Walks, CoRR 2020

Page 19: Machine Learning and AI for the sciences – Towards

19/20Müller, Montavon, Samek | ICML Workshop XAI

Higher-Order DTD Cast into an LRP Algorithm

Higher-order DTD can be cast into multiple LRP computations, one-per walk into the input graph and we call the algorithm GNN-LRP.

Schnake et al. Higher-Order Explanations of Graph Neural Networks via Relevant Walks, CoRR 2020

Page 20: Machine Learning and AI for the sciences – Towards

20/20Müller, Montavon, Samek | ICML Workshop XAI

Example of a GNN-LRP Explanation

Schnake et al. Higher-Order Explanations of Graph Neural Networks via Relevant Walks, CoRR 2020

GNN-LRP explanation explains the prediction in terms of relevant walks into the input graph.