Tuesday 2:10 PM–2:55 PM in Central Park West (6501)

tf-explain: Interpretability for Tensorflow 2.0

Raphaƫl Meudec

Audience level:
Intermediate

Description

Deep learning models now emerge in multiple domains. The question data scientists and users always ask is "Why does it work?". Explaining decisions from neural networks is vital for model improvements and analysis, and users' adoption. In this talk, I will explain interpretability methods implementations with TF2.0 and introduce tf-explain, a TF2.0 library for interpretability.

Abstract

We will explore some research papers on interpretability of neural networks, at different scale: from the ultra-specific with analysis of convolutional filters to more user-friendly input visualizations.

For each method, I'll provide some theoretical explanations (what mathematical operations we are performing), and a Tensorflow 2 implementation to examine in details how to proceed.

Finally, we will go through tf-explain usage, from offline model inspection to training monitoring.

Roadmap :

  1. Convolutional Kernel Filter Visualization
  2. Saliency Maps (Vanilla Gradients, SmoothGrad)
  3. Class Activation Maps
  4. Occlusion Sensitivity
  5. TF-explain Usage

Subscribe to Receive PyData Updates

Subscribe