↓ Skip to main content

Explainable AI: Interpreting, Explaining and Visualizing Deep Learning

Overview of attention for book
Cover of 'Explainable AI: Interpreting, Explaining and Visualizing Deep Learning'

Table of Contents

  1. Altmetric Badge
    Book Overview
  2. Altmetric Badge
    Chapter 1 Towards Explainable Artificial Intelligence
  3. Altmetric Badge
    Chapter 2 Transparency: Motivations and Challenges
  4. Altmetric Badge
    Chapter 3 Interpretability in Intelligent Systems – A New Concept?
  5. Altmetric Badge
    Chapter 4 Understanding Neural Networks via Feature Visualization: A Survey
  6. Altmetric Badge
    Chapter 5 Interpretable Text-to-Image Synthesis with Hierarchical Semantic Layout Generation
  7. Altmetric Badge
    Chapter 6 Unsupervised Discrete Representation Learning
  8. Altmetric Badge
    Chapter 7 Towards Reverse-Engineering Black-Box Neural Networks
  9. Altmetric Badge
    Chapter 8 Explanations for Attributing Deep Neural Network Predictions
  10. Altmetric Badge
    Chapter 9 Gradient-Based Attribution Methods
  11. Altmetric Badge
    Chapter 10 Layer-Wise Relevance Propagation: An Overview
  12. Altmetric Badge
    Chapter 11 Explaining and Interpreting LSTMs
  13. Altmetric Badge
    Chapter 12 Comparing the Interpretability of Deep Networks via Network Dissection
  14. Altmetric Badge
    Chapter 13 Gradient-Based Vs. Propagation-Based Explanations: An Axiomatic Comparison
  15. Altmetric Badge
    Chapter 14 The (Un)reliability of Saliency Methods
  16. Altmetric Badge
    Chapter 15 Visual Scene Understanding for Autonomous Driving Using Semantic Segmentation
  17. Altmetric Badge
    Chapter 16 Understanding Patch-Based Learning of Video Data by Explaining Predictions
  18. Altmetric Badge
    Chapter 17 Quantum-Chemical Insights from Interpretable Atomistic Neural Networks
  19. Altmetric Badge
    Chapter 18 Interpretable Deep Learning in Drug Discovery
  20. Altmetric Badge
    Chapter 19 NeuralHydrology – Interpreting LSTMs in Hydrology
  21. Altmetric Badge
    Chapter 20 Feature Fallacy: Complications with Interpreting Linear Decoding Weights in fMRI
  22. Altmetric Badge
    Chapter 21 Current Advances in Neural Decoding
  23. Altmetric Badge
    Chapter 22 Software and Application Patterns for Explanation Methods
Attention for Chapter 14: The (Un)reliability of Saliency Methods
Altmetric Badge

Mentioned by

news
1 news outlet

Citations

dimensions_citation
587 Dimensions

Readers on

mendeley
482 Mendeley
You are seeing a free-to-access but limited selection of the activity Altmetric has collected about this research output. Click here to find out more.
Chapter title
The (Un)reliability of Saliency Methods
Chapter number 14
Book title
Explainable AI: Interpreting, Explaining and Visualizing Deep Learning
Published by
Springer, Cham, September 2019
DOI 10.1007/978-3-030-28954-6_14
Book ISBNs
978-3-03-028953-9, 978-3-03-028954-6
Authors

Pieter-Jan Kindermans, Sara Hooker, Julius Adebayo, Maximilian Alber, Kristof T. Schütt, Sven Dähne, Dumitru Erhan, Been Kim

Mendeley readers

Mendeley readers

The data shown below were compiled from readership statistics for 482 Mendeley readers of this research output. Click here to see the associated Mendeley record.

Geographical breakdown

Country Count As %
Unknown 482 100%

Demographic breakdown

Readers by professional status Count As %
Student > Ph. D. Student 123 26%
Student > Master 94 20%
Researcher 66 14%
Student > Bachelor 43 9%
Other 16 3%
Other 34 7%
Unknown 106 22%
Readers by discipline Count As %
Computer Science 228 47%
Engineering 58 12%
Mathematics 10 2%
Physics and Astronomy 8 2%
Medicine and Dentistry 7 1%
Other 48 10%
Unknown 123 26%