BOOK-CHAPTER

Chapter 15. Human-Centered Concept Explanations for Neural Networks

Abstract

Understanding complex machine learning models such as deep neural networks with explanations is crucial in various applications. Many explanations stem from the model perspective, and may not necessarily effectively communicate why the model is making its predictions at the right level of abstraction. For example, providing importance weights to individual pixels in an image can only express which parts of that particular image is important to the model, but humans may prefer an explanation which explains the prediction by concept-based thinking. In this work, we review the emerging area of concept based explanations. We start by introducing concept explanations including the class of Concept Activation Vectors (CAV) which characterize concepts using vectors in appropriate spaces of neural activations, and discuss different properties of useful concepts, and approaches to measure the usefulness of concept vectors. We then discuss approaches to automatically extract concepts, and approaches to address some of their caveats. Finally, we discuss some case studies that showcase the utility of such concept-based explanations in synthetic settings and real world applications.

Keywords:
Abstraction Computer science Perspective (graphical) Artificial intelligence Class (philosophy) Artificial neural network Cognitive science Machine learning Epistemology Psychology

Metrics

8
Cited By
2.01
FWCI (Field Weighted Citation Impact)
0
Refs
0.90
Citation Normalized Percentile
Is in top 1%
Is in top 10%

Citation History

Topics

Explainable Artificial Intelligence (XAI)
Physical Sciences →  Computer Science →  Artificial Intelligence
Machine Learning in Materials Science
Physical Sciences →  Materials Science →  Materials Chemistry
© 2026 ScienceGate Book Chapters — All rights reserved.