Unveiling Neural Network Concepts: A Comprehensive Survey on Explainable Artificial Intelligence
Neural networks can learn complex concepts that are often not easily interpretable. This survey reviews recent methods for explaining the concepts learned by neural networks, ranging from analyzing individual neurons to learning classifiers for entire layers, in order to make neural networks more transparent and easier to control.