Deep Neural Networks have demonstrated remarkable success across various disciplines, primarily due to their ability to learn intricate data representations. However, the inherent semantic nature of these representations remains elusive, posing challenges for the responsible application of Deep Learning methods, particularly in safety-critical domains. In response to this challenge, this special track delves into the critical aspects of global explainability, a subfield of Explainable AI. Generally, the global explainability methods aim to interpret what abstractions have been learned by the network. This can be achieved by analyzing the network’s reliance on specific concepts in general or by examining individual neurons and their functional roles within models. This approach facilitates the elucidation of abstractions learned by the networks. It extends to identifying and interpreting circuits—computational subgraphs within the models that elucidate information flow within complex architectures. Furthermore, global explainability could be employed to explain the local decision-making of machines, termed glocal explainability.

Topics

quantification of interpretability of deep visual representations via network dissection xAI methods
compositional explanations of neurons
labelling neural representations with inverse recognition
automatic description XAI methods for neuron representations in deep vision networks
natural language-based descriptions of deep visual features for XAI
identification and analysis of interpretable subspaces in image representations
magnitude constrained optimization methods for feature visualization in deep neural networks
human-understandable explanations through concept relevance propagation
attribution maps for enhancing the explainability of concept-based features
concept recursive activation factorization methods for xAI
quantitative testing via concept activation vectors
completeness-aware concept-based explanations in deep learning
non-negative concept activation vectors for invertible concept-based explanations in convolutional neural networks
multi-dimensional concept discovery methods for XAI
mechanistic interpretability for automated circuit discovery
brain-inspired modular training for mechanistic interpretability
vision-language mechanistic interpretability xAI methods
mechanistic interpretability for grokking measures

Supported by