Explainable Artificial Intelligence (XAI)

Report on Current Developments in Explainable Artificial Intelligence (XAI)

General Direction of the Field

The field of Explainable Artificial Intelligence (XAI) is witnessing a significant shift towards more sophisticated and context-specific interpretability methods. Researchers are increasingly focusing on developing frameworks that not only enhance the transparency of deep neural networks (DNNs) but also align more closely with human understanding. This trend is evident in the integration of multimodal approaches, leveraging both visual and linguistic explanations to bridge the semantic gap between machine outputs and human interpretation.

Another notable direction is the automation of complex tasks that traditionally relied on manual intervention, such as the identification of stages of decomposition in forensic science and the classification of underwater SONAR images. These advancements are driven by the need for scalable, reliable, and real-time solutions that can operate in high-stakes environments.

Furthermore, there is a growing emphasis on revisiting and refining evaluation metrics and benchmarks for XAI methods. This includes the development of new metrics to assess the faithfulness and understandability of explanations, as well as the reevaluation of existing benchmarks to ensure they accurately reflect the capabilities of different explainability techniques.

Innovative Work and Results

The field is also seeing innovative approaches to address the limitations of existing XAI methods. For instance, the introduction of model-agnostic surrogate explanations that adapt locality dynamically, ensuring more accurate and consistent explanations without the need for sensitive hyperparameters. Additionally, there is a push towards real-time incremental explanations for object detectors, which significantly reduce computation time while maintaining explanation quality.

Noteworthy Papers

  1. LCE: A Framework for Explainability of DNNs for Ultrasound Image Based on Concept Discovery - Introduces a novel framework that combines attribution and concept-based methods, enhancing the explainability of ultrasound image DNNs.
  2. Towards Automation of Human Stage of Decay Identification: An Artificial Intelligence Approach - Demonstrates the potential of AI models to automate SOD identification with reliability comparable to human experts.
  3. MASALA: Model-Agnostic Surrogate Explanations by Locality Adaptation - Proposes a novel method that automatically determines the appropriate local region for impactful model behavior, improving explanation fidelity and consistency.
  4. VALE: A Multimodal Visual and Language Explanation Framework for Image Classifiers using eXplainable AI and Language Models - Integrates XAI techniques with advanced language models to provide comprehensive and understandable explanations for image classification tasks.

These developments underscore the dynamic and innovative nature of the XAI field, pushing the boundaries of interpretability and reliability in AI systems.

Sources

LCE: A Framework for Explainability of DNNs for Ultrasound Image Based on Concept Discovery

Towards Automation of Human Stage of Decay Identification: An Artificial Intelligence Approach

MASALA: Model-Agnostic Surrogate Explanations by Locality Adaptation

Revisiting FunnyBirds evaluation framework for prototypical parts networks

Real-Time Incremental Explanations for Object Detectors

Underwater SONAR Image Classification and Analysis using LIME-based Explainable Artificial Intelligence

VALE: A Multimodal Visual and Language Explanation Framework for Image Classifiers using eXplainable AI and Language Models

Interpretable breast cancer classification using CNNs on mammographic images

Perturbation on Feature Coalition: Towards Interpretable Deep Neural Networks

Modeling of Terrain Deformation by a Grouser Wheel for Lunar Rover Simulation

Explainable Concept Generation through Vision-Language Preference Learning

Explainable Convolutional Networks for Crater Detection and Lunar Landing Navigation

Explanatory Model Monitoring to Understand the Effects of Feature Shifts on Performance