Explainable AI (XAI)

Report on Current Developments in Explainable AI (XAI)

General Direction of the Field

The field of Explainable Artificial Intelligence (XAI) is currently witnessing a shift towards more sophisticated and versatile methods for interpreting and understanding the decisions made by machine learning models. This shift is driven by the need for greater transparency, trustworthiness, and robustness in AI systems, particularly as these systems are increasingly deployed in critical applications. The recent developments in XAI can be broadly categorized into several key areas:

  1. Time-Domain Explanations for Audio Classifiers: There is a growing emphasis on producing explanations that are not only faithful but also directly interpretable in the domain of the input data. This is particularly evident in audio classification, where methods are being developed to generate explanations in the time domain, enhancing both the quality and interpretability of the explanations.

  2. Explanation-Driven Adversarial Attacks: A novel and critical area of research is the development of adversarial attacks that leverage explainability methods to exploit vulnerabilities in black-box models. These attacks, while concerning, also highlight the importance of robust explainability techniques that do not inadvertently expose models to adversarial threats.

  3. Hybrid Models for Image Classification: The integration of post-hoc and intrinsic methods is gaining traction, particularly in image classification. Hybrid models that combine the strengths of both approaches are being proposed to provide more detailed and interpretable insights into the decision-making processes of deep networks.

  4. Optimal Ablation for Interpretability: The concept of optimal ablation is emerging as a powerful tool for understanding the importance of model components. This method offers both theoretical and empirical advantages over traditional ablation techniques, potentially benefiting a range of interpretability tasks.

  5. Gradient-Free Post-Hoc Explainability: With the rise of large models that are only accessible via query interfaces, there is a growing need for gradient-free explainability methods. Recent advancements in this area focus on distillation-based approaches that can generate saliency-based explanations without requiring access to model gradients.

  6. Abductive Explanations Under Constraints: The field is also addressing the complexities of abductive explanations in the presence of feature constraints. New methods are being developed to handle these constraints more effectively, reducing redundancy and improving the efficiency of explanation generation.

Noteworthy Papers

  • LMAC-TD: Introduces a time-domain explanation method for audio classifiers, significantly improving audio quality while maintaining faithfulness.
  • XSub: Proposes an explanation-driven adversarial attack, demonstrating effectiveness and stealthiness with minimal query costs.
  • InfoDisent: Combines post-hoc and intrinsic methods for image classification, providing detailed atomic components of classification decisions.
  • Optimal Ablation: Proposes a new method for quantifying component importance, showing theoretical and empirical advantages in interpretability tasks.
  • DAX: Develops a gradient-free post-hoc explainability framework, outperforming existing methods across multiple modalities and evaluation metrics.
  • Abductive Explanations: Addresses the complexity and properties of abductive explanations under feature constraints, offering a comprehensive catalogue of explanation types.

Sources

LMAC-TD: Producing Time Domain Explanations for Audio Classifiers

XSub: Explanation-Driven Adversarial Attack against Blackbox Classifiers via Feature Substitution

InfoDisent: Explainability of Image Classification Models by Information Disentanglement

Optimal ablation for interpretability

Gradient-free Post-hoc Explainability Using Distillation Aided Learnable Approach

Abductive explanations of classifiers under constraints: Complexity and properties

Built with on top of