Causal Inference and Bayesian Network Learning

Report on Current Developments in Causal Inference and Bayesian Network Learning

General Direction of the Field

The field of causal inference and Bayesian network learning is witnessing a significant shift towards more robust, scalable, and computationally efficient methods. Recent advancements are characterized by a blend of theoretical innovations and practical implementations that address long-standing challenges in both causal effect estimation and network structure learning.

  1. Model-Based Causal Inference: There is a growing emphasis on learning causal models directly from observational data, rather than relying solely on estimands derived from predefined causal diagrams. This approach, often referred to as "model completion," leverages probabilistic graphical models (PGMs) to infer causal relationships, offering a more flexible and scalable solution, especially for complex models where traditional estimand methods become computationally intractable.

  2. Machine Learning Integration: The integration of machine learning techniques into causal mediation analysis is gaining traction. This includes the development of non-parametric estimation methods that can handle multiple, continuous, or high-dimensional mediators. These methods often rely on advanced machine learning algorithms, such as Riesz learning, to estimate complex density ratios, thereby enabling more accurate and interpretable mediation analyses.

  3. Information-Theoretic Criteria: New information-theoretic criteria for Bayesian network structure learning are emerging, focusing on parsimonious models with good predictive accuracy. These criteria, such as the quotient normalized maximum likelihood (qNML), aim to balance model complexity and predictive performance without the need for adjustable hyperparameters.

  4. Simulation-Based Inference: The field is also seeing advancements in simulation-based inference, particularly in the data-poor regime. Bayesian neural networks are being employed to account for computational uncertainty, leading to well-calibrated posteriors even with limited simulations. This approach is particularly promising for scenarios where simulations are computationally expensive.

  5. Gradient-Free Variational Learning: There is a move towards gradient-free variational learning methods that balance computational efficiency with robust predictive performance. Conditional mixture networks (CMNs) are being explored as a probabilistic alternative to traditional deep learning models, offering advantages in terms of uncertainty quantification and scalability.

  6. Targeted Cause Discovery: Novel machine learning approaches are being developed for targeted cause discovery, aiming to identify both direct and indirect causes of a target variable. These methods leverage neural networks and local-inference strategies to efficiently scale to large-scale systems, outperforming existing causal discovery methods.

  7. Sufficient Representation Learning: The estimation of conditional average treatment effects (CATE) is being advanced through sufficient representation learning. New neural network approaches, such as CrossNet, are being proposed to ensure that learned representations satisfy unconfoundedness assumptions, leading to more effective estimates of treatment effects.

  8. Gaussian Noise Handling: Algorithms for learning linear acyclic causal models are being refined to handle Gaussian noise more effectively. These algorithms aim to reduce time complexity while maintaining the ability to identify distribution-equivalence patterns in linear causal models.

Noteworthy Papers

  • Estimating Causal Effects from Learned Causal Networks: Introduces a model completion approach that outperforms traditional estimand methods, especially for large models.
  • General targeted machine learning for modern causal mediation analysis: Proposes a one-step estimation algorithm that integrates machine learning with non-parametric mediation analysis, addressing high-dimensional mediators.
  • Quotient Normalized Maximum Likelihood Criterion for Learning Bayesian Network Structures: Introduces qNML, a criterion that leads to parsimonious models with good predictive accuracy, free of adjustable hyperparameters.
  • Low-Budget Simulation-Based Inference with Bayesian Neural Networks: Demonstrates the effectiveness of Bayesian neural networks in simulation-based inference, even with limited simulations.
  • Gradient-free variational learning with conditional mixture networks: Shows that conditional mixture networks can achieve competitive predictive accuracy while maintaining computational efficiency.
  • Targeted Cause Discovery with Data-Driven Learning: Proposes a neural network-based approach for identifying both direct and indirect causes, outperforming existing methods in large-scale systems.
  • Estimating Conditional Average Treatment Effects via Sufficient Representation Learning: Introduces CrossNet, a novel neural network approach that ensures unconfoundedness in learned representations for CATE estimation.
  • Learning linear acyclic causal model including Gaussian noise using ancestral relationships: Proposes an algorithm with lower time complexity for learning linear causal models with Gaussian noise.

Sources

Estimating Causal Effects from Learned Causal Networks

General targeted machine learning for modern causal mediation analysis

Quotient Normalized Maximum Likelihood Criterion for Learning Bayesian Network Structures

Low-Budget Simulation-Based Inference with Bayesian Neural Networks

Generalized Naive Bayes

Gradient-free variational learning with conditional mixture networks

Targeted Cause Discovery with Data-Driven Learning

Estimating Conditional Average Treatment Effects via Sufficient Representation Learning

Learning linear acyclic causal model including Gaussian noise using ancestral relationships