kth.sePublications KTH
Change search
Link to record
Permanent link

Direct link
Annadani, Yashas
Publications (3 of 3) Show all publications
Annadani, Y., Pawlowski, N., Jennings, J., Bauer, S., Zhang, C. & Gong, W. (2023). BayesDAG: Gradient-Based Posterior Inference for Causal Discovery. In: Advances in Neural Information Processing Systems 36 - 37th Conference on Neural Information Processing Systems, NeurIPS 2023: . Paper presented at 37th Conference on Neural Information Processing Systems, NeurIPS 2023, New Orleans, United States of America, Dec 10 2023 - Dec 16 2023. Neural Information Processing Systems Foundation
Open this publication in new window or tab >>BayesDAG: Gradient-Based Posterior Inference for Causal Discovery
Show others...
2023 (English)In: Advances in Neural Information Processing Systems 36 - 37th Conference on Neural Information Processing Systems, NeurIPS 2023, Neural Information Processing Systems Foundation , 2023Conference paper, Published paper (Refereed)
Abstract [en]

Bayesian causal discovery aims to infer the posterior distribution over causal models from observed data, quantifying epistemic uncertainty and benefiting downstream tasks. However, computational challenges arise due to joint inference over combinatorial space of Directed Acyclic Graphs (DAGs) and nonlinear functions. Despite recent progress towards efficient posterior inference over DAGs, existing methods are either limited to variational inference on node permutation matrices for linear causal models, leading to compromised inference accuracy, or continuous relaxation of adjacency matrices constrained by a DAG regularizer, which cannot ensure resulting graphs are DAGs. In this work, we introduce a scalable Bayesian causal discovery framework based on a combination of stochastic gradient Markov Chain Monte Carlo (SG-MCMC) and Variational Inference (VI) that overcomes these limitations. Our approach directly samples DAGs from the posterior without requiring any DAG regularization, simultaneously draws function parameter samples and is applicable to both linear and nonlinear causal models. To enable our approach, we derive a novel equivalence to the permutation-based DAG learning, which opens up possibilities of using any relaxed gradient estimator defined over permutations. To our knowledge, this is the first framework applying gradient-based MCMC sampling for causal discovery. Empirical evaluation on synthetic and real-world datasets demonstrate our approach's effectiveness compared to state-of-the-art baselines.

Place, publisher, year, edition, pages
Neural Information Processing Systems Foundation, 2023
National Category
Computer Sciences
Identifiers
urn:nbn:se:kth:diva-350180 (URN)2-s2.0-85186355915 (Scopus ID)
Conference
37th Conference on Neural Information Processing Systems, NeurIPS 2023, New Orleans, United States of America, Dec 10 2023 - Dec 16 2023
Note

QC 20240708

Available from: 2024-07-08 Created: 2024-07-08 Last updated: 2024-07-08Bibliographically approved
Tigas, P., Annadani, Y., Ivanova, D. R., Jesson, A., Gal, Y., Foster, A. & Bauer, S. (2023). Differentiable Multi-Target Causal Bayesian Experimental Design. In: Proceedings of the 40th International Conference on Machine Learning, ICML 2023: . Paper presented at 40th International Conference on Machine Learning, ICML 2023, Honolulu, United States of America, Jul 23 2023 - Jul 29 2023 (pp. 34263-34279). ML Research Press
Open this publication in new window or tab >>Differentiable Multi-Target Causal Bayesian Experimental Design
Show others...
2023 (English)In: Proceedings of the 40th International Conference on Machine Learning, ICML 2023, ML Research Press , 2023, p. 34263-34279Conference paper, Published paper (Refereed)
Abstract [en]

We introduce a gradient-based approach for the problem of Bayesian optimal experimental design to learn causal models in a batch setting - a critical component for causal discovery from finite data where interventions can be costly or risky. Existing methods rely on greedy approximations to construct a batch of experiments while using black-box methods to optimize over a single target-state pair to intervene with. In this work, we completely dispose of the black-box optimization techniques and greedy heuristics and instead propose a conceptually simple end-to-end gradient-based optimization procedure to acquire a set of optimal intervention target-state pairs. Such a procedure enables parameterization of the design space to efficiently optimize over a batch of multi-target-state interventions, a setting which has hitherto not been explored due to its complexity. We demonstrate that our proposed method outperforms baselines and existing acquisition strategies in both single-target and multi-target settings across a number of synthetic datasets.

Place, publisher, year, edition, pages
ML Research Press, 2023
National Category
Probability Theory and Statistics
Identifiers
urn:nbn:se:kth:diva-350008 (URN)2-s2.0-85174413460 (Scopus ID)
Conference
40th International Conference on Machine Learning, ICML 2023, Honolulu, United States of America, Jul 23 2023 - Jul 29 2023
Note

QC 20240704

Available from: 2024-07-04 Created: 2024-07-04 Last updated: 2024-07-04Bibliographically approved
Tigas, P., Annadani, Y., Jesson, A., Schölkopf, B., Gal, Y. & Bauer, S. (2022). Interventions, Where and How?: Experimental Design for Causal Models at Scale. In: Advances in Neural Information Processing Systems 35 - 36th Conference on Neural Information Processing Systems, NeurIPS 2022: . Paper presented at 36th Conference on Neural Information Processing Systems, NeurIPS 2022, New Orleans, United States of America, Nov 28 2022 - Dec 9 2022. Neural information processing systems foundation
Open this publication in new window or tab >>Interventions, Where and How?: Experimental Design for Causal Models at Scale
Show others...
2022 (English)In: Advances in Neural Information Processing Systems 35 - 36th Conference on Neural Information Processing Systems, NeurIPS 2022, Neural information processing systems foundation , 2022Conference paper, Published paper (Refereed)
Abstract [en]

Causal discovery from observational and interventional data is challenging due to limited data and non-identifiability: factors that introduce uncertainty in estimating the underlying structural causal model (SCM). Selecting experiments (interventions) based on the uncertainty arising from both factors can expedite the identification of the SCM. Existing methods in experimental design for causal discovery from limited data either rely on linear assumptions for the SCM or select only the intervention target. This work incorporates recent advances in Bayesian causal discovery into the Bayesian optimal experimental design framework, allowing for active causal discovery of large, nonlinear SCMs while selecting both the interventional target and the value. We demonstrate the performance of the proposed method on synthetic graphs (Erdos-Rènyi, Scale Free) for both linear and nonlinear SCMs as well as on the in-silico single-cell gene regulatory network dataset, DREAM.

Place, publisher, year, edition, pages
Neural information processing systems foundation, 2022
National Category
Signal Processing Computer Sciences
Identifiers
urn:nbn:se:kth:diva-333403 (URN)2-s2.0-85148435478 (Scopus ID)
Conference
36th Conference on Neural Information Processing Systems, NeurIPS 2022, New Orleans, United States of America, Nov 28 2022 - Dec 9 2022
Note

Part of ISBN 9781713871088

QC 20230801

Available from: 2023-08-01 Created: 2023-08-01 Last updated: 2023-08-01Bibliographically approved
Organisations

Search in DiVA

Show all publications