Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
research-article
Free access

Inference for a large directed acyclic graph with unspecified interventions

Published: 01 January 2023 Publication History

Abstract

Statistical inference of directed relations given some unspecified interventions (i.e., the intervention targets are unknown) is challenging. In this article, we test hypothesized directed relations with unspecified interventions. First, we derive conditions to yield an identifiable model. Unlike classical inference, testing directed relations requires identifying the ancestors and relevant interventions of hypothesis-specific primary variables. To this end, we propose a peeling algorithm based on nodewise regressions to establish a topological order of primary variables. Moreover, we prove that the peeling algorithm yields a consistent estimator in low-order polynomial time. Second, we propose a likelihood ratio test integrated with a data perturbation scheme to account for the uncertainty of identifying ancestors and interventions. Also, we show that the distribution of a data perturbation test statistic converges to the target distribution. Numerical examples demonstrate the utility and effectiveness of the proposed methods, including an application to infer gene regulatory networks. The R implementation is available at https://github.com/chunlinli/intdag.

References

[1]
Joshua D Angrist, Guido W Imbens, and Donald B Rubin. Identification of causal effects using instrumental variables. Journal of the American Statistical Association, 91(434): 444-455, 1996.
[2]
Til Ole Bergmann and Gesa Hartwigsen. Inferring causality from noninvasive brain stimulation in cognitive neuroscience. Journal of Cognitive Neuroscience, 33(2):195-225, 2021.
[3]
Peter J Bickel, Ya'acov Ritov, and Alexandre B Tsybakov. Simultaneous analysis of Lasso and Dantzig selector. The Annals of Statistics, 37(4):1705-1732, 2009.
[4]
Patrick Breheny and Jian Huang. Coordinate descent algorithms for nonconvex penalized regression, with applications to biological feature selection. The Annals of Applied Statistics, 5(1):232-253, 2011.
[5]
Leo Breiman. The little bootstrap and other methods for dimensionality selection in regression: X-fixed prediction error. Journal of the American Statistical Association, 87(419): 738-754, 1992.
[6]
Brielin C Brown and David A Knowles. Phenome-scale causal network discovery with bidirectional mediated Mendelian randomization. bioRxiv, 2020.
[7]
Guojun Bu. Apolipoprotein E and its receptors in Alzheimer's disease: pathways, pathogenesis and therapy. Nature Reviews Neuroscience, 10(5):333-344, 2009.
[8]
Emmanuel Candes, Yingying Fan, Lucas Janson, and Jinchi Lv. Panning for gold: 'model-X' knockoffs for high dimensional controlled variable selection. Journal of the Royal Statistical Society: Series B (Statistical Methodology), 80(3):551-577, 2018.
[9]
Chen Chen, Min Ren, Min Zhang, and Dabao Zhang. A two-stage penalized least squares method for constructing large systems of structural equations. Journal of Machine Learning Research, 19(1):40-73, 2018.
[10]
Daniel Eaton and Kevin Murphy. Exact Bayesian structure learning from uncertain interventions. In International Conference on Artificial Intelligence and Statistics, pages 107-114. PMLR, 2007.
[11]
Bradley Efron, Trevor Hastie, Iain Johnstone, and Robert Tibshirani. Least angle regression. The Annals of Statistics, 32(2):407-499, 2004.
[12]
Jianqing Fan, Lingzhou Xue, and Hui Zou. Strong oracle optimality of folded concave penalized estimation. The Annals of Statistics, 42(3):819-849, 2014.
[13]
Jerome H. Friedman, Trevor Hastie, and Rob Tibshirani. Regularization paths for generalized linear models via coordinate descent. Journal of Statistical Software, 33(1):1-22, 2010.
[14]
Nir Friedman and Daphne Koller. Being Bayesian about network structure. A Bayesian approach to structure discovery in Bayesian networks. Machine Learning, 50(1):95-125, 2003.
[15]
Asish Ghoshal and Jean Honorio. Learning linear structural equation models in polynomial time and sample complexity. In International Conference on Artificial Intelligence and Statistics, pages 1466-1475. PMLR, 2018.
[16]
Moritz Grosse-Wentrup, Dominik Janzing, Markus Siegel, and Bernhard Schölkopf. Identification of causal relations in neuroimaging data with latent confounders: An instrumental variable approach. NeuroImage, 125:825-833, 2016.
[17]
Christina Heinze-Deml, Marloes H Maathuis, and Nicolai Meinshausen. Causal structure learning. Annual Review of Statistics and Its Application, 5:371-391, 2018.
[18]
Aimee L Jackson, Steven R Bartz, Janell Schelter, Sumire V Kobayashi, Julja Burchard, Mao Mao, Bin Li, Guy Cavet, and Peter S Linsley. Expression profiling reveals off-target gene regulation by RNAi. Nature Biotechnology, 21(6):635-637, 2003.
[19]
Jana Janková and Sara van de Geer. Inference in high-dimensional graphical models. In Handbook of Graphical Models, pages 325-350. CRC Press, 2018.
[20]
TCW Julia and Alison M Goate. Genetics of β-amyloid precursor protein in Alzheimer's disease. Cold Spring Harbor Perspectives in Medicine, 7(6), 2017.
[21]
Minoru Kanehisa and Susumu Goto. KEGG: Kyoto encyclopedia of genes and genomes. Nucleic Acids Research, 28(1):27-30, 2000.
[22]
Raymond J Kelleher III and Jie Shen. Presenilin-1 mutations and Alzheimer's disease. Proceedings of the National Academy of Sciences, 114(4):629-631, 2017.
[23]
Ariel Kleiner, Ameet Talwalkar, Purnamrita Sarkar, and Michael I Jordan. The big data bootstrap. In Proceedings of the 29th International Coference on International Conference on Machine Learning, pages 1787-1794, 2012.
[24]
Anna Kremer, Justin V Louis, Tomasz Jaworski, and Fred Van Leuven. GSK3 and Alzheimer's disease: facts and fiction. Frontiers in Molecular Neuroscience, 4:17, 2011.
[25]
Meghana M Kulkarni, Matthew Booker, Serena J Silver, Adam Friedman, Pengyu Hong, Norbert Perrimon, and Bernard Mathey-Prevot. Evidence of off-target effects associated with long dsRNAs in Drosophila melanogaster cell-based assays. Nature Methods, 3(10): 833-838, 2006.
[26]
Beatrice Laurent and Pascal Massart. Adaptive estimation of a quadratic functional by model selection. The Annals of Statistics, 28(5):1302-1338, 2000.
[27]
Chunlin Li, Xiaotong Shen, and Wei Pan. Likelihood ratio tests for a large directed acyclic graph. Journal of the American Statistical Association, 115(531):1304-1319, 2020.
[28]
Chunlin Li, Xiaotong Shen, and Wei Pan. Nonlinear causal discovery with confounders. Journal of the American Statistical Association, pages 1-32, 2023.
[29]
Zhian Liu, Ming Zhang, Gongcheng Xu, Congcong Huo, Qitao Tan, Zengyong Li, and Quan Yuan. Effective connectivity analysis of the brain network in drivers during actual driving using near-infrared spectroscopy. Frontiers in Behavioral Neuroscience, 11:211, 2017.
[30]
Richard Lockhart, Jonathan Taylor, Ryan J Tibshirani, and Robert Tibshirani. A significance test for the lasso. The Annals of Statistics, 42(2):413-468, 2014.
[31]
Po-Ling Loh and Martin J Wainwright. Support recovery without incoherence: A case for nonconvex regularization. The Annals of Statistics, 45(6):2455-2482, 2017.
[32]
Ruiyan Luo and Hongyu Zhao. Bayesian hierarchical modeling for signaling pathway inference from single cell interventional data. The Annals of Applied Statistics, 5(2A):725-745, 2011.
[33]
Toshifumi Matsui, Martin Ingelsson, Hiroaki Fukumoto, Karunya Ramasamy, Hisatomo Kowa, Matthew P Frosch, Michael C Irizarry, and Bradley T Hyman. Expression of APP pathway mRNAs and proteins in Alzheimer's disease. Brain Research, 1161:116-123, 2007.
[34]
Nicolai Meinshausen and Peter Bühlmann. Stability selection. Journal of the Royal Statistical Society: Series B (Statistical Methodology), 72(4):417-473, 2010.
[35]
Aaron J Molstad, Wei Sun, and Li Hsu. A covariance-enhanced approach to multitissue joint eQTL mapping with application to transcriptome-wide association studies. The Annals of Applied Statistics, 15(2):998-1016, 2021.
[36]
Michael Murray. Avoiding invalid instruments and coping with weak instruments. Journal of Economic Perspectives, 20(4):111-132, 2006.
[37]
Chris J Oates, Jim Q Smith, and Sach Mukherjee. Estimating causal structure using conditional DAG models. Journal of Machine Learning Research, 17(1):1880-1903, 2016.
[38]
Roxana Pamfil, Nisara Sriwattanaworachai, Shaan Desai, Philip Pilgerstorfer, Konstantinos Georgatzis, Paul Beaumont, and Bryon Aragam. DYNOTEARS: Structure learning from time-series data. In International Conference on Artificial Intelligence and Statistics, pages 1595-1605. PMLR, 2020.
[39]
Judea Pearl. Causality. Cambridge University Press, 2009.
[40]
Jonas Peters and Peter Bühlmann. Identifiability of Gaussian structural equation models with equal error variances. Biometrika, 101(1):219-228, 2014.
[41]
Jonas Peters, Joris M Mooij, Dominik Janzing, and Bernhard Schölkopf. Causal discovery with continuous additive noise models. Journal of Machine Learning Research, 15(1): 2009-2053, 2014.
[42]
Jonas Peters, Peter Bühlmann, and Nicolai Meinshausen. Causal inference by using invariant prediction: identification and confidence intervals. Journal of the Royal Statistical Society: Series B (Statistical Methodology), 78(5):947-1012, 2016.
[43]
Goutham Rajendran, Bohdan Kivva, Ming Gao, and Bryon Aragam. Structure learning in polynomial time: Greedy algorithms, Bregman information, and exponential families. In Advances in Neural Information Processing Systems, volume 34, pages 18660-18672, 2021.
[44]
Paul Rolland, Volkan Cevher, Matthäus Kleindessner, Chris Russell, Dominik Janzing, Bernhard Schölkopf, and Francesco Locatello. Score matching enables causal discovery of nonlinear additive noise models. In International Conference on Machine Learning, pages 18741-18753. PMLR, 2022.
[45]
Dominik Rothenhäusler, Peter Bühlmann, and Nicolai Meinshausen. Causal Dantzig: fast inference in linear structural equation models with hidden variables under additive interventions. The Annals of Statistics, 47(3):1688-1722, 2019.
[46]
Mark Rudelson and Shuheng Zhou. Reconstruction from anisotropic random measurements. IEEE Transactions on Information Theory, 59(6):3434-3447, 2013.
[47]
Karen Sachs, Omar Perez, Dana Pe'er, Douglas A Lauffenburger, and Garry P Nolan. Causal protein-signaling networks derived from multiparameter single-cell data. Science, 308(5721):523-529, 2005.
[48]
Xiaotong Shen and Jianming Ye. Adaptive model selection. Journal of the American Statistical Association, 97(457):210-221, 2002.
[49]
Xiaotong Shen, Wei Pan, and Yunzhang Zhu. Likelihood-based selection and sharp parameter estimation. Journal of the American Statistical Association, 107(497):223-232, 2012.
[50]
Xiaotong Shen, Wei Pan, Yunzhang Zhu, and Hui Zhou. On constrained and regularized high-dimensional regression. Annals of the Institute of Statistical Mathematics, 65(5): 807-832, 2013.
[51]
Chengchun Shi, Rui Song, Zhao Chen, and Runze Li. Linear hypothesis testing for high dimensional generalized linear models. The Annals of Statistics, 47(5):2671-2703, 2019.
[52]
Shohei Shimizu, Patrik O Hoyer, Aapo Hyvärinen, and Antti Kerminen. A linear non-Gaussian acyclic model for causal discovery. Journal of Machine Learning Research, 7: 2003-2030, 2006.
[53]
Ali Shojaie and George Michailidis. Penalized likelihood methods for estimation of sparse high-dimensional directed acyclic graphs. Biometrika, 97(3):519-538, 2010.
[54]
Christopher G Small. Expansions and Asymptotics for Statistics. Chapman and Hall/CRC, 2010.
[55]
Peter Spirtes, Clark Glymour, and Richard Scheines. Causation, Prediction, and Search. MIT Press, 2000.
[56]
Chandler Squires, Yuhao Wang, and Caroline Uhler. Permutation-based causal structure learning with unknown intervention targets. In Conference on Uncertainty in Artificial Intelligence, pages 1039-1048. PMLR, 2020.
[57]
Joseph H Su, Ming Zhao, Aileen J Anderson, Anu Srinivasan, and Carl W Cotman. Activated caspase-3 expression in Alzheimer's and aged control brain: correlation with Alzheimer pathology. Brain Research, 898(2):350-357, 2001.
[58]
Alexander Teumer. Common methods for performing Mendelian randomization. Frontiers in Cardiovascular Medicine, 5:51, 2018.
[59]
Caroline Uhler, Garvesh Raskutti, Peter Bühlmann, and Bin Yu. Geometry of the faithfulness assumption in causal inference. The Annals of Statistics, 41(2):436-463, 2013.
[60]
Sara van de Geer and Peter Bühlmann. ℓ0-penalized maximum likelihood for sparse directed acyclic graphs. The Annals of Statistics, 41(2):536-567, 2013.
[61]
Jussi Viinikka, Antti Hyttinen, Johan Pensar, and Mikko Koivisto. Towards scalable Bayesian learning of causal DAGs. In Proceedings of the 34th International Conference on Neural Information Processing Systems, pages 6584-6594, 2020.
[62]
Larry Wasserman and Kathryn Roeder. High-dimensional variable selection. The Annals of Statistics, 37(5A):2178-2201, 2009.
[63]
Yiping Yuan, Xiaotong Shen, Wei Pan, and Zizhuo Wang. Constrained likelihood for reconstructing a directed acyclic Gaussian graph. Biometrika, 106(1):109-125, 2019.
[64]
Yuchen Zhang, Martin J Wainwright, and Michael I Jordan. Lower bounds on the performance of polynomial-time algorithms for sparse linear regression. In Conference on Learning Theory, pages 921-948. PMLR, 2014.
[65]
Tuo Zhao, Han Liu, and Tong Zhang. Pathwise coordinate optimization for sparse learning: Algorithm and theory. The Annals of Statistics, 46(1):180-218, 2018.
[66]
Xun Zheng, Bryon Aragam, Pradeep Ravikumar, and Eric P Xing. DAGs with NO TEARS: continuous optimization for structure learning. In Proceedings of the 32nd International Conference on Neural Information Processing Systems, pages 9492-9503, 2018.
[67]
Yunzhang Zhu, Xiaotong Shen, and Wei Pan. On high-dimensional constrained maximum likelihood inference. Journal of the American Statistical Association, 115(529):217-230, 2020.

Index Terms

  1. Inference for a large directed acyclic graph with unspecified interventions
          Index terms have been assigned to the content through auto-classification.

          Recommendations

          Comments

          Information & Contributors

          Information

          Published In

          cover image The Journal of Machine Learning Research
          The Journal of Machine Learning Research  Volume 24, Issue 1
          January 2023
          18881 pages
          ISSN:1532-4435
          EISSN:1533-7928
          Issue’s Table of Contents
          CC-BY 4.0

          Publisher

          JMLR.org

          Publication History

          Accepted: 01 February 2023
          Published: 01 January 2023
          Revised: 01 January 2023
          Received: 01 July 2021
          Published in JMLR Volume 24, Issue 1

          Author Tags

          1. high-dimensional inference
          2. data perturbation
          3. structure learning
          4. peeling algorithm
          5. identifiability

          Qualifiers

          • Research-article

          Contributors

          Other Metrics

          Bibliometrics & Citations

          Bibliometrics

          Article Metrics

          • 0
            Total Citations
          • 68
            Total Downloads
          • Downloads (Last 12 months)68
          • Downloads (Last 6 weeks)10
          Reflects downloads up to 26 Jan 2025

          Other Metrics

          Citations

          View Options

          View options

          PDF

          View or Download as a PDF file.

          PDF

          eReader

          View online with eReader.

          eReader

          Login options

          Full Access

          Figures

          Tables

          Media

          Share

          Share

          Share this Publication link

          Share on social media