Journal
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE
Volume 43, Issue 9, Pages 3196-3213Publisher
IEEE COMPUTER SOC
DOI: 10.1109/TPAMI.2020.2980542
Keywords
Correlation; Covariance matrices; Measurement; Graphical models; Gaussian distribution; Sparse representation; Alzheimer's disease; Gaussian graphical models; model selection; high dimension low sample size; sparse matrices; maximum likelihood estimation
Funding
- European Research Council (ERC) [678304]
- European Union [666992, 826421]
- French government under management of Agence Nationale de la Recherche as part of the Investissements d'avenir Program [ANR-19P3IA-0001, ANR-10IAIHU-06]
Ask authors/readers for more resources
In this paper, two families of Gaussian graphical model (GGM) inference methods, nodewise approach and penalised likelihood maximisation, are compared. The study demonstrates that when the sample size is small, both methods may result in graphs with either too few or too many edges compared to the real one. A composite procedure is proposed to address this issue, which explores a family of graphs with a nodewise numerical scheme and selects a candidate based on an overall likelihood criterion. This selection method yields graphs closer to the truth and corresponding to distributions with better KL divergence compared to the other two methods, particularly when the number of observations is small.
Gaussian graphical models (GGM) are often used to describe the conditional correlations between the components of a random vector. In this article, we compare two families of GGM inference methods: the nodewise approach and the penalised likelihood maximisation. We demonstrate on synthetic data that, when the sample size is small, the two methods produce graphs with either too few or too many edges when compared to the real one. As a result, we propose a composite procedure that explores a family of graphs with a nodewise numerical scheme and selects a candidate among them with an overall likelihood criterion. We demonstrate that, when the number of observations is small, this selection method yields graphs closer to the truth and corresponding to distributions with better KL divergence with regards to the real distribution than the other two. Finally, we show the interest of our algorithm on two concrete cases: first on brain imaging data, then on biological nephrology data. In both cases our results are more in line with current knowledge in each field.
Authors
I am an author on this paper
Click your name to claim this paper and add it to your profile.
Reviews
Recommended
No Data Available