4.8 Article

Gaussian Graphical Model Exploration and Selection in High Dimension Low Sample Size Setting

Journal

Publisher

IEEE COMPUTER SOC
DOI: 10.1109/TPAMI.2020.2980542

Keywords

Correlation; Covariance matrices; Measurement; Graphical models; Gaussian distribution; Sparse representation; Alzheimer's disease; Gaussian graphical models; model selection; high dimension low sample size; sparse matrices; maximum likelihood estimation

Funding

  1. European Research Council (ERC) [678304]
  2. European Union [666992, 826421]
  3. French government under management of Agence Nationale de la Recherche as part of the Investissements d'avenir Program [ANR-19P3IA-0001, ANR-10IAIHU-06]

Ask authors/readers for more resources

In this paper, two families of Gaussian graphical model (GGM) inference methods, nodewise approach and penalised likelihood maximisation, are compared. The study demonstrates that when the sample size is small, both methods may result in graphs with either too few or too many edges compared to the real one. A composite procedure is proposed to address this issue, which explores a family of graphs with a nodewise numerical scheme and selects a candidate based on an overall likelihood criterion. This selection method yields graphs closer to the truth and corresponding to distributions with better KL divergence compared to the other two methods, particularly when the number of observations is small.
Gaussian graphical models (GGM) are often used to describe the conditional correlations between the components of a random vector. In this article, we compare two families of GGM inference methods: the nodewise approach and the penalised likelihood maximisation. We demonstrate on synthetic data that, when the sample size is small, the two methods produce graphs with either too few or too many edges when compared to the real one. As a result, we propose a composite procedure that explores a family of graphs with a nodewise numerical scheme and selects a candidate among them with an overall likelihood criterion. We demonstrate that, when the number of observations is small, this selection method yields graphs closer to the truth and corresponding to distributions with better KL divergence with regards to the real distribution than the other two. Finally, we show the interest of our algorithm on two concrete cases: first on brain imaging data, then on biological nephrology data. In both cases our results are more in line with current knowledge in each field.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.8
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available