Journal
SANKHYA-SERIES A-MATHEMATICAL STATISTICS AND PROBABILITY
Volume 82, Issue 2, Pages 426-447Publisher
SPRINGER
DOI: 10.1007/s13171-019-00191-2
Keywords
Bayes regularization; root Lasso; Convolution; Lasso; Logistic; Quantile
Categories
Funding
- US National Science Foundation [DMS-1613063]
Ask authors/readers for more resources
Global-local mixtures, including Gaussian scale mixtures, have gained prominence in recent times, both as a sparsity inducing prior in p >> n problems as well as default priors for non-linear many-to-one functionals of high-dimensional parameters. Here we propose a unifying framework for global-local scale mixtures using the Cauchy-Schlomilch and Liouville integral transformation identities, and use the framework to build a new Bayesian sparse signal recovery method. This new method is a Bayesian counterpart of the root Lasso (Belloni et al., Biometrika 98, 4, 791-806, 2011) that adapts to unknown error variance. Our framework also characterizes well-known scale mixture distributions including the Laplace density used in Bayesian Lasso, logit and quantile via a single integral identity. Finally, we derive a few convolutions that commonly arise in Bayesian inference and posit a conjecture concerning bridge and uniform correlation mixtures.
Authors
I am an author on this paper
Click your name to claim this paper and add it to your profile.
Reviews
Recommended
No Data Available