glasso
The graphical lasso: new insights and alternatives. The graphical lasso [5] is an algorithm for learning the structure in an undirected Gaussian graphical model, using ℓ 1 regularization to control the number of zeros in the precision matrix Θ=Σ -1 [2, 11]. The R package glasso [5] is popular, fast, and allows one to efficiently build a path of models for different values of the tuning parameter. Convergence of glasso can be tricky; the converged precision matrix might not be the inverse of the estimated covariance, and occasionally it fails to converge with warm starts. In this paper we explain this behavior, and propose new algorithms that appear to outperform glasso. By studying the “normal equations” we see that, glasso is solving the dual of the graphical lasso penalized likelihood, by block coordinate ascent; a result which can also be found in [2]. In this dual, the target of estimation is Σ, the covariance matrix, rather than the precision matrix Θ. We propose similar primal algorithms p-glasso and dp-glasso, that also operate by block-coordinate descent, where Θ is the optimization target. We study all of these algorithms, and in particular different approaches to solving their coordinate sub-problems. We conclude that dp-glasso is superior from several points of view.
Keywords for this software
References in zbMATH (referenced in 399 articles , 1 standard article )
Showing results 1 to 20 of 399.
Sorted by year (- Kashlak, Adam B.: Non-asymptotic error controlled sparse high dimensional precision matrix estimation (2021)
- Kereta, Željko; Klock, Timo: Estimating covariance and precision matrices along subspaces (2021)
- Andrade, Daniel; Takeda, Akiko; Fukumizu, Kenji: Robust Bayesian model selection for variable clustering with the Gaussian graphical model (2020)
- An, Ziwen; Nott, David J.; Drovandi, Christopher: Robust Bayesian synthetic likelihood via a semi-parametric approach (2020)
- Augugliaro, Luigi; Sottile, Gianluca; Vinciotti, Veronica: The conditional censored graphical Lasso estimator (2020)
- Barber, Rina Foygel; Candès, Emmanuel J.; Samworth, Richard J.: Robust inference with knockoffs (2020)
- Bertsimas, Dimitris; Lamperski, Jourdain; Pauphilet, Jean: Certifiably optimal sparse inverse covariance estimation (2020)
- Bian, Lingbin; Cui, Tiangang; Sofronov, Georgy; Keith, Jonathan: Network structure change point detection by posterior predictive discrepancy (2020)
- Bing, Xin; Bunea, Florentina; Ning, Yang; Wegkamp, Marten: Adaptive estimation in structured factor models with applications to overlapping clustering (2020)
- Boudt, Kris; Rousseeuw, Peter J.; Vanduffel, Steven; Verdonck, Tim: The minimum regularized covariance determinant estimator (2020)
- Chen, Jingnan; Dai, Gengling; Zhang, Ning: An application of sparse-group Lasso regularization to equity portfolio optimization and sector selection (2020)
- Chen, Zehua; Jiang, Yiwei: A two-stage sequential conditional selection approach to sparse high-dimensional multivariate regression models (2020)
- Córdoba, Irene; Bielza, Concha; Larrañaga, Pedro: A review of Gaussian Markov models for conditional independence (2020)
- De Canditiis, Daniela: A global approach for learning sparse Ising models (2020)
- Evans, Robin J.: Model selection and local geometry (2020)
- Fang, Qian; Yu, Chen; Weiping, Zhang: Regularized estimation of precision matrix for high-dimensional multivariate longitudinal data (2020)
- Fan, Jianqing; Feng, Yang; Xia, Lucy: A projection-based conditional dependence measure with applications to high-dimensional undirected graphical models (2020)
- Farnè, Matteo; Montanari, Angela: A large covariance matrix estimator under intermediate spikiness regimes (2020)
- Fontanella, Lara; Fontanella, Sara; Ignaccolo, Rosaria; Ippoliti, Luigi; Valentini, Pasquale: G-Lasso network analysis for functional data (2020)
- Fu, Anqi; Zhang, Junzi; Boyd, Stephen: Anderson accelerated Douglas-Rachford splitting (2020)