SDPNAL+
SDPNAL+: a majorized semismooth Newton-CG augmented Lagrangian method for semidefinite programming with nonnegative constraints.In this paper, we present a majorized semismooth Newton-CG augmented Lagrangian method, called SDPNAL+, for semidefinite programming (SDP) with partial or full nonnegative constraints on the matrix variable. SDPNAL+ is a much enhanced version of SDPNAL introduced by X.-Y. Zhao et al. [SIAM J. Optim. 20, No. 4, 1737–1765 (2010; Zbl 1213.90175)] for solving generic SDPs. SDPNAL works very efficiently for nondegenerate SDPs but may encounter numerical difficulty for degenerate ones. Here we tackle this numerical difficulty by employing a majorized semismooth Newton-CG augmented Lagrangian method coupled with a convergent 3-block alternating direction method of multipliers introduced recently by D. Sun et al. [SIAM J. Optim. 25, No. 2, 882–915 (2015; Zbl 06444987)]. Numerical results for various large scale SDPs with or without nonnegative constraints show that the proposed method is not only fast but also robust in obtaining accurate solutions. It outperforms, by a significant margin, two other competitive publicly available first order methods based codes: (1) an alternating direction method of multipliers based solver called SDPAD by Z. Wen et al. [Math. Program. Comput. 2, No. 3–4, 203–230 (2010; Zbl 1206.90088)] and (2) a two-easy-block-decomposition hybrid proximal extragradient method called 2EBD-HPE by R. Monteiro et al. [“A first-order block-decomposition method for solving two-easy-block structured semidefinite programs”, Math. Program. Comput. 6, No. 2, 103–150 (2014; doi:10.1007/s12532-013-0062-7)]. In contrast to these two codes, we are able to solve all the 95 difficult SDP problems arising from the relaxations of quadratic assignment problems tested in SDPNAL to an accuracy of 10 -6 efficiently, while SDPAD and 2EBD-HPE successfully solve 30 and 16 problems, respectively. In addition, SDPNAL+ appears to be the only viable method currently available to solve large scale SDPs arising from rank-1 tensor approximation problems constructed by J. Nie and L. Wang [SIAM J. Matrix Anal. Appl. 35, No. 3, 1155–1179 (2014; Zbl 1305.65134)]. The largest rank-1 tensor approximation problem we solved (in about 14.5 h) is nonsym(21,4), in which its resulting SDP problem has matrix dimension n=9261 and the number of equality constraints m=12,326,390.
Keywords for this software
References in zbMATH (referenced in 9 articles , 1 standard article )
Showing results 1 to 9 of 9.
Sorted by year (- Li, Xudong; Sun, Defeng; Toh, Kim-Chuan: A highly efficient semismooth Newton augmented Lagrangian method for solving lasso problems (2018)
- Arima, Naohiko; Kim, Sunyoung; Kojima, Masakazu; Toh, Kim-Chuan: A robust Lagrangian-DNN method for a class of quadratic optimization problems (2017)
- Cui, Ying; Ding, Chao; Zhao, Xinyuan: Quadratic growth conditions for convex matrix optimization problems associated with spectral functions (2017)
- Mixon, Dustin G.; Villar, Soledad; Ward, Rachel: Clustering subgaussian mixtures by semidefinite programming (2017)
- Wang, Chengjing; Tang, Peipei: A primal majorized semismooth Newton-CG augmented Lagrangian method for large-scale linearly constrained convex programming (2017)
- Li, Xudong; Sun, Defeng; Toh, Kim-Chuan: A Schur complement based semi-proximal ADMM for convex quadratic conic programming and extensions (2016)
- Qi, Hou-Duo: A convex matrix optimization for the additive constant problem in multidimensional scaling with application to locally linear embedding (2016)
- Sun, Defeng; Toh, Kim-Chuan; Yang, Liuqin: An efficient inexact ABCD method for least squares semidefinite programming (2016)
- Yang, Liuqin; Sun, Defeng; Toh, Kim-Chuan: SDPNAL+: a majorized semismooth Newton-CG augmented Lagrangian method for semidefinite programming with nonnegative constraints (2015)