SDPNAL+: a majorized semismooth Newton-CG augmented Lagrangian method for semidefinite programming with nonnegative constraints.In this paper, we present a majorized semismooth Newton-CG augmented Lagrangian method, called SDPNAL+, for semidefinite programming (SDP) with partial or full nonnegative constraints on the matrix variable. SDPNAL+ is a much enhanced version of SDPNAL introduced by X.-Y. Zhao et al. [SIAM J. Optim. 20, No. 4, 1737–1765 (2010; Zbl 1213.90175)] for solving generic SDPs. SDPNAL works very efficiently for nondegenerate SDPs but may encounter numerical difficulty for degenerate ones. Here we tackle this numerical difficulty by employing a majorized semismooth Newton-CG augmented Lagrangian method coupled with a convergent 3-block alternating direction method of multipliers introduced recently by D. Sun et al. [SIAM J. Optim. 25, No. 2, 882–915 (2015; Zbl 06444987)]. Numerical results for various large scale SDPs with or without nonnegative constraints show that the proposed method is not only fast but also robust in obtaining accurate solutions. It outperforms, by a significant margin, two other competitive publicly available first order methods based codes: (1) an alternating direction method of multipliers based solver called SDPAD by Z. Wen et al. [Math. Program. Comput. 2, No. 3–4, 203–230 (2010; Zbl 1206.90088)] and (2) a two-easy-block-decomposition hybrid proximal extragradient method called 2EBD-HPE by R. Monteiro et al. [“A first-order block-decomposition method for solving two-easy-block structured semidefinite programs”, Math. Program. Comput. 6, No. 2, 103–150 (2014; doi:10.1007/s12532-013-0062-7)]. In contrast to these two codes, we are able to solve all the 95 difficult SDP problems arising from the relaxations of quadratic assignment problems tested in SDPNAL to an accuracy of 10 -6 efficiently, while SDPAD and 2EBD-HPE successfully solve 30 and 16 problems, respectively. In addition, SDPNAL+ appears to be the only viable method currently available to solve large scale SDPs arising from rank-1 tensor approximation problems constructed by J. Nie and L. Wang [SIAM J. Matrix Anal. Appl. 35, No. 3, 1155–1179 (2014; Zbl 1305.65134)]. The largest rank-1 tensor approximation problem we solved (in about 14.5 h) is nonsym(21,4), in which its resulting SDP problem has matrix dimension n=9261 and the number of equality constraints m=12,326,390.

References in zbMATH (referenced in 53 articles , 2 standard articles )

Showing results 1 to 20 of 53.
Sorted by year (citations)

1 2 3 next

  1. Chen, Liang; Li, Xudong; Sun, Defeng; Toh, Kim-Chuan: On the equivalence of inexact proximal ALM and ADMM for a class of convex composite programming (2021)
  2. Ding, Lijun; Yurtsever, Alp; Cevher, Volkan; Tropp, Joel A.; Udell, Madeleine: An optimal-storage approach to semidefinite programming using approximate complementarity (2021)
  3. Hu, Shenglong; Li, Guoyin: (\mathrmB)-subdifferentials of the projection onto the matrix simplex (2021)
  4. Jiang, Zhuoxuan; Zhao, Xinyuan; Ding, Chao: A proximal DC approach for quadratic assignment problem (2021)
  5. Kim, Sunyoung; Kojima, Masakazu; Toh, Kim-Chuan: A Newton-bracketing method for a simple conic optimization problem (2021)
  6. Lin, Tianyi; Ma, Shiqian; Ye, Yinyu; Zhang, Shuzhong: An ADMM-based interior-point method for large-scale linear programming (2021)
  7. Rauhut, Holger; Stojanac, Željka: Tensor theta norms and low rank recovery (2021)
  8. Shinde, Nimita; Narayanan, Vishnu; Saunderson, James: Memory-efficient structured convex optimization via extreme point sampling (2021)
  9. Sun, Defeng; Toh, Kim-Chuan; Yuan, Yancheng: Convex clustering: model, theoretical guarantee and efficient algorithm (2021)
  10. Yan, Bowei; Sarkar, Purnamrita: Covariate regularized community detection in sparse graphs (2021)
  11. Yurtsever, Alp; Tropp, Joel A.; Fercoq, Olivier; Udell, Madeleine; Cevher, Volkan: Scalable semidefinite programming (2021)
  12. Zhao, Ping-Fan; Li, Qing-Na; Chen, Wei-Kun; Liu, Ya-Feng: An efficient quadratic programming relaxation based algorithm for large-scale MIMO detection (2021)
  13. Zheng, Meng-Meng; Huang, Zheng-Hai; Wang, Yong: T-positive semidefiniteness of third-order symmetric tensors and T-semidefinite programming (2021)
  14. Chen, Liang; Chang, Xiaokai; Liu, Sanyang: A three-operator splitting perspective of a three-block ADMM for convex quadratic semidefinite programming and beyond (2020)
  15. Chen, Shixiang; Ma, Shiqian; Man-Cho So, Anthony; Zhang, Tong: Proximal gradient method for nonsmooth optimization over the Stiefel manifold (2020)
  16. Ding, Chao; Sun, Defeng; Sun, Jie; Toh, Kim-Chuan: Spectral operators of matrices: semismoothness and characterizations of the generalized Jacobian (2020)
  17. Gaar, Elisabeth; Rendl, Franz: A computational study of exact subgraph based SDP bounds for max-cut, stable set and coloring (2020)
  18. Goulart, Paul J.; Nakatsukasa, Yuji; Rontsis, Nikitas: Accuracy of approximate projection to the semidefinite cone (2020)
  19. Li, Xiaodong; Li, Yang; Ling, Shuyang; Strohmer, Thomas; Wei, Ke: When do birds of a feather flock together? (k)-means, proximity, and conic programming (2020)
  20. Li, Xudong; Sun, Defeng; Toh, Kim-Chuan: An asymptotically superlinearly convergent semismooth Newton augmented Lagrangian method for linear programming (2020)

1 2 3 next