Skip to main content
Log in

A block coordinate gradient descent method for regularized convex separable optimization and covariance selection

  • Full Length Paper
  • Series B
  • Published:
Mathematical Programming Submit manuscript

Abstract

We consider a class of unconstrained nonsmooth convex optimization problems, in which the objective function is the sum of a convex smooth function on an open subset of matrices and a separable convex function on a set of matrices. This problem includes the covariance selection problem that can be expressed as an 1-penalized maximum likelihood estimation problem. In this paper, we propose a block coordinate gradient descent method (abbreviated as BCGD) for solving this class of nonsmooth separable problems with the coordinate block chosen by a Gauss-Seidel rule. The method is simple, highly parallelizable, and suited for large-scale problems. We establish global convergence and, under a local Lipschizian error bound assumption, linear rate of convergence for this method. For the covariance selection problem, the method can terminate in \({O(n^3/\epsilon)}\) iterations with an \({\epsilon}\)-optimal solution. We compare the performance of the BCGD method with the first-order methods proposed by Lu (SIAM J Optim 19:1807–1827, 2009; SIAM J Matrix Anal Appl 31:2000–2016, 2010) for solving the covariance selection problem on randomly generated instances. Our numerical experience suggests that the BCGD method can be efficient for large-scale covariance selection problems with constraints.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Institutional subscriptions

Similar content being viewed by others

References

  1. Borwein J.M., Lewis A.S.: Convex Analysis and Nonlinear Optimization: Theory and Examples. Springer, New York (2000)

    MATH  Google Scholar 

  2. Dahl J., Vandenberghe L., Roychowdhury V.: Covariance selection for non-chordal graphs via chordal embedding. Optim. Methods Softw. 23, 501–520 (2008)

    Article  MathSciNet  MATH  Google Scholar 

  3. D’Aspremont A., Banerjee O., El Ghaoui L.: First-order methods for sparse covariance selection. SIAM J. Matrix Anal. Appl. 30, 56–66 (2008)

    Article  MathSciNet  MATH  Google Scholar 

  4. Friedman J., Hastie T., Tibshirani R.: Sparse inverse covariance estimation with the graphical lasso. Biostatistics 9, 432–441 (2008)

    Article  MATH  Google Scholar 

  5. Goebel R., Rockafellar R.T.: Local strong convexity and local Lipschitz continuity of the gradients of convex functions. J. Convex Anal. 15, 263–270 (2008)

    MathSciNet  MATH  Google Scholar 

  6. Horn R., Johnson C.: Matrix Analysis. Cambridge University Press, Cambridge (1999)

    Google Scholar 

  7. Lu Z.: Smooth optimization approach for covariance selection. SIAM J. Optim. 19, 1807–1827 (2009)

    Article  MATH  Google Scholar 

  8. Lu Z.: Adaptive first-order methods for general sparse inverse covariance selection. SIAM J. Matrix Anal. Appl. 31, 2000–2016 (2010)

    Article  MATH  Google Scholar 

  9. Nesterov Y.: A method of solving a convex programming problem with convergence rate O(1/k 2). Sov. Math. Doklady 27, 372–376 (1983)

    MATH  Google Scholar 

  10. Nesterov Y.: Smooth minimization of nonsmooth functions. Math. Prog. 103, 127–152 (2005)

    Article  MathSciNet  MATH  Google Scholar 

  11. Nocedal J., Wright S.J.: Numerical Optimization. Springer, New York (1999)

    Book  MATH  Google Scholar 

  12. Rockafellar R.T.: Convex Analysis. Princeton University Press, Princeton (1970)

    MATH  Google Scholar 

  13. Rockafellar R.T., Wets R.J.-B.: Variational Analysis. Springer, New York (1998)

    Book  MATH  Google Scholar 

  14. Saad Y.: Iterative Methods for Sparse Linear Systems. PWS Publishing Company, Boston (1996)

    MATH  Google Scholar 

  15. Scheinberg, K., Rish, I.: Learning sparse Gaussian markov networks using a greedy coordinate ascent approach. In: Machine Learning and Knowledge Discovery in Databases. Lecture Notes in Computer Science, vol. 6323, pp. 196–212 (2010)

  16. Tseng P.: Convergence of block coordinate descent method for nondifferentiable minimization. J. Optim. Theory Appl. 109, 473–492 (2001)

    Article  MathSciNet  Google Scholar 

  17. Tseng P., Yun S.: A coordinate gradient descent method for nonsmooth separable minimization. Math. Prog. Ser. B 117, 387–423 (2009)

    Article  MathSciNet  MATH  Google Scholar 

  18. Wang C., Sun D., Toh K.-C.: Solving log-determinant optimization problems by a Newton-CG primal proximal point algorithm. SIAM J. Optim. 20, 2994–3013 (2010)

    Article  MathSciNet  MATH  Google Scholar 

  19. Wong F., Carter C.K., Kohn R.: Efficient estimation of covariance selection models. Biometrika 90, 809–830 (2003)

    Article  MathSciNet  Google Scholar 

  20. Yuan, X.: Alternating direction methods for sparse covariance selection. http://www.optimization-online.org/DB_FILE/2009/09/2390.pdf

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Sangwoon Yun.

Additional information

In memory of Paul Tseng, who went missing while on a kayaking trip along the Jinsha river, China, on August 13, 2009, for his friendship and contributions to large-scale continuous optimization.

Rights and permissions

Reprints and permissions

About this article

Cite this article

Yun, S., Tseng, P. & Toh, KC. A block coordinate gradient descent method for regularized convex separable optimization and covariance selection. Math. Program. 129, 331–355 (2011). https://doi.org/10.1007/s10107-011-0471-1

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s10107-011-0471-1

Keywords

Mathematics Subject Classification (2000)

Navigation