Skip to main content
Log in

Models and algorithms for distributionally robust least squares problems

  • Full Length Paper
  • Series A
  • Published:
Mathematical Programming Submit manuscript

Abstract

We present three different robust frameworks using probabilistic ambiguity descriptions of the data in least squares problems. These probability ambiguity descriptions are given by: (1) confidence region over the first two moments; (2) bounds on the probability measure with moments constraints; (3) the Kantorovich probability distance from a given measure. For the first case, we give an equivalent formulation and show that the optimization problem can be solved using a semidefinite optimization reformulation or polynomial time algorithms. For the second case, we derive the equivalent Lagrangian problem and show that it is a convex stochastic programming problem. We further analyze three special subcases: (i) finite support; (ii) measure bounds by a reference probability measure; (iii) measure bounds by two reference probability measures with known density functions. We show that case (i) has an equivalent semidefinite programming reformulation and the sample average approximations of case (ii) and (iii) have equivalent semidefinite programming reformulations. For ambiguity description (3), we show that the finite support case can be solved by using an equivalent second order cone programming reformulation.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Similar content being viewed by others

References

  1. Anstreicher, K.M.: On Vaidya’s volumetric cutting plane method for convex programming. Math. Oper. Res. 22(1), 63–89 (1997)

    Article  MATH  MathSciNet  Google Scholar 

  2. Bertsekas, Dimitri P.: Convex Analysis and Optimization. Athena Scientific, Belmont (2003)

    MATH  Google Scholar 

  3. Bertsimas, D., Doan, X.V., Natarajan, K., Teo, C.: Models for minimax stochastic linear optimization problems with risk aversion. Math. Oper. Res. 35(3), 580–602 (2010)

    Article  MATH  MathSciNet  Google Scholar 

  4. Boyd, S., Vandenberghe, L.: Convex Optimization. Cambridge University Press, Cambridge (2004)

    Book  MATH  Google Scholar 

  5. Delage, E., Ye, Y.: Distributionally robust optimization under moment uncertainty with application to data-driven problems. Oper. Res. 58(3), 595–612 (2010)

    Google Scholar 

  6. Dupacova, J.: The minimax approach to stochastic programming and an illustrative application. Stochastics 20, 73–88 (1987)

    Article  MATH  MathSciNet  Google Scholar 

  7. Fortin, C., Wolkowicz, H.: The trust region subproblem and semidefinite programming. Optim. Methods Softw. 19(1), 41–67 (2004)

    Article  MATH  MathSciNet  Google Scholar 

  8. Ghaoui, L.E., Lebret, H.: Robust solution to least-squares problem with uncertain data. SIAM J. Matrix Anal. Appl. 18(4), 1035–1064 (1997)

    Article  MATH  MathSciNet  Google Scholar 

  9. Greene William, H.: Econometric Analysis, 6th edn. Prentice Hall, Upper Saddle River (2007)

    Google Scholar 

  10. Lyandres Olga, Van Duyne, Richard P., Glucksberg, Joseph T., Sanjay Mehrotra, : Prediction range estimation from noisy raman spectra with robust optimization. Analyst 135(8), 2111–2118 (2010)

    Article  Google Scholar 

  11. Pflug, G., Wozabal, D.: Ambiguity in portfolio selection. Quant. Finance 7(4), 435–442 (2007)

    Article  MATH  MathSciNet  Google Scholar 

  12. Rachev, S.T.: Probability Metrics and the Stability of Stochastic Models. Willey Series in Probability and Mathematical Statistics. Wiley, New York (1991)

    Google Scholar 

  13. Calyampudi Radhakrishna Rao: The theory of least squares when the parameters are stochastic and its application to the analysis of growth curves. Biometrika 52(3–4), 447–458 (1965)

    MathSciNet  Google Scholar 

  14. Rendl, F., Wolkowicz, H.: A semidefinite framework for trust region subproblems with applications to large scale minimization. Math. Program. 77, 273–299 (1997)

    MATH  MathSciNet  Google Scholar 

  15. Rockafellar R.T.: Conjugate Duality and Optimization. CBMS-NSF Regional Conference Series in Applied Mathematics 16. SIAM, Philadelphia (1974)

  16. Ruszczynski, A., Shapiro, A. (eds.): Stochastic Programming. Handbooks in Operations Research and Management Science 10. North-Holland, Amsterdam (2003)

  17. Scarf, H.: A min-max Solution of an Inventory Problem. Stanford University Press, Stanford (1958)

    Google Scholar 

  18. Shapiro, A., Ahmed, S.: On a class of minimax stochastic programs. SIAM J. Optim. 14(4), 1237–1249 (2004)

    Article  MATH  MathSciNet  Google Scholar 

  19. Shapiro, A., Homem de mello, T.: On the rate of convergence of optimal solutions of monte carlo approximations of stochastic programs. SIAM J. Optim. 11(1), 70–86 (2000)

    Article  MATH  MathSciNet  Google Scholar 

  20. Sturm Jos, F.: Using SeDuMi 1.02, a MATLAB* toolbox for optimization over symmetric cones. Optim. Methods Softw. 11–12, 625–653 (1999)

    Article  Google Scholar 

  21. Sturm Jos, F.: Implementation of interior point methods for mixed semidefinite and second order cone optimization problems. Optim. Methods Softw. 17, 1105–1154 (2002)

    Article  MATH  MathSciNet  Google Scholar 

  22. Vaidya, P.M.: A new algorithm for minimizing convex functions over convex sets. Math. Program. 73, 291–341 (1996)

    MATH  MathSciNet  Google Scholar 

  23. Vallander, S.S.: Calculation of the wasserstein distance between probability distributions on the line. Theory Probab. Appl. 18, 784–786 (1973)

    Article  MATH  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Sanjay Mehrotra.

Additional information

Research partially supported by NSF grant CMMI-1100868 and ONR grant N00014-09-10518 and N00014210051.

Appendix

Appendix

In Sect. 3 we use SAA method to approximate the stochastic programming problems (3.13) and (3.18). The convergence results for the SAA method from [16] are summarized here. Consider a stochastic programming problem of the form

$$\begin{aligned} \min _\mathbf{x \in \mathcal X }\{ f(\mathbf x ):=\mathbb E _\mathbb{P }[F(\mathbf x ,\pmb {\xi })] \}, \end{aligned}$$
(6.1)

where \(F(\mathbf x ,\pmb {\xi })\) is a function of variables \(\mathbf x \in \mathbb R ^n\) and parameters \(\pmb {\xi } \in \mathbb R ^d\). \(\mathcal X \subset \mathbb R ^n\) is a given set, and \(\pmb {\xi }=\pmb {\xi }(\omega )\) is a random vector. The expectation in (6.1) is taken with respect to the probability distribution of \(\pmb {\xi }\) which is assumed to be known as \(\mathbb P \). Denote by \(\Xi \subset \mathbb R ^d\) the support of the probability distribution of \(\pmb {\xi }\), that is, \(\Xi \) is the smallest closed set in \(\mathbb R ^d\) such that the probability of the event \(\pmb {\xi } \in \mathbb R ^d \setminus \Xi \) is zero. Also denote by \(\mathbb P (A)\) the probability of an event \(A\). With the generated sample \(\xi ^1,\ldots ,\xi ^K\), we associate the sample average function

$$\begin{aligned} \hat{f}_K(\mathbf x ):=\frac{1}{K} \sum _{i=1}^K F(\mathbf x ,\xi ^i). \end{aligned}$$
(6.2)

The stochastic programming problem (6.1) is approximated by the optimization problem

$$\begin{aligned} \min _\mathbf{x \in \mathcal X } \left\{ \hat{f}_K(\mathbf x ):=\frac{1}{K} \sum _{i=1}^K F(\mathbf x ,\xi ^i) \right\} . \end{aligned}$$
(6.3)

Let the optimal value of (6.1) be \(\nu \) and it’s optimal solution set be \(S\). Let \(\hat{\nu }_K\) and \(\hat{S}_K\) be the optimal value and the set of optimal solutions of the SAA problem (6.3). For sets \(A,B \subset \mathbf R ^n\), denote \(\mathrm{dist}(x,A):=\inf _{x^{\prime } \in A}\left| \left| x-x^{\prime }\right| \right| \) to be the distance from \(x \in \mathbb X ^n\) to \(A\), and

$$\begin{aligned} \mathbb D (A,B):=\sup _{x \in A} \mathrm{dist}(x,B). \end{aligned}$$
(6.4)

Also, define the function \((\mathbf x ,\xi ) \mapsto F(\mathbf x ,\xi )\) to be a random lower semicontinuous function if the associated epigraphical multifunction \(\xi \mapsto \mathrm{epi}F(\cdot ,\xi )\) is closed valued and measurable. We say that the Law of Large Numbers (LLN) holds, for \(\hat{f}_K(\mathbf x )\), pointwise if \(\hat{f}_K(\mathbf x )\) converges w.p.1 to \(f(\mathbf x )\), as \(K \rightarrow \infty \), for any fixed \(\mathbf x \in \mathbb R ^n\). The following convergence theorem ensures that a solution of SAA problem (6.2) converges to that of (6.1) as the sample size increases.

Theorem 8

[16, Chapter 6, Theorem 4.] Suppose that: (i) the integrand function \(F\) is random lower semicontinuous, (ii) for almost every \(\xi \in \Xi \) the function \(F(\cdot ,\xi )\) is convex, (iii) the set \(\mathcal X \) is closed and convex, (iv) the expected value function \(f\) is lower semicontinuous and there exists a point \(\hat{\mathbf{x }} \in \mathcal X \) such that \(f(\mathbf x ) \le +\infty \) for all \(x\) in a neighborhood of \(\hat{\mathbf{x }}\), (v) the set \(S\) of optimal solutions of the original problem (6.1) is nonempty and bounded, (vi) the LLN holds pointwise. Then \(\hat{\nu }_K \rightarrow \nu ^*\) and \(\mathbb D (\hat{S}_K,S) \rightarrow 0\) w.p.1 as \(K \rightarrow \infty \).

There are also results about the exponential convergence rate of the SAA method. Please refer to [16, 19] for details of such results. We note that the assumptions in Theorem 8 are satisfied for the models (3.13) and (3.18). Hence the convergence of the solution of SAA of these problems to a solution of the original problem is guaranteed.

Rights and permissions

Reprints and permissions

About this article

Cite this article

Mehrotra, S., Zhang, H. Models and algorithms for distributionally robust least squares problems. Math. Program. 146, 123–141 (2014). https://doi.org/10.1007/s10107-013-0681-9

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s10107-013-0681-9

Mathematics Subject Classification

Navigation