Abstract
We present three different robust frameworks using probabilistic ambiguity descriptions of the data in least squares problems. These probability ambiguity descriptions are given by: (1) confidence region over the first two moments; (2) bounds on the probability measure with moments constraints; (3) the Kantorovich probability distance from a given measure. For the first case, we give an equivalent formulation and show that the optimization problem can be solved using a semidefinite optimization reformulation or polynomial time algorithms. For the second case, we derive the equivalent Lagrangian problem and show that it is a convex stochastic programming problem. We further analyze three special subcases: (i) finite support; (ii) measure bounds by a reference probability measure; (iii) measure bounds by two reference probability measures with known density functions. We show that case (i) has an equivalent semidefinite programming reformulation and the sample average approximations of case (ii) and (iii) have equivalent semidefinite programming reformulations. For ambiguity description (3), we show that the finite support case can be solved by using an equivalent second order cone programming reformulation.
Similar content being viewed by others
References
Anstreicher, K.M.: On Vaidya’s volumetric cutting plane method for convex programming. Math. Oper. Res. 22(1), 63–89 (1997)
Bertsekas, Dimitri P.: Convex Analysis and Optimization. Athena Scientific, Belmont (2003)
Bertsimas, D., Doan, X.V., Natarajan, K., Teo, C.: Models for minimax stochastic linear optimization problems with risk aversion. Math. Oper. Res. 35(3), 580–602 (2010)
Boyd, S., Vandenberghe, L.: Convex Optimization. Cambridge University Press, Cambridge (2004)
Delage, E., Ye, Y.: Distributionally robust optimization under moment uncertainty with application to data-driven problems. Oper. Res. 58(3), 595–612 (2010)
Dupacova, J.: The minimax approach to stochastic programming and an illustrative application. Stochastics 20, 73–88 (1987)
Fortin, C., Wolkowicz, H.: The trust region subproblem and semidefinite programming. Optim. Methods Softw. 19(1), 41–67 (2004)
Ghaoui, L.E., Lebret, H.: Robust solution to least-squares problem with uncertain data. SIAM J. Matrix Anal. Appl. 18(4), 1035–1064 (1997)
Greene William, H.: Econometric Analysis, 6th edn. Prentice Hall, Upper Saddle River (2007)
Lyandres Olga, Van Duyne, Richard P., Glucksberg, Joseph T., Sanjay Mehrotra, : Prediction range estimation from noisy raman spectra with robust optimization. Analyst 135(8), 2111–2118 (2010)
Pflug, G., Wozabal, D.: Ambiguity in portfolio selection. Quant. Finance 7(4), 435–442 (2007)
Rachev, S.T.: Probability Metrics and the Stability of Stochastic Models. Willey Series in Probability and Mathematical Statistics. Wiley, New York (1991)
Calyampudi Radhakrishna Rao: The theory of least squares when the parameters are stochastic and its application to the analysis of growth curves. Biometrika 52(3–4), 447–458 (1965)
Rendl, F., Wolkowicz, H.: A semidefinite framework for trust region subproblems with applications to large scale minimization. Math. Program. 77, 273–299 (1997)
Rockafellar R.T.: Conjugate Duality and Optimization. CBMS-NSF Regional Conference Series in Applied Mathematics 16. SIAM, Philadelphia (1974)
Ruszczynski, A., Shapiro, A. (eds.): Stochastic Programming. Handbooks in Operations Research and Management Science 10. North-Holland, Amsterdam (2003)
Scarf, H.: A min-max Solution of an Inventory Problem. Stanford University Press, Stanford (1958)
Shapiro, A., Ahmed, S.: On a class of minimax stochastic programs. SIAM J. Optim. 14(4), 1237–1249 (2004)
Shapiro, A., Homem de mello, T.: On the rate of convergence of optimal solutions of monte carlo approximations of stochastic programs. SIAM J. Optim. 11(1), 70–86 (2000)
Sturm Jos, F.: Using SeDuMi 1.02, a MATLAB* toolbox for optimization over symmetric cones. Optim. Methods Softw. 11–12, 625–653 (1999)
Sturm Jos, F.: Implementation of interior point methods for mixed semidefinite and second order cone optimization problems. Optim. Methods Softw. 17, 1105–1154 (2002)
Vaidya, P.M.: A new algorithm for minimizing convex functions over convex sets. Math. Program. 73, 291–341 (1996)
Vallander, S.S.: Calculation of the wasserstein distance between probability distributions on the line. Theory Probab. Appl. 18, 784–786 (1973)
Author information
Authors and Affiliations
Corresponding author
Additional information
Research partially supported by NSF grant CMMI-1100868 and ONR grant N00014-09-10518 and N00014210051.
Appendix
Appendix
In Sect. 3 we use SAA method to approximate the stochastic programming problems (3.13) and (3.18). The convergence results for the SAA method from [16] are summarized here. Consider a stochastic programming problem of the form
where \(F(\mathbf x ,\pmb {\xi })\) is a function of variables \(\mathbf x \in \mathbb R ^n\) and parameters \(\pmb {\xi } \in \mathbb R ^d\). \(\mathcal X \subset \mathbb R ^n\) is a given set, and \(\pmb {\xi }=\pmb {\xi }(\omega )\) is a random vector. The expectation in (6.1) is taken with respect to the probability distribution of \(\pmb {\xi }\) which is assumed to be known as \(\mathbb P \). Denote by \(\Xi \subset \mathbb R ^d\) the support of the probability distribution of \(\pmb {\xi }\), that is, \(\Xi \) is the smallest closed set in \(\mathbb R ^d\) such that the probability of the event \(\pmb {\xi } \in \mathbb R ^d \setminus \Xi \) is zero. Also denote by \(\mathbb P (A)\) the probability of an event \(A\). With the generated sample \(\xi ^1,\ldots ,\xi ^K\), we associate the sample average function
The stochastic programming problem (6.1) is approximated by the optimization problem
Let the optimal value of (6.1) be \(\nu \) and it’s optimal solution set be \(S\). Let \(\hat{\nu }_K\) and \(\hat{S}_K\) be the optimal value and the set of optimal solutions of the SAA problem (6.3). For sets \(A,B \subset \mathbf R ^n\), denote \(\mathrm{dist}(x,A):=\inf _{x^{\prime } \in A}\left| \left| x-x^{\prime }\right| \right| \) to be the distance from \(x \in \mathbb X ^n\) to \(A\), and
Also, define the function \((\mathbf x ,\xi ) \mapsto F(\mathbf x ,\xi )\) to be a random lower semicontinuous function if the associated epigraphical multifunction \(\xi \mapsto \mathrm{epi}F(\cdot ,\xi )\) is closed valued and measurable. We say that the Law of Large Numbers (LLN) holds, for \(\hat{f}_K(\mathbf x )\), pointwise if \(\hat{f}_K(\mathbf x )\) converges w.p.1 to \(f(\mathbf x )\), as \(K \rightarrow \infty \), for any fixed \(\mathbf x \in \mathbb R ^n\). The following convergence theorem ensures that a solution of SAA problem (6.2) converges to that of (6.1) as the sample size increases.
Theorem 8
[16, Chapter 6, Theorem 4.] Suppose that: (i) the integrand function \(F\) is random lower semicontinuous, (ii) for almost every \(\xi \in \Xi \) the function \(F(\cdot ,\xi )\) is convex, (iii) the set \(\mathcal X \) is closed and convex, (iv) the expected value function \(f\) is lower semicontinuous and there exists a point \(\hat{\mathbf{x }} \in \mathcal X \) such that \(f(\mathbf x ) \le +\infty \) for all \(x\) in a neighborhood of \(\hat{\mathbf{x }}\), (v) the set \(S\) of optimal solutions of the original problem (6.1) is nonempty and bounded, (vi) the LLN holds pointwise. Then \(\hat{\nu }_K \rightarrow \nu ^*\) and \(\mathbb D (\hat{S}_K,S) \rightarrow 0\) w.p.1 as \(K \rightarrow \infty \).
There are also results about the exponential convergence rate of the SAA method. Please refer to [16, 19] for details of such results. We note that the assumptions in Theorem 8 are satisfied for the models (3.13) and (3.18). Hence the convergence of the solution of SAA of these problems to a solution of the original problem is guaranteed.
Rights and permissions
About this article
Cite this article
Mehrotra, S., Zhang, H. Models and algorithms for distributionally robust least squares problems. Math. Program. 146, 123–141 (2014). https://doi.org/10.1007/s10107-013-0681-9
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10107-013-0681-9