首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 31 毫秒
1.
We provide a refined convergence analysis for the SAA (sample average approximation) method applied to stochastic optimization problems with either single or mixed CVaR (conditional value-at-risk) measures. Under certain regularity conditions, it is shown that any accumulation point of the weak GKKT (generalized Karush-Kuhn-Tucker) points produced by the SAA method is almost surely a weak stationary point of the original CVaR or mixed CVaR optimization problems. In addition, it is shown that, as the sample size increases, the difference of the optimal values between the SAA problems and the original problem tends to zero with probability approaching one exponentially fast.  相似文献   

2.
Journal of Optimization Theory and Applications - We study a first-order primal-dual subgradient method to optimize risk-constrained risk-penalized optimization problems, where risk is modeled via...  相似文献   

3.
In this paper, we consider the stochastic second-order cone complementarity problems (SSOCCP). We first formulate the SSOCCP contained expectation as an optimization problem using the so-called second-order cone complementarity function. We then use sample average approximation method and smoothing technique to obtain the approximation problems for solving this reformulation. In theory, we show that any accumulation point of the global optimal solutions or stationary points of the approximation problems are global optimal solution or stationary point of the original problem under suitable conditions. Finally, some numerical examples are given to explain that the proposed methods are feasible.  相似文献   

4.
We investigate one stage stochastic multiobjective optimization problems where the objectives are the expected values of random functions. Assuming that the closed form of the expected values is difficult to obtain, we apply the well known Sample Average Approximation (SAA) method to solve it. We propose a smoothing infinity norm scalarization approach to solve the SAA problem and analyse the convergence of efficient solution of the SAA problem to the original problem as sample sizes increase. Under some moderate conditions, we show that, with probability approaching one exponentially fast with the increase of sample size, an ϵ-optimal solution to the SAA problem becomes an ϵ-optimal solution to its true counterpart. Moreover, under second order growth conditions, we show that an efficient point of the smoothed problem approximates an efficient solution of the true problem at a linear rate. Finally, we describe some numerical experiments on some stochastic multiobjective optimization problems and report preliminary results.  相似文献   

5.
6.
We study sample approximations of chance constrained problems. In particular, we consider the sample average approximation (SAA) approach and discuss the convergence properties of the resulting problem. We discuss how one can use the SAA method to obtain good candidate solutions for chance constrained problems. Numerical experiments are performed to correctly tune the parameters involved in the SAA. In addition, we present a method for constructing statistical lower bounds for the optimal value of the considered problem and discuss how one should tune the underlying parameters. We apply the SAA to two chance constrained problems. The first is a linear portfolio selection problem with returns following a multivariate lognormal distribution. The second is a joint chance constrained version of a simple blending problem. B.K. Pagnoncelli’s research was supported by CAPES and FUNENSEG. S. Ahmed’s research was partly supported by the NSF Award DMI-0133943. A. Shapiro’s research was partly supported by the NSF Award DMI-0619977.  相似文献   

7.
In the matter of Portfolio selection, we consider an extended version of the Mean-Absolute Deviation (MAD) model, which includes discrete asset choice constraints (threshold and cardinality constraints) and one is allowed to sell assets short if it leads to a better risk-return tradeoff. Cardinality constraints limit the number of assets in the optimal portfolio and threshold constraints limit the amount of capital to be invested in (or sold short from) each asset and prevent very small investments in (or short selling from) any asset. The problem is formulated as a mixed 0–1 programming problem, which is known to be NP-hard. Attempting to use DC (Difference of Convex functions) programming and DCA (DC Algorithms), an efficient approach in non-convex programming framework, we reformulate the problem in terms of a DC program, and investigate a DCA scheme to solve it. Some computational results carried out on benchmark data sets show that DCA has a better performance in comparison to the standard solver IBM CPLEX.  相似文献   

8.
资产组合的CVaR风险的敏感度分析   总被引:6,自引:0,他引:6       下载免费PDF全文
基于CVaR风险计量技术,分别给出了正态和t分布情形下资产组合的CVaR值,对一般情形下风险资产组合的CVaR风险关于头寸的敏感度进行了分析,研究了其经济意义。  相似文献   

9.
In this paper, by using the notion of strong subdifferential and epsilon-subdifferential, necessary optimality conditions are established firstly for an epsilon-weak Pareto minimal point and an epsilon-proper Pareto minimal point of a vector optimization problem, where its objective function and constraint set are denoted by using differences of two vector-valued maps, respectively. Then, by using the concept of approximate pseudo-dissipativity, sufficient optimality conditions are obtained. As an application of these results, sufficient and necessary optimality conditions are also given for an epsilon-weak Pareto minimal point and an epsilon-proper Pareto minimal point of a vector fractional mathematical programming.  相似文献   

10.
11.
This paper presents new approximation bounds for trilinear and biquadratic optimization problems over nonconvex constraints. We first consider the partial semidefinite relaxation of the original problem, and show that there is a bounded approximation solution to it. This will be achieved by determining the diameters of certain convex bodies. We then show that there is also a bounded approximation solution to the original problem via extracting the approximation solution of its semidefinite relaxation. Under some conditions, the approximation bounds obtained in this paper improve those in the literature.  相似文献   

12.
We consider convex stochastic optimization problems under different assumptions on the properties of available stochastic subgradient. It is known that, if the value of the objective function is available, one can obtain, in parallel, several independent approximate solutions in terms of the objective residual expectation. Then, choosing the solution with the minimum function value, one can control the probability of large deviation of the objective residual. On the contrary, in this short paper, we address the situation, when the value of the objective function is unavailable or is too expensive to calculate. Under "‘light-tail"’ assumption for stochastic subgradient and in general case with moderate large deviation probability, we show that parallelization combined with averaging gives bounds for probability of large deviation similar to a serial method. Thus, in these cases, one can benefit from parallel computations and reduce the computational time without loss in the solution quality.  相似文献   

13.
Using the concept of a subdifferential of a vector-valued convex mapping, we provide duality formulas for the minimization of nonconvex composite functions and related optimization problems such as the minimization of a convex function over a vectorial DC constraint.  相似文献   

14.
We investigate sample average approximation of a general class of one-stage stochastic mathematical programs with equilibrium constraints. By using graphical convergence of unbounded set-valued mappings, we demonstrate almost sure convergence of a sequence of stationary points of sample average approximation problems to their true counterparts as the sample size increases. In particular we show the convergence of M(Mordukhovich)-stationary point and C(Clarke)-stationary point of the sample average approximation problem to those of the true problem. The research complements the existing work in the literature by considering a general constraint to be represented by a stochastic generalized equation and exploiting graphical convergence of coderivative mappings.  相似文献   

15.
In this paper, we study optimality conditions for vector optimization problems of a difference of convex mappings
where is a closed convex cone in a Banach space Z, l is a mapping Q-convex from a Banach space X into Z, A is a continuous linear operator from X into a Banach space and are respectively the nonnegative orthants of and , C is a nonempty closed convex subset of X, bW, and the functions fi,gi,hj and kj are convex for i=1,...,p and j=1,ldots,m. Necessary optimality conditions for (VP) are established in terms of Lagrange-Fritz-John multipliers. When the set of constraints for (VP) is convex and under the generalized Slater constraint qualification introduced in Jeyakumar and Wolkowicz [11] , we derive necessary optimality conditions in terms of Lagrange-Karush-Kuhn-Tucker multipliers which are also sufficient whenever the functions gi,i=1,...,p are polyhedrals. Our approach consists in using a special scalarization function. A necessary optimality condition for convex vector maximization problem is derived. Also an application to vector fractional mathematical programming is given. Our contribution extends the results obtained in scalar optimization by Hiriart-Urruty [9] and improve substantially the few results known in vector case (see for instance: [11], [12] and [14]).Mathematics Subject Classification (1991). Primary: 90C29; Secondary 49K30  相似文献   

16.
The DC programming and its DC algorithm (DCA) address the problem of minimizing a function f=gh (with g,h being lower semicontinuous proper convex functions on R n ) on the whole space. Based on local optimality conditions and DC duality, DCA was successfully applied to a lot of different and various nondifferentiable nonconvex optimization problems to which it quite often gave global solutions and proved to be more robust and more efficient than related standard methods, especially in the large scale setting. The computational efficiency of DCA suggests to us a deeper and more complete study on DC programming, using the special class of DC programs (when either g or h is polyhedral convex) called polyhedral DC programs. The DC duality is investigated in an easier way, which is more convenient to the study of optimality conditions. New practical results on local optimality are presented. We emphasize regularization techniques in DC programming in order to construct suitable equivalent DC programs to nondifferentiable nonconvex optimization problems and new significant questions which have to be answered. A deeper insight into DCA is introduced which really sheds new light on DCA and could partly explain its efficiency. Finally DC models of real world nonconvex optimization are reported.  相似文献   

17.
本文首先对IPA算法进行了修正,并证明了修正IPA算法的收敛性,然后将修正后的IPA应用到不等式约束凸优化问题中得到新的内点算法,并与传统的障碍函数法作了比较,从理论上体现了新算法的优势,并给出了其工程解求解法以及收敛性的证明.  相似文献   

18.
We derive an implementable algorithm for solving nonlinear stochastic optimization problems with failure probability constraints using sample average approximations. The paper extends prior results dealing with a failure probability expressed by a single measure to the case of failure probability expressed in terms of multiple performance measures. We also present a new formula for the failure probability gradient. A numerical example addressing the optimal design of a reinforced concrete highway bridge illustrates the algorithm. This work was sponsored by the Research Associateship Program, National Research Council. The authors are grateful for the valuable insight from Professors Alexander Shapiro, Evarist Gine, and Jon A. Wellner. The authors also thank Professor Tito Homem-de-Mello for commenting on an early draft of this paper.  相似文献   

19.
In this paper, we consider the computation of a rigorous lower error bound for the optimal value of convex optimization problems. A discussion of large-scale problems, degenerate problems, and quadratic programming problems is included. It is allowed that parameters, whichdefine the convex constraints and the convex objective functions, may be uncertain and may vary between given lower and upper bounds. The error bound is verified for the family of convex optimization problems which correspond to these uncertainties. It can be used to perform a rigorous sensitivity analysis in convex programming, provided the width of the uncertainties is not too large. Branch and bound algorithms can be made reliable by using such rigorous lower bounds.  相似文献   

20.
Doklady Mathematics - We consider a number of optimal control problems with 2-dimensional control lying in an arbitrary convex compact set $$\Omega $$ . Solutions to these problems are obtained...  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号