首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 15 毫秒
1.
In this paper, we introduce an augmented Lagrangian function for a multiobjective optimization problem with an extended vector-valued function. On the basis of this augmented Lagrangian, set-valued dual maps and dual optimization problems are constructed. Weak and strong duality results are obtained. Necessary and sufficient conditions for uniformly exact penalization and exact penalization are established. Finally, comparisons of saddle-point properties are made between a class of augmented Lagrangian functions and nonlinear Lagrangian functions for a constrained multiobjective optimization problem.  相似文献   

2.
Many polynomial and discrete optimization problems can be reduced to multiextremal quadratic type models of nonlinear programming. For solving these problems one may use Lagrangian bounds in combination with branch and bound techniques. The Lagrangian bounds may be improved for some important examples by adding in a model the so-called superfluous quadratic constraints which modify Lagrangian bounds. Problems of finding Lagrangian bounds as a rule can be reduced to minimization of nonsmooth convex functions and may be successively solved by modern methods of nondifferentiable optimization. This approach is illustrated by examples of solving polynomial-type problems and some discrete optimization problems on graphs.  相似文献   

3.
向量映射的鞍点和Lagrange对偶问题   总被引:4,自引:0,他引:4  
本文研究拓扑向量空间广义锥-次类凸映射向量优化问题的鞍点最优性条件和Lagrange对偶问题,建立向量优化问题的Fritz John鞍点和Kuhn-Tucker鞍点的最优性条件及其与向量优化问题的有效解和弱有效解之间的联系。通过对偶问题和向量优化问题的标量化刻画各解之间的关系,给出目标映射是广义锥-次类凸的向量优化问题在其约束映射满足广义Slater约束规格的条件下的对偶定理。  相似文献   

4.
基于一个含有控制参数的修正Lagrangian函数,该文建立了一个求解非线性约束优化问题的修正Lagrangian算法.在一些适当的条件下,证明了控制参数存在一个阀值,当控制参数小于这一阀值时,由这一算法产生的序列解局部收敛于问题的Kuhn-Tucker点,并且建立了解的误差上界.最后给出一些约束优化问题的数值结果.  相似文献   

5.
The paper at hand presents a combination of optimal control approaches for PDEs with meshless discretizations. Applying a classical Lagrangian type particle method to optimization problems with hyperbolic constraints, several adjoint-based strategies differing in the sequential order of optimization and discretization of the Lagrangian or Eulerian problem formulation are proposed and compared. The numerical results confirm the theoretically predicted independence principle of the optimization approaches and show the expected convergence behavior. Moreover, they exemplify the superiority of meshless methods over the conventional mesh-based approaches for the numerical handling and optimization of problems with time-dependent geometries and freely moving boundaries.  相似文献   

6.
Sufficient and necessary optimality conditions are given for weakly minimized optimization problems in terms of a vector valued Lagrangian. Lagrangian and Wolfe type duals are constructed and duality established using an ordering that accords with the definition of a weak minimum. The results for differentiable problems continue to hold under weakened convexity assumptions and for problems which quasiminimize rather than minimize.  相似文献   

7.
A solution concept in optimization problems with interval-valued objective functions, which is essentially similar to the concept of nondominated solution in vector optimization problems, is introduced by imposing a partial ordering on the set of all closed intervals. The interval-valued Lagrangian function and interval-valued Lagrangian dual function are also proposed to formulate the dual problem of the interval-valued optimization problem. Under this setting, weak and strong duality theorems can be obtained.  相似文献   

8.
In this paper, we present a necessary and sufficient condition for a zero duality gap between a primal optimization problem and its generalized augmented Lagrangian dual problems. The condition is mainly expressed in the form of the lower semicontinuity of a perturbation function at the origin. For a constrained optimization problem, a general equivalence is established for zero duality gap properties defined by a general nonlinear Lagrangian dual problem and a generalized augmented Lagrangian dual problem, respectively. For a constrained optimization problem with both equality and inequality constraints, we prove that first-order and second-order necessary optimality conditions of the augmented Lagrangian problems with a convex quadratic augmenting function converge to that of the original constrained program. For a mathematical program with only equality constraints, we show that the second-order necessary conditions of general augmented Lagrangian problems with a convex augmenting function converge to that of the original constrained program.This research is supported by the Research Grants Council of Hong Kong (PolyU B-Q359.)  相似文献   

9.
Lagrangian Duality and Cone Convexlike Functions   总被引:1,自引:0,他引:1  
In this paper, we consider first the most important classes of cone convexlike vector-valued functions and give a dual characterization for some of these classes. It turns out that these characterizations are strongly related to the closely convexlike and Ky Fan convex bifunctions occurring within minimax problems. Applying the Lagrangian perturbation approach, we show that some of these classes of cone convexlike vector-valued functions show up naturally in verifying strong Lagrangian duality for finite-dimensional optimization problems. This is achieved by extending classical convexity results for biconjugate functions to the class of so-called almost convex functions. In particular, for a general class of finite-dimensional optimization problems, strong Lagrangian duality holds if some vector-valued function related to this optimization problem is closely K-convexlike and satisfies some additional regularity assumptions. For K a full-dimensional convex cone, it turns out that the conditions for strong Lagrangian duality simplify. Finally, we compare the results obtained by the Lagrangian perturbation approach worked out in this paper with the results achieved by the so-called image space approach initiated by Giannessi.  相似文献   

10.
Augmented Lagrangian function is one of the most important tools used in solving some constrained optimization problems. In this article, we study an augmented Lagrangian objective penalty function and a modified augmented Lagrangian objective penalty function for inequality constrained optimization problems. First, we prove the dual properties of the augmented Lagrangian objective penalty function, which are at least as good as the traditional Lagrangian function's. Under some conditions, the saddle point of the augmented Lagrangian objective penalty function satisfies the first-order Karush-Kuhn-Tucker condition. This is especially so when the Karush-Kuhn-Tucker condition holds for convex programming of its saddle point existence. Second, we prove the dual properties of the modified augmented Lagrangian objective penalty function. For a global optimal solution, when the exactness of the modified augmented Lagrangian objective penalty function holds, its saddle point exists. The sufficient and necessary stability conditions used to determine whether the modified augmented Lagrangian objective penalty function is exact for a global solution is proved. Based on the modified augmented Lagrangian objective penalty function, an algorithm is developed to find a global solution to an inequality constrained optimization problem, and its global convergence is also proved under some conditions. Furthermore, the sufficient and necessary calmness condition on the exactness of the modified augmented Lagrangian objective penalty function is proved for a local solution. An algorithm is presented in finding a local solution, with its convergence proved under some conditions.  相似文献   

11.
The Lagrangian function in the conventional theory for solving constrained optimization problems is a linear combination of the cost and constraint functions. Typically, the optimality conditions based on linear Lagrangian theory are either necessary or sufficient, but not both unless the underlying cost and constraint functions are also convex.We propose a somewhat different approach for solving a nonconvex inequality constrained optimization problem based on a nonlinear Lagrangian function. This leads to optimality conditions which are both sufficient and necessary, without any convexity assumption. Subsequently, under appropriate assumptions, the optimality conditions derived from the new nonlinear Lagrangian approach are used to obtain an equivalent root-finding problem. By appropriately defining a dual optimization problem and an alternative dual problem, we show that zero duality gap will hold always regardless of convexity, contrary to the case of linear Lagrangian duality.  相似文献   

12.
A generally nonconvex optimization problem with equality constraints is studied. The problem is introduced as an “inf sup” of a generalized augmented Lagrangian function. A dual problem is defined as the “sup inf” of the same generalized augmented Lagrangian. Sufficient conditions are derived for constructing the augmented Lagrangian function such that the extremal values of the primal and dual problems are equal. Characterization of a class of augmented Lagrangian functions which satisfy the sufficient conditions for strong duality is presented. Finally, some examples of functions and primal-dual problems in the above-mentioned class are presented.  相似文献   

13.
In this paper, foundations of a new approach for solving vector optimization problems are introduced. Generalized Lagrangian duality, related for the first time with vector optimization, provides new scalarization techniques and allows for the generation of efficient solutions for problems which are not required to satisfy any convexity assumptions.  相似文献   

14.
A solution concept of fuzzy optimization problems, which is essentially similar to the notion of Pareto optimal solution (nondominated solution) in multiobjective programming problems, is introduced by imposing a partial ordering on the set of all fuzzy numbers. We also introduce a concept of fuzzy scalar (inner) product based on the positive and negative parts of fuzzy numbers. Then the fuzzy-valued Lagrangian function and the fuzzy-valued Lagrangian dual function for the fuzzy optimization problem are proposed via the concept of fuzzy scalar product. Under these settings, the weak and strong duality theorems for fuzzy optimization problems can be elicited. We show that there is no duality gap between the primal and dual fuzzy optimization problems under suitable assumptions for fuzzy-valued functions.  相似文献   

15.
The augmented Lagrangian method is a classical method for solving constrained optimization.Recently,the augmented Lagrangian method attracts much attention due to its applications to sparse optimization in compressive sensing and low rank matrix optimization problems.However,most Lagrangian methods use first order information to update the Lagrange multipliers,which lead to only linear convergence.In this paper,we study an update technique based on second order information and prove that superlinear convergence can be obtained.Theoretical properties of the update formula are given and some implementation issues regarding the new update are also discussed.  相似文献   

16.
Convex relaxations can be used to obtain lower bounds on the optimal objective function value of nonconvex quadratically constrained quadratic programs. However, for some problems, significantly better bounds can be obtained by minimizing the restricted Lagrangian function for a given estimate of the Lagrange multipliers. The difficulty in utilizing Lagrangian duality within a global optimization context is that the restricted Lagrangian is often nonconvex. Minimizing a convex underestimate of the restricted Lagrangian overcomes this difficulty and facilitates the use of Lagrangian duality within a global optimization framework. A branch-and-bound algorithm is presented that relies on these Lagrangian underestimates to provide lower bounds and on the interval Newton method to facilitate convergence in the neighborhood of the global solution. Computational results show that the algorithm compares favorably to the Reformulation–Linearization Technique for problems with a favorable structure.  相似文献   

17.
<正>Analysis on a Superlinearly Convergent Augmented Lagrangian Method Ya Xiang YUAN Abstract The augmented Lagrangian method is a classical method for solving constrained optimization.Recently,the augmented Lagrangian method attracts much attention due to its applications to sparse optimization in compressive sensing and low rank matrix optimization problems.However,most Lagrangian methods use first order information to update the Lagrange multipliers,which lead to only linear convergence.In this paper,we study an update technique  相似文献   

18.
Refinement of Lagrangian bounds in optimization problems   总被引:1,自引:0,他引:1  
Lagrangian constraint relaxation and the corresponding bounds for the optimal value of an original optimization problem are examined. Techniques for the refinement of the classical Lagrangian bounds are investigated in the case where the complementary slackness conditions are not fulfilled because either the original formulation is nonconvex or the Lagrange multipliers are nonoptimal. Examples are given of integer and convex problems for which the modified bounds improve the classical Lagrangian bounds.  相似文献   

19.
B. Jin 《Optimization》2016,65(6):1151-1166
In this paper, we revisit the augmented Lagrangian method for a class of nonsmooth convex optimization. We present the Lagrange optimality system of the augmented Lagrangian associated with the problems, and establish its connections with the standard optimality condition and the saddle point condition of the augmented Lagrangian, which provides a powerful tool for developing numerical algorithms: we derive a Lagrange–Newton algorithm for the nonsmooth convex optimization, and establish the nonsingularity of the Newton system and the local convergence of the algorithm.  相似文献   

20.
Lagrangian bounds, i.e. bounds computed by Lagrangian relaxation, have been used successfully in branch and bound bound methods for solving certain classes of nonconvex optimization problems by reducing the duality gap. We discuss this method for the class of partly linear and partly convex optimization problems and, incidentally, point out incorrect results in the recent literature on this subject.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号