𝔖 Bobbio Scriptorium
✦   LIBER   ✦

A simple feasible SQP method for inequality constrained optimization with global and superlinear convergence

✍ Scribed by Zhong Jin; Yuqing Wang


Publisher
Elsevier Science
Year
2010
Tongue
English
Weight
750 KB
Volume
233
Category
Article
ISSN
0377-0427

No coin nor oath required. For personal study only.

✦ Synopsis


In this paper, a simple feasible SQP method for nonlinear inequality constrained optimization is presented. At each iteration, we need to solve one QP subproblem only. After solving a system of linear equations, a new feasible descent direction is designed. The Maratos effect is avoided by using a high-order corrected direction. Under some suitable conditions the global and superlinear convergence can be induced. In the end, numerical experiments show that the method in this paper is effective.


πŸ“œ SIMILAR VOLUMES


A generalized super-memory gradient proj
✍ Jin-Bao Jian; You-Fang Zeng; Chun-Ming Tang πŸ“‚ Article πŸ“… 2007 πŸ› Elsevier Science 🌐 English βš– 362 KB

In this work, combining the properties of the generalized super-memory gradient projection methods with the ideas of the strongly sub-feasible directions methods, we present a new algorithm with strong convergence for nonlinear inequality constrained optimization. At each iteration, the proposed alg

A superlinearly convergent method of qua
✍ Jin-bao Jian; Yi Liu πŸ“‚ Article πŸ“… 2011 πŸ› Elsevier Science 🌐 English βš– 297 KB

Combining the norm-relaxed sequential quadratic programming (SQP) method and the idea of method of quasi-strongly sub-feasible directions (MQSSFD) with active set identification technique, a new SQP algorithm for solving nonlinear inequality constrained optimization is proposed. Unlike the previous

A combined global and local search metho
✍ Xi Chen; Jing Yang; Zhaohua Li; Daqing Tian; Zhijiang Shao πŸ“‚ Article πŸ“… 2008 πŸ› John Wiley and Sons 🌐 English βš– 269 KB

## Abstract Heuristic methods, such as tabu search, are efficient for global optimizations. Most studies, however, have focused on constraint‐free optimizations. Penalty functions are commonly used to deal with constraints for global optimization algorithms in dealing with constraints. This is some