Repository logo
 

Constrained NLP via gradient flow penalty continuation: Towards self-tuning robust penalty schemes

Accepted version
Peer-reviewed

Type

Article

Change log

Authors

Scott, F 
Conejeros, R 
Vassiliadis, VS 

Abstract

This work presents a new numerical solution approach to nonlinear constrained optimization problems based on a gradient flow reformulation. The proposed solution schemes use self-tuning penalty parameters where the updating of the penalty parameter is directly embedded in the system of ODEs used in the reformulation, and its growth rate is linked to the violation of the constraints and variable bounds. The convergence properties of these schemes are analyzed, and it is shown that they converge to a local minimum asymptotically. Numerical experiments using a set of test problems, ranging from a few to several hundred variables, show that the proposed schemes are robust and converge to feasible points and local minima. Moreover, results suggest that the GF formulations were able to find the optimal solution to problems where conventional NLP solvers fail, and in less integration steps and time compared to a previously reported GF formulation.

Description

Keywords

gradient flow, nonlinear programming problem, convergence analysis

Journal Title

Computers and Chemical Engineering

Conference Name

Journal ISSN

0098-1354
1873-4375

Volume Title

101

Publisher

Elsevier