Results 1  10
of
17
SNOPT: An SQP Algorithm For LargeScale Constrained Optimization
, 2002
"... Sequential quadratic programming (SQP) methods have proved highly effective for solving constrained optimization problems with smooth nonlinear functions in the objective and constraints. Here we consider problems with general inequality constraints (linear and nonlinear). We assume that first deriv ..."
Abstract

Cited by 597 (24 self)
 Add to MetaCart
(Show Context)
Sequential quadratic programming (SQP) methods have proved highly effective for solving constrained optimization problems with smooth nonlinear functions in the objective and constraints. Here we consider problems with general inequality constraints (linear and nonlinear). We assume that first derivatives are available, and that the constraint gradients are sparse. We discuss
User's Guide For SNOPT 5.3: A Fortran Package For LargeScale Nonlinear Programming
, 1999
"... SNOPT is a generalpurpose system for solving optimization problems involving many variables and constraints. It minimizes a linear or nonlinear function subject to bounds on the variables and sparse linear or nonlinear constraints. It is suitable for largescale linear and quadratic programming ..."
Abstract

Cited by 96 (2 self)
 Add to MetaCart
(Show Context)
SNOPT is a generalpurpose system for solving optimization problems involving many variables and constraints. It minimizes a linear or nonlinear function subject to bounds on the variables and sparse linear or nonlinear constraints. It is suitable for largescale linear and quadratic programming and for linearly constrained optimization, as well as for general nonlinear programs. SNOPT finds solutions that are locally optimal , and ideally any nonlinear functions should be smooth and users should provide gradients. It is often more widely useful. For example, local optima are often global solutions, and discontinuities in the function gradients can often be tolerated if they are not too close to an optimum. Unknown gradients are estimated by finite differences. SNOPT uses a sequential quadratic programming (SQP) algorithm that obtains search directions from a sequence of quadratic programming subproblems. Each QP subproblem minimizes a quadratic model of a certain Lagrangian function subject to a linearization of the constraints. An augmented Lagrangian merit function is reduced along each search direction to ensure convergence from any starting point. SNOPT is most efficient if only some of the variables enter nonlinearly, or if the number of active constraints (including simple bounds) is nearly as large as the number of variables. SNOPT requires relatively few evaluations of the problem functions. Hence it is especially effective if the objective or constraint functions (and their gradients) are expensive to evaluate. The source code for SNOPT is suitable for any machine with a Fortran compiler. SNOPT may be called from a driver program (typically in Fortran, C or MATLAB). SNOPT can also be used as a standalone package, reading data in the MPS ...
User’s Guide for SNOPT Version 7: Software for LargeScale Nonlinear Programming
"... SNOPT is a generalpurpose system for constrained optimization. It minimizes a linear or nonlinear function subject to bounds on the variables and sparse linear or nonlinear constraints. It is suitable for largescale linear and quadratic programming and for linearly constrained optimization, as wel ..."
Abstract

Cited by 49 (1 self)
 Add to MetaCart
SNOPT is a generalpurpose system for constrained optimization. It minimizes a linear or nonlinear function subject to bounds on the variables and sparse linear or nonlinear constraints. It is suitable for largescale linear and quadratic programming and for linearly constrained optimization, as well as for general nonlinear programs. SNOPT finds solutions that are locally optimal, and ideally any nonlinear functions should be smooth and users should provide gradients. It is often more widely useful. For example, local optima are often global solutions, and discontinuities in the function gradients can often be tolerated if they are not too close to an optimum. Unknown gradients are estimated by finite differences. SNOPT uses a sequential quadratic programming (SQP) algorithm. Search directions are obtained from QP subproblems that minimize a quadratic model of the Lagrangian function subject to linearized constraints. An augmented Lagrangian merit function is reduced along each search direction to ensure convergence from any starting point.
A reduced Hessian method for largescale constrained optimization
 SIAM JOURNAL ON OPTIMIZATION
, 1995
"... ..."
The ULagrangian of a convex function
 TRANS. AMER. MATH. SOC
, 1999
"... At a given point p, a convex function f is differentiable in a certain subspace U (the subspace along which ∂f(p) has 0breadth). This property opens the way to defining a suitably restricted second derivative of f at p. We do this via an intermediate function, convex on U. We call this function t ..."
Abstract

Cited by 34 (7 self)
 Add to MetaCart
(Show Context)
At a given point p, a convex function f is differentiable in a certain subspace U (the subspace along which ∂f(p) has 0breadth). This property opens the way to defining a suitably restricted second derivative of f at p. We do this via an intermediate function, convex on U. We call this function the ULagrangian; it coincides with the ordinary Lagrangian in composite cases: exact penalty, semidefinite programming. Also, we use this new theory to design a conceptual pattern for superlinearly convergent minimization algorithms. Finally, we establish a connection with the MoreauYosida regularization.
A Newton Barrier method for Minimizing a Sum of Euclidean Norms subject to linear equality constraints
, 1995
"... An algorithm for minimizing a sum of Euclidean Norms subject to linear equality constraints is described. The algorithm is based on a recently developed Newton barrier method for the unconstrained minimization of a sum of Euclidean norms (MSN ). The linear equality constraints are handled using an e ..."
Abstract

Cited by 27 (2 self)
 Add to MetaCart
An algorithm for minimizing a sum of Euclidean Norms subject to linear equality constraints is described. The algorithm is based on a recently developed Newton barrier method for the unconstrained minimization of a sum of Euclidean norms (MSN ). The linear equality constraints are handled using an exact L 1 penalty function which is made smooth in the same way as the Euclidean norms. It is shown that the dual problem is to maximize a linear objective function subject to homogeneous linear equality constraints and quadratic inequalities. Hence the suggested method also solves such problems efficiently. In fact such a problem from plastic collapse analysis motivated this work. Numerical results are presented for large sparse problems, demonstrating the extreme efficiency of the method. Keywords: Sum of Norms, Nonsmooth Optimization, Duality, Newton Barrier Method. AMS(MOS) subject classification: 65K05, 90C06, 90C25, 90C90. Abbreviated title: A Newton barrier method. Supported by the ...
Users Guide to SNOPT Version 6: A Fortran Package for LargeScale Nonlinear Programming
, 2002
"... SNOPT is a generalpurpose system for solving optimization problems involving many variables and constraints. It minimizes a linear or nonlinear function subject to bounds on the variables and sparse linear or nonlinear constraints. It is suitable for largescale linear and quadratic programming and ..."
Abstract

Cited by 11 (0 self)
 Add to MetaCart
(Show Context)
SNOPT is a generalpurpose system for solving optimization problems involving many variables and constraints. It minimizes a linear or nonlinear function subject to bounds on the variables and sparse linear or nonlinear constraints. It is suitable for largescale linear and quadratic programming and for linearly constrained optimization, as well as for general nonlinear programs. SNOPT finds solutions that are locally optimal, and ideally any nonlinear functions should be smooth and users should provide gradients. It is often more widely useful. For example, local optima are often global solutions, and discontinuities in the function gradients can often be tolerated if they are not too close to an optimum. Unknown gradients are estimated by finite differences. SNOPT uses a sequential quadratic programming (SQP) algorithm that obtains search directions from a sequence of quadratic programming subproblems. Each QP subproblem minimizes a quadratic model of a certain Lagrangian function subject to a linearization of the constraints. An augmented Lagrangian merit function is reduced
Exact Penalty Methods
 In I. Ciocco (Ed.), Algorithms for Continuous Optimization
, 1994
"... . Exact penalty methods for the solution of constrained optimization problems are based on the construction of a function whose unconstrained minimizing points are also solution of the constrained problem. In the first part of this paper we recall some definitions concerning exactness properties of ..."
Abstract

Cited by 4 (2 self)
 Add to MetaCart
(Show Context)
. Exact penalty methods for the solution of constrained optimization problems are based on the construction of a function whose unconstrained minimizing points are also solution of the constrained problem. In the first part of this paper we recall some definitions concerning exactness properties of penalty functions, of barrier functions, of augmented Lagrangian functions, and discuss under which assumptions on the constrained problem these properties can be ensured. In the second part of the paper we consider algorithmic aspects of exact penalty methods; in particular we show that, by making use of continuously differentiable functions that possess exactness properties, it is possible to define implementable algorithms that are globally convergent with superlinear convergence rate towards KKT points of the constrained problem. 1 Introduction "It would be a major theoretic breakthrough in nonlinear programming if a simple continuously differentiable function could be exhibited with th...