Results 1  10
of
13
Snopt: An SQP Algorithm For LargeScale Constrained Optimization
, 1997
"... Sequential quadratic programming (SQP) methods have proved highly effective for solving constrained optimization problems with smooth nonlinear functions in the objective and constraints. Here we consider problems with general inequality constraints (linear and nonlinear). We assume that first deriv ..."
Abstract

Cited by 328 (18 self)
 Add to MetaCart
Sequential quadratic programming (SQP) methods have proved highly effective for solving constrained optimization problems with smooth nonlinear functions in the objective and constraints. Here we consider problems with general inequality constraints (linear and nonlinear). We assume that first derivatives are available, and that the constraint gradients are sparse.
User's Guide For SNOPT 5.3: A Fortran Package For LargeScale Nonlinear Programming
, 1999
"... SNOPT is a generalpurpose system for solving optimization problems involving many variables and constraints. It minimizes a linear or nonlinear function subject to bounds on the variables and sparse linear or nonlinear constraints. It is suitable for largescale linear and quadratic programming ..."
Abstract

Cited by 75 (1 self)
 Add to MetaCart
SNOPT is a generalpurpose system for solving optimization problems involving many variables and constraints. It minimizes a linear or nonlinear function subject to bounds on the variables and sparse linear or nonlinear constraints. It is suitable for largescale linear and quadratic programming and for linearly constrained optimization, as well as for general nonlinear programs. SNOPT finds solutions that are locally optimal , and ideally any nonlinear functions should be smooth and users should provide gradients. It is often more widely useful. For example, local optima are often global solutions, and discontinuities in the function gradients can often be tolerated if they are not too close to an optimum. Unknown gradients are estimated by finite differences. SNOPT uses a sequential quadratic programming (SQP) algorithm that obtains search directions from a sequence of quadratic programming subproblems. Each QP subproblem minimizes a quadratic model of a certain Lagrangian function subject to a linearization of the constraints. An augmented Lagrangian merit function is reduced along each search direction to ensure convergence from any starting point. SNOPT is most efficient if only some of the variables enter nonlinearly, or if the number of active constraints (including simple bounds) is nearly as large as the number of variables. SNOPT requires relatively few evaluations of the problem functions. Hence it is especially effective if the objective or constraint functions (and their gradients) are expensive to evaluate. The source code for SNOPT is suitable for any machine with a Fortran compiler. SNOPT may be called from a driver program (typically in Fortran, C or MATLAB). SNOPT can also be used as a standalone package, reading data in the MPS ...
A reduced Hessian method for largescale constrained optimization
 SIAM JOURNAL ON OPTIMIZATION
, 1995
"... ..."
The ULagrangian of a convex function
 Trans. Amer. Math. Soc
"... Abstract. At a given point p, a convex function f is differentiable in a certain subspace U (the subspace along which ∂f(p) has 0breadth). This property opens the way to defining a suitably restricted second derivative of f at p. We do this via an intermediate function, convex on U. We call this fu ..."
Abstract

Cited by 25 (7 self)
 Add to MetaCart
Abstract. At a given point p, a convex function f is differentiable in a certain subspace U (the subspace along which ∂f(p) has 0breadth). This property opens the way to defining a suitably restricted second derivative of f at p. We do this via an intermediate function, convex on U. We call this function the ULagrangian; it coincides with the ordinary Lagrangian in composite cases: exact penalty, semidefinite programming. Also, we use this new theory to design a conceptual pattern for superlinearly convergent minimization algorithms. Finally, we establish a connection with the MoreauYosida regularization. 1.
A Newton Barrier method for Minimizing a Sum of Euclidean Norms subject to linear equality constraints
, 1995
"... An algorithm for minimizing a sum of Euclidean Norms subject to linear equality constraints is described. The algorithm is based on a recently developed Newton barrier method for the unconstrained minimization of a sum of Euclidean norms (MSN ). The linear equality constraints are handled using an e ..."
Abstract

Cited by 18 (2 self)
 Add to MetaCart
An algorithm for minimizing a sum of Euclidean Norms subject to linear equality constraints is described. The algorithm is based on a recently developed Newton barrier method for the unconstrained minimization of a sum of Euclidean norms (MSN ). The linear equality constraints are handled using an exact L 1 penalty function which is made smooth in the same way as the Euclidean norms. It is shown that the dual problem is to maximize a linear objective function subject to homogeneous linear equality constraints and quadratic inequalities. Hence the suggested method also solves such problems efficiently. In fact such a problem from plastic collapse analysis motivated this work. Numerical results are presented for large sparse problems, demonstrating the extreme efficiency of the method. Keywords: Sum of Norms, Nonsmooth Optimization, Duality, Newton Barrier Method. AMS(MOS) subject classification: 65K05, 90C06, 90C25, 90C90. Abbreviated title: A Newton barrier method. Supported by the ...
User’s Guide for SNOPT Version 7: Software for LargeScale Nonlinear Programming
"... SNOPT is a generalpurpose system for constrained optimization. It minimizes a linear or nonlinear function subject to bounds on the variables and sparse linear or nonlinear constraints. It is suitable for largescale linear and quadratic programming and for linearly constrained optimization, as wel ..."
Abstract

Cited by 16 (0 self)
 Add to MetaCart
SNOPT is a generalpurpose system for constrained optimization. It minimizes a linear or nonlinear function subject to bounds on the variables and sparse linear or nonlinear constraints. It is suitable for largescale linear and quadratic programming and for linearly constrained optimization, as well as for general nonlinear programs. SNOPT finds solutions that are locally optimal, and ideally any nonlinear functions should be smooth and users should provide gradients. It is often more widely useful. For example, local optima are often global solutions, and discontinuities in the function gradients can often be tolerated if they are not too close to an optimum. Unknown gradients are estimated by finite differences. SNOPT uses a sequential quadratic programming (SQP) algorithm. Search directions are obtained from QP subproblems that minimize a quadratic model of the Lagrangian function subject to linearized constraints. An augmented Lagrangian merit function is reduced along each search direction to ensure convergence from any starting point.
GAMS/SNOPT: An SQP Algorithm For LargeScale Constrained Optimization
, 2000
"... Introduction This section describes the GAMS interface to the generalpurpose NLP solver SNOPT, (Sparse Nonlinear Optimizer) which implements a sequential quadratic programming (SQP) method for solving constrained optimization problems with smooth nonlinear functions in the objective and constraints ..."
Abstract

Cited by 3 (0 self)
 Add to MetaCart
Introduction This section describes the GAMS interface to the generalpurpose NLP solver SNOPT, (Sparse Nonlinear Optimizer) which implements a sequential quadratic programming (SQP) method for solving constrained optimization problems with smooth nonlinear functions in the objective and constraints. The optimization problem is assumed to be stated in the form NP minimize x or maximize x f(x) subject to F (x) # b 1 Gx # b 2 l # x # u, (1) where x # # n , f(x) is a linear or nonlinear smooth objective function, l and u are constant lower and upper bounds, F (x) is a set of nonlinear constraint functions, G is a sparse ma
Exact Penalty Methods
 In I. Ciocco (Ed.), Algorithms for Continuous Optimization
, 1994
"... . Exact penalty methods for the solution of constrained optimization problems are based on the construction of a function whose unconstrained minimizing points are also solution of the constrained problem. In the first part of this paper we recall some definitions concerning exactness properties of ..."
Abstract

Cited by 3 (1 self)
 Add to MetaCart
. Exact penalty methods for the solution of constrained optimization problems are based on the construction of a function whose unconstrained minimizing points are also solution of the constrained problem. In the first part of this paper we recall some definitions concerning exactness properties of penalty functions, of barrier functions, of augmented Lagrangian functions, and discuss under which assumptions on the constrained problem these properties can be ensured. In the second part of the paper we consider algorithmic aspects of exact penalty methods; in particular we show that, by making use of continuously differentiable functions that possess exactness properties, it is possible to define implementable algorithms that are globally convergent with superlinear convergence rate towards KKT points of the constrained problem. 1 Introduction "It would be a major theoretic breakthrough in nonlinear programming if a simple continuously differentiable function could be exhibited with th...
A Reduced Hessian Method for LargeScale Constrained Optimization
 SIAM J. Optimization
, 1993
"... We propose a quasiNewton algorithm for solving large optimization problems with nonlinear equality constraints. It is designed for problems with few degrees of freedom and is motivated by the need to use sparse matrix factorizations. The algorithm incorporates a correction vector that approximates ..."
Abstract

Cited by 1 (0 self)
 Add to MetaCart
We propose a quasiNewton algorithm for solving large optimization problems with nonlinear equality constraints. It is designed for problems with few degrees of freedom and is motivated by the need to use sparse matrix factorizations. The algorithm incorporates a correction vector that approximates the cross term Z T I/VYp in order to estimate the curvature in both the range and null spaces of the constraints. The algorithm can be considered to be, in some sense, a practical implementation of an algorithm of Coleman and Conn. We give conditions under which local and superlinear convergence is obtained.