Results 1  10
of
191
Snopt: An SQP Algorithm For LargeScale Constrained Optimization
, 1997
"... Sequential quadratic programming (SQP) methods have proved highly effective for solving constrained optimization problems with smooth nonlinear functions in the objective and constraints. Here we consider problems with general inequality constraints (linear and nonlinear). We assume that first deriv ..."
Abstract

Cited by 328 (18 self)
 Add to MetaCart
Sequential quadratic programming (SQP) methods have proved highly effective for solving constrained optimization problems with smooth nonlinear functions in the objective and constraints. Here we consider problems with general inequality constraints (linear and nonlinear). We assume that first derivatives are available, and that the constraint gradients are sparse.
Probing the Pareto frontier for basis pursuit solutions
, 2008
"... The basis pursuit problem seeks a minimum onenorm solution of an underdetermined leastsquares problem. Basis pursuit denoise (BPDN) fits the leastsquares problem only approximately, and a single parameter determines a curve that traces the optimal tradeoff between the leastsquares fit and the ..."
Abstract

Cited by 160 (2 self)
 Add to MetaCart
The basis pursuit problem seeks a minimum onenorm solution of an underdetermined leastsquares problem. Basis pursuit denoise (BPDN) fits the leastsquares problem only approximately, and a single parameter determines a curve that traces the optimal tradeoff between the leastsquares fit and the onenorm of the solution. We prove that this curve is convex and continuously differentiable over all points of interest, and show that it gives an explicit relationship to two other optimization problems closely related to BPDN. We describe a rootfinding algorithm for finding arbitrary points on this curve; the algorithm is suitable for problems that are large scale and for those that are in the complex domain. At each iteration, a spectral gradientprojection method approximately minimizes a leastsquares problem with an explicit onenorm constraint. Only matrixvector operations are required. The primaldual solution of this problem gives function and derivative information needed for the rootfinding method. Numerical experiments on a comprehensive set of test problems demonstrate that the method scales well to large problems.
Choosing the Forcing Terms in an Inexact Newton Method
 SIAM J. Sci. Comput
, 1994
"... An inexact Newton method is a generalization of Newton's method for solving F(x) = 0, F:/ /, in which, at the kth iteration, the step sk from the current approximate solution xk is required to satisfy a condition ]lF(x) + F'(x)s]l _< /]lF(xk)]l for a "forcing term" / [0,1). In typical applications, ..."
Abstract

Cited by 94 (2 self)
 Add to MetaCart
An inexact Newton method is a generalization of Newton's method for solving F(x) = 0, F:/ /, in which, at the kth iteration, the step sk from the current approximate solution xk is required to satisfy a condition ]lF(x) + F'(x)s]l _< /]lF(xk)]l for a "forcing term" / [0,1). In typical applications, the choice of the forcing terms is critical to the efficiency of the method and can affect robustness as well. Promising choices of the forcing terms arc given, their local convergence properties are analyzed, and their practical performance is shown on a representative set of test problems.
Theory of Algorithms for Unconstrained Optimization
, 1992
"... this article I will attempt to review the most recent advances in the theory of unconstrained optimization, and will also describe some important open questions. Before doing so, I should point out that the value of the theory of optimization is not limited to its capacity for explaining the behavio ..."
Abstract

Cited by 83 (1 self)
 Add to MetaCart
this article I will attempt to review the most recent advances in the theory of unconstrained optimization, and will also describe some important open questions. Before doing so, I should point out that the value of the theory of optimization is not limited to its capacity for explaining the behavior of the most widely used techniques. The question
Convergence analysis of pseudotransient continuation
 SIAM J. Num. Anal
, 1998
"... Abstract. Pseudotransient continuation (Ψtc) is a wellknown and physically motivated technique for computation of steady state solutions of timedependent partial differential equations. Standard globalization strategies such as line search or trust region methods often stagnate at local minima. Ψ ..."
Abstract

Cited by 61 (25 self)
 Add to MetaCart
Abstract. Pseudotransient continuation (Ψtc) is a wellknown and physically motivated technique for computation of steady state solutions of timedependent partial differential equations. Standard globalization strategies such as line search or trust region methods often stagnate at local minima. Ψtc succeeds in many of these cases by taking advantage of the underlying PDE structure of the problem. Though widely employed, the convergence of Ψtc is rarely discussed. In this paper we prove convergence for a generic form of Ψtc and illustrate it with two practical strategies.
Theory of inexact Krylov subspace methods and applications to scientific computing
, 2002
"... Abstract. We provide a general frameworkfor the understanding of inexact Krylov subspace methods for the solution of symmetric and nonsymmetric linear systems of equations, as well as for certain eigenvalue calculations. This frameworkallows us to explain the empirical results reported in a series o ..."
Abstract

Cited by 48 (6 self)
 Add to MetaCart
Abstract. We provide a general frameworkfor the understanding of inexact Krylov subspace methods for the solution of symmetric and nonsymmetric linear systems of equations, as well as for certain eigenvalue calculations. This frameworkallows us to explain the empirical results reported in a series of CERFACS technical reports by Bouras, Frayssé, and Giraud in 2000. Furthermore, assuming exact arithmetic, our analysis can be used to produce computable criteria to bound the inexactness of the matrixvector multiplication in such a way as to maintain the convergence of the Krylov subspace method. The theory developed is applied to several problems including the solution of Schur complement systems, linear systems which depend on a parameter, and eigenvalue problems. Numerical experiments for some of these scientific applications are reported.
Preconditioning indefinite systems in interior point methods for optimization
 Computational Optimization and Applications
, 2004
"... Abstract. Every Newton step in an interiorpoint method for optimization requires a solution of a symmetric indefinite system of linear equations. Most of today’s codes apply direct solution methods to perform this task. The use of logarithmic barriers in interior point methods causes unavoidable il ..."
Abstract

Cited by 44 (13 self)
 Add to MetaCart
Abstract. Every Newton step in an interiorpoint method for optimization requires a solution of a symmetric indefinite system of linear equations. Most of today’s codes apply direct solution methods to perform this task. The use of logarithmic barriers in interior point methods causes unavoidable illconditioning of linear systems and, hence, iterative methods fail to provide sufficient accuracy unless appropriately preconditioned. Two types of preconditioners which use some form of incomplete Cholesky factorization for indefinite systems are proposed in this paper. Although they involve significantly sparser factorizations than those used in direct approaches they still capture most of the numerical properties of the preconditioned system. The spectral analysis of the preconditioned matrix is performed: for convex optimization problems all the eigenvalues of this matrix are strictly positive. Numerical results are given for a set of public domain large linearly constrained convex quadratic programming problems with sizes reaching tens of thousands of variables. The analysis of these results reveals that the solution times for such problems on a modern PC are measured in minutes when direct methods are used and drop to seconds when iterative methods with appropriate preconditioners are used. Keywords: interiorpoint methods, iterative solvers, preconditioners 1.
The Adaptive Multilevel Finite Element Solution of the PoissonBoltzmann Equation on Massively Parallel Computers
 J. COMPUT. CHEM
, 2000
"... Using new methods for the parallel solution of elliptic partial differential equations, the teraflops computing power of massively parallel computers can be leveraged to perform electrostatic calculations on large biological systems. This paper describes the adaptive multilevel finite element soluti ..."
Abstract

Cited by 43 (14 self)
 Add to MetaCart
Using new methods for the parallel solution of elliptic partial differential equations, the teraflops computing power of massively parallel computers can be leveraged to perform electrostatic calculations on large biological systems. This paper describes the adaptive multilevel finite element solution of the PoissonBoltzmann equation for a microtubule on the NPACI IBM Blue Horizon supercomputer. The microtubule system is 40 nm in length and 24 nm in diameter, consists of roughly 600,000 atoms, and has a net charge of1800 e. PoissonBoltzmann calculations are performed for several processor configurations and the algorithm shows excellent parallel scaling.
Adaptive numerical treatment of elliptic systems on manifolds
 Advances in Computational Mathematics, 15(1):139
, 2001
"... ABSTRACT. Adaptive multilevel finite element methods are developed and analyzed for certain elliptic systems arising in geometric analysis and general relativity. This class of nonlinear elliptic systems of tensor equations on manifolds is first reviewed, and then adaptive multilevel finite element ..."
Abstract

Cited by 40 (24 self)
 Add to MetaCart
ABSTRACT. Adaptive multilevel finite element methods are developed and analyzed for certain elliptic systems arising in geometric analysis and general relativity. This class of nonlinear elliptic systems of tensor equations on manifolds is first reviewed, and then adaptive multilevel finite element methods for approximating solutions to this class of problems are considered in some detail. Two a posteriori error indicators are derived, based on local residuals and on global linearized adjoint or dual problems. The design of Manifold Code (MC) is then discussed; MC is an adaptive multilevel finite element software package for 2 and 3manifolds developed over several years at Caltech and UC San Diego. It employs a posteriori error estimation, adaptive simplex subdivision, unstructured algebraic multilevel methods, global inexact Newton methods, and numerical continuation methods for the numerical solution of nonlinear covariant elliptic systems on 2 and 3manifolds. Some of the more interesting features of MC are described in detail, including some new ideas for topology and geometry representation in simplex meshes, and an unusual partition of unitybased method for exploiting parallel computers. A short example is then given which involves the Hamiltonian and momentum constraints in the Einstein equations, a representative nonlinear 4component covariant elliptic system on a Riemannian 3manifold which arises in general relativity. A number of operator properties and solvability results recently established are first summarized, making possible two quasioptimal a priori error estimates for Galerkin approximations which are then derived. These two results complete the theoretical framework for effective use of adaptive multilevel finite element methods. A sample calculation using the MC software is then presented.
On the implementation of an algorithm for largescale equality constrained optimization
 SIAM Journal on Optimization
, 1998
"... Abstract. This paper describes a software implementation of Byrd and Omojokun’s trust region algorithm for solving nonlinear equality constrained optimization problems. The code is designed for the efficient solution of large problems and provides the user with a variety of linear algebra techniques ..."
Abstract

Cited by 38 (11 self)
 Add to MetaCart
Abstract. This paper describes a software implementation of Byrd and Omojokun’s trust region algorithm for solving nonlinear equality constrained optimization problems. The code is designed for the efficient solution of large problems and provides the user with a variety of linear algebra techniques for solving the subproblems occurring in the algorithm. Second derivative information can be used, but when it is not available, limited memory quasiNewton approximations are made. The performance of the code is studied using a set of difficult test problems from the CUTE collection.