Results 1  10
of
79
Global Optimization with Polynomials and the Problem of Moments
 SIAM Journal on Optimization
, 2001
"... We consider the problem of finding the unconstrained global minimum of a realvalued polynomial p(x) : R R, as well as the global minimum of p(x), in a compact set K defined by polynomial inequalities. It is shown that this problem reduces to solving an (often finite) sequence of convex linear mat ..."
Abstract

Cited by 309 (33 self)
 Add to MetaCart
We consider the problem of finding the unconstrained global minimum of a realvalued polynomial p(x) : R R, as well as the global minimum of p(x), in a compact set K defined by polynomial inequalities. It is shown that this problem reduces to solving an (often finite) sequence of convex linear matrix inequality (LMI) problems. A notion of KarushKuhnTucker polynomials is introduced in a global optimality condition. Some illustrative examples are provided. Key words. global optimization, theory of moments and positive polynomials, semidefinite programming AMS subject classifications. 90C22, 90C25 PII. S1052623400366802 1.
Semidefinite Programming Relaxations for Semialgebraic Problems
, 2001
"... A hierarchy of convex relaxations for semialgebraic problems is introduced. For questions reducible to a finite number of polynomial equalities and inequalities, it is shown how to construct a complete family of polynomially sized semidefinite programming conditions that prove infeasibility. The mai ..."
Abstract

Cited by 220 (19 self)
 Add to MetaCart
A hierarchy of convex relaxations for semialgebraic problems is introduced. For questions reducible to a finite number of polynomial equalities and inequalities, it is shown how to construct a complete family of polynomially sized semidefinite programming conditions that prove infeasibility. The main tools employed are a semidefinite programming formulation of the sum of squares decomposition for multivariate polynomials, and some results from real algebraic geometry. The techniques provide a constructive approach for finding bounded degree solutions to the Positivstellensatz, and are illustrated with examples from diverse application fields.
Structured Semidefinite Programs and Semialgebraic Geometry Methods in Robustness and Optimization
, 2000
"... ..."
GloptiPoly: Global Optimization over Polynomials with Matlab and SeDuMi
 ACM Trans. Math. Soft
, 2002
"... GloptiPoly is a Matlab/SeDuMi addon to build and solve convex linear matrix inequality relaxations of the (generally nonconvex) global optimization problem of minimizing a multivariable polynomial function subject to polynomial inequality, equality or integer constraints. It generates a series of ..."
Abstract

Cited by 96 (18 self)
 Add to MetaCart
GloptiPoly is a Matlab/SeDuMi addon to build and solve convex linear matrix inequality relaxations of the (generally nonconvex) global optimization problem of minimizing a multivariable polynomial function subject to polynomial inequality, equality or integer constraints. It generates a series of lower bounds monotonically converging to the global optimum. Global optimality is detected and isolated optimal solutions are extracted automatically. Numerical experiments show that for most of the small and mediumscale problems described in the literature, the global optimum is reached at low computational cost. 1
Lectures on modern convex optimization
 Society for Industrial and Applied Mathematics (SIAM
, 2001
"... Mathematical Programming deals with optimization programs of the form and includes the following general areas: minimize f(x) subject to gi(x) ≤ 0, i = 1,..., m, [x ⊂ R n] 1. Modelling: methodologies for posing various applied problems as optimization programs; 2. Optimization Theory, focusing on e ..."
Abstract

Cited by 92 (7 self)
 Add to MetaCart
Mathematical Programming deals with optimization programs of the form and includes the following general areas: minimize f(x) subject to gi(x) ≤ 0, i = 1,..., m, [x ⊂ R n] 1. Modelling: methodologies for posing various applied problems as optimization programs; 2. Optimization Theory, focusing on existence, uniqueness and on characterization of optimal solutions to optimization programs; 3. Optimization Methods: development and analysis of computational algorithms for various classes of optimization programs; 4. Implementation, testing and application of modelling methodologies and computational algorithms. Essentially, Mathematical Programming was born in 1948, when George Dantzig has invented Linear Programming – the class of optimization programs (P) with linear objective f(·) and
A comparison of the SheraliAdams, LovászSchrijver and Lasserre relaxations for 01 programming
 Mathematics of Operations Research
, 2001
"... ..."
Proving Program Invariance and Termination by Parametric Abstraction, Lagrangian Relaxation and Semidefinite Programming
 IN VMCAI’2005: VERIFICATION, MODEL CHECKING, AND ABSTRACT INTERPRETATION, VOLUME 3385 OF LNCS
, 2005
"... In order to verify semialgebraic programs, we automatize the Floyd/Naur/Hoare proof method. The main task is to automatically infer valid invariants and rank functions. First we express the program semantics in polynomial form. Then the unknown rank function and invariants are abstracted in parametr ..."
Abstract

Cited by 63 (1 self)
 Add to MetaCart
In order to verify semialgebraic programs, we automatize the Floyd/Naur/Hoare proof method. The main task is to automatically infer valid invariants and rank functions. First we express the program semantics in polynomial form. Then the unknown rank function and invariants are abstracted in parametric form. The implication in the Floyd/Naur/Hoare verification conditions is handled by abstraction into numerical constraints by Lagrangian relaxation. The remaining universal quantification is handled by semidefinite programming relaxation. Finally the parameters are computed using semidefinite programming solvers. This new approach exploits the recent progress in the numerical resolution of linear or bilinear matrix inequalities by semidefinite programming using efficient polynomial primal/dual interior point methods generalizing those wellknown in linear programming to convex optimization. The framework is applied to invariance and termination proof of sequential, nondeterministic, concurrent, and fair parallel imperative polynomial programs and can easily be extended to other safety and liveness properties.
On Cones of Nonnegative Quadratic Functions
, 2001
"... We derive LMIcharacterizations and dual decomposition algorithms for certain matrix cones which are generated by a given set using generalized copositivity. These matrix cones are in fact cones of nonconvex quadratic functions that are nonnegative on a certain domain. As a domain, we consider for ..."
Abstract

Cited by 38 (10 self)
 Add to MetaCart
We derive LMIcharacterizations and dual decomposition algorithms for certain matrix cones which are generated by a given set using generalized copositivity. These matrix cones are in fact cones of nonconvex quadratic functions that are nonnegative on a certain domain. As a domain, we consider for instance the intersection of a (upper) levelset of a quadratic function and a halfplane. We arrive at a generalization of Yakubovich's Sprocedure result. As an application we show that optimizing a general quadratic function over the intersection of an ellipsoid and a halfplane can be formulated as SDP, thus proving the polynomiality of this class of optimization problems, which arise, e.g., from the application of the trust region method for nonlinear programming. Other applications are in control theory and robust optimization. Keywords: LMI, SDP, CoPositive Cones, Quadratic Functions, SProcedure, Matrix Decomposition.
Introducing SOSTOOLS: A General Purpose Sum of Squares Programming Solver
 Proceedings of the IEEE Conference on Decision and Control (CDC), Las Vegas, NV
, 2002
"... SOSTOOLS is a MATLAB toolbox for constructing and solving sum of squares programs. It can be used in combination with semidefinite programming software, such as SeDuMi, to solve many continuous and combinatorial optimization problems, as well as various controlrelated problems. This paper provides ..."
Abstract

Cited by 37 (11 self)
 Add to MetaCart
SOSTOOLS is a MATLAB toolbox for constructing and solving sum of squares programs. It can be used in combination with semidefinite programming software, such as SeDuMi, to solve many continuous and combinatorial optimization problems, as well as various controlrelated problems. This paper provides an overview on sum of squares programming, describes the primary features of SOSTOOLS, and shows how SOSTOOLS is used to solve sum of squares programs. Some applications from different areas are presented to show the wide applicability of sum of squares programming in general and SOSTOOLS in particular. 1
Semidefinite Representations for Finite Varieties
 MATHEMATICAL PROGRAMMING
, 2002
"... We consider the problem of minimizing a polynomial over a semialgebraic set defined by polynomial equalities and inequalities. When the polynomial equalities have a finite number of complex solutions and define a radical ideal we can reformulate this problem as a semidefinite programming prob ..."
Abstract

Cited by 36 (6 self)
 Add to MetaCart
We consider the problem of minimizing a polynomial over a semialgebraic set defined by polynomial equalities and inequalities. When the polynomial equalities have a finite number of complex solutions and define a radical ideal we can reformulate this problem as a semidefinite programming problem. This semidefinite program involves combinatorial moment matrices, which are matrices indexed by a basis of the quotient vector space R[x 1 , . . . , x n ]/I. Our arguments are elementary and extend known facts for the grid case including 0/1 and polynomial programming. They also relate to known algebraic tools for solving polynomial systems of equations with finitely many complex solutions. Semidefinite approximations can be constructed by considering truncated combinatorial moment matrices; rank conditions are given (in a grid case) that ensure that the approximation solves the original problem at optimality.