Results 11  20
of
175
SDPA (Semidefinite Programming Algorithm)  User's Manual
, 1995
"... Abstract. The SDPA (SemiDefinite Programming Algorithm) [5] is a software package for solving semidefinite programs (SDPs). It is based on a Mehrotratype predictorcorrector infeasible primaldual interiorpoint method. The SDPA handles the standard form SDP and its dual. It is implemented in C++ l ..."
Abstract

Cited by 95 (28 self)
 Add to MetaCart
Abstract. The SDPA (SemiDefinite Programming Algorithm) [5] is a software package for solving semidefinite programs (SDPs). It is based on a Mehrotratype predictorcorrector infeasible primaldual interiorpoint method. The SDPA handles the standard form SDP and its dual. It is implemented in C++ language utilizing the LAPACK [1] for matrix computations. The SDPA version 7.0.5 enjoys the following features: • Efficient method for computing the search directions when the SDP to be solved is large scale and sparse [4]. • Block diagonal matrix structure and sparse matrix structure are supported for data matrices. • Sparse or dense Cholesky factorization for the Schur matrix is automatically selected. • An initial point can be specified. • Some information on infeasibility of the SDP is provided. This manual and the SDPA can be downloaded from the WWW site
The Mathematics Of Eigenvalue Optimization
, 2003
"... Optimization problems involving the eigenvalues of symmetric and nonsymmetric matrices present a fascinating mathematical challenge. Such problems arise often in theory and practice, particularly in engineering design, and are amenable to a rich blend of classical mathematical techniques and contemp ..."
Abstract

Cited by 92 (13 self)
 Add to MetaCart
Optimization problems involving the eigenvalues of symmetric and nonsymmetric matrices present a fascinating mathematical challenge. Such problems arise often in theory and practice, particularly in engineering design, and are amenable to a rich blend of classical mathematical techniques and contemporary optimization theory. This essay presents a personal choice of some central mathematical ideas, outlined for the broad optimization community. I discuss the convex analysis of spectral functions and invariant matrix norms, touching briey on semide nite representability, and then outlining two broader algebraic viewpoints based on hyperbolic polynomials and Lie algebra. Analogous nonconvex notions lead into eigenvalue perturbation theory. The last third of the article concerns stability, for polynomials, matrices, and associated dynamical systems, ending with a section on robustness. The powerful and elegant language of nonsmooth analysis appears throughout, as a unifying narrative thread.
Semidefinite Programming Relaxations For The Quadratic Assignment Problem
, 1998
"... Semidefinite programming (SDP) relaxations for the quadratic assignment problem (QAP) are derived using the dual of the (homogenized) Lagrangian dual of appropriate equivalent representations of QAP. These relaxations result in the interesting, special, case where only the dual problem of the SDP re ..."
Abstract

Cited by 72 (25 self)
 Add to MetaCart
Semidefinite programming (SDP) relaxations for the quadratic assignment problem (QAP) are derived using the dual of the (homogenized) Lagrangian dual of appropriate equivalent representations of QAP. These relaxations result in the interesting, special, case where only the dual problem of the SDP relaxation has strict interior, i.e. the Slater constraint qualification always fails for the primal problem. Although there is no duality gap in theory, this indicates that the relaxation cannot be solved in a numerically stable way. By exploring the geometrical structure of the relaxation, we are able to find projected SDP relaxations. These new relaxations, and their duals, satisfy the Slater constraint qualification, and so can be solved numerically using primaldual interiorpoint methods. For one of our models, a preconditioned conjugate gradient method is used for solving the large linear systems which arise when finding the Newton direction. The preconditioner is found by exploiting th...
Solving Euclidean Distance Matrix Completion Problems Via Semidefinite Programming
, 1997
"... Given a partial symmetric matrix A with only certain elements specified, the Euclidean distance matrix completion problem (IgDMCP) is to find the unspecified elements of A that make A a Euclidean distance matrix (IgDM). In this paper, we follow the successful approach in [20] and solve the IgDMCP by ..."
Abstract

Cited by 69 (14 self)
 Add to MetaCart
Given a partial symmetric matrix A with only certain elements specified, the Euclidean distance matrix completion problem (IgDMCP) is to find the unspecified elements of A that make A a Euclidean distance matrix (IgDM). In this paper, we follow the successful approach in [20] and solve the IgDMCP by generalizing the completion problem to allow for approximate completions. In particular, we introduce a primaldual interiorpoint algorithm that solves an equivalent (quadratic objective function) semidefinite programming problem (SDP). Numerical results are included which illustrate the efficiency and robustness of our approach. Our randomly generated problems consistently resulted in low dimensional solutions when no completion existed.
Exploiting Sparsity in PrimalDual InteriorPoint Methods for Semidefinite Programming
 Mathematical Programming
, 1997
"... Abstract. The HelmbergRendlVanderbeiWolkowicz/KojimaShindohHara/Monteiro and the NesterovTodd search directions have been used in many primaldual interiorpoint methods for semidefinite programs. This paper proposes an efficient method for computing the two directions when a semidefinite prog ..."
Abstract

Cited by 68 (21 self)
 Add to MetaCart
Abstract. The HelmbergRendlVanderbeiWolkowicz/KojimaShindohHara/Monteiro and the NesterovTodd search directions have been used in many primaldual interiorpoint methods for semidefinite programs. This paper proposes an efficient method for computing the two directions when a semidefinite program to be solved is large scale and sparse.
Exploiting Sparsity in Semidefinite Programming via Matrix Completion I: General Framework
 SIAM JOURNAL ON OPTIMIZATION
, 1999
"... A critical disadvantage of primaldual interiorpoint methods against dual interiorpoint methods for large scale SDPs (semidefinite programs) has been that the primal positive semidefinite variable matrix becomes fully dense in general even when all data matrices are sparse. Based on some fundamenta ..."
Abstract

Cited by 62 (27 self)
 Add to MetaCart
A critical disadvantage of primaldual interiorpoint methods against dual interiorpoint methods for large scale SDPs (semidefinite programs) has been that the primal positive semidefinite variable matrix becomes fully dense in general even when all data matrices are sparse. Based on some fundamental results about positive semidefinite matrix completion, this article proposes a general method of exploiting the aggregate sparsity pattern over all data matrices to overcome this disadvantage. Our method is used in two ways. One is a conversion of a sparse SDP having a large scale positive semidefinite variable matrix into an SDP having multiple but smaller size positive semidefinite variable matrices to which we can effectively apply any interiorpoint method for SDPs employing a standard blockdiagonal matrix data structure. The other way is an incorporation of our method into primaldual interiorpoint methods which we can apply directly to a given SDP. In Part II of this article, we wi...
A semidefinite framework for trust region subproblems with applications to large scale minimization
 Math. Programming
, 1997
"... This is an abbreviated revision of the University of Waterloo research report CORR 9432. y ..."
Abstract

Cited by 59 (8 self)
 Add to MetaCart
This is an abbreviated revision of the University of Waterloo research report CORR 9432. y
On Extending Some PrimalDual InteriorPoint Algorithms From Linear Programming to Semidefinite Programming
 SIAM Journal on Optimization
, 1998
"... This work concerns primaldual interiorpoint methods for semidefinite programming (SDP) that use a search direction originally proposed by HelmbergRendlVanderbeiWolkowicz [5] and KojimaShindohHara [11], and recently rediscovered by Monteiro [15] in a more explicit form. In analyzing these meth ..."
Abstract

Cited by 55 (1 self)
 Add to MetaCart
This work concerns primaldual interiorpoint methods for semidefinite programming (SDP) that use a search direction originally proposed by HelmbergRendlVanderbeiWolkowicz [5] and KojimaShindohHara [11], and recently rediscovered by Monteiro [15] in a more explicit form. In analyzing these methods, a number of basic equalities and inequalities were developed in [11] and also in [15] through different means and in different forms. In this paper, we give a concise derivation of the key equalities and inequalities for complexity analysis along the exact line used in linear programming (LP), producing basic relationships that have compact forms almost identical to their counterparts in LP. We also introduce a new formulation of the central path and variablemetric measures of centrality. These results provide convenient tools for deriving polynomiality results for primaldual algorithms extended from LP to SDP using the aforementioned and related search directions. We present examples...
Solving Quadratic (0,1)Problems by Semidefinite Programs and Cutting Planes
, 1996
"... We present computational experiments for solving quadratic (0, 1) problems. Our approach combines a semidefinite relaxation with a cutting plane technique, and is applied in a Branch and Bound setting. Our experiments indicate that this type of approach is very robust, and allows to solve many moder ..."
Abstract

Cited by 51 (7 self)
 Add to MetaCart
We present computational experiments for solving quadratic (0, 1) problems. Our approach combines a semidefinite relaxation with a cutting plane technique, and is applied in a Branch and Bound setting. Our experiments indicate that this type of approach is very robust, and allows to solve many moderately sized problems, having say, less than 100 binary variables, in a routine manner.