Results 1  10
of
278,007
CONSTRAINED MATRIX SYLVESTER EQUATIONS*
"... gratitude for his tradition of fruitful research in linear algebra, inspired by applications. Abstract. The problem of finding matrices L and T satisfying TA FT LC and TB 0 is considered. Existence conditions for the solution are established and an algorithm for computing the solution is derived. C ..."
Abstract
 Add to MetaCart
. Conditions under which the matrix [CT, TT] is full rank are also discussed. The problem arises in control theory in the design of reducedorder observers that achieve loop transfer recovery. Key words. Sylvester operator, matrix Liapunov equation, loop transfer recovery AMS(MOS) subject classifications. 93B
SNOPT: An SQP Algorithm For LargeScale Constrained Optimization
, 2002
"... Sequential quadratic programming (SQP) methods have proved highly effective for solving constrained optimization problems with smooth nonlinear functions in the objective and constraints. Here we consider problems with general inequality constraints (linear and nonlinear). We assume that first deriv ..."
Abstract

Cited by 582 (23 self)
 Add to MetaCart
Sequential quadratic programming (SQP) methods have proved highly effective for solving constrained optimization problems with smooth nonlinear functions in the objective and constraints. Here we consider problems with general inequality constraints (linear and nonlinear). We assume that first
Constrained model predictive control: Stability and optimality
 AUTOMATICA
, 2000
"... Model predictive control is a form of control in which the current control action is obtained by solving, at each sampling instant, a finite horizon openloop optimal control problem, using the current state of the plant as the initial state; the optimization yields an optimal control sequence and t ..."
Abstract

Cited by 696 (15 self)
 Add to MetaCart
important because efficiency demands operating points on or close to the boundary of the set of admissible states and controls. In this review, we focus on model predictive control of constrained systems, both linear and nonlinear and discuss only briefly model predictive control of unconstrained nonlinear
A Limited Memory Algorithm for Bound Constrained Optimization
 SIAM Journal on Scientific Computing
, 1994
"... An algorithm for solving large nonlinear optimization problems with simple bounds is described. ..."
Abstract

Cited by 557 (9 self)
 Add to MetaCart
An algorithm for solving large nonlinear optimization problems with simple bounds is described.
Numerical integration of the Cartesian equations of motion of a system with constraints: molecular dynamics of nalkanes
 J. Comput. Phys
, 1977
"... A numerical algorithm integrating the 3N Cartesian equations of motion of a system of N points subject to holonomic constraints is formulated. The relations of constraint remain perfectly fulfilled at each step of the trajectory despite the approximate character of numerical integration. The method ..."
Abstract

Cited by 682 (6 self)
 Add to MetaCart
A numerical algorithm integrating the 3N Cartesian equations of motion of a system of N points subject to holonomic constraints is formulated. The relations of constraint remain perfectly fulfilled at each step of the trajectory despite the approximate character of numerical integration. The method
Stochastic Perturbation Theory
, 1988
"... . In this paper classical matrix perturbation theory is approached from a probabilistic point of view. The perturbed quantity is approximated by a firstorder perturbation expansion, in which the perturbation is assumed to be random. This permits the computation of statistics estimating the variatio ..."
Abstract

Cited by 886 (35 self)
 Add to MetaCart
. In this paper classical matrix perturbation theory is approached from a probabilistic point of view. The perturbed quantity is approximated by a firstorder perturbation expansion, in which the perturbation is assumed to be random. This permits the computation of statistics estimating
The information bottleneck method
 University of Illinois
, 1999
"... We define the relevant information in a signal x ∈ X as being the information that this signal provides about another signal y ∈ Y. Examples include the information that face images provide about the names of the people portrayed, or the information that speech sounds provide about the words spoken. ..."
Abstract

Cited by 545 (38 self)
 Add to MetaCart
about Y through a ‘bottleneck ’ formed by a limited set of codewords ˜X. This constrained optimization problem can be seen as a generalization of rate distortion theory in which the distortion measure d(x, ˜x) emerges from the joint statistics of X and Y. This approach yields an exact set of self
Maximum Likelihood Linear Transformations for HMMBased Speech Recognition
 Computer Speech and Language
, 1998
"... This paper examines the application of linear transformations for speaker and environmental adaptation in an HMMbased speech recognition system. In particular, transformations that are trained in a maximum likelihood sense on adaptation data are investigated. Other than in the form of a simple bias ..."
Abstract

Cited by 538 (65 self)
 Add to MetaCart
) constrained, which requires the variance transform to have the same form as the mean transform (sometimes referred to as featurespace transforms). Reestimation formulae for all appropriate cases of transform are given. This includes a new and efficient "full" variance transform and the extension
A Fast and Elitist MultiObjective Genetic Algorithm: NSGAII
, 2000
"... Multiobjective evolutionary algorithms which use nondominated sorting and sharing have been mainly criticized for their (i) O(MN computational complexity (where M is the number of objectives and N is the population size), (ii) nonelitism approach, and (iii) the need for specifying a sharing param ..."
Abstract

Cited by 1707 (58 self)
 Add to MetaCart
to solve constrained multiobjective problems eciently. Simulation results of the constrained NSGAII on a number of test problems, including a fiveobjective, sevenconstraint nonlinear problem, are compared with another constrained multiobjective optimizer and much better performance of NSGA
Training Support Vector Machines: an Application to Face Detection
, 1997
"... We investigate the application of Support Vector Machines (SVMs) in computer vision. SVM is a learning technique developed by V. Vapnik and his team (AT&T Bell Labs.) that can be seen as a new method for training polynomial, neural network, or Radial Basis Functions classifiers. The decision sur ..."
Abstract

Cited by 728 (1 self)
 Add to MetaCart
surfaces are found by solving a linearly constrained quadratic programming problem. This optimization problem is challenging because the quadratic form is completely dense and the memory requirements grow with the square of the number of data points. We present a decomposition algorithm that guarantees
Results 1  10
of
278,007