Results 1  10
of
31,330
The selfduality equations on a Riemann surface
 Proc. Lond. Math. Soc., III. Ser
, 1987
"... In this paper we shall study a special class of solutions of the selfdual YangMills equations. The original selfduality equations which arose in mathematical physics were defined on Euclidean 4space. The physically relevant solutions were the ones with finite action—the socalled 'instanton ..."
Abstract

Cited by 524 (6 self)
 Add to MetaCart
In this paper we shall study a special class of solutions of the selfdual YangMills equations. The original selfduality equations which arose in mathematical physics were defined on Euclidean 4space. The physically relevant solutions were the ones with finite action—the socalled 'instantons'. The same equations may be
Global Optimization with Polynomials and the Problem of Moments
 SIAM Journal on Optimization
, 2001
"... We consider the problem of finding the unconstrained global minimum of a realvalued polynomial p(x) : R R, as well as the global minimum of p(x), in a compact set K defined by polynomial inequalities. It is shown that this problem reduces to solving an (often finite) sequence of convex linear mat ..."
Abstract

Cited by 569 (47 self)
 Add to MetaCart
We consider the problem of finding the unconstrained global minimum of a realvalued polynomial p(x) : R R, as well as the global minimum of p(x), in a compact set K defined by polynomial inequalities. It is shown that this problem reduces to solving an (often finite) sequence of convex linear matrix inequality (LMI) problems. A notion of KarushKuhnTucker polynomials is introduced in a global optimality condition. Some illustrative examples are provided. Key words. global optimization, theory of moments and positive polynomials, semidefinite programming AMS subject classifications. 90C22, 90C25 PII. S1052623400366802 1.
Graphical models, exponential families, and variational inference
, 2008
"... The formalism of probabilistic graphical models provides a unifying framework for capturing complex dependencies among random variables, and building largescale multivariate statistical models. Graphical models have become a focus of research in many statistical, computational and mathematical fiel ..."
Abstract

Cited by 800 (26 self)
 Add to MetaCart
The formalism of probabilistic graphical models provides a unifying framework for capturing complex dependencies among random variables, and building largescale multivariate statistical models. Graphical models have become a focus of research in many statistical, computational and mathematical fields, including bioinformatics, communication theory, statistical physics, combinatorial optimization, signal and image processing, information retrieval and statistical machine learning. Many problems that arise in specific instances — including the key problems of computing marginals and modes of probability distributions — are best studied in the general setting. Working with exponential family representations, and exploiting the conjugate duality between the cumulant function and the entropy for exponential families, we develop general variational representations of the problems of computing likelihoods, marginal probabilities and most probable configurations. We describe how a wide varietyof algorithms — among them sumproduct, cluster variational methods, expectationpropagation, mean field methods, maxproduct and linear programming relaxation, as well as conic programming relaxations — can all be understood in terms of exact or approximate forms of these variational representations. The variational approach provides a complementary alternative to Markov chain Monte Carlo as a general source of approximation methods for inference in largescale statistical models.
Learning the Kernel Matrix with SemiDefinite Programming
, 2002
"... Kernelbased learning algorithms work by embedding the data into a Euclidean space, and then searching for linear relations among the embedded data points. The embedding is performed implicitly, by specifying the inner products between each pair of points in the embedding space. This information ..."
Abstract

Cited by 780 (22 self)
 Add to MetaCart
Kernelbased learning algorithms work by embedding the data into a Euclidean space, and then searching for linear relations among the embedded data points. The embedding is performed implicitly, by specifying the inner products between each pair of points in the embedding space. This information is contained in the socalled kernel matrix, a symmetric and positive definite matrix that encodes the relative positions of all points. Specifying this matrix amounts to specifying the geometry of the embedding space and inducing a notion of similarity in the input spaceclassical model selection problems in machine learning. In this paper we show how the kernel matrix can be learned from data via semidefinite programming (SDP) techniques. When applied
Wireless Communications
, 2005
"... Copyright c ○ 2005 by Cambridge University Press. This material is in copyright. Subject to statutory exception and to the provisions of relevant collective licensing agreements, no reproduction of any part may take place without the written permission of Cambridge University ..."
Abstract

Cited by 1129 (32 self)
 Add to MetaCart
Copyright c ○ 2005 by Cambridge University Press. This material is in copyright. Subject to statutory exception and to the provisions of relevant collective licensing agreements, no reproduction of any part may take place without the written permission of Cambridge University
Planning Algorithms
, 2004
"... This book presents a unified treatment of many different kinds of planning algorithms. The subject lies at the crossroads between robotics, control theory, artificial intelligence, algorithms, and computer graphics. The particular subjects covered include motion planning, discrete planning, planning ..."
Abstract

Cited by 1108 (51 self)
 Add to MetaCart
This book presents a unified treatment of many different kinds of planning algorithms. The subject lies at the crossroads between robotics, control theory, artificial intelligence, algorithms, and computer graphics. The particular subjects covered include motion planning, discrete planning, planning under uncertainty, sensorbased planning, visibility, decisiontheoretic planning, game theory, information spaces, reinforcement learning, nonlinear systems, trajectory planning, nonholonomic planning, and kinodynamic planning.
Gravity coupled with matter and the foundation of non commutative geometry
, 1996
"... We first exhibit in the commutative case the simple algebraic relations between the algebra of functions on a manifold and its infinitesimal length element ds. Its unitary representations correspond to Riemannian metrics and Spin structure while ds is the Dirac propagator ds = ×— × = D −1 where D i ..."
Abstract

Cited by 354 (18 self)
 Add to MetaCart
work with Ali Chamseddine) to give the SM Lagrangian coupled to gravity. The internal fluctuations of the non commutative geometry are trivial in the commutative case but yield the full bosonic sector of SM with all correct quantum numbers in the slightly non commutative case. The group of local gauge
Liouville Correlation Functions from Fourdimensional Gauge Theories
 SIMONS CENTER FOR GEOMETRY AND PHYSICS, STONY BROOK UNIVERSITY
, 2009
"... We conjecture an expression for the Liouville theory conformal blocks and correlation functions on a Riemann surface of genus g and n punctures as the Nekrasov partition function of a certain class of N = 2 SCFTs recently defined by one of the authors. We conduct extensive tests of the conjecture ..."
Abstract

Cited by 394 (22 self)
 Add to MetaCart
We conjecture an expression for the Liouville theory conformal blocks and correlation functions on a Riemann surface of genus g and n punctures as the Nekrasov partition function of a certain class of N = 2 SCFTs recently defined by one of the authors. We conduct extensive tests of the conjecture at genus 0, 1.
Results 1  10
of
31,330