Results 1  10
of
387,624
MATRIX PNORMS ARE NPHARD TO APPROXIMATE IF P ̸ = 1, 2, ∞.
, 908
"... Abstract. We show that for any rational p ∈ [1, ∞) except p = 1,2, unless P = NP, there is no polynomialtime algorithm for approximating the matrix pnorm to arbitrary relative precision. We also show that for any rational p ∈ [1, ∞) including p = 1, 2, unless P = NP, there is no polynomialtime al ..."
Abstract

Cited by 1 (0 self)
 Add to MetaCart
Abstract. We show that for any rational p ∈ [1, ∞) except p = 1,2, unless P = NP, there is no polynomialtime algorithm for approximating the matrix pnorm to arbitrary relative precision. We also show that for any rational p ∈ [1, ∞) including p = 1, 2, unless P = NP, there is no polynomial
Approximating Matrix pnorms
"... We consider the problem of computing the q ↦ → p norm of a matrix A, which is defined for p, q ≥ 1, as ‖Ax‖p ‖A‖q↦→p = max. x̸=⃗0 ‖x‖q This is in general a nonconvex optimization problem, and is a natural generalization of the wellstudied question of computing singular values (this corresponds to ..."
Abstract

Cited by 6 (0 self)
 Add to MetaCart
(such as nonnegativity), we prove that the problem is NPhard to approximate to any constant factor, for 2 < p ≤ q and p ≤ q < 2 (these are precisely the ranges of p, q with p ≤ q where constant factor approximations are not known). In this range, our techniques also show that if NP /∈ DTIME
A Threshold of ln n for Approximating Set Cover
 JOURNAL OF THE ACM
, 1998
"... Given a collection F of subsets of S = f1; : : : ; ng, set cover is the problem of selecting as few as possible subsets from F such that their union covers S, and max kcover is the problem of selecting k subsets from F such that their union has maximum cardinality. Both these problems are NPhar ..."
Abstract

Cited by 778 (5 self)
 Add to MetaCart
o(1)) ln n), and previous results of Lund and Yannakakis, that showed hardness of approximation within a ratio of (log 2 n)=2 ' 0:72 lnn. For max kcover we show an approximation threshold of (1 \Gamma 1=e) (up to low order terms), under the assumption that P != NP .
Determining the Number of Factors in Approximate Factor Models
, 2000
"... In this paper we develop some statistical theory for factor models of large dimensions. The focus is the determination of the number of factors, which is an unresolved issue in the rapidly growing literature on multifactor models. We propose a panel Cp criterion and show that the number of factors c ..."
Abstract

Cited by 538 (29 self)
 Add to MetaCart
In this paper we develop some statistical theory for factor models of large dimensions. The focus is the determination of the number of factors, which is an unresolved issue in the rapidly growing literature on multifactor models. We propose a panel Cp criterion and show that the number of factors can be consistently estimated using the criterion. The theory is developed under the framework of large crosssections (N) and large time dimensions (T). No restriction is imposed on the relation between N and T. Simulations show that the proposed criterion yields almost precise estimates of the number of factors for configurations of the panel data encountered in practice. The idea that variations in a large number of economic variables can be modelled bya small number of reference variables is appealing and is used in manyeconomic analysis. In the finance literature, the arbitrage pricing theory(APT) of Ross (1976) assumes that a small number of factors can be used to explain a large number of asset returns.
Improved Approximation Algorithms for Maximum Cut and Satisfiability Problems Using Semidefinite Programming
 Journal of the ACM
, 1995
"... We present randomized approximation algorithms for the maximum cut (MAX CUT) and maximum 2satisfiability (MAX 2SAT) problems that always deliver solutions of expected value at least .87856 times the optimal value. These algorithms use a simple and elegant technique that randomly rounds the solution ..."
Abstract

Cited by 1231 (13 self)
 Add to MetaCart
We present randomized approximation algorithms for the maximum cut (MAX CUT) and maximum 2satisfiability (MAX 2SAT) problems that always deliver solutions of expected value at least .87856 times the optimal value. These algorithms use a simple and elegant technique that randomly rounds
A Guided Tour to Approximate String Matching
 ACM COMPUTING SURVEYS
, 1999
"... We survey the current techniques to cope with the problem of string matching allowing errors. This is becoming a more and more relevant issue for many fast growing areas such as information retrieval and computational biology. We focus on online searching and mostly on edit distance, explaining t ..."
Abstract

Cited by 584 (38 self)
 Add to MetaCart
We survey the current techniques to cope with the problem of string matching allowing errors. This is becoming a more and more relevant issue for many fast growing areas such as information retrieval and computational biology. We focus on online searching and mostly on edit distance, explaining the problem and its relevance, its statistical behavior, its history and current developments, and the central ideas of the algorithms and their complexities. We present a number of experiments to compare the performance of the different algorithms and show which are the best choices according to each case. We conclude with some future work directions and open problems.
Property Testing and its connection to Learning and Approximation
"... We study the question of determining whether an unknown function has a particular property or is fflfar from any function with that property. A property testing algorithm is given a sample of the value of the function on instances drawn according to some distribution, and possibly may query the fun ..."
Abstract

Cited by 498 (68 self)
 Add to MetaCart
w.r.t the vertex set). Our graph property testing algorithms are probabilistic and make assertions which are correct with high probability, utilizing only poly(1=ffl) edgequeries into the graph, where ffl is the distance parameter. Moreover, the property testing algorithms can be used
Learning the Kernel Matrix with SemiDefinite Programming
, 2002
"... Kernelbased learning algorithms work by embedding the data into a Euclidean space, and then searching for linear relations among the embedded data points. The embedding is performed implicitly, by specifying the inner products between each pair of points in the embedding space. This information ..."
Abstract

Cited by 780 (22 self)
 Add to MetaCart
is contained in the socalled kernel matrix, a symmetric and positive definite matrix that encodes the relative positions of all points. Specifying this matrix amounts to specifying the geometry of the embedding space and inducing a notion of similarity in the input spaceclassical model selection
Global Optimization with Polynomials and the Problem of Moments
 SIAM Journal on Optimization
, 2001
"... We consider the problem of finding the unconstrained global minimum of a realvalued polynomial p(x) : R R, as well as the global minimum of p(x), in a compact set K defined by polynomial inequalities. It is shown that this problem reduces to solving an (often finite) sequence of convex linear mat ..."
Abstract

Cited by 569 (47 self)
 Add to MetaCart
matrix inequality (LMI) problems. A notion of KarushKuhnTucker polynomials is introduced in a global optimality condition. Some illustrative examples are provided. Key words. global optimization, theory of moments and positive polynomials, semidefinite programming AMS subject classifications. 90C22
Results 1  10
of
387,624