Results 1  10
of
3,756
Where the REALLY Hard Problems Are
 IN J. MYLOPOULOS AND R. REITER (EDS.), PROCEEDINGS OF 12TH INTERNATIONAL JOINT CONFERENCE ON AI (IJCAI91),VOLUME 1
, 1991
"... It is well known that for many NPcomplete problems, such as KSat, etc., typical cases are easy to solve; so that computationally hard cases must be rare (assuming P != NP). This paper shows that NPcomplete problems can be summarized by at least one "order parameter", and that the hard p ..."
Abstract

Cited by 681 (1 self)
 Add to MetaCart
It is well known that for many NPcomplete problems, such as KSat, etc., typical cases are easy to solve; so that computationally hard cases must be rare (assuming P != NP). This paper shows that NPcomplete problems can be summarized by at least one "order parameter", and that the hard problems occur at a critical value of such a parameter. This critical value separates two regions of characteristically different properties. For example, for Kcolorability, the critical value separates overconstrained from underconstrained random graphs, and it marks the value at which the probability of a solution changes abruptly from near 0 to near 1. It is the high density of wellseparated almost solutions (local minima) at this boundary that cause search algorithms to "thrash". This boundary is a type of phase transition and we show that it is preserved under mappings between problems. We show that for some P problems either there is no phase transition or it occurs for bounded N (and so bound...
Parameterized Complexity
, 1998
"... the rapidly developing systematic connections between FPT and useful heuristic algorithms  a new and exciting bridge between the theory of computing and computing in practice. The organizers of the seminar strongly believe that knowledge of parameterized complexity techniques and results belongs ..."
Abstract

Cited by 1218 (75 self)
 Add to MetaCart
the rapidly developing systematic connections between FPT and useful heuristic algorithms  a new and exciting bridge between the theory of computing and computing in practice. The organizers of the seminar strongly believe that knowledge of parameterized complexity techniques and results belongs into the toolkit of every algorithm designer. The purpose of the seminar was to bring together leading experts from all over the world, and from the diverse areas of computer science that have been attracted to this new framework. The seminar was intended as the rst larger international meeting with a specic focus on parameterized complexity, and it hopefully serves as a driving force in the development of the eld. 1 We had 49 participants from Australia, Canada, India, Israel, New Zealand, USA, and various European countries. During the workshop 25 lectures were given. Moreover, one night session was devoted to open problems and Thursday was basically used for problem discussion
Graphical models, exponential families, and variational inference
, 2008
"... The formalism of probabilistic graphical models provides a unifying framework for capturing complex dependencies among random variables, and building largescale multivariate statistical models. Graphical models have become a focus of research in many statistical, computational and mathematical fiel ..."
Abstract

Cited by 800 (26 self)
 Add to MetaCart
The formalism of probabilistic graphical models provides a unifying framework for capturing complex dependencies among random variables, and building largescale multivariate statistical models. Graphical models have become a focus of research in many statistical, computational and mathematical fields, including bioinformatics, communication theory, statistical physics, combinatorial optimization, signal and image processing, information retrieval and statistical machine learning. Many problems that arise in specific instances — including the key problems of computing marginals and modes of probability distributions — are best studied in the general setting. Working with exponential family representations, and exploiting the conjugate duality between the cumulant function and the entropy for exponential families, we develop general variational representations of the problems of computing likelihoods, marginal probabilities and most probable configurations. We describe how a wide varietyof algorithms — among them sumproduct, cluster variational methods, expectationpropagation, mean field methods, maxproduct and linear programming relaxation, as well as conic programming relaxations — can all be understood in terms of exact or approximate forms of these variational representations. The variational approach provides a complementary alternative to Markov chain Monte Carlo as a general source of approximation methods for inference in largescale statistical models.
Books in graphs
, 2008
"... A set of q triangles sharing a common edge is called a book of size q. We write β (n, m) for the the maximal q such that every graph G (n, m) contains a book of size q. In this note 1) we compute β ( n, cn 2) for infinitely many values of c with 1/4 < c < 1/3, 2) we show that if m ≥ (1/4 − α) ..."
Abstract

Cited by 2380 (22 self)
 Add to MetaCart
A set of q triangles sharing a common edge is called a book of size q. We write β (n, m) for the the maximal q such that every graph G (n, m) contains a book of size q. In this note 1) we compute β ( n, cn 2) for infinitely many values of c with 1/4 < c < 1/3, 2) we show that if m ≥ (1/4 − α) n 2 with 0 < α < 17 −3 (), and G has no book of size at least graph G1 of order at least
The Great Reversals: The Politics of Financial Development in the 20th Century
, 2001
"... Indicators of the development of the financial sector do not improve monotonically over time. In particular, we find that by most measures, countries were more financially developed in 1913 than in 1980 and only recently have they surpassed their 1913 levels. This pattern cannot be explained by stru ..."
Abstract

Cited by 527 (13 self)
 Add to MetaCart
Indicators of the development of the financial sector do not improve monotonically over time. In particular, we find that by most measures, countries were more financially developed in 1913 than in 1980 and only recently have they surpassed their 1913 levels. This pattern cannot be explained by structural theories that attribute crosscountry differences in financial development to timeinvariant factors, such as a country's legal origin or culture. We propose an "interest group" theory of financial development where incumbents oppose financial development because it breeds competition. The theory predicts that incumbents' opposition will be weaker when an economy allows both crossborder trade and capital flows. This theory can go some way in accounting for the crosscountry differences and the time series variation of financial development. When we recognize that different kinds of institutional heritages afford different scope for private interests to express themselves, we obtain a...
Statistical mechanics of complex networks
 Rev. Mod. Phys
"... Complex networks describe a wide range of systems in nature and society, much quoted examples including the cell, a network of chemicals linked by chemical reactions, or the Internet, a network of routers and computers connected by physical links. While traditionally these systems were modeled as ra ..."
Abstract

Cited by 2083 (10 self)
 Add to MetaCart
Complex networks describe a wide range of systems in nature and society, much quoted examples including the cell, a network of chemicals linked by chemical reactions, or the Internet, a network of routers and computers connected by physical links. While traditionally these systems were modeled as random graphs, it is increasingly recognized that the topology and evolution of real
The DLV System for Knowledge Representation and Reasoning
 ACM Transactions on Computational Logic
, 2002
"... Disjunctive Logic Programming (DLP) is an advanced formalism for knowledge representation and reasoning, which is very expressive in a precise mathematical sense: it allows to express every property of finite structures that is decidable in the complexity class ΣP 2 (NPNP). Thus, under widely believ ..."
Abstract

Cited by 455 (100 self)
 Add to MetaCart
Disjunctive Logic Programming (DLP) is an advanced formalism for knowledge representation and reasoning, which is very expressive in a precise mathematical sense: it allows to express every property of finite structures that is decidable in the complexity class ΣP 2 (NPNP). Thus, under widely believed assumptions, DLP is strictly more expressive than normal (disjunctionfree) logic programming, whose expressiveness is limited to properties decidable in NP. Importantly, apart from enlarging the class of applications which can be encoded in the language, disjunction often allows for representing problems of lower complexity in a simpler and more natural fashion. This paper presents the DLV system, which is widely considered the stateoftheart implementation of disjunctive logic programming, and addresses several aspects. As for problem solving, we provide a formal definition of its kernel language, functionfree disjunctive logic programs (also known as disjunctive datalog), extended by weak constraints, which are a powerful tool to express optimization problems. We then illustrate the usage of DLV as a tool for knowledge representation and reasoning, describing a new declarative programming methodology which allows one to encode complex problems (up to ∆P 3complete problems) in a declarative fashion. On the foundational side, we provide a detailed analysis of the computational complexity of the language of
Results 1  10
of
3,756