Results 1  10
of
70
Knowledge compilation and theory approximation
 Journal of the ACM
, 1996
"... Computational efficiency is a central concern in the design of knowledge representation systems. In order to obtain efficient systems, it has been suggested that one should limit the form of the statements in the knowledge base or use an incomplete inference mechanism. The former approach is often t ..."
Abstract

Cited by 184 (5 self)
 Add to MetaCart
Computational efficiency is a central concern in the design of knowledge representation systems. In order to obtain efficient systems, it has been suggested that one should limit the form of the statements in the knowledge base or use an incomplete inference mechanism. The former approach is often too restrictive for practical applications, whereas the latter leads to uncertainty about exactly what can and cannot be inferred from the knowledge base. We present a third alternative, in which knowledge given in a general representation language is translated (compiled) into a tractable form — allowing for efficient subsequent query answering. We show how propositional logical theories can be compiled into Horn theories that approximate the original information. The approximations bound the original theory from below and above in terms of logical strength. The procedures are extended to other tractable languages (for example, binary clauses) and to the firstorder case. Finally, we demonstrate the generality of our approach by compiling concept descriptions in a general framebased language into a tractable form.
On Hiding Information from an Oracle
, 1989
"... : We consider the problem of computing with encrypted data. Player A wishes to know the value f(x) for some x but lacks the power to compute it. Player B has the power to compute f and is willing to send f(y) to A if she sends him y, for any y. Informally, an encryption scheme for the problem f is a ..."
Abstract

Cited by 144 (15 self)
 Add to MetaCart
: We consider the problem of computing with encrypted data. Player A wishes to know the value f(x) for some x but lacks the power to compute it. Player B has the power to compute f and is willing to send f(y) to A if she sends him y, for any y. Informally, an encryption scheme for the problem f is a method by which A, using her inferior resources, can transform the cleartext instance x into an encrypted instance y, obtain f(y) from B, and infer f(x) from f(y) in such a way that B cannot infer x from y. When such an encryption scheme exists, we say that f is encryptable. The framework defined in this paper enables us to prove precise statements about what an encrypted instance hides and what it leaks, in an informationtheoretic sense. Our definitions are cast in the language of probability theory and do not involve assumptions such as the intractability of factoring or the existence of oneway functions. We use our framework to describe encryption schemes for some wellknown function...
Analog Computation via Neural Networks
 THEORETICAL COMPUTER SCIENCE
, 1994
"... We pursue a particular approach to analog computation, based on dynamical systems of the type used in neural networks research. Our systems have a fixed structure, invariant in time, corresponding to an unchanging number of "neurons". If allowed exponential time for computation, they turn ..."
Abstract

Cited by 96 (9 self)
 Add to MetaCart
We pursue a particular approach to analog computation, based on dynamical systems of the type used in neural networks research. Our systems have a fixed structure, invariant in time, corresponding to an unchanging number of "neurons". If allowed exponential time for computation, they turn out to have unbounded power. However, under polynomialtime constraints there are limits on their capabilities, though being more powerful than Turing Machines. (A similar but more restricted model was shown to be polynomialtime equivalent to classical digital computation in the previous work [20].) Moreover, there is a precise correspondence between nets and standard nonuniform circuits with equivalent resources, and as a consequence one has lower bound constraints on what they can compute. This relationship is perhaps surprising since our analog devices do not change in any manner with input size. We note that these networks are not likely to solve polynomially NPhard problems, as the equality ...
On the compilability and expressive power of propositional planning formalisms
, 1998
"... The recent approaches of extending the GRAPHPLAN algorithm to handle more expressive planning formalisms raise the question of what the formal meaning of “expressive power ” is. We formalize the intuition that expressive power is a measure of how concisely planning domains and plans can be expressed ..."
Abstract

Cited by 87 (10 self)
 Add to MetaCart
(Show Context)
The recent approaches of extending the GRAPHPLAN algorithm to handle more expressive planning formalisms raise the question of what the formal meaning of “expressive power ” is. We formalize the intuition that expressive power is a measure of how concisely planning domains and plans can be expressed in a particular formalism by introducing the notion of “compilation schemes ” between planning formalisms. Using this notion, we analyze the expressiveness of a large family of propositional planning formalisms, ranging from basic STRIPS to a formalism with conditional effects, partial state specifications, and propositional formulae in the preconditions. One of the results is that conditional effects cannot be compiled away if plan size should grow only linearly but can be compiled away if we allow for polynomial growth of the resulting plans. This result confirms that the recently proposed extensions to the GRAPHPLAN algorithm concerning conditional effects are optimal with respect to the “compilability ” framework. Another result is that general propositional formulae cannot be compiled into conditional effects if the plan size should be preserved linearly. This implies that allowing general propositional formulae in preconditions and effect conditions adds another level of difficulty in generating a plan.
Forming Concepts for Fast Inference
 In Proceedings of the Tenth National Conference on Artificial Intelligence (AAAI92
, 1992
"... Knowledge compilation speeds inference by creating tractable approximations of a knowledge base, but this advantage is lost if the approximations are too large. We show how learning concept generalizations can allow for a more compact representation of the tractable theory. We also give a general in ..."
Abstract

Cited by 50 (2 self)
 Add to MetaCart
Knowledge compilation speeds inference by creating tractable approximations of a knowledge base, but this advantage is lost if the approximations are too large. We show how learning concept generalizations can allow for a more compact representation of the tractable theory. We also give a general induction rule for generating such concept generalizations. Finally, we prove that unless NP ` nonuniform P, not all theories have small Horn least upperbound approximations. 1 Introduction Work in machine learning has traditionally been divided into two main camps: concept learning (e.g. [ Kearns, 1990 ] ) and speedup learning (e.g. [ Minton, 1988 ] ). The work reported in this paper bridges these two areas by showing how concept learning can be used to speed up inference by allowing a more compact and efficient representation of a knowledge base. We have been studying techniques for boosting the performance of knowledge representation systems by compiling expressive but intractable repre...
Infeasibility of instance compression and succinct PCPs for NP
 Electronic Colloquium on Computational Complexity (ECCC
"... The ORSAT problem asks, given Boolean formulae φ1,..., φm each of size at most n, whether at least one of the φi’s is satisfiable. We show that there is no reduction from ORSAT to any set A where the length of the output is bounded by a polynomial in n, unless NP ⊆ coNP/poly, and the PolynomialTi ..."
Abstract

Cited by 49 (1 self)
 Add to MetaCart
(Show Context)
The ORSAT problem asks, given Boolean formulae φ1,..., φm each of size at most n, whether at least one of the φi’s is satisfiable. We show that there is no reduction from ORSAT to any set A where the length of the output is bounded by a polynomial in n, unless NP ⊆ coNP/poly, and the PolynomialTime Hierarchy collapses. This result settles an open problem proposed by Bodlaender et. al. [4] and Harnik and Naor [15] and has a number of implications. • A number of parametric NP problems, including Satisfiability, Clique, Dominating Set and Integer Programming, are not instance compressible or polynomially kernelizable unless NP ⊆ coNP/poly. • Satisfiability does not have PCPs of size polynomial in the number of variables unless NP ⊆ coNP/poly. • An approach of Harnik and Naor to constructing collisionresistant hash functions from oneway functions is unlikely to be viable in its present form. • (BuhrmanHitchcock) There are no subexponentialsize hard sets for NP unless NP is in coNP/poly. We also study probabilistic variants of compression, and show various results about and connections between these variants. To this end, we introduce a new strong derandomization hypothesis, the Oracle Derandomization Hypothesis, and discuss how it relates to traditional derandomization assumptions. Categories and Subject Descriptors
NPcomplete problems and physical reality
 ACM SIGACT News Complexity Theory Column, March. ECCC
, 2005
"... Can NPcomplete problems be solved efficiently in the physical universe? I survey proposals including soap bubbles, protein folding, quantum computing, quantum advice, quantum adiabatic algorithms, quantummechanical nonlinearities, hidden variables, relativistic time dilation, analog computing, Mal ..."
Abstract

Cited by 47 (5 self)
 Add to MetaCart
(Show Context)
Can NPcomplete problems be solved efficiently in the physical universe? I survey proposals including soap bubbles, protein folding, quantum computing, quantum advice, quantum adiabatic algorithms, quantummechanical nonlinearities, hidden variables, relativistic time dilation, analog computing, MalamentHogarth spacetimes, quantum gravity, closed timelike curves, and “anthropic computing. ” The section on soap bubbles even includes some “experimental ” results. While I do not believe that any of the proposals will let us solve NPcomplete problems efficiently, I argue that by studying them, we can learn something not only about computation but also about physics. 1
Fixpoint logic vs. infinitary logic in finitemodel theory
 In Proc. 7th IEEE Symp. on Logic in Computer Science
, 1992
"... In recent years several extensions of firstorder logic have been investigated in the context of finitemodel theory. Fixpoint logic and the infinitary logic Lω∞ω with a finite number of variables have turned out to be of particular importance. The study of fixpoint logic generated interactions with ..."
Abstract

Cited by 34 (1 self)
 Add to MetaCart
(Show Context)
In recent years several extensions of firstorder logic have been investigated in the context of finitemodel theory. Fixpoint logic and the infinitary logic Lω∞ω with a finite number of variables have turned out to be of particular importance. The study of fixpoint logic generated interactions with both database theory and complexity theory, while the infinitary logic Lω∞ω proved to be a useful tool for analyzing the expressive power of fixpoint logic. In addition to being a proper extension of fixpoint logic, Lω∞ω enjoys a gametheoretic characterization and possesses interesting structural properties, such as the 01 law. In this paper we pursue further the study of the relationship between Lω∞ω and fixpoint logic. We observe that equivalence of two finite structures with respect to Lω∞ω is expressible in fixpoint logic. As a first application of this, we obtain a normalform theorem for Lω∞ω on finite structures. We then focus on the relative expressive power of firstorder logic, fixpoint logic, and Lω∞ω on arbitrary classes of finite structures. Our second main result characterizes when Lω∞ω collapses to firstorder logic on an arbitrary class of finite structures. This resolves affirmatively a conjecture of G.L. McColm. 1
The Complexity of Planarity Testing
, 2000
"... We clarify the computational complexity of planarity testing, by showing that planarity testing is hard for L, and lies in SL. This nearly settles the question, since it is widely conjectured that L = SL [25]. The upper bound of SL matches the lower bound of L in the context of (nonuniform) circ ..."
Abstract

Cited by 29 (8 self)
 Add to MetaCart
We clarify the computational complexity of planarity testing, by showing that planarity testing is hard for L, and lies in SL. This nearly settles the question, since it is widely conjectured that L = SL [25]. The upper bound of SL matches the lower bound of L in the context of (nonuniform) circuit complexity, since L/poly is equal to SL/poly. Similarly, we show that a planar embedding, when one exists, can be found in FL SL . Previously, these problems were known to reside in the complexity class AC 1 , via a O(log n) time CRCW PRAM algorithm [22], although planarity checking for degreethree graphs had been shown to be in SL [23, 20].
TimeSpace Tradeoffs in the Counting Hierarchy
, 2001
"... We extend the lower bound techniques of [14], to the unboundederror probabilistic model. A key step in the argument is a generalization of Nepomnjasci's theorem from the Boolean setting to the arithmetic setting. This generalization is made possible, due to the recent discovery of logspaceuni ..."
Abstract

Cited by 18 (3 self)
 Add to MetaCart
We extend the lower bound techniques of [14], to the unboundederror probabilistic model. A key step in the argument is a generalization of Nepomnjasci's theorem from the Boolean setting to the arithmetic setting. This generalization is made possible, due to the recent discovery of logspaceuniform TC 0 circuits for iterated multiplication [9]. Here is an