Results 1  10
of
141
Monotone Complexity
, 1990
"... We give a general complexity classification scheme for monotone computation, including monotone spacebounded and Turing machine models not previously considered. We propose monotone complexity classes including mAC i , mNC i , mLOGCFL, mBWBP , mL, mNL, mP , mBPP and mNP . We define a simple ..."
Abstract

Cited by 2350 (12 self)
 Add to MetaCart
We give a general complexity classification scheme for monotone computation, including monotone spacebounded and Turing machine models not previously considered. We propose monotone complexity classes including mAC i , mNC i , mLOGCFL, mBWBP , mL, mNL, mP , mBPP and mNP . We define a simple notion of monotone reducibility and exhibit complete problems. This provides a framework for stating existing results and asking new questions. We show that mNL (monotone nondeterministic logspace) is not closed under complementation, in contrast to Immerman's and Szelepcs 'enyi's nonmonotone result [Imm88, Sze87] that NL = coNL; this is a simple extension of the monotone circuit depth lower bound of Karchmer and Wigderson [KW90] for stconnectivity. We also consider mBWBP (monotone bounded width branching programs) and study the question of whether mBWBP is properly contained in mNC 1 , motivated by Barrington's result [Bar89] that BWBP = NC 1 . Although we cannot answer t...
Expander Codes
 IEEE Transactions on Information Theory
, 1996
"... We present a new class of asymptotically good, linear errorcorrecting codes based upon expander graphs. These codes have linear time sequential decoding algorithms, logarithmic time parallel decoding algorithms with a linear number of processors, and are simple to understand. We present both random ..."
Abstract

Cited by 275 (10 self)
 Add to MetaCart
We present a new class of asymptotically good, linear errorcorrecting codes based upon expander graphs. These codes have linear time sequential decoding algorithms, logarithmic time parallel decoding algorithms with a linear number of processors, and are simple to understand. We present both randomized and explicit constructions for some of these codes. Experimental results demonstrate the extremely good performance of the randomly chosen codes. 1. Introduction We present a new class of error correcting codes derived from expander graphs. These codes have the advantage that they can be decoded very efficiently. That makes them particularly suitable for devices which must decode cheaply, such as compact disk players and remote satellite receivers. We hope that the connection we draw between expander graphs and error correcting codes will stimulate research in both fields. 1.1. Error correcting codes An error correcting code is a mapping from messages to codewords such that the mappi...
Knowledge compilation and theory approximation
 Journal of the ACM
, 1996
"... Computational efficiency is a central concern in the design of knowledge representation systems. In order to obtain efficient systems, it has been suggested that one should limit the form of the statements in the knowledge base or use an incomplete inference mechanism. The former approach is often t ..."
Abstract

Cited by 157 (5 self)
 Add to MetaCart
Computational efficiency is a central concern in the design of knowledge representation systems. In order to obtain efficient systems, it has been suggested that one should limit the form of the statements in the knowledge base or use an incomplete inference mechanism. The former approach is often too restrictive for practical applications, whereas the latter leads to uncertainty about exactly what can and cannot be inferred from the knowledge base. We present a third alternative, in which knowledge given in a general representation language is translated (compiled) into a tractable form — allowing for efficient subsequent query answering. We show how propositional logical theories can be compiled into Horn theories that approximate the original information. The approximations bound the original theory from below and above in terms of logical strength. The procedures are extended to other tractable languages (for example, binary clauses) and to the firstorder case. Finally, we demonstrate the generality of our approach by compiling concept descriptions in a general framebased language into a tractable form.
Which Problems Have Strongly Exponential Complexity?
 Journal of Computer and System Sciences
, 1998
"... For several NPcomplete problems, there have been a progression of better but still exponential algorithms. In this paper, we address the relative likelihood of subexponential algorithms for these problems. We introduce a generalized reduction which we call SubExponential Reduction Family (SERF) t ..."
Abstract

Cited by 128 (5 self)
 Add to MetaCart
For several NPcomplete problems, there have been a progression of better but still exponential algorithms. In this paper, we address the relative likelihood of subexponential algorithms for these problems. We introduce a generalized reduction which we call SubExponential Reduction Family (SERF) that preserves subexponential complexity. We show that CircuitSAT is SERFcomplete for all NPsearch problems, and that for any fixed k, kSAT, kColorability, kSet Cover, Independent Set, Clique, Vertex Cover, are SERFcomplete for the class SNP of search problems expressible by second order existential formulas whose first order part is universal. In particular, subexponential complexity for any one of the above problems implies the same for all others. We also look at the issue of proving strongly exponential lower bounds for AC 0 ; that is, bounds of the form 2 \Omega\Gamma n) . This problem is even open for depth3 circuits. In fact, such a bound for depth3 circuits with even l...
On span programs
 In Proc. of the 8th IEEE Structure in Complexity Theory
, 1993
"... We introduce a linear algebraic model of computation, the Span Program, and prove several upper and lower bounds on it. These results yield the following applications in complexity and cryptography: • SL ⊆ ⊕L (a weak Logspace analogue of N P ⊆ ⊕P). • The first superlinear size lower bounds on branc ..."
Abstract

Cited by 120 (6 self)
 Add to MetaCart
We introduce a linear algebraic model of computation, the Span Program, and prove several upper and lower bounds on it. These results yield the following applications in complexity and cryptography: • SL ⊆ ⊕L (a weak Logspace analogue of N P ⊆ ⊕P). • The first superlinear size lower bounds on branching programs that count. • A broader class of functions which posses informationtheoretic secret sharing schemes. The proof of the main connection, between span programs and counting branching programs, uses a variant of Razborov’s general approximation method. 1
The DLLite family and relations
 JOURNAL OF ARTIFICIAL INTELLIGENCE RESEARCH (JAIR)
, 2009
"... The recently introduced series of description logics under the common moniker ‘DLLite ’ has attracted attention of the description logic and semantic web communities due to the low computational complexity of inference, on the one hand, and the ability to represent conceptual modeling formalisms, o ..."
Abstract

Cited by 118 (51 self)
 Add to MetaCart
The recently introduced series of description logics under the common moniker ‘DLLite ’ has attracted attention of the description logic and semantic web communities due to the low computational complexity of inference, on the one hand, and the ability to represent conceptual modeling formalisms, on the other. The main aim of this article is to carry out a thorough and systematic investigation of inference in extensions of the original DLLite logics along five axes: by (i) adding the Boolean connectives and (ii) number restrictions to concept constructs, (iii) allowing role hierarchies, (iv) allowing role disjointness, symmetry, asymmetry, reflexivity, irreflexivity and transitivity constraints, and (v) adopting or dropping the unique name assumption. We analyze the combined complexity of satisfiability for the resulting logics, as well as the data complexity of instance checking and answering positive existential queries. Our approach is based on embedding DLLite logics in suitable fragments of the onevariable firstorder logic, which provides useful insights into their properties and, in particular, computational behavior.
Wrappers For Performance Enhancement And Oblivious Decision Graphs
, 1995
"... In this doctoral dissertation, we study three basic problems in machine learning and two new hypothesis spaces with corresponding learning algorithms. The problems we investigate are: accuracy estimation, feature subset selection, and parameter tuning. The latter two problems are related and are stu ..."
Abstract

Cited by 107 (8 self)
 Add to MetaCart
In this doctoral dissertation, we study three basic problems in machine learning and two new hypothesis spaces with corresponding learning algorithms. The problems we investigate are: accuracy estimation, feature subset selection, and parameter tuning. The latter two problems are related and are studied under the wrapper approach. The hypothesis spaces we investigate are: decision tables with a default majority rule (DTMs) and oblivious readonce decision graphs (OODGs).
Algorithmic SelfAssembly of DNA
, 1998
"... How can molecules compute? In his early studies of reversible computation, Bennett imagined an enzymatic Turing Machine which modified a heteropolymer (such as DNA) to perform computation with asymptotically low energy expenditures. Adleman's recent experimental demonstration of a DNA computation, ..."
Abstract

Cited by 104 (6 self)
 Add to MetaCart
How can molecules compute? In his early studies of reversible computation, Bennett imagined an enzymatic Turing Machine which modified a heteropolymer (such as DNA) to perform computation with asymptotically low energy expenditures. Adleman's recent experimental demonstration of a DNA computation, using an entirely different approach, has led to a wealth of ideas for how to build DNAbased computers in the laboratory, whose energy efficiency, information density, and parallelism may have potential to surpass conventional electronic computers for some purposes. In this thesis, I examine one mechanism used in all designs for DNAbased computer  the selfassembly of DNA by hybridization and formation of the double helix  and show that this mechanism alone in theory can perform universal computation. To do so, I borrow an important result in the mathematical theory of tiling: Wang showed how jigsawshaped tiles can be designed to simulate the operation of any Turing Machine. I propose...
Some Connections between Bounded Query Classes and NonUniform Complexity
 In Proceedings of the 5th Structure in Complexity Theory Conference
, 1990
"... This paper is dedicated to the memory of Ronald V. Book, 19371997. ..."
Abstract

Cited by 71 (23 self)
 Add to MetaCart
This paper is dedicated to the memory of Ronald V. Book, 19371997.
The Polynomial Method in Circuit Complexity
 In Proceedings of the 8th IEEE Structure in Complexity Theory Conference
, 1993
"... The representation of functions as lowdegree polynomials over various rings has provided many insights in the theory of smalldepth circuits. We survey some of the closure properties, upper bounds, and lower bounds obtained via this approach. 1. Introduction There is a long history of using polyno ..."
Abstract

Cited by 70 (4 self)
 Add to MetaCart
The representation of functions as lowdegree polynomials over various rings has provided many insights in the theory of smalldepth circuits. We survey some of the closure properties, upper bounds, and lower bounds obtained via this approach. 1. Introduction There is a long history of using polynomials in order to prove complexity bounds. Minsky and Papert [39] used polynomials to prove early lower bounds on the order of perceptrons. Razborov [46] and Smolensky [49] used them to prove lower bounds on the size of ANDOR circuits. Other lower bounds via polynomials are due to [50, 4, 10, 51, 9, 55]. Paturi and Saks [44] discovered that rational functions could be used for lower bounds on the size of threshold circuits. Toda [53] used polynomials to prove upper bounds on the power of the polynomial hierarchy. This led to a series of upper bounds on the power of the polynomial hierarchy [54, 52], AC 0 [2, 3, 52, 19], and ACC [58, 20, 30, 37], and related classes [21, 42]. Beigel and Gi...