Results 1  10
of
62
Relational Queries Computable in Polynomial Time
 Information and Control
, 1986
"... We characterize the polynomial time computable queries as those expressible in relational calculus plus a least fixed point operator and a total ordering on the universe. We also show that even without the ordering one application of fixed point suffices to express any query expressible with several ..."
Abstract

Cited by 269 (17 self)
 Add to MetaCart
We characterize the polynomial time computable queries as those expressible in relational calculus plus a least fixed point operator and a total ordering on the universe. We also show that even without the ordering one application of fixed point suffices to express any query expressible with several alternations of fixed point and negation. This proves that the fixed point query hierarchy suggested by Chandra and Harel collapses at the first fixed point level. It is also a general result showing that in finite model theory one application of fixed point suffices. Introduction and Summary Query languages for relational databases have received considerable attention. In 1972 Codd showed that two natural languages for queries  one algebraic and the other a version of first order predicate calculus  have identical powers of expressibility, [Cod72]. Query languages which are as expressive as Codd's Relational Calculus are sometimes called complete. This term is misleading however becau...
Almost Optimal Lower Bounds for Small Depth Circuits
 RANDOMNESS AND COMPUTATION
, 1989
"... We give improved lower bounds for the size of small depth circuits computing several functions. In particular we prove almost optimal lower bounds for the size of parity circuits. Further we show that there are functions computable in polynomial size and depth k but requires exponential size when ..."
Abstract

Cited by 237 (7 self)
 Add to MetaCart
We give improved lower bounds for the size of small depth circuits computing several functions. In particular we prove almost optimal lower bounds for the size of parity circuits. Further we show that there are functions computable in polynomial size and depth k but requires exponential size when the depth is restricted to k1. Our main lemma which is of independent interest states that by using a random restriction we can convert an AND of small ORs to an OR of small ANDs and conversely.
Languages That Capture Complexity Classes
 SIAM Journal of Computing
, 1987
"... this paper a series of languages adequate for expressing exactly those properties checkable in a series of computational complexity classes. For example, we show that a property of graphs (respectively groups, binary strings, etc.) is in polynomial time if and only if it is expressible in the first ..."
Abstract

Cited by 230 (21 self)
 Add to MetaCart
this paper a series of languages adequate for expressing exactly those properties checkable in a series of computational complexity classes. For example, we show that a property of graphs (respectively groups, binary strings, etc.) is in polynomial time if and only if it is expressible in the first order language of graphs (respectively groups, binary strings, etc.) together with a least fixed point operator. As another example, a property is in logspace if and only if it is expressible in first order logic together with a deterministic transitive closure operator. The roots of our approach to complexity theory go back to 1974 when Fagin showed that the NP properties are exactly those expressible in second order existential sentences. It follows that second order logic expresses exactly those properties which are in the polynomial time hierarchy. We show that adding suitable transitive closure operators to second order logic results in languages capturing polynomial space and exponential time, respectively. The existence of such natural languages for each important complexity class sheds a new light on complexity theory. These languages reaffirm the importance of the complexity classes as much more than machine dependent issues. Furthermore a whole new approach is suggested. Upper bounds (algorithms) can be produced by expressing the property of interest in one of our languages. Lower bounds may be demonstrated by showing that such expression is impossible.
The NPcompleteness column: an ongoing guide
 Journal of Algorithms
, 1985
"... This is the nineteenth edition of a (usually) quarterly column that covers new developments in the theory of NPcompleteness. The presentation is modeled on that used by M. R. Garey and myself in our book ‘‘Computers and Intractability: A Guide to the Theory of NPCompleteness,’ ’ W. H. Freeman & Co ..."
Abstract

Cited by 188 (0 self)
 Add to MetaCart
This is the nineteenth edition of a (usually) quarterly column that covers new developments in the theory of NPcompleteness. The presentation is modeled on that used by M. R. Garey and myself in our book ‘‘Computers and Intractability: A Guide to the Theory of NPCompleteness,’ ’ W. H. Freeman & Co., New York, 1979 (hereinafter referred to as ‘‘[G&J]’’; previous columns will be referred to by their dates). A background equivalent to that provided by [G&J] is assumed, and, when appropriate, crossreferences will be given to that book and the list of problems (NPcomplete and harder) presented there. Readers who have results they would like mentioned (NPhardness, PSPACEhardness, polynomialtimesolvability, etc.) or open problems they would like publicized, should
Logic and the Challenge of Computer Science
, 1988
"... Nowadays computer science is surpassing mathematics as the primary field of logic applications, but logic is not tuned properly to the new role. In particular, classical logic is preoccupied mostly with infinite static structures whereas many objects of interest in computer science are dynamic objec ..."
Abstract

Cited by 153 (16 self)
 Add to MetaCart
Nowadays computer science is surpassing mathematics as the primary field of logic applications, but logic is not tuned properly to the new role. In particular, classical logic is preoccupied mostly with infinite static structures whereas many objects of interest in computer science are dynamic objects with bounded resources. This chapter consists of two independent parts. The first part is devoted to finite model theory; it is mostly a survey of logics tailored for computational complexity. The second part is devoted to dynamic structures with bounded resources. In particular, we use dynamic structures with bounded resources to model Pascal.
On the power of smalldepth threshold circuits
 Proceedings 31st Annual IEEE Symposium on Foundations of Computer Science
, 1990
"... Abstract. Weinvestigate the power of threshold circuits of small depth. In particular, we give functions that require exponential size unweighted threshold circuits of depth 3 when we restrict the bottom fanin. We also prove that there are monotone functions fk that can be computed in depth k and li ..."
Abstract

Cited by 103 (2 self)
 Add to MetaCart
Abstract. Weinvestigate the power of threshold circuits of small depth. In particular, we give functions that require exponential size unweighted threshold circuits of depth 3 when we restrict the bottom fanin. We also prove that there are monotone functions fk that can be computed in depth k and linear size ^ � _circuits but require exponential size to compute by a depth k; 1 monotone weighted threshold circuit. Key words. Circuit complexity, monotone circuits, threshold circuits, lower bounds Subject classi cations. 68Q15, 68Q99 1.
Lower Bounds to the Size of ConstantDepth Propositional Proofs
, 1994
"... 1 LK is a natural modification of Gentzen sequent calculus for propositional logic with connectives : and V ; W (both of unbounded arity). Then for every d 0 and n 2, there is a set T d n of depth d sequents of total size O(n 3+d ) which are refutable in LK by depth d + 1 proof of size exp ..."
Abstract

Cited by 53 (6 self)
 Add to MetaCart
1 LK is a natural modification of Gentzen sequent calculus for propositional logic with connectives : and V ; W (both of unbounded arity). Then for every d 0 and n 2, there is a set T d n of depth d sequents of total size O(n 3+d ) which are refutable in LK by depth d + 1 proof of size exp(O(log 2 n)) but such that every depth d refutation must have the size at least exp(n\Omega\Gamma21 ). The sets T d n express a weaker form of the pigeonhole principle. It is a fundamental problem of mathematical logic and complexity theory whether there exists a proof system for propositional logic in which every tautology has a short proof, where the length (equivalently the size) of a proof is measured essentially by the total number of symbols in it and short means polynomial in the length of the tautology. Equivalently one can ask whether for every theory T there is another theory S (both first order and reasonably axiomatized, e.g. by schemes) having the property that if a statement...
The History and Status of the P versus NP Question
, 1992
"... this article, I have attempted to organize and describe this literature, including an occasional opinion about the most fruitful directions, but no technical details. In the first half of this century, work on the power of formal systems led to the formalization of the notion of algorithm and the re ..."
Abstract

Cited by 50 (0 self)
 Add to MetaCart
this article, I have attempted to organize and describe this literature, including an occasional opinion about the most fruitful directions, but no technical details. In the first half of this century, work on the power of formal systems led to the formalization of the notion of algorithm and the realization that certain problems are algorithmically unsolvable. At around this time, forerunners of the programmable computing machine were beginning to appear. As mathematicians contemplated the practical capabilities and limitations of such devices, computational complexity theory emerged from the theory of algorithmic unsolvability. Early on, a particular type of computational task became evident, where one is seeking an object which lies
Optimal bounds for decision problems on the CRCW PRAM
 In Proceedings of the 19th ACM Symposium on Theory of Computing (New
"... Abstract. Optimal Q(logn/log logn) lower bounds on the time for CRCW PRAMS with polynomially bounded numbers of processors or memory cells to compute parity and a number of related problems are proven. A strict time hierarchy of explicit Boolean functions of n bits on such machines that holds up to ..."
Abstract

Cited by 47 (3 self)
 Add to MetaCart
Abstract. Optimal Q(logn/log logn) lower bounds on the time for CRCW PRAMS with polynomially bounded numbers of processors or memory cells to compute parity and a number of related problems are proven. A strict time hierarchy of explicit Boolean functions of n bits on such machines that holds up to O(logn/loglogn) time is also exhibited. That is, for every time bound T within this range a function is exhibited that can be easily computed using polynomial resources in time T but requires more than polynomial resources to be computed in time T 1. Finally, it is shown that almost all Boolean functions of n bits require logn loglogn + fi ( 1) time when the number of processors is at most polynomial in n. The bounds do not place restrictions on the uniformity of the algorithms nor on the instruction sets of the machines.
A Switching Lemma for Small Restrictions and Lower Bounds for kDNF Resolution (Extended Abstract)
 SIAM J. Comput
, 2002
"... We prove a new switching lemma that works for restrictions that set only a small fraction of the variables and is applicable to DNFs with small conjunctions. We use this to prove lower bounds for the Res(k) propositional proof system, an extension of resolution which works with kDNFs instead of cla ..."
Abstract

Cited by 45 (7 self)
 Add to MetaCart
We prove a new switching lemma that works for restrictions that set only a small fraction of the variables and is applicable to DNFs with small conjunctions. We use this to prove lower bounds for the Res(k) propositional proof system, an extension of resolution which works with kDNFs instead of clauses. We also obtain an exponential separation between depth d circuits of k + 1.