Results 1  10
of
29
Monotone Complexity
, 1990
"... We give a general complexity classification scheme for monotone computation, including monotone spacebounded and Turing machine models not previously considered. We propose monotone complexity classes including mAC i , mNC i , mLOGCFL, mBWBP , mL, mNL, mP , mBPP and mNP . We define a simple ..."
Abstract

Cited by 2350 (12 self)
 Add to MetaCart
We give a general complexity classification scheme for monotone computation, including monotone spacebounded and Turing machine models not previously considered. We propose monotone complexity classes including mAC i , mNC i , mLOGCFL, mBWBP , mL, mNL, mP , mBPP and mNP . We define a simple notion of monotone reducibility and exhibit complete problems. This provides a framework for stating existing results and asking new questions. We show that mNL (monotone nondeterministic logspace) is not closed under complementation, in contrast to Immerman's and Szelepcs 'enyi's nonmonotone result [Imm88, Sze87] that NL = coNL; this is a simple extension of the monotone circuit depth lower bound of Karchmer and Wigderson [KW90] for stconnectivity. We also consider mBWBP (monotone bounded width branching programs) and study the question of whether mBWBP is properly contained in mNC 1 , motivated by Barrington's result [Bar89] that BWBP = NC 1 . Although we cannot answer t...
The NPcompleteness column: an ongoing guide
 Journal of Algorithms
, 1985
"... This is the nineteenth edition of a (usually) quarterly column that covers new developments in the theory of NPcompleteness. The presentation is modeled on that used by M. R. Garey and myself in our book ‘‘Computers and Intractability: A Guide to the Theory of NPCompleteness,’ ’ W. H. Freeman & Co ..."
Abstract

Cited by 188 (0 self)
 Add to MetaCart
This is the nineteenth edition of a (usually) quarterly column that covers new developments in the theory of NPcompleteness. The presentation is modeled on that used by M. R. Garey and myself in our book ‘‘Computers and Intractability: A Guide to the Theory of NPCompleteness,’ ’ W. H. Freeman & Co., New York, 1979 (hereinafter referred to as ‘‘[G&J]’’; previous columns will be referred to by their dates). A background equivalent to that provided by [G&J] is assumed, and, when appropriate, crossreferences will be given to that book and the list of problems (NPcomplete and harder) presented there. Readers who have results they would like mentioned (NPhardness, PSPACEhardness, polynomialtimesolvability, etc.) or open problems they would like publicized, should
On Uniformity within NC¹
 JOURNAL OF COMPUTER AND SYSTEM SCIENCES
, 1990
"... In order to study circuit complexity classes within NC¹ in a uniform setting, we need a uniformity condition which is more restrictive than those in common use. Two such conditions, stricter than NC¹ uniformity [Ru81,Co85], have appeared in recent research: Immerman's families of circuits defined by ..."
Abstract

Cited by 127 (19 self)
 Add to MetaCart
In order to study circuit complexity classes within NC¹ in a uniform setting, we need a uniformity condition which is more restrictive than those in common use. Two such conditions, stricter than NC¹ uniformity [Ru81,Co85], have appeared in recent research: Immerman's families of circuits defined by firstorder formulas [Im87a,Im87b] and a uniformity corresponding to Buss' deterministic logtime reductions [Bu87]. We show that these two notions are equivalent, leading to a natural notion of uniformity for lowlevel circuit complexity classes. We show that recent results on the structure of NC¹ [Ba89] still hold true in this very uniform setting. Finally, we investigate a parallel notion of uniformity, still more restrictive, based on the regular languages. Here we give characterizations of subclasses of the regular languages based on their logical expressibility, extending recent work of Straubing, Th'erien, and Thomas [STT88]. A preliminary version of this work appeared as [BIS88].
On relating time and space to size and depth
 SIAM Journal on Computing
, 1977
"... Abstract. Turing machine space complexity is related to circuit depth complexity. The relationship complements the known connection between Turing machine time and circuit size, thus enabling us to expose the related nature of some important open problems concerning Turing machine and circuit comple ..."
Abstract

Cited by 97 (1 self)
 Add to MetaCart
Abstract. Turing machine space complexity is related to circuit depth complexity. The relationship complements the known connection between Turing machine time and circuit size, thus enabling us to expose the related nature of some important open problems concerning Turing machine and circuit complexity. We are also able to show some connection between Turing machine complexity and arithmetic complexity.
The Boolean formula value problem is in ALOGTIME
 in Proceedings of the 19th Annual ACM Symposium on Theory of Computing
, 1987
"... The Boolean formula value problem is in alternating log time and, more generally, parenthesis contextfree languages are in alternating log time. The evaluation of reverse Polish notation Boolean formulas is also in alternating log time. These results are optimal since the Boolean formula value ..."
Abstract

Cited by 66 (7 self)
 Add to MetaCart
The Boolean formula value problem is in alternating log time and, more generally, parenthesis contextfree languages are in alternating log time. The evaluation of reverse Polish notation Boolean formulas is also in alternating log time. These results are optimal since the Boolean formula value problem is complete for alternating log time under deterministic log time reductions. Consequently, it is also complete for alternating log time under AC reductions.
Lower Bounds for Deterministic and Nondeterministic Branching Programs
 in Proceedings of the FCT'91, Lecture Notes in Computer Science
, 1991
"... We survey lower bounds established for the complexity of computing explicitly given Boolean functions by switchingandrectifier networks, branching programs and switching networks. We first consider the unrestricted case and then proceed to various restricted models. Among these are monotone networ ..."
Abstract

Cited by 57 (4 self)
 Add to MetaCart
We survey lower bounds established for the complexity of computing explicitly given Boolean functions by switchingandrectifier networks, branching programs and switching networks. We first consider the unrestricted case and then proceed to various restricted models. Among these are monotone networks, boundedwidth devices , oblivious devices and readk times only devices. 1 Introduction The main goal of the Boolean complexity theory is to prove lower bounds on the complexity of computing "explicitly given" Boolean functions in interesting computational models. By "explicitly given" researchers usually mean "belonging to the class NP ". This is a very plausible interpretation since on the one hand this class contains the overwhelming majority of interesting Boolean functions and on the other hand it is small enough to prevent us from the necessity to take into account counting arguments. To illustrate the second point, let me remind the reader that already the class \Delta p 2 ,...
Lower Bounds to the Size of ConstantDepth Propositional Proofs
, 1994
"... 1 LK is a natural modification of Gentzen sequent calculus for propositional logic with connectives : and V ; W (both of unbounded arity). Then for every d 0 and n 2, there is a set T d n of depth d sequents of total size O(n 3+d ) which are refutable in LK by depth d + 1 proof of size exp ..."
Abstract

Cited by 53 (6 self)
 Add to MetaCart
1 LK is a natural modification of Gentzen sequent calculus for propositional logic with connectives : and V ; W (both of unbounded arity). Then for every d 0 and n 2, there is a set T d n of depth d sequents of total size O(n 3+d ) which are refutable in LK by depth d + 1 proof of size exp(O(log 2 n)) but such that every depth d refutation must have the size at least exp(n\Omega\Gamma21 ). The sets T d n express a weaker form of the pigeonhole principle. It is a fundamental problem of mathematical logic and complexity theory whether there exists a proof system for propositional logic in which every tautology has a short proof, where the length (equivalently the size) of a proof is measured essentially by the total number of symbols in it and short means polynomial in the length of the tautology. Equivalently one can ask whether for every theory T there is another theory S (both first order and reasonably axiomatized, e.g. by schemes) having the property that if a statement...
An Optimal Parallel Algorithm for Formula Evaluation
, 1992
"... A new approach to Buss’s NC¹ algorithm [Proc. 19thACM Symposium on Theory of Computing, Association for Computing Machinery, New York, 1987, pp. 123131] for evaluation of Boolean formulas is presented. This problem is shown to be complete for NC¹ over AC¬ reductions. This approach is then used to s ..."
Abstract

Cited by 43 (6 self)
 Add to MetaCart
A new approach to Buss’s NC¹ algorithm [Proc. 19thACM Symposium on Theory of Computing, Association for Computing Machinery, New York, 1987, pp. 123131] for evaluation of Boolean formulas is presented. This problem is shown to be complete for NC¹ over AC¬ reductions. This approach is then used to solve the more general problem of evaluating arithmetic formulas by using arithmetic circuits.
On Frege and Extended Frege Proof Systems
, 1993
"... We propose a framework for proving lower bounds to the size of EF  proofs (equivalently, to the number of proofsteps in Fproofs) in terms of boolean valuations . The concept is motivated by properties of propositional provability in models of bounded arithmetic and it is a finitisation of a parti ..."
Abstract

Cited by 21 (2 self)
 Add to MetaCart
We propose a framework for proving lower bounds to the size of EF  proofs (equivalently, to the number of proofsteps in Fproofs) in terms of boolean valuations . The concept is motivated by properties of propositional provability in models of bounded arithmetic and it is a finitisation of a particular forcing construction explained also in the paper. It reduces the question of proving a lower bound to the question of constructing a partial boolean algebra and a map of formulas into that algebra with particular properties. We show that in principle one can obtain via this method optimal lower bounds (up to a polynomial increase). Introduction A propositional proof system is any polynomial time function P whose range is exactly the set of tautologies TAUT, cf. [17]. For ø a tautology any string ß such that P (ß) = ø is called a P proof of ø . Any usual propositional calculus, be it resolution or extended resolution, a Hilbert style system based on finitely many axiom schemes and inf...
The quantum adversary method and classical formula size lower bounds
 Computational Complexity
"... We introduce two new complexity measures for Boolean functions, or more generally for functions of the form f: S → T, where S ⊆ Σ n for some alphabet Σ and T an arbitrary set. We call these measures sumPI and maxPI. The quantity sumPI has been emerging through a line of research on quantum query com ..."
Abstract

Cited by 21 (9 self)
 Add to MetaCart
We introduce two new complexity measures for Boolean functions, or more generally for functions of the form f: S → T, where S ⊆ Σ n for some alphabet Σ and T an arbitrary set. We call these measures sumPI and maxPI. The quantity sumPI has been emerging through a line of research on quantum query complexity lower bounds via the socalled quantum adversary method [Amb02, Amb03, BSS03, Zha04, LM04], culminating in [ ˇ SS04] with the realization that these many different formulations are in fact equivalent. Given that sumPI turns out to be such a robust invariant of a function, we begin to investigate this quantity in its own right and see that it also has applications to classical complexity theory. As a surprising application we show that sumPI 2 (f) is a lower bound on the formula size, and even, up to a constant multiplicative factor, the probabilistic formula size of f. We show that several formula size lower bounds in the literature, specifically Khrapchenko and its extensions [Khr71, Kou93], including a key lemma of [H˚as98], are in fact special cases of our method. The second quantity we introduce, maxPI(f), is always at least as large as sumPI(f), and is