Results 1  10
of
140
The quantitative structure of exponential time
 Complexity theory retrospective II
, 1997
"... ABSTRACT Recent results on the internal, measuretheoretic structure of the exponential time complexity classes E and EXP are surveyed. The measure structure of these classes is seen to interact in informative ways with biimmunity, complexity cores, polynomialtime reductions, completeness, circuit ..."
Abstract

Cited by 90 (13 self)
 Add to MetaCart
ABSTRACT Recent results on the internal, measuretheoretic structure of the exponential time complexity classes E and EXP are surveyed. The measure structure of these classes is seen to interact in informative ways with biimmunity, complexity cores, polynomialtime reductions, completeness, circuitsize complexity, Kolmogorov complexity, natural proofs, pseudorandom generators, the density of hard languages, randomized complexity, and lowness. Possible implications for the structure of NP are also discussed. 1
The complexity of relational query languages (extended abstract
 In Proceedings of the fourteenth annual ACM symposium on Theory of computing (STOC ’82
, 1982
"... Two complexity measures for query languages are proposed. Data complexity is the complexity of evaluating a query in the language as a function of the size of the database, and expression complexity is the complexity of ewduating a query in the language as a function of the size of the expression de ..."
Abstract

Cited by 67 (0 self)
 Add to MetaCart
Two complexity measures for query languages are proposed. Data complexity is the complexity of evaluating a query in the language as a function of the size of the database, and expression complexity is the complexity of ewduating a query in the language as a function of the size of the expression defining the query. We study the data and expression complexity of logical langnages relational calculus and its extensions by transitive closure, fixpoint and second order existential quantification and algebraic languages relational algebra and its extensions by bounded and unbounded looping. The pattern which will bc shown is that the expression complexity of the investigated languages is one exponential higher then their data complexity, and for both types of complexity we show completeness in some complexity class. Research supported by a Weizrnann Postdoctoral Fellowship,
Infinitary Logic and Inductive Definability over Finite Structures
 Information and Computation
, 1995
"... The extensions of firstorder logic with a least fixed point operator (FO + LFP) and with a partial fixed point operator (FO + PFP) are known to capture the complexity classes P and PSPACE respectively in the presence of an ordering relation over finite structures. Recently, Abiteboul and Vianu [Abi ..."
Abstract

Cited by 56 (6 self)
 Add to MetaCart
The extensions of firstorder logic with a least fixed point operator (FO + LFP) and with a partial fixed point operator (FO + PFP) are known to capture the complexity classes P and PSPACE respectively in the presence of an ordering relation over finite structures. Recently, Abiteboul and Vianu [Abiteboul and Vianu, 1991b] investigated the relationship of these two logics in the absence of an ordering, using a machine model of generic computation. In particular, they showed that the two languages have equivalent expressive power if and only if P = PSPACE. These languages can also be seen as fragments of an infinitary logic where each formula has a bounded number of variables, L ! 1! (see, for instance, [Kolaitis and Vardi, 1990]). We investigate this logic of finite structures and provide a normal form for it. We also present a treatment of the results in [Abiteboul and Vianu, 1991b] from this point of view. In particular, we show that we can write a formula of FO + LFP that defines ...
Two–way simulation of multitape Turing machines
, 1966
"... Abstract, It has long been known that increasing the number of tapes used by a Turing machine does not provide the ability to compute any new functions. On the other hand, the use of extra tapes does make it possible to speed up the computation of certain functions. It is known that a square factor ..."
Abstract

Cited by 54 (0 self)
 Add to MetaCart
Abstract, It has long been known that increasing the number of tapes used by a Turing machine does not provide the ability to compute any new functions. On the other hand, the use of extra tapes does make it possible to speed up the computation of certain functions. It is known that a square factor is sometimes required for a onetape machine to behave as a twotape machine and that a square factor is always sufficient. The purpose of this paper is to show that, if a given function requires computation time T for a ktape realization, then it requires at most computation time T log T for a twotape realization. The proof of this fact is constructive; given any ktape machine, it is shown how to design an equivalent twotape machine that operates within the stated time bounds. In ~ddition to being interesting in its own right, the tr~deoff relation between number of tapes and speed of computation can be used in a diagonalization argument to show that if T(n) and U(n) are two time functions such that T(n) log T(n) inf U(n) 0 then there exists a function that can be computed within the time bound U(n) but not within the time bound T(n). 1.
The Computational Power and Complexity of Constraint Handling Rules
 In Second Workshop on Constraint Handling Rules, at ICLP05
, 2005
"... Constraint Handling Rules (CHR) is a highlevel rulebased programming language which is increasingly used for general purposes. We introduce the CHR machine, a model of computation based on the operational semantics of CHR. Its computational power and time complexity properties are compared to thos ..."
Abstract

Cited by 50 (21 self)
 Add to MetaCart
Constraint Handling Rules (CHR) is a highlevel rulebased programming language which is increasingly used for general purposes. We introduce the CHR machine, a model of computation based on the operational semantics of CHR. Its computational power and time complexity properties are compared to those of the wellunderstood Turing machine and Random Access Memory machine. This allows us to prove the interesting result that every algorithm can be implemented in CHR with the best known time and space complexity. We also investigate the practical relevance of this result and the constant factors involved. Finally we expand the scope of the discussion to other (declarative) programming languages.
The History and Status of the P versus NP Question
, 1992
"... this article, I have attempted to organize and describe this literature, including an occasional opinion about the most fruitful directions, but no technical details. In the first half of this century, work on the power of formal systems led to the formalization of the notion of algorithm and the re ..."
Abstract

Cited by 50 (0 self)
 Add to MetaCart
this article, I have attempted to organize and describe this literature, including an occasional opinion about the most fruitful directions, but no technical details. In the first half of this century, work on the power of formal systems led to the formalization of the notion of algorithm and the realization that certain problems are algorithmically unsolvable. At around this time, forerunners of the programmable computing machine were beginning to appear. As mathematicians contemplated the practical capabilities and limitations of such devices, computational complexity theory emerged from the theory of algorithmic unsolvability. Early on, a particular type of computational task became evident, where one is seeking an object which lies
The Role of Relativization in Complexity Theory
 Bulletin of the European Association for Theoretical Computer Science
, 1994
"... Several recent nonrelativizing results in the area of interactive proofs have caused many people to review the importance of relativization. In this paper we take a look at how complexity theorists use and misuse oracle results. We pay special attention to the new interactive proof systems and progr ..."
Abstract

Cited by 40 (9 self)
 Add to MetaCart
Several recent nonrelativizing results in the area of interactive proofs have caused many people to review the importance of relativization. In this paper we take a look at how complexity theorists use and misuse oracle results. We pay special attention to the new interactive proof systems and program checking results and try to understand why they do not relativize. We give some new results that may help us to understand these questions better.
Feasible Computation through Model Theory
, 1993
"... The computational complexity of a problem is usually defined in terms of the resources required on some machine model of computation. An alternative view looks at the complexity of describing the problem (seen as a collection of relational structures) in a logic, measuring logical resources such as ..."
Abstract

Cited by 36 (7 self)
 Add to MetaCart
The computational complexity of a problem is usually defined in terms of the resources required on some machine model of computation. An alternative view looks at the complexity of describing the problem (seen as a collection of relational structures) in a logic, measuring logical resources such as the number of variables, quantifiers, operators, etc. A close correspondence has been observed between these two, with many natural logics corresponding exactly to independently defined complexity classes. For the complexity classes that are generally identified with feasible computation, such characterizations require the presence of a linear order on the domain of every structure, in which case the class PTIME is characterized by an extension of firstorder logic by means of an inductive operator. No logical characterization of feasible computation is known for unordered structures. We approach this question from two directions. On the one hand, we seek to accurately characterize the expre...
On the complexity of algebraic numbers I. Expansions in integer bases
, 2005
"... Let b ≥ 2 be an integer. We prove that the badic expansion of every irrational algebraic number cannot have low complexity. Furthermore, we establish that irrational morphic numbers are transcendental, for a wide class of morphisms. In particular, irrational automatic numbers are transcendental. O ..."
Abstract

Cited by 33 (21 self)
 Add to MetaCart
Let b ≥ 2 be an integer. We prove that the badic expansion of every irrational algebraic number cannot have low complexity. Furthermore, we establish that irrational morphic numbers are transcendental, for a wide class of morphisms. In particular, irrational automatic numbers are transcendental. Our main tool is a new, combinatorial transcendence criterion.
Averagecase computational complexity theory
 Complexity Theory Retrospective II
, 1997
"... ABSTRACT Being NPcomplete has been widely interpreted as being computationally intractable. But NPcompleteness is a worstcase concept. Some NPcomplete problems are \easy on average", but some may not be. How is one to know whether an NPcomplete problem is \di cult on average"? The the ..."
Abstract

Cited by 31 (2 self)
 Add to MetaCart
ABSTRACT Being NPcomplete has been widely interpreted as being computationally intractable. But NPcompleteness is a worstcase concept. Some NPcomplete problems are \easy on average", but some may not be. How is one to know whether an NPcomplete problem is \di cult on average"? The theory of averagecase computational complexity, initiated by Levin about ten years ago, is devoted to studying this problem. This paper is an attempt to provide an overview of the main ideas and results in this important new subarea of complexity theory. 1