Results 1  10
of
148
An automatatheoretic approach to linear temporal logic
 Logics for Concurrency: Structure versus Automata, volume 1043 of Lecture Notes in Computer Science
, 1996
"... Abstract. The automatatheoretic approach to linear temporal logic uses the theory of automata as a unifying paradigm for program specification, verification, and synthesis. Both programs and specifications are in essence descriptions of computations. These computations can be viewed as words over s ..."
Abstract

Cited by 218 (22 self)
 Add to MetaCart
Abstract. The automatatheoretic approach to linear temporal logic uses the theory of automata as a unifying paradigm for program specification, verification, and synthesis. Both programs and specifications are in essence descriptions of computations. These computations can be viewed as words over some alphabet. Thus,programs and specificationscan be viewed as descriptions of languagesover some alphabet. The automatatheoretic perspective considers the relationships between programs and their specifications as relationships between languages.By translating programs and specifications to automata, questions about programs and their specifications can be reduced to questions about automata. More specifically, questions such as satisfiability of specifications and correctness of programs with respect to their specifications can be reduced to questions such as nonemptiness and containment of automata. Unlike classical automata theory, which focused on automata on finite words, the applications to program specification, verification, and synthesis, use automata on infinite words, since the computations in which we are interested are typically infinite. This paper provides an introduction to the theory of automata on infinite words and demonstrates its applications to program specification, verification, and synthesis. 1
The NPcompleteness column: an ongoing guide
 Journal of Algorithms
, 1985
"... This is the nineteenth edition of a (usually) quarterly column that covers new developments in the theory of NPcompleteness. The presentation is modeled on that used by M. R. Garey and myself in our book ‘‘Computers and Intractability: A Guide to the Theory of NPCompleteness,’ ’ W. H. Freeman & Co ..."
Abstract

Cited by 188 (0 self)
 Add to MetaCart
This is the nineteenth edition of a (usually) quarterly column that covers new developments in the theory of NPcompleteness. The presentation is modeled on that used by M. R. Garey and myself in our book ‘‘Computers and Intractability: A Guide to the Theory of NPCompleteness,’ ’ W. H. Freeman & Co., New York, 1979 (hereinafter referred to as ‘‘[G&J]’’; previous columns will be referred to by their dates). A background equivalent to that provided by [G&J] is assumed, and, when appropriate, crossreferences will be given to that book and the list of problems (NPcomplete and harder) presented there. Readers who have results they would like mentioned (NPhardness, PSPACEhardness, polynomialtimesolvability, etc.) or open problems they would like publicized, should
The Complexity of LogicBased Abduction
, 1993
"... Abduction is an important form of nonmonotonic reasoning allowing one to find explanations for certain symptoms or manifestations. When the application domain is described by a logical theory, we speak about logicbased abduction. Candidates for abductive explanations are usually subjected to minima ..."
Abstract

Cited by 162 (26 self)
 Add to MetaCart
Abduction is an important form of nonmonotonic reasoning allowing one to find explanations for certain symptoms or manifestations. When the application domain is described by a logical theory, we speak about logicbased abduction. Candidates for abductive explanations are usually subjected to minimality criteria such as subsetminimality, minimal cardinality, minimal weight, or minimality under prioritization of individual hypotheses. This paper presents a comprehensive complexity analysis of relevant decision and search problems related to abduction on propositional theories. Our results indicate that abduction is harder than deduction. In particular, we show that with the most basic forms of abduction the relevant decision problems are complete for complexity classes at the second level of the polynomial hierarchy, while the use of prioritization raises the complexity to the third level in certain cases.
Model Checking of Safety Properties
, 1999
"... Of special interest in formal verification are safety properties, which assert that the system always stays within some allowed region. Proof rules for the verification of safety properties have been developed in the proofbased approach to verification, making verification of safety properties simp ..."
Abstract

Cited by 102 (16 self)
 Add to MetaCart
Of special interest in formal verification are safety properties, which assert that the system always stays within some allowed region. Proof rules for the verification of safety properties have been developed in the proofbased approach to verification, making verification of safety properties simpler than verification of general properties. In this paper we consider model checking of safety properties. A computation that violates a general linear property reaches a bad cycle, which witnesses the violation of the property. Accordingly, current methods and tools for model checking of linear properties are based on a search for bad cycles. A symbolic implementation of such a search involves the calculation of a nested fixedpoint expression over the system's state space, and is often impossible. Every computation that violates a safety property has a finite prefix along which the property is violated. We use this fact in order to base model checking of safety properties on a search for ...
Improvements to the evaluation of quantified Boolean formulae
 In Proceedings of the Sixteenth International Joint Conference on Artificial Intelligence (IJCAI'99), July 31August 6
, 1999
"... We present a theoremprover for quantified Boolean formulae and evaluate it on random quantified formulae and formulae that represent problems from automated planning. Even though the notion of quantified Boolean formula is theoretically important, automated reasoning with QBF has not been thoroughl ..."
Abstract

Cited by 73 (3 self)
 Add to MetaCart
We present a theoremprover for quantified Boolean formulae and evaluate it on random quantified formulae and formulae that represent problems from automated planning. Even though the notion of quantified Boolean formula is theoretically important, automated reasoning with QBF has not been thoroughly investigated. Universal quantifiers are needed in representing many computational problems that cannot be easily translated to the propositional logic and solved by satisfiability algorithms. Therefore efficient reasoning with QBF is important. The DavisPutnam procedure can be extended to evaluate quantified Boolean formulae. A straightforward algorithm of this kind is not very efficient. We identify universal quantifiers as the main area where improvements to the basic algorithm can be made. We present a number of techniques for reducing the amount of search that is needed, and evaluate their effectiveness by running the algorithm on a collection of formulae obtained from planning and generated randomly. For the structured problems we consider, the techniques lead to a dramatic speedup. 1
Default Reasoning System DeReS
, 1996
"... In this paper, we describe an automated reasoning system, called DeReS. DeReS implements default logic of Reiter by supporting several basic reasoning tasks such as testing whether extensions exist, finding one or all extensions (if at least one exists) and querying if a formula belongs to one ..."
Abstract

Cited by 71 (5 self)
 Add to MetaCart
In this paper, we describe an automated reasoning system, called DeReS. DeReS implements default logic of Reiter by supporting several basic reasoning tasks such as testing whether extensions exist, finding one or all extensions (if at least one exists) and querying if a formula belongs to one or all extensions. If an input theory is a logic program, DeReS computes stable models of this program and supports queries on membership of an atom in some or all stable models. The paper contains an account of our preliminary experiments with DeReS and a discussion of the results. We show that a choice of a propositional prover is critical for the efficiency of DeReS. We also present a general technique that eliminates the need for some global consistency checks and results in substantial speedups. We experimentally demonstrate the potential of the concept of relaxed stratification for making automated reasoning systems practical. 1 INTRODUCTION The area of nonmonotonic l...
Counting Classes: Thresholds, Parity, Mods, and Fewness
, 1996
"... Counting classes consist of languages defined in terms of the number of accepting computations of nondeterministic polynomialtime Turing machines. Well known examples of counting classes are NP, coNP, \PhiP, and PP. Every counting class is a subset of P #P[1] , the class of languages computable ..."
Abstract

Cited by 61 (13 self)
 Add to MetaCart
Counting classes consist of languages defined in terms of the number of accepting computations of nondeterministic polynomialtime Turing machines. Well known examples of counting classes are NP, coNP, \PhiP, and PP. Every counting class is a subset of P #P[1] , the class of languages computable in polynomial time using a single call to an oracle capable of determining the number of accepting paths of an NP machine. Using closure properties of #P, we systematically develop a complexity theory for counting classes defined in terms of thresholds and moduli. An unexpected result is that MOD k iP = MOD k P for prime k. Finally, we improve a result of Cai and Hemachandra by showing that recognizing languages in the class Few is as easy as distinguishing uniquely satisfiable formulas from unsatisfiable formulas (or detecting unique solutions, as in [28]). 1. Introduction Valiant [27] defined the class #P of functions whose values equal the number of accepting paths of polynomialtime bo...
Exact analysis of Dodgson elections: Lewis Carroll’s 1876 voting system is complete for parallel access to NP
 Journal of the ACM
, 1997
"... Abstract. In 1876, Lewis Carroll proposed a voting system in which the winner is the candidate who with the fewest changes in voters ’ preferences becomes a Condorcet winner—a candidate who beats all other candidates in pairwise majorityrule elections. Bartholdi, Tovey, and Trick provided a lower b ..."
Abstract

Cited by 54 (13 self)
 Add to MetaCart
Abstract. In 1876, Lewis Carroll proposed a voting system in which the winner is the candidate who with the fewest changes in voters ’ preferences becomes a Condorcet winner—a candidate who beats all other candidates in pairwise majorityrule elections. Bartholdi, Tovey, and Trick provided a lower bound—NPhardness—on the computational complexity of determining the election winner in Carroll’s system. We provide a stronger lower bound and an upper bound that matches our lower bound. In particular, determining the winner in Carroll’s system is complete for parallel access to NP, that is, it is complete for � 2 p, for which it becomes the most natural complete problem known. It
The shortest vector in a lattice is hard to approximate to within some constant
 in Proc. 39th Symposium on Foundations of Computer Science
, 1998
"... Abstract. We show that approximating the shortest vector problem (in any ℓp norm) to within any constant factor less than p √ 2 is hardfor NP under reverse unfaithful random reductions with inverse polynomial error probability. In particular, approximating the shortest vector problem is not in RP (r ..."
Abstract

Cited by 51 (4 self)
 Add to MetaCart
Abstract. We show that approximating the shortest vector problem (in any ℓp norm) to within any constant factor less than p √ 2 is hardfor NP under reverse unfaithful random reductions with inverse polynomial error probability. In particular, approximating the shortest vector problem is not in RP (random polynomial time), unless NP equals RP. We also prove a proper NPhardness result (i.e., hardness under deterministic manyone reductions) under a reasonable number theoretic conjecture on the distribution of squarefree smooth numbers. As part of our proof, we give an alternative construction of Ajtai’s constructive variant of Sauer’s lemma that greatly simplifies Ajtai’s original proof. Key words. NPhardness, shortest vector problem, point lattices, geometry of numbers, sphere packing
The History and Status of the P versus NP Question
, 1992
"... this article, I have attempted to organize and describe this literature, including an occasional opinion about the most fruitful directions, but no technical details. In the first half of this century, work on the power of formal systems led to the formalization of the notion of algorithm and the re ..."
Abstract

Cited by 51 (0 self)
 Add to MetaCart
this article, I have attempted to organize and describe this literature, including an occasional opinion about the most fruitful directions, but no technical details. In the first half of this century, work on the power of formal systems led to the formalization of the notion of algorithm and the realization that certain problems are algorithmically unsolvable. At around this time, forerunners of the programmable computing machine were beginning to appear. As mathematicians contemplated the practical capabilities and limitations of such devices, computational complexity theory emerged from the theory of algorithmic unsolvability. Early on, a particular type of computational task became evident, where one is seeking an object which lies