Results 1  10
of
19
A Knowledge Compilation Map
 Journal of Artificial Intelligence Research
, 2002
"... We propose a perspective on knowledge compilation which calls for analyzing different compilation approaches according to two key dimensions: the succinctness of the target compilation language, and the class of queries and transformations that the language supports in polytime. ..."
Abstract

Cited by 174 (26 self)
 Add to MetaCart
We propose a perspective on knowledge compilation which calls for analyzing different compilation approaches according to two key dimensions: the succinctness of the target compilation language, and the class of queries and transformations that the language supports in polytime.
A Survey on Knowledge Compilation
, 1998
"... this paper we survey recent results in knowledge compilation of propositional knowledge bases. We first define and limit the scope of such a technique, then we survey exact and approximate knowledge compilation methods. We include a discussion of compilation for nonmonotonic knowledge bases. Keywor ..."
Abstract

Cited by 100 (3 self)
 Add to MetaCart
this paper we survey recent results in knowledge compilation of propositional knowledge bases. We first define and limit the scope of such a technique, then we survey exact and approximate knowledge compilation methods. We include a discussion of compilation for nonmonotonic knowledge bases. Keywords: Knowledge Representation, Efficiency of Reasoning
A Perspective on Knowledge Compilation
 In Proc. International Joint Conference on Artificial Intelligence (IJCAI
, 2001
"... We provide a perspective on knowledge compilation which calls for analyzing different compilation approaches according to two key dimensions: the succinctness of the target compilation language, and the class of queries and transformations that the language supports in polytime. We argue that ..."
Abstract

Cited by 29 (9 self)
 Add to MetaCart
We provide a perspective on knowledge compilation which calls for analyzing different compilation approaches according to two key dimensions: the succinctness of the target compilation language, and the class of queries and transformations that the language supports in polytime. We argue that such analysis is necessary for placing new compilation approaches within the context of existing ones.
Tractable cover compilations
 in IJCAI’97
, 1997
"... Tractable covers are introduced as a new approach to equivalencepreserving compilation of propositional knowledge bases. First, a general framework is presented. Then, two specific cases are considered. In the first one, partial interpretations are used to shape the knowledge base into tractable fo ..."
Abstract

Cited by 21 (11 self)
 Add to MetaCart
Tractable covers are introduced as a new approach to equivalencepreserving compilation of propositional knowledge bases. First, a general framework is presented. Then, two specific cases are considered. In the first one, partial interpretations are used to shape the knowledge base into tractable formulas from several possible classes. In the second case, they are used to derive renamable Horn formulas. This last case is proved less spaceconsuming than prime implicants cover compilations for every knowledge base. Finally, experimental results show that the new approaches can prove efficient w.r.t. direct query answering and offer significant time and space savings w.r.t. prime implicants covers. I
Compilation of Propositional Weighted Bases
 ARTIFICIAL INTELLIGENCE
, 2004
"... In this paper, we investigate the extent to which knowledge compilation can be used to improve inference from propositional weighted bases. We present a general notion of compilation of a weighted base that is parametrized by any equivalencepreserving compilation function. ..."
Abstract

Cited by 21 (7 self)
 Add to MetaCart
In this paper, we investigate the extent to which knowledge compilation can be used to improve inference from propositional weighted bases. We present a general notion of compilation of a weighted base that is parametrized by any equivalencepreserving compilation function.
Using possibilistic logic for modeling qualitative decision: ATMSbased algorithms
, 1999
"... This paper describes a logical machinery for computing decisions, where the available knowledge on the state of the world is described by a possibilistic propositional logic base (i.e., a collection of logical statements associated with qualitative certainty levels), and where the preferences of the ..."
Abstract

Cited by 13 (6 self)
 Add to MetaCart
This paper describes a logical machinery for computing decisions, where the available knowledge on the state of the world is described by a possibilistic propositional logic base (i.e., a collection of logical statements associated with qualitative certainty levels), and where the preferences of the user are also described by another possibilistic logic base whose formula weights are interpreted in terms of priorities. Two attitudes are allowed for the decision maker: a pessimistic riskaverse one and an optimistic one. The computed decisions are in agreement with a qualitative counterpart to the classical theory of expected utility, recently developed by three of the authors. A link is established between this logical view of qualitative decision making and an ATMSbased computation procedure. Efficient algorithms for computing pessimistic and optimistic optimal decisions are finally given in this logical setting (using some previous work of the fourth author).
Algorithms for propositional KB approximation
 in Proc. of AAAI’98
, 1998
"... One of the obstacles to the effective compilation of propositional knowledge bases (KBs) using Horn approximations, as introduced by (Selman & Kautz 1991), is the lack of computationally feasible methods for generating Horn bounds. In this paper new algorithms for generating Horn Greatest Low ..."
Abstract

Cited by 11 (0 self)
 Add to MetaCart
One of the obstacles to the effective compilation of propositional knowledge bases (KBs) using Horn approximations, as introduced by (Selman & Kautz 1991), is the lack of computationally feasible methods for generating Horn bounds. In this paper new algorithms for generating Horn Greatest Lower Bounds (GLB) that can apply to large size KBs, are presented. The approach is extended through a more general target language: the renamable Horn class. The conditions under which a renamable Horn formula is a renamable Horn GLB of a KB are established and algorithms for computing it are derived. These algorithms can be used in the other approaches based on computation of Horn or renamable lower bounds as (Boufkhad et al. 1997). The efficiency of these algorithms and the tightness with respect to the KB in terms of number of models of the bounds, are experimentally evaluated. The renamable Horn GLB proves to be closer to the KB than the Horn GLB.
On Stratified Belief Base Compilation
, 2004
"... In this paper, we investigate the extent to which knowledge compilation can be used to circumvent the complexity of skeptical inference from a stratified belief base (SBB). We first analyze the compilability of skeptical inference from an SBB, under various requirements concerning both the selection ..."
Abstract

Cited by 6 (3 self)
 Add to MetaCart
In this paper, we investigate the extent to which knowledge compilation can be used to circumvent the complexity of skeptical inference from a stratified belief base (SBB). We first analyze the compilability of skeptical inference from an SBB, under various requirements concerning both the selection policy under consideration, the possibility to make the stratification vary at the online query answering stage and the expected complexity of inference from the compiled form. Not surprisingly, the results are mainly negative. However, since they concern the worst case situation only, they do not prevent a compilationbased approach from being practically useful for some families of instances. While many approaches to compile an SBB can be designed, we are primarily interested in those which take advantage of existing knowledge compilation techniques for classical inference. Specifically, we present a general framework for compiling SBBs into socalled Cnormal SBBs, where C is any tractable class for clausal entailment which is the target class of a compilation function. Another major advantage of the proposed approach lies in the flexibility of the Cnormal belief bases obtained, which means that changing the stratification does not require to recompile the SBB. For several families of compiled SBBs and several selection policies, the complexity of skeptical inference is identified. Some tractable restrictions are exhibited for each policy. Finally, some empirical results are presented.
New compilation languages based on structured decomposability
, 2008
"... We introduce in this paper two new, complete propositional languages and study their properties in terms of (1) their support for polytime operations and (2) their ability to represent boolean functions compactly. The new languages are based on a structured version of decomposability—a property tha ..."
Abstract

Cited by 6 (1 self)
 Add to MetaCart
We introduce in this paper two new, complete propositional languages and study their properties in terms of (1) their support for polytime operations and (2) their ability to represent boolean functions compactly. The new languages are based on a structured version of decomposability—a property that underlies a number of tractable languages. The key characteristic of structured decomposability is its support for a polytime conjoin operation, which is known to be intractable for unstructured decomposability. We show that any CNF can be compiled into formulas in the new languages, whose size is only exponential in the treewidth of the CNF. Our study also reveals that one of the languages we identify is as powerful as OBDDs in terms of answering key inference queries, yet is more succinct than OBDDs.
Exact Knowledge Compilation in Predicate Calculus: the Partial Achievement Case
 In Proceedings of CADE14
, 1997
"... . Forward chaining is an inference algorithm based on modus ponens which is widely used but unfortunately incomplete with respect to the usual boolean logic. A solution to this incompleteness is to compile the knowledge base so that forward chaining becomes complete for any base of facts. This compi ..."
Abstract

Cited by 3 (0 self)
 Add to MetaCart
. Forward chaining is an inference algorithm based on modus ponens which is widely used but unfortunately incomplete with respect to the usual boolean logic. A solution to this incompleteness is to compile the knowledge base so that forward chaining becomes complete for any base of facts. This compilation is called achievement and was so far restricted to propositional calculus. In this paper, we extend this compilation method to the predicate calculus. To our knowledge, this is the first exact knowledge compilation for first order logic. 1 Introduction Forward chaining is an algorithm which is widely used because it is very simple and quite efficient. It aims at producing the set of literals which are implied by the knowledge (a set of rules 1 or clauses) and a set of facts 2 . This is a production algorithm which differs from query algorithms because it produces a set of implied literals instead of just answering whether or not a literal is implied. We use the notation Fwch(B [ F...