Results 1  10
of
47
Logical Models of Argument
 ACM COMPUTING SURVEYS
, 2000
"... Logical models of argument formalize commonsense reasoning while taking process and computation seriously. This survey discusses the main ideas which characterize different logical models of argument. It presents the formal features of a few main approaches to the modeling of argumentation. We trace ..."
Abstract

Cited by 144 (33 self)
 Add to MetaCart
Logical models of argument formalize commonsense reasoning while taking process and computation seriously. This survey discusses the main ideas which characterize different logical models of argument. It presents the formal features of a few main approaches to the modeling of argumentation. We trace the
Probabilistic Default Reasoning with Conditional Constraints
 ANN. MATH. ARTIF. INTELL
, 2000
"... We present an approach to reasoning from statistical and subjective knowledge, which is based on a combination of probabilistic reasoning from conditional constraints with approaches to default reasoning from conditional knowledge bases. More precisely, we introduce the notions of , lexicographic, ..."
Abstract

Cited by 35 (20 self)
 Add to MetaCart
We present an approach to reasoning from statistical and subjective knowledge, which is based on a combination of probabilistic reasoning from conditional constraints with approaches to default reasoning from conditional knowledge bases. More precisely, we introduce the notions of , lexicographic, and conditional entailment for conditional constraints, which are probabilistic generalizations of Pearl's entailment in system , Lehmann's lexicographic entailment, and Geffner's conditional entailment, respectively. We show that the new formalisms have nice properties. In particular, they show a similar behavior as referenceclass reasoning in a number of uncontroversial examples. The new formalisms, however, also avoid many drawbacks of referenceclass reasoning. More precisely, they can handle complex scenarios and even purely probabilistic subjective knowledge as input. Moreover, conclusions are drawn in a global way from all the available knowledge as a whole. We then show that the new formalisms also have nice general nonmonotonic properties. In detail, the new notions of , lexicographic, and conditional entailment have similar properties as their classical counterparts. In particular, they all satisfy the rationality postulates proposed by Kraus, Lehmann, and Magidor, and they have some general irrelevance and direct inference properties. Moreover, the new notions of  and lexicographic entailment satisfy the property of rational monotonicity. Furthermore, the new notions of , lexicographic, and conditional entailment are proper generalizations of both their classical counterparts and the classical notion of logical entailment for conditional constraints. Finally, we provide algorithms for reasoning under the new formalisms, and we analyze its computational com...
Possibility theory and statistical reasoning
 Computational Statistics & Data Analysis Vol
, 2006
"... Numerical possibility distributions can encode special convex families of probability measures. The connection between possibility theory and probability theory is potentially fruitful in the scope of statistical reasoning when uncertainty due to variability of observations should be distinguished f ..."
Abstract

Cited by 26 (2 self)
 Add to MetaCart
Numerical possibility distributions can encode special convex families of probability measures. The connection between possibility theory and probability theory is potentially fruitful in the scope of statistical reasoning when uncertainty due to variability of observations should be distinguished from uncertainty due to incomplete information. This paper proposes an overview of numerical possibility theory. Its aim is to show that some notions in statistics are naturally interpreted in the language of this theory. First, probabilistic inequalites (like Chebychev’s) offer a natural setting for devising possibility distributions from poor probabilistic information. Moreover, likelihood functions obey the laws of possibility theory when no prior probability is available. Possibility distributions also generalize the notion of confidence or prediction intervals, shedding some light on the role of the mode of asymmetric probability densities in the derivation of maximally informative interval substitutes of probabilistic information. Finally, the simulation of fuzzy sets comes down to selecting a probabilistic representation of a possibility distribution, which coincides with the Shapley value of the corresponding consonant capacity. This selection process is in agreement with Laplace indifference principle and is closely connected with the mean interval of a fuzzy interval. It sheds light on the “defuzzification ” process in fuzzy set theory and provides a natural definition of a subjective possibility distribution that sticks to the Bayesian framework of exchangeable bets. Potential applications to risk assessment are pointed out. 1
Probabilistic Logic under Coherence: Complexity and Algorithms
 In Proceedings ISIPTA01
, 2001
"... We study probabilistic logic under the viewpoint of the coherence principle of de Finetti. In detail, we explore the relationship between coherencebased and classical modeltheoretic probabilistic logic. Interestingly, we show that the notions of gcoherence and of gcoherent entailment can be expre ..."
Abstract

Cited by 22 (11 self)
 Add to MetaCart
We study probabilistic logic under the viewpoint of the coherence principle of de Finetti. In detail, we explore the relationship between coherencebased and classical modeltheoretic probabilistic logic. Interestingly, we show that the notions of gcoherence and of gcoherent entailment can be expressed by combining notions in modeltheoretic probabilistic logic with concepts from default reasoning. Using these results, we analyze the computational complexity of probabilistic reasoning under coherence. Moreover, we present new algorithms for deciding gcoherence and for computing tight gcoherent intervals, which reduce these tasks to standard reasoning tasks in modeltheoretic probabilistic logic. Thus, efficient techniques for modeltheoretic probabilistic reasoning can immediately be applied for probabilistic reasoning under coherence, for example, column generation techniques. We then describe two other interesting techniques for efficient modeltheoretic probabilistic reasoning in the conjunctive case.
Probabilistic Logic under Coherence, ModelTheoretic Probabilistic Logic, and Default Reasoning
 Journal of Applied NonClassical Logics
"... We study probabilistic logic under the viewpoint of the coherence principle of de Finetti. In detail, we explore the relationship between coherencebased and modeltheoretic probabilistic logic. Interestingly, we show that the notions of gcoherence and of gcoherent entailment can be expressed by co ..."
Abstract

Cited by 22 (9 self)
 Add to MetaCart
We study probabilistic logic under the viewpoint of the coherence principle of de Finetti. In detail, we explore the relationship between coherencebased and modeltheoretic probabilistic logic. Interestingly, we show that the notions of gcoherence and of gcoherent entailment can be expressed by combining notions in modeltheoretic probabilistic logic with concepts from default reasoning. Crucially, we even show that probabilistic reasoning under coherence is a probabilistic generalization of default reasoning in system P. That is, we provide a new probabilistic semantics for system P, which is neither based on infinitesimal probabilities nor on atomicbound (or also bigstepped) probabilities. These results also give new insight into default reasoning with conditional objects.
Default Reasoning from Conditional Knowledge Bases: Complexity and Tractable Cases
 Artif. Intell
, 2000
"... Conditional knowledge bases have been proposed as belief bases that include defeasible rules (also called defaults) of the form " ! ", which informally read as "generally, if then ." Such rules may have exceptions, which can be handled in different ways. A number of entailment semantics for condi ..."
Abstract

Cited by 21 (13 self)
 Add to MetaCart
Conditional knowledge bases have been proposed as belief bases that include defeasible rules (also called defaults) of the form " ! ", which informally read as "generally, if then ." Such rules may have exceptions, which can be handled in different ways. A number of entailment semantics for conditional knowledge bases have been proposed in the literature. However, while the semantic properties and interrelationships of these formalisms are quite well understood, about their computational properties only partial results are known so far. In this paper, we fill these gaps and first draw a precise picture of the complexity of default reasoning from conditional knowledge bases: Given a conditional knowledge base KB and a default ! , does KB entail ! ? We classify the complexity of this problem for a number of wellknown approaches (including Goldszmidt et al.'s maximum entropy approach and Geffner's conditional entailment), where we consider the general propositional case as well as natural syntactic restrictions (in particular, to Horn and literalHorn conditional knowledge bases). As we show, the more sophisticated semantics for conditional knowledge bases are plagued with intractability in all these fragments. We thus explore cases in which these semantics are tractable, and find that most of them enjoy this property on feedbackfree Horn conditional knowledge bases, which constitute a new, meaningful class of conditional knowledge bases. Furthermore, we generalize previous tractability results from Horn to qHorn conditional knowledge bases, which allow for a limited use of disjunction. Our results complement and extend previous results, and contribute in refining the tractability/intractability frontier of default reasoning from conditional know...
Weak nonmonotonic probabilistic logics
"... Towards probabilistic formalisms for resolving local inconsistencies under modeltheoretic probabilistic entailment, we present probabilistic generalizations of Pearl’s entailment in System Z and Lehmann’s lexicographic entailment. We then analyze the nonmonotonic and semantic properties of the new ..."
Abstract

Cited by 21 (6 self)
 Add to MetaCart
Towards probabilistic formalisms for resolving local inconsistencies under modeltheoretic probabilistic entailment, we present probabilistic generalizations of Pearl’s entailment in System Z and Lehmann’s lexicographic entailment. We then analyze the nonmonotonic and semantic properties of the new notions of entailment. In particular, we show that they satisfy the rationality postulates of System P and the property of Rational Monotonicity. Moreover, we show that modeltheoretic probabilistic entailment is stronger than the new notion of lexicographic entailment, which in turn is stronger than the new notion of entailment in System Z. As an important feature of the new notions of entailment in System Z and lexicographic entailment, we show that they coincide with modeltheoretic probabilistic entailment whenever there are no local inconsistencies. We also show that the new notions of entailment in System Z and lexicographic entailment are proper generalizations of their classical counterparts. Finally, we present algorithms for reasoning under the new formalisms, and we give a precise picture of its computational complexity.
Logical Representation and Computation of Optimal Decisions in a Qualitative Setting
 In Proceedings of the Fifteenth National Conference on Artificial Intelligence
, 1998
"... This paper describes a logical machinery for computing decisions based on an ATMS procedure, where the available knowledge on the state of the world is described by a possibilistic propositional logic base (i.e., a collection of logical statements associated with qualitative certainty levels). ..."
Abstract

Cited by 16 (6 self)
 Add to MetaCart
This paper describes a logical machinery for computing decisions based on an ATMS procedure, where the available knowledge on the state of the world is described by a possibilistic propositional logic base (i.e., a collection of logical statements associated with qualitative certainty levels). The preferences of the user are also described by another possibilistic logic base whose formula weights are interpreted in terms of priorities and formulas express goals. Two attitudes are allowed for the decision maker: a pessimistic uncertaintyaverse one and an optimistic one. The computed decisions are in agreement with a qualitative counterpart to classical expected utility theory for decision under uncertainty.
A stratificationbased approach for handling conflicts in access control
 In 8th ACM Symposium on Access Control Models and Technologies (SACMAT’03
, 2003
"... ..."
Ordinal and probabilistic representations of acceptance
 J. Artificial Intelligence Research
, 2004
"... An accepted belief is a proposition considered likely enough by an agent, to be inferred from as if it were true. This paper bridges the gap between probabilistic and logical representations of accepted beliefs. To this end, natural properties of relations on propositions, describing relative streng ..."
Abstract

Cited by 11 (4 self)
 Add to MetaCart
An accepted belief is a proposition considered likely enough by an agent, to be inferred from as if it were true. This paper bridges the gap between probabilistic and logical representations of accepted beliefs. To this end, natural properties of relations on propositions, describing relative strength of belief are augmented with some conditions ensuring that accepted beliefs form a deductively closed set. This requirement turns out to be very restrictive. In particular, it is shown that the sets of accepted belief of an agent can always be derived from a family of possibility rankings of states. An agent accepts a proposition in a given context if this proposition is considered more possible than its negation in this context, for all possibility rankings in the family. These results are closely connected to the nonmonotonic 'preferential' inference system of Kraus, Lehmann and Magidor and the socalled plausibility functions of Friedman and Halpern. The extent to which probability theory is compatible with acceptance relations is laid bare. A solution to the lottery paradox, which is considered as a major impediment to the use of nonmonotonic inference is proposed using a special kind of probabilities (called lexicographic, or bigstepped). The setting of acceptance relations also proposes another way of approaching the theory of belief change after the works of Gärdenfors and colleagues. Our view considers the acceptance relation as a primitive object from which belief sets are derived in various contexts. 1.