• Documents
  • Authors
  • Tables
  • Log in
  • Sign up
  • MetaCart
  • DMCA
  • Donate

CiteSeerX logo

Advanced Search Include Citations
Advanced Search Include Citations

Bucket elimination: A unifying framework for reasoning (1999)

by R Dechter
Venue:Artif. Intell
Add To MetaCart

Tools

Sorted by:
Results 1 - 10 of 298
Next 10 →

Multiagent Planning with Factored MDPs

by Carlos Guestrin, Daphne Koller, Ronald Parr - In NIPS-14 , 2001
"... We present a principled and efficient planning algorithm for cooperative multiagent dynamic systems. A striking feature of our method is that the coordination and communication between the agents is not imposed, but derived directly from the system dynamics and function approximation architecture ..."
Abstract - Cited by 176 (15 self) - Add to MetaCart
We present a principled and efficient planning algorithm for cooperative multiagent dynamic systems. A striking feature of our method is that the coordination and communication between the agents is not imposed, but derived directly from the system dynamics and function approximation architecture. We view the entire multiagent system as a single, large Markov decision process (MDP), which we assume can be represented in a factored way using a dynamic Bayesian network (DBN). The action space of the resulting MDP is the joint action space of the entire set of agents. Our approach is based on the use of factored linear value functions as an approximation to the joint value function. This factorization of the value function allows the agents to coordinate their actions at runtime using a natural message passing scheme. We provide a simple and efficient method for computing such an approximate value function by solving a single linear program, whose size is determined by the interaction between the value function structure and the DBN. We thereby avoid the exponential blowup in the state and action space. We show that our approach compares favorably with approaches based on reward sharing. We also show that our algorithm is an efficient alternative to more complicated algorithms even in the single agent case.
(Show Context)

Citation Context

...ular Q i . Fig. 1(a) shows the coordination graph for an example where Q = Q 1 (a 1 ; a 2 ) +Q 2 (a 2 ; a 4 ) +Q 3 (a 1 ; a 3 ) +Q 4 (a 3 ; a 4 ). A graph structure suggests the use of a cost network =-=[5]-=-, which can be solved using non-serial dynamic programming [1] or a variable elimination algorithm which is virtually identical to variable elimination in a Bayesian network. The key idea is that, rat...

Lifted first-order probabilistic inference

by Rodrigo De Salvo Braz, Eyal Amir, Dan Roth - In Proceedings of IJCAI-05, 19th International Joint Conference on Artificial Intelligence , 2005
"... Most probabilistic inference algorithms are specified and processed on a propositional level. In the last decade, many proposals for algorithms accepting first-order specifications have been presented, but in the inference stage they still operate on a mostly propositional representation level. [Poo ..."
Abstract - Cited by 126 (8 self) - Add to MetaCart
Most probabilistic inference algorithms are specified and processed on a propositional level. In the last decade, many proposals for algorithms accepting first-order specifications have been presented, but in the inference stage they still operate on a mostly propositional representation level. [Poole, 2003] presented a method to perform inference directly on the first-order level, but this method is limited to special cases. In this paper we present the first exact inference algorithm that operates directly on a first-order level, and that can be applied to any first-order model (specified in a language that generalizes undirected graphical models). Our experiments show superior performance in comparison with propositional exact inference. 1

AND/OR Search Spaces for Graphical Models

by Rina Dechter , 2004
"... The paper introduces an AND/OR search space perspective for graphical models that include probabilistic networks (directed or undirected) and constraint networks. In contrast to the traditional (OR) search space view, the AND/OR search tree displays some of the independencies present in the gr ..."
Abstract - Cited by 119 (44 self) - Add to MetaCart
The paper introduces an AND/OR search space perspective for graphical models that include probabilistic networks (directed or undirected) and constraint networks. In contrast to the traditional (OR) search space view, the AND/OR search tree displays some of the independencies present in the graphical model explicitly and may sometime reduce the search space exponentially. Indeed, most

Exact Bayesian structure discovery in Bayesian networks

by Mikko Koivisto - J. of Machine Learning Research , 2004
"... We consider a Bayesian method for learning the Bayesian network structure from complete data. Recently, Koivisto and Sood (2004) presented an algorithm that for any single edge computes its marginal posterior probability in O(n2 n) time, where n is the number of attributes; the number of parents per ..."
Abstract - Cited by 115 (12 self) - Add to MetaCart
We consider a Bayesian method for learning the Bayesian network structure from complete data. Recently, Koivisto and Sood (2004) presented an algorithm that for any single edge computes its marginal posterior probability in O(n2 n) time, where n is the number of attributes; the number of parents per attribute is bounded by a constant. In this paper we show that the posterior probabilities for all the n(n−1) potential edges can be computed in O(n2 n) total time. This result is achieved by a forward–backward technique and fast Möbius transform algorithms, which are of independent interest. The resulting speedup by a factor of about n 2 allows us to experimentally study the statistical power of learning moderate-size networks. We report results from a simulation study that covers data sets with 20 to 10,000 records over 5 to 25 discrete attributes. 1

Coordinated Reinforcement Learning

by Carlos Guestrin, Michail Lagoudakis, Ronald Parr - In Proceedings of the ICML-2002 The Nineteenth International Conference on Machine Learning , 2002
"... We present several new algorithms for multiagent reinforcement learning. A common feature of these algorithms is a parameterized, structured representation of a policy or value function. This structure is leveraged in an approach we call coordinated reinforcement learning, by which agents coordinate ..."
Abstract - Cited by 113 (6 self) - Add to MetaCart
We present several new algorithms for multiagent reinforcement learning. A common feature of these algorithms is a parameterized, structured representation of a policy or value function. This structure is leveraged in an approach we call coordinated reinforcement learning, by which agents coordinate both their action selection activities and their parameter updates. Within the limits of our parametric representations, the agents will determine a jointly optimal action without explicitly considering every possible action in their exponentially large joint action space. Our methods differ from many previous reinforcement learning approaches to multiagent coordination in that structured communication and coordination between agents appears at the core of both the learning algorithm and the execution architecture. Our experimental results, comparing our approach to other RL methods, illustrate both the quality of the policies obtained and the additional benefits of coordination. 1.
(Show Context)

Citation Context

...heir actions to optimize some particular Qj. Fig. 1 shows the coordination graph for an example where Q = Q1(a1,a2)+Q2(a2,a4)+Q3(a1,a3)+Q4(a3,a4). A graph structure suggests the use of a cost network =-=[6]-=-, which can be solved using non-serial dynamic programming [3] or a variable elimination algorithm which is virtually identical to variable elimination in a Bayesian network. We review this constructi...

Policy Iteration for Factored MDPs

by Daphne Koller, Ronald Parr - In Proceedings of the Sixteenth Conference on Uncertainty in Artificial Intelligence (UAI-00 , 2000
"... Many large MDPs can be represented compactly using a dynamic Bayesian network. Although the structure of the value function does not retain the structure of the process, recent work has suggested that value functions in factored MDPs can often be approximated well using a factored value functi ..."
Abstract - Cited by 94 (6 self) - Add to MetaCart
Many large MDPs can be represented compactly using a dynamic Bayesian network. Although the structure of the value function does not retain the structure of the process, recent work has suggested that value functions in factored MDPs can often be approximated well using a factored value function: a linear combination of restricted basis functions, each of which refers only to a small subset of variables. An approximate factored value function for a particular policy can be computed using approximate dynamic programming, but this approach (and others) can only produce an approximation relative to a distance metric which is weighted by the stationary distribution of the current policy. This type of weighted projection is ill-suited to policy improvement.
(Show Context)

Citation Context

... functions. Hence each inner maximization is over a linear combination of functions, each of which is restricted to some small subset of variables. This type of optimization problem is a cost network =-=[8]-=-, and can be solved using standard variable elimination algorithms. The computational cost, as for other related structures, is exponential in the induced width of the graph induced by the hyper-edges...

Max-norm Projections for Factored MDPs

by Carlos Guestrin, Daphne Koller, Ronald Parr - In IJCAI , 2001
"... Markov Decision Processes (MDPs) provide a coherent mathematical framework for planning under uncertainty. ..."
Abstract - Cited by 77 (10 self) - Add to MetaCart
Markov Decision Processes (MDPs) provide a coherent mathematical framework for planning under uncertainty.

Conditional random fields for activity recognition

by Douglas L. Vail, Carlos Guestrin - In Proceedings of the Sixth International Joint Conference on Autonomous Agents and Multiagent Systems (AAMAS 2007 , 2007
"... of any sponsoring institution, the U.S. government or any other entity. ..."
Abstract - Cited by 76 (0 self) - Add to MetaCart
of any sponsoring institution, the U.S. government or any other entity.
(Show Context)

Citation Context

...is section, we describe how to restructure the computation to eliminate this inefficiency. In the general case, the procedure that we describe is known as variable elimination [1], bucket elimination =-=[20]-=-, or the sum-product algorithm [53]. Our particular derivation for linear-chain CRFs is equivalent to the backward pass of the forward-backward algorithm for HMMs [91] and similar to the discussion in...

Lifted probabilistic inference with counting formulas

by Brian Milch, Luke S. Zettlemoyer, Kristian Kersting, Michael Haimes, Leslie Pack Kaelbling - Proceedings of the Twenty-Third AAAI Conference on Artificial Intelligence (AAAI-2008 , 2008
"... Lifted inference algorithms exploit repeated structure in probabilistic models to answer queries efficiently. Previous work such as de Salvo Braz et al.’s first-order variable elimination (FOVE) has focused on the sharing of potentials across interchangeable random variables. In this paper, we also ..."
Abstract - Cited by 73 (11 self) - Add to MetaCart
Lifted inference algorithms exploit repeated structure in probabilistic models to answer queries efficiently. Previous work such as de Salvo Braz et al.’s first-order variable elimination (FOVE) has focused on the sharing of potentials across interchangeable random variables. In this paper, we also exploit interchangeability within individual potentials by introducing counting formulas, which indicate how many of the random variables in a set have each possible value. We present a new lifted inference algorithm, C-FOVE, that not only handles counting formulas in its input, but also creates counting formulas for use in intermediate potentials. C-FOVE can be described succinctly in terms of six operators, along with heuristics for when to apply them. Because counting formulas capture dependencies among large numbers of variables compactly, C-FOVE achieves asymptotic speed improvements compared to FOVE.
(Show Context)

Citation Context

...hical model is usually measured by its treewidth — a graphtheoretic concept that determines the size of the largest potential created while executing the best possible sequence of sum-out operations (=-=Dechter 1999-=-). An appropriate measure in the lifted setting is the size of the largest potential created by the optimal sequence of C-FOVE operations (of which sum-out is just one). In models with sufficient symm...

Constraint Satisfaction, Bounded Treewidth, and Finite-Variable Logics

by Victor Dalmau, Phokion G. Kolaitis, Moshe Y. Vardi , 2002
"... We systematically investigate the connections between constraint satisfaction problems, structures of bounded treewidth, and definability in logics with a finite number of variables. We first show that constraint satisfaction problems on inputs of treewidth less than k are definable using Datalog ..."
Abstract - Cited by 71 (12 self) - Add to MetaCart
We systematically investigate the connections between constraint satisfaction problems, structures of bounded treewidth, and definability in logics with a finite number of variables. We first show that constraint satisfaction problems on inputs of treewidth less than k are definable using Datalog programs with at most k variables; this provides a new explanation for the tractability of these classes of problems. After this, we investigate constraint satisfaction on inputs that are homomorphically equivalent to structures of bounded treewidth.
Powered by: Apache Solr
  • About CiteSeerX
  • Submit and Index Documents
  • Privacy Policy
  • Help
  • Data
  • Source
  • Contact Us

Developed at and hosted by The College of Information Sciences and Technology

© 2007-2019 The Pennsylvania State University