• Documents
  • Authors
  • Tables
  • Log in
  • Sign up
  • MetaCart
  • DMCA
  • Donate

CiteSeerX logo

Advanced Search Include Citations
Advanced Search Include Citations | Disambiguate

DMCA

Being Bayesian about network structure (2000)

Cached

  • Download as a PDF

Download Links

  • [www.cs.iastate.edu]
  • [people.ee.duke.edu]
  • [people.ee.duke.edu]
  • [ai.stanford.edu]
  • [robotics.stanford.edu]
  • [robotics.stanford.edu]
  • [robotics.stanford.edu]
  • [ai.stanford.edu]
  • [robotics.stanford.edu]
  • [robotics.stanford.edu]
  • [www.cs.huji.ac.il]
  • [robotics.stanford.edu]
  • [ringil.cis.ksu.edu]
  • [www.kddresearch.org]
  • [www.kddresearch.org]
  • [www.cs.huji.ac.il]
  • [www.cs.huji.ac.il]
  • [robotics.stanford.edu]
  • [robotics.stanford.edu]
  • [robotics.stanford.edu]
  • [www-robotics.stanford.edu]
  • [www.cs.huji.ac.il]
  • [www.cs.huji.ac.il]
  • [robotics.stanford.edu]
  • [robotics.stanford.edu]
  • [robotics.stanford.edu]
  • [ai.stanford.edu]
  • [www-robotics.stanford.edu]
  • [ai.stanford.edu]
  • [robotics.stanford.edu]
  • [robotics.stanford.edu]
  • [ai.stanford.edu]
  • [robotics.stanford.edu]

  • Other Repositories/Bibliography

  • DBLP
  • Save to List
  • Add to Collection
  • Correct Errors
  • Monitor Changes
by Nir Friedman
Venue:Machine Learning
Citations:299 - 3 self
  • Summary
  • Citations
  • Active Bibliography
  • Co-citation
  • Clustered Documents
  • Version History

Citations

3620 Equation of State Calculations by Fast Computing Machines - Metropolis, Rosenbluth, et al. - 1953 (Show Context)

Citation Context

...y, we have no elegant tricks that allow a closed form solution. Therefore, we provide a solution which uses our closed form solution of Eq. (6) as a subroutine in a Markov Chain Monte Carlo algorithm =-=[15]-=-. 4.1 The basic algorithm We introduce a uniform prior over orderings ffi , and define ffi to be of the same nature as the priors we used in the previous section. It is important to note that ...

1397 A Bayesian method for the induction of probabilistic networks from data - Cooper, Herskovits - 1992 (Show Context)

Citation Context

...ring ffi . In other words, we restrict attention to structures where if )! #%$ "! then #$ffi . This assumption was a standard one in the early work on learning Bayesian networks from data =-=[4]-=-. 3.1 Computing marginal likelihood We first consider the problem of computing the probability of the data given the ordering: ffiss ffiss(5) Note that this summation...

1343 Reversible jump Markov chain Monte Carlo computation and Bayesian model determination. Biometrika 82 - Green - 1995
1157 Learning Bayesian networks: The combination of knowledge and statistical data. - Heckerman, Geiger, et al. - 1995 (Show Context)

Citation Context

... each instantiation " to its parents [8]. In Gaussian networks, we might use a Wishart prior [9]. For our purpose, we need only require that the prior satisfies two basic assumptions, as presented in =-=[10]-=-: 9 Global parameter independence: Let & fi ffi >@?=AffBC fiED be the parameters specifying the behavior of the variable ! given the various instantiations to its parents. Then we require that ...

1088 Using Bayesian networks to analyze expression data - Friedman, Linial, et al. - 2000 (Show Context)

Citation Context

...ight want to use Bayesian network learning to understand the mechanisms by which genes in a cell produce proteins, which in turn cause other genes to express themselves, or prevent them from doing so =-=[6]-=-. In this case, our main goal is to discover the causal and dependence relations between the expression levels of different genes [12]. The common approach to discovering structure is to use learning ...

1068 A tutorial on learning with Bayesian networks. - Heckerman - 1998 (Show Context)

Citation Context

...rk structures, and to a non-Bayesian bootstrap approach. 1 Introduction In the last decade there has been a great deal of research focused on the problem of learning Bayesian networks (BNs) from data =-=[3, 8]-=-. An obvious motivation for this problem is to learn a model that we can then use for inference or decision making, as a substitute for a model constructed by a human expert. In certain cases, however...

855 UCI Repository of machine learning databases. - Murphy, Aha - 1994 (Show Context)

Citation Context

...orderings, this enumeration is possible for as many as 10 variables, but for structures, we are limited to domains with 5–6 variables. We took two data sets — Vote and Flare — from the UCI repository =-=[16]-=- and selected five variables from each. We generated datasets of sizes ands , and computed the full Bayesian averaging posterior for these datasets using both methods. Figure 1 compares the res...

520 Sampling based approaches to calculating marginal densities. - Gelfand, Smith - 1990
370 Model selection and accounting for model uncertainly in graphical models using Occam’s window. - Madigan, Raftery - 1994 (Show Context)

Citation Context

...onal constraints that restrict the space (as in [11]). Alternatively, this summation can be approximated by considering only a subset of possible structures. Several approximations have been proposed =-=[13, 14]-=-. One theoretically well-founded approach is to use Markov Chain Monte Carlo (MCMC) methods: we define a Markov chain over structures whose stationary distribution is the posterior s, we then...

354 Bayesian graphical models for discrete data - Madigan, York - 1995 (Show Context)

Citation Context

...onal constraints that restrict the space (as in [11]). Alternatively, this summation can be approximated by considering only a subset of possible structures. Several approximations have been proposed =-=[13, 14]-=-. One theoretically well-founded approach is to use Markov Chain Monte Carlo (MCMC) methods: we define a Markov chain over structures whose stationary distribution is the posterior s, we then...

285 The ALARM monitoring system: A case study with two probablistic inference techniques for belief networks - Beinlich, Suermondt, et al. - 1988 (Show Context)

Citation Context

...he exact computation for both types of features. We then considered larger datasets, where exhaustive enumeration is not an option. For this purpose we used synthetic data generated from the Alarm BN =-=[1]-=-, a network with 37 nodes. Here, our computational tricks are necessary. We used the following settings: (max. number of parents in a family) ; (max. number of potential parents) s ; (...

255 Theory refinement on Bayesian networks. - Buntine - 1991 (Show Context)

Citation Context

...e families for a node ! is at most * - . 6 - . Hence, the total cost of computing Eq. (6) is at most 66 - 6 - . We note that the decomposition of Eq. (6) was first mentioned by Buntine =-=[2]-=-, but the ramifications for Bayesian model averaging were not pursued. The concept of Bayesian model averaging using a closed-form summation over an exponentially large set of structures was proposed ...

246 Learning Bayesian network structure from massive datasets: The “sparse candidate” algorithm. - Friedman, Nachman, et al. - 1999
242 Covariance structure of the Gibbs sampler with applications to comparisions of estimators and augmentation schemes. - Liu, Wong, et al. - 1994
205 A guide to the literature on learning probabilistic networks from data - Buntine - 1996 (Show Context)

Citation Context

...rk structures, and to a non-Bayesian bootstrap approach. 1 Introduction In the last decade there has been a great deal of research focused on the problem of learning Bayesian networks (BNs) from data =-=[3, 8]-=-. An obvious motivation for this problem is to learn a model that we can then use for inference or decision making, as a substitute for a model constructed by a human expert. In certain cases, however...

192 Rao-Blackwellisation of sampling schemes,” - Casella, Robert - 1996
98 A Bayesian Approach to Causal Discovery, - Heckerman, Meek, et al. - 1997 (Show Context)

Citation Context

...r model for the structure of the domain. Indeed, in small domains with a substantial amount of data, it has been shown that the highest scoring model is orders of magnitude more likely than any other =-=[11]-=-. In such cases, model selection is a good approximation. Unfortunately, there are many domains of interest where this situation does not hold. In our gene expression example, it is now possible to me...

71 Data analysis with bayesian networks: A bootstrap approach. - Friedman, Goldszmidt, et al. - 1999 (Show Context)

Citation Context

...tes. We also present results showing that our approach accurately captures dominant features even with sparse data, and that it outperforms both MCMC over structures and the non-Bayesian bootstrap of =-=[5]-=-. 2 Bayesian learning of Bayesian networks Consider the problem of analyzing the distribution over some set of random variables , based on a fully observed data set ss ...

70 Array of hope. - Lander - 1999 (Show Context)

Citation Context

...e other genes to express themselves, or prevent them from doing so [6]. In this case, our main goal is to discover the causal and dependence relations between the expression levels of different genes =-=[12]-=-. The common approach to discovering structure is to use learning with model selection to provide us with a single high-scoring model. We then use that model (or its equivalence class) as our model fo...

68 Learning Bayesian network structures by searching for the best ordering with genetic algorithms. - Larranaga, Kuijpers, et al. - 1996
52 Learning Bayesian networks: a unification for discrete and Gaussian domains - Heckerman, Geiger - 1995 (Show Context)

Citation Context

.... In discrete networks, the standard assumption is a Dirichlet prior over &flfiffisfor each node )! and each instantiation " to its parents [8]. In Gaussian networks, we might use a Wishart prior =-=[9]-=-. For our purpose, we need only require that the prior satisfies two basic assumptions, as presented in [10]: 9 Global parameter independence: Let & fi ffi >@?=AffBC fiED be the parameters specifyi...

46 Bayesian model averaging and model selection for Markov equivalence classes of acyclic digraphs,” - Madigan, Andersson, et al. - 1996
27 An efficient extension to mixture techniques for prediction and decision trees - Pereira, Singer - 1999 (Show Context)

Citation Context

...Bayesian model averaging were not pursued. The concept of Bayesian model averaging using a closed-form summation over an exponentially large set of structures was proposed (in a different setting) in =-=[17]-=-. The computation of ffi is useful in and of itself; as we show in the next section, computing the probability ffi is a key step in our MCMC algorithm. 3.2 Probabilities of features Fo...

13 Markov chain Monte Carlo Methods in Practice - Gilks, Richardson, et al. - 1996 (Show Context)

Citation Context

...his move with probabilitys ffi H * ffi ffi Hs ffi * ffi H ffis It is well known that the resulting chain is reversible and has the desired stationary distribution =-=[7]-=-. We consider several specific constructions for the proposal distribution, based on different neighborhoods in the space of orderings. In one very simple construction, we consider only operators that...

2 Botstein D and Futcher B (1998): Comprehensive identification of cell cycle-regulated genes of the yeast saccharomyces cerevisiae by microarray hybridization - PT, Sherlock, et al. - 1982
1 Tarantola: 2000, `Efficient model determination for discrete graphical models'. Biometrika - Giudici, Green, et al.
1 Aha: 1995, `UCI Repository of Machine Learning Databases'. http://www.ics.uci.edu/mlearn/MLRepository.html - Murphy, W
Powered by: Apache Solr
  • About CiteSeerX
  • Submit and Index Documents
  • Privacy Policy
  • Help
  • Data
  • Source
  • Contact Us

Developed at and hosted by The College of Information Sciences and Technology

© 2007-2019 The Pennsylvania State University