Results 1  10
of
247,557
Probabilistic principal surfaces classifier
 FSKD 2005, LNAI 3614
, 2005
"... Abstract. In this paper we propose using manifolds modeled by probabilistic principle surfaces (PPS) to characterize and classify highD data. The PPS can be thought of as a nonlinear probabilistic generalization of principal components, as it is designed to pass through the “middle ” of the data. I ..."
Abstract

Cited by 1 (0 self)
 Add to MetaCart
Abstract. In this paper we propose using manifolds modeled by probabilistic principle surfaces (PPS) to characterize and classify highD data. The PPS can be thought of as a nonlinear probabilistic generalization of principal components, as it is designed to pass through the “middle ” of the data
Mixtures of Probabilistic Principal Component Analysers
, 1998
"... Principal component analysis (PCA) is one of the most popular techniques for processing, compressing and visualising data, although its effectiveness is limited by its global linearity. While nonlinear variants of PCA have been proposed, an alternative paradigm is to capture data complexity by a com ..."
Abstract

Cited by 537 (6 self)
 Add to MetaCart
maximumlikelihood framework, based on a specific form of Gaussian latent variable model. This leads to a welldefined mixture model for probabilistic principal component analysers, whose parameters can be determined using an EM algorithm. We discuss the advantages of this model in the context
Estimating Continuous Distributions in Bayesian Classifiers
 In Proceedings of the Eleventh Conference on Uncertainty in Artificial Intelligence
, 1995
"... When modeling a probability distribution with a Bayesian network, we are faced with the problem of how to handle continuous variables. Most previous work has either solved the problem by discretizing, or assumed that the data are generated by a single Gaussian. In this paper we abandon the normality ..."
Abstract

Cited by 489 (2 self)
 Add to MetaCart
the normality assumption and instead use statistical methods for nonparametric density estimation. For a naive Bayesian classifier, we present experimental results on a variety of natural and artificial domains, comparing two methods of density estimation: assuming normality and modeling each conditional
The Perceptron: A Probabilistic Model for Information Storage and Organization in The Brain
 Psychological Review
, 1958
"... If we are eventually to understand the capability of higher organisms for perceptual recognition, generalization, recall, and thinking, we must first have answers to three fundamental questions: 1. How is information about the physical world sensed, or detected, by the biological system? 2. In what ..."
Abstract

Cited by 1143 (0 self)
 Add to MetaCart
If we are eventually to understand the capability of higher organisms for perceptual recognition, generalization, recall, and thinking, we must first have answers to three fundamental questions: 1. How is information about the physical world sensed, or detected, by the biological system? 2. In what form is information stored, or remembered? 3. How does information contained in storage, or in memory, influence recognition and behavior? The first of these questions is in the
The selfduality equations on a Riemann surface
 Proc. Lond. Math. Soc., III. Ser
, 1987
"... In this paper we shall study a special class of solutions of the selfdual YangMills equations. The original selfduality equations which arose in mathematical physics were defined on Euclidean 4space. The physically relevant solutions were the ones with finite action—the socalled 'instanton ..."
Abstract

Cited by 524 (6 self)
 Add to MetaCart
In this paper we shall study a special class of solutions of the selfdual YangMills equations. The original selfduality equations which arose in mathematical physics were defined on Euclidean 4space. The physically relevant solutions were the ones with finite action—the socalled 'instantons'. The same equations may be
New empirical relationships among magnitude, rupture length, rupture width, rupture area, and surface
, 1994
"... Abstract Source parameters for historical earthquakes worldwide are compiled to develop a series of empirical relationships among moment magnitude (M), surface rupture length, subsurface rupture length, downdip rupture width, rupture area, and maximum and average displacement per event. The resultin ..."
Abstract

Cited by 524 (0 self)
 Add to MetaCart
. The resulting data base is a significant update of previous compilations and includes the additional source parameters of seismic moment, moment magnitude, subsurface rupture length, downdip rupture width, and average surface displacement. Each source parameter is classified as reliable or unreliable, based
Surface deformation due to shear and tensile faults in a halfspace
, 1985
"... A complete set of closed analytical expressions is presented in a unified manner for the internal displacements and strains due to shear and tensile faults in a halfspace for both point and finite rectangular sources. These expressions are particularly compact and systematically composed of terms r ..."
Abstract

Cited by 698 (1 self)
 Add to MetaCart
representing deformations in an infinite medium, a term related to surface deformation and that is multiplied by the depth of observation point. Several practical suggestions to avoid mathematical singularities and computational instabilities are also presented. The expressions derived here represent powerful
Text Classification from Labeled and Unlabeled Documents using EM
 MACHINE LEARNING
, 1999
"... This paper shows that the accuracy of learned text classifiers can be improved by augmenting a small number of labeled training documents with a large pool of unlabeled documents. This is important because in many text classification problems obtaining training labels is expensive, while large qua ..."
Abstract

Cited by 1033 (19 self)
 Add to MetaCart
, and probabilistically labels the unlabeled documents. It then trains a new classifier using the labels for all the documents, and iterates to convergence. This basic EM procedure works well when the data conform to the generative assumptions of the model. However these assumptions are often violated in practice
A Survey of Medical Image Registration
, 1998
"... The purpose of this chapter is to present a survey of recent publications concerning medical image registration techniques. These publications will be classified according to a model based on nine salient criteria, the main dichotomy of which is extrinsic versus intrinsic methods The statistics of t ..."
Abstract

Cited by 540 (5 self)
 Add to MetaCart
The purpose of this chapter is to present a survey of recent publications concerning medical image registration techniques. These publications will be classified according to a model based on nine salient criteria, the main dichotomy of which is extrinsic versus intrinsic methods The statistics
Ensemble Methods in Machine Learning
 MULTIPLE CLASSIFIER SYSTEMS, LBCS1857
, 2000
"... Ensemble methods are learning algorithms that construct a set of classifiers and then classify new data points by taking a (weighted) vote of their predictions. The original ensemble method is Bayesian averaging, but more recent algorithms include errorcorrecting output coding, Bagging, and boostin ..."
Abstract

Cited by 607 (3 self)
 Add to MetaCart
Ensemble methods are learning algorithms that construct a set of classifiers and then classify new data points by taking a (weighted) vote of their predictions. The original ensemble method is Bayesian averaging, but more recent algorithms include errorcorrecting output coding, Bagging
Results 1  10
of
247,557