Results 1  10
of
46
Constructive Incremental Learning from Only Local Information
, 1998
"... ... This article illustrates the potential learning capabilities of purely local learning and offers an interesting and powerful approach to learning with receptive fields. ..."
Abstract

Cited by 160 (37 self)
 Add to MetaCart
... This article illustrates the potential learning capabilities of purely local learning and offers an interesting and powerful approach to learning with receptive fields.
Learning Machines
, 1965
"... This book is about machines that learn to discover hidden relationships in data. A constant sfream of data bombards our senses and millions of sensory channels carry information into our brains. Brains are also learning machines that condition, ..."
Abstract

Cited by 150 (0 self)
 Add to MetaCart
This book is about machines that learn to discover hidden relationships in data. A constant sfream of data bombards our senses and millions of sensory channels carry information into our brains. Brains are also learning machines that condition,
Hoeffding Races: Accelerating Model Selection Search for Classification and Function Approximation
 In Advances in neural information processing systems 6
, 1994
"... Selecting a good model of a set of input points by cross validation is a computationally intensive process, especially if the number of possible models or the number of training points is high. Techniques such as gradient descent are helpful in searching through the space of models, but problems suc ..."
Abstract

Cited by 100 (9 self)
 Add to MetaCart
Selecting a good model of a set of input points by cross validation is a computationally intensive process, especially if the number of possible models or the number of training points is high. Techniques such as gradient descent are helpful in searching through the space of models, but problems such as local minima, and more importantly, lack of a distance metric between various models reduce the applicability of these search methods. Hoeffding Races is a technique for finding a good model for the data by quickly discarding bad models, and concentrating the computational effort at differentiating between the better ones. This paper focuses on the special case of leaveoneout cross validation applied to memorybased learning algorithms, but we also argue that it is applicable to any class of model selection problems. 1 Introduction Model selection addresses "high level" decisions about how best to tune learning algorithm architectures for particular tasks. Such decisions include which...
Improving Regression Estimation: Averaging Methods for Variance Reduction with Extensions to General Convex Measure Optimization
, 1993
"... ..."
Prediction risk and architecture selection for neural networks
, 1994
"... Abstract. We describe two important sets of tools for neural network modeling: prediction risk estimation and network architecture selection. Prediction risk is defined as the expected performance of an estimator in predicting new observations. Estimated prediction risk can be used both for estimati ..."
Abstract

Cited by 75 (2 self)
 Add to MetaCart
Abstract. We describe two important sets of tools for neural network modeling: prediction risk estimation and network architecture selection. Prediction risk is defined as the expected performance of an estimator in predicting new observations. Estimated prediction risk can be used both for estimating the quality of model predictions and for model selection. Prediction risk estimation and model selection are especially important for problems with limited data. Techniques for estimating prediction risk include data resampling algorithms such as nonlinear crossâ€“validation (NCV) and algebraic formulae such as the predicted squared error (PSE) and generalized prediction error (GPE). We show that exhaustive search over the space of network architectures is computationally infeasible even for networks of modest size. This motivates the use of heuristic strategies that dramatically reduce the search complexity. These strategies employ directed search algorithms, such as selecting the number of nodes via sequential network construction (SNC) and pruning inputs and weights via sensitivity based pruning (SBP) and optimal brain damage (OBD) respectively.
Adaptive spectral methods for simulation output analysis
 IBM Journal of Research and Development
, 1981
"... This paper addresses two central problems in simulation methodology: the generation of conjidence intervals for the steady state means of the output sequences and the sequential of use these conjidence intervals to control the run length. The variance of the sample mean of a covariance stationary pr ..."
Abstract

Cited by 15 (0 self)
 Add to MetaCart
This paper addresses two central problems in simulation methodology: the generation of conjidence intervals for the steady state means of the output sequences and the sequential of use these conjidence intervals to control the run length. The variance of the sample mean of a covariance stationary process is given approximately by p(O)lN, where p(f) is the spectral density at frequency f and N is the sample size. In earlier an paper we developed a method of confidence interval generation based on the estimation of p(0) through the least squares jit of a quadratic to the logarithm of the periodogram. This method was applied in a run length Control procedure to a sequence of batched means. As the run length increased the batch means were rebatched into larger batch sizes so as to limit storage requirements. In this rebatching the shape of the spectral density changes, gradually becoming flat as N increases. Quadratics were chosen as a compromise between small sample bias and large sample stability. In this paper we consider smoothing techniques which adapt to the changing spectral shape in an attempt to improve both the small and large sample behavior of the method. The techniques considered are polynomial smoothing with the degree selected sequentially using standard regression statistics, polynomial smoothing with the degree selected by cross validation, and smoothing splines with the amount of smoothing determined by cross validation. These techniques were empirically evaluated both for fixed sample sizes and when incorporated into the sequential run length control procedure. Forjixed sample sizes they did not improve the small sample behavior and only marginally improved the large sample behavior when compared with the quadratic method. Their performance in the sequential procedure was unsatisfactory. Hence, the straightforward quadratic technique recommended in the earlier paper is still recommended as an effective, practical technique for simulation conjidence interval generation and run length control. 1.
Objectbased 3D reconstruction of arterial trees from magnetic resonance angiograms
 IEEE Transactions on Medical Imaging
, 1991
"... This paper describes an objectbased approach to the problem of reconstructing threedimensional descriptions of arterial trees from a few angiographic projections. The method incorporates aprioriknowledge of the structure of branching arteries into a natural optimality criterion that encompasses ..."
Abstract

Cited by 15 (3 self)
 Add to MetaCart
This paper describes an objectbased approach to the problem of reconstructing threedimensional descriptions of arterial trees from a few angiographic projections. The method incorporates aprioriknowledge of the structure of branching arteries into a natural optimality criterion that encompasses the entire arterial tree. This global approach enables reconstruction from a few noisy projection images. We present an efficient optimization algorithm for object estimation, and demonstrate its performance on simulated, phantom, and in vivo magnetic resonance angiograms.
A Survey on Spherical Spline Approximation
 Surveys Math. Indust
, 1997
"... Spline functions that approximate data given on the sphere are developed in a weighted Sobolev space setting. The flexibility of the weights makes possible the choice of the approximating function in a way which emphasizes attributes desirable for the particular application area. Examples show that ..."
Abstract

Cited by 14 (1 self)
 Add to MetaCart
Spline functions that approximate data given on the sphere are developed in a weighted Sobolev space setting. The flexibility of the weights makes possible the choice of the approximating function in a way which emphasizes attributes desirable for the particular application area. Examples show that certain choices of the weight sequences yield known methods. A convergence theorem containing explicit constants yields a usable error bound. Our survey ends with the discussion of spherical splines in geodetically relevant pseudodifferential equations. (submitted to "Surveys on Mathematics for Industry") AMS classification: 41A05, 43A90, 65D07, 86A30 Keywords: spherical splines, scattered data interpolation, smoothing, geoid determination Contents 1 Introduction 3 2 Preliminaries 4 3 Sobolev Spaces and Pseudodifferential Operators 7 4 Sobolev Lemma and Reproducing Kernel Sobolev Spaces 10 5 Examples of Radial Basis Functions 13 5.1 Green's Kernels Corresponding to Iterated Beltrami Oper...
Robust Full Bayesian Learning for Neural Networks
, 1999
"... In this paper, we propose a hierarchical full Bayesian model for neural networks. This model treats the model dimension (number of neurons), model parameters, regularisation parameters and noise parameters as random variables that need to be estimated. We develop a reversible jump Markov chain Monte ..."
Abstract

Cited by 12 (9 self)
 Add to MetaCart
In this paper, we propose a hierarchical full Bayesian model for neural networks. This model treats the model dimension (number of neurons), model parameters, regularisation parameters and noise parameters as random variables that need to be estimated. We develop a reversible jump Markov chain Monte Carlo (MCMC) method to perform the necessary computations. We find that the results obtained using this method are not only better than the ones reported previously, but also appear to be robust with respect to the prior specification. In addition, we propose a novel and computationally efficient reversible jump MCMC simulated annealing algorithm to optimise neural networks. This algorithm enables us to maximise the joint posterior distribution of the network parameters and the number of basis function. It performs a global search in the joint space of the parameters and number of parameters, thereby surmounting the problem of local minima. We show that by calibrating the full hierarchical ...