• Documents
  • Authors
  • Tables
  • Log in
  • Sign up
  • MetaCart
  • DMCA
  • Donate

CiteSeerX logo

Advanced Search Include Citations
Advanced Search Include Citations | Disambiguate

Combined genetic algorithm optimization and regularized orthogonal least squares learning for radial basis function networks (1999)

by S Chen, Y Wu, B Luk
Venue:IEEE Trans. Neural Networks
Add To MetaCart

Tools

Sorted by:
Results 1 - 10 of 40
Next 10 →

Sparse modelling using orthogonal forward regression with press statistic and regularization

by Sheng Chen, Xia Hong, Chris J. Harris, Paul M. Sharkey - IEEE TRANS. SYSTEMS, MAN AND CYBERNETICS, PART B , 2004
"... The paper introduces an efficient construction algorithm for obtaining sparse linear-in-the-weights regression models based on an approach of directly optimizing model generalization capability. This is achieved by utilizing the delete-1 cross validation concept and the associated leave-one-out tes ..."
Abstract - Cited by 49 (23 self) - Add to MetaCart
The paper introduces an efficient construction algorithm for obtaining sparse linear-in-the-weights regression models based on an approach of directly optimizing model generalization capability. This is achieved by utilizing the delete-1 cross validation concept and the associated leave-one-out test error also known as the predicted residual sums of squares (PRESS) statistic, without resorting to any other validation data set for model evaluation in the model construction process. Computational efficiency is ensured using an orthogonal forward regression, but the algorithm incrementally minimizes the PRESS statistic instead of the usual sum of the squared training errors. A local regularization method can naturally be incorporated into the model selection procedure to further enforce model sparsity. The proposed algorithm is fully automatic, and the user is not required to specify any criterion to terminate the model construction procedure. Comparisons with some of the existing state-of-art modeling methods are given, and several examples are included to demonstrate the ability of the proposed algorithm to effectively construct sparse models that generalize well.

Probabilistic neural-network structure determination for pattern classi

by K. Z. Mao, K. -c. Tan, W. Ser - IEEE Transactions on Neural Networks , 2000
"... Abstract—Network structure determination is an important issue in pattern classification based on a probabilistic neural network. In this study, a supervised network structure determi-nation algorithm is proposed. The proposed algorithm consists of two parts and runs in an iterative way. The first p ..."
Abstract - Cited by 26 (0 self) - Add to MetaCart
Abstract—Network structure determination is an important issue in pattern classification based on a probabilistic neural network. In this study, a supervised network structure determi-nation algorithm is proposed. The proposed algorithm consists of two parts and runs in an iterative way. The first part identifies an appropriate smoothing parameter using a genetic algorithm, while the second part determines suitable pattern layer neurons using a forward regression orthogonal algorithm. The proposed algorithm is capable of offering a fairly small network structure with satisfactory classification accuracy. Index Terms—Genetic algorithms, orthogonal algorithm, pattern classification, probabilistic neural network (PNN). I.
(Show Context)

Citation Context

...e second part performs pattern layer neurons selection using an orthogonal algorithm [4]. Recently, an algorithm with similar architecture for RBF neural networks structure determination was proposed =-=[5]-=-. However, we developed our algorithm independently, and submitted our work for review before the publication of [5]. The difference bewteen our algorithm and that in [5] is discussed in Section III-A...

A kernel-based two-class classifier for imbalanced data sets

by X Hong, S Chen, Harris CJ - IEEE Trans Neural Netw. 2007; 18: 28–41. PMID: 17278459
"... ..."
Abstract - Cited by 23 (4 self) - Add to MetaCart
Abstract not found

Evolutionary optimization of radial basis function classifiers for data mining applications

by Oliver Buchtala, Manuel Klimek, Bernhard Sick - IEEE Transactions on Systems Man and Cybernetics Part B-Cybernetics , 2005
"... Abstract—In many data mining applications that address classification problems, feature and model selection are considered as key tasks. That is, appropriate input features of the classifier must be selected from a given (and often large) set of possible features and structure parameters of the clas ..."
Abstract - Cited by 16 (0 self) - Add to MetaCart
Abstract—In many data mining applications that address classification problems, feature and model selection are considered as key tasks. That is, appropriate input features of the classifier must be selected from a given (and often large) set of possible features and structure parameters of the classifier must be adapted with respect to these features and a given data set. This paper describes an evolutionary algorithm (EA) that performs feature and model selection simultaneously for radial basis function (RBF) classifiers. In order to reduce the optimization effort, various techniques are integrated that accelerate and improve the EA significantly: hybrid training of RBF networks, lazy evaluation, consideration of soft constraints by means of penalty terms, and temperature-based adaptive control of the EA. The feasibility and the benefits of the approach are demonstrated by means of four data mining problems: intrusion detection in computer networks, biometric signature verification, customer acquisition with direct marketing methods, and optimization of chemical production processes. It is shown that, compared to earlier EA-based RBF optimization techniques, the runtime is reduced by up to 99% while error rates are lowered by up to 86%, depending on the application. The algorithm is independent of specific applications so that many ideas and solutions can be transferred to other classifier paradigms. Index Terms—Data mining, evolutionary algorithm (EA), feature selection, model selection, radial basis function (RBF) network. I.
(Show Context)

Citation Context

... [71], [73], [75], and [77], for instance. Other architecture parameters are the type of basis functions [57], [58], the training time (epoch number) [36], and parameters of training algorithms [26], =-=[27]-=-, [36], [63]. Examples of such parameters are learning rate and momentum or a regularization parameter used for training with regularized orthogonal least squares training. 3) Feature selection for RB...

Neuroevolutionary reinforcement learning for generalized helicopter control

by Shimon Whiteson , 2009
"... Helicopter hovering is an important challenge problem in the field of reinforcement learning. This paper considers several neuroevolutionary approaches to discovering robust controllers for a generalized version of the problem used in the 2008 Reinforcement Learning Competition, in which wind in the ..."
Abstract - Cited by 11 (2 self) - Add to MetaCart
Helicopter hovering is an important challenge problem in the field of reinforcement learning. This paper considers several neuroevolutionary approaches to discovering robust controllers for a generalized version of the problem used in the 2008 Reinforcement Learning Competition, in which wind in the helicopter’s environment varies from run to run. We present the simple model-free strategy that won first place in the competition and also describe several more complex model-based approaches. Our empirical results demonstrate that neuroevolution is effective at optimizing the weights of multi-layer perceptrons, that linear regression is faster and more effective than evolution for learning models, and that model-based approaches can outperform the simple modelfree strategy, especially if prior knowledge is used to aid model learning.

Particle swarm optimization aided orthogonal forward regression for unified data modelling

by Sheng Chen, Xia Hong, Chris J. Harris - IEEE TRANS. EVOLUTION. COMPUT , 2010
"... We propose a unified data modeling approach that is equally applicable to supervised regression and classification applications, as well as to unsupervised probability density function estimation. A particle swarm optimization (PSO) aided orthogonal forward regression (OFR) algorithm based on leave ..."
Abstract - Cited by 8 (4 self) - Add to MetaCart
We propose a unified data modeling approach that is equally applicable to supervised regression and classification applications, as well as to unsupervised probability density function estimation. A particle swarm optimization (PSO) aided orthogonal forward regression (OFR) algorithm based on leave-one-out (LOO) criteria is developed to construct parsimonious radial basis function (RBF) networks with tunable nodes. Each stage of the construction process determines the center vector and diagonal covariance matrix of one RBF node by minimizing the LOO statistics. For regression applications, the LOO criterion is chosen to be the LOO mean square error, while the LOO misclassification rate is adopted in two-class classification appli-cations. By adopting the Parzen window estimate as the desired response, the unsupervised density estimation problem is trans-formed into a constrained regression problem. This PSO aided OFR algorithm for tunable-node RBF networks is capable of constructing very parsimonious RBF models that generalize well, and our analysis and experimental results demonstrate that the algorithm is computationally even simpler than the efficient regularization assisted orthogonal least square algorithm based on LOO criteria for selecting fixed-node RBF models. Another significant advantage of the proposed learning procedure is that it does not have learning hyperparameters that have to be tuned using costly cross validation. The effectiveness of the proposed PSO aided OFR construction procedure is illustrated using several examples taken from regression and classification, as well as density estimation applications.

Construction of Tunable Radial Basis Function Networks Using Orthogonal Forward Selection

by Sheng Chen, Xia Hong, Bing L. Luk, Chris J. Harris
"... Abstract—An orthogonal forward selection (OFS) algorithm based on leave-one-out (LOO) criteria is proposed for the construction of radial basis function (RBF) networks with tunable nodes. Each stage of the construction process determines an RBF node, namely, its center vector and diagonal covariance ..."
Abstract - Cited by 6 (4 self) - Add to MetaCart
Abstract—An orthogonal forward selection (OFS) algorithm based on leave-one-out (LOO) criteria is proposed for the construction of radial basis function (RBF) networks with tunable nodes. Each stage of the construction process determines an RBF node, namely, its center vector and diagonal covariance matrix, by minimizing the LOO statistics. For regression application, the LOO criterion is chosen to be the LOO mean-square error, while the LOO misclassification rate is adopted in two-class classification application. This OFS-LOO algorithm is computationally efficient, and it is capable of constructing parsimonious RBF networks that generalize well. Moreover, the proposed algorithm is fully automatic, and the user does not need to specify a termination criterion for the construction process. The effectiveness of the proposed RBF network construction procedure is demonstrated using examples taken from both regression and classification applications. Index Terms—Classification, leave-one-out (LOO) statistics, orthogonal forward selection (OFS), radial basis function (RBF) network, regression, tunable node. I.

Modeling systems with internal state using Evolino

by Daan Wierstra, Faustino J. Gomez, Jürgen Schmidhuber - In Proc. of the 2005 conference on genetic and evolutionary computation (GECCO , 2005
"... Existing Recurrent Neural Networks (RNNs) are limited in their ability to model dynamical systems with nonlinearities and hidden internal states. Here we use our general framework for sequence learning, EVOlution of recurrent systems with LINear Outputs (Evolino), to discover good RNN hidden node we ..."
Abstract - Cited by 6 (1 self) - Add to MetaCart
Existing Recurrent Neural Networks (RNNs) are limited in their ability to model dynamical systems with nonlinearities and hidden internal states. Here we use our general framework for sequence learning, EVOlution of recurrent systems with LINear Outputs (Evolino), to discover good RNN hidden node weights through evolution, while using linear regression to compute an optimal linear mapping from hidden state to output. Using the Long Short-Term Memory RNN Architecture, Evolino outperforms previous state-of-the-art methods on several tasks: 1) context-sensitive languages, 2) multiple superimposed sine waves. Categories and Subject Descriptors I.2.6 [Artificial Intelligence]: Learning—Connectionism and neural nets
(Show Context)

Citation Context

...edictors. That is, predictors that can detect temporal dependencies in the input data that span an arbitrary number of time steps. The MackeyGlass time-series that is often used to test these methods =-=[21, 25, 3, 5, 13]-=-, for instance, can be predicted very accurately using a feedforward network with a relatively short time-delay window on the input. Recurrent Neural Networks (RNNs; [20, 18, 23]) can potentially impl...

Robust nonlinear model identification method forward regression

by X. Hong, C. J. Harris, S. Chen, P. M. Sharkey - IEEE Transactions on Systems, Man and Cybernetics, Part A , 2003
"... Abstract—In this correspondence new robust nonlinear model con-struction algorithms for a large class of linear-in-the-parameters models are introduced to enhance model robustness via combined parameter regularization and new robust structural selective criteria. In parallel to parameter regularizat ..."
Abstract - Cited by 6 (4 self) - Add to MetaCart
Abstract—In this correspondence new robust nonlinear model con-struction algorithms for a large class of linear-in-the-parameters models are introduced to enhance model robustness via combined parameter regularization and new robust structural selective criteria. In parallel to parameter regularization, we use two classes of robust model selection criteria based on either experimental design criteria that optimizes model adequacy, or the predicted residual sums of squares (PRESS) statistic that optimizes model generalization capability, respectively. Three robust identification algorithms are introduced, i.e., combined A- and D-opti-mality with regularized orthogonal least squares algorithm, respectively; and combined PRESS statistic with regularized orthogonal least squares algorithm. A common characteristic of these algorithms is that the inherent computation efficiency associated with the orthogonalization scheme in orthogonal least squares or regularized orthogonal least squares has been extended such that the new algorithms are computationally efficient. Numerical examples are included to demonstrate effectiveness of the algorithms. Index Terms—Cross validation, experimental design, forward regres-sion, generalization, structure identification. I.

M-Estimator and D-Optimality Model Construction Using Orthogonal Forward Regression

by Xia Hong, Senior Member, Sheng Chen, Senior Member
"... Abstract—This correspondence introduces a new orthogonal forward regression (OFR) model identification algorithm using D-optimality for model structure selection and is based on an M-estimators of parameter estimates. M-estimator is a classical robust parameter estimation technique to tackle bad dat ..."
Abstract - Cited by 4 (1 self) - Add to MetaCart
Abstract—This correspondence introduces a new orthogonal forward regression (OFR) model identification algorithm using D-optimality for model structure selection and is based on an M-estimators of parameter estimates. M-estimator is a classical robust parameter estimation technique to tackle bad data conditions such as outliers. Computationally, The M-estimator can be derived using an iterative reweighted least squares (IRLS) algorithm. D-optimality is a model structure robustness criterion in experimental design to tackle ill-conditioning in model structure. The orthogonal forward regression (OFR), often based on the modified Gram–Schmidt procedure, is an efficient method incorporating structure selection and parameter estimation simultaneously. The basic idea of the proposed approach is to incorporate an IRLS inner loop into the modified Gram–Schmidt procedure. In this manner, the OFR algorithm for parsimonious model structure determination is extended to bad data conditions with improved performance via the derivation of parameter M-estimators with inherent robustness to outliers. Numerical examples are included to demonstrate the effectiveness of the proposed algorithm. Index Terms—Forward regression, Gram–Schmidt, identification, M-estimator, model structure selection. I.
(Show Context)

Citation Context

...by requirements for improved model generalization, a few variants of OFR have been introduced in order to tackle ill-conditioning problem that may be associated with least squares parameter estimates =-=[7]-=-–[11]. Recently, variants of the forward OFR algorithms have been introduced by modifying the selective criteria to include A- and D-optimality in forward regression [12], [13] to form hybrid approach...

Powered by: Apache Solr
  • About CiteSeerX
  • Submit and Index Documents
  • Privacy Policy
  • Help
  • Data
  • Source
  • Contact Us

Developed at and hosted by The College of Information Sciences and Technology

© 2007-2019 The Pennsylvania State University