Results 1 
4 of
4
A Unifying Review of Linear Gaussian Models
, 1999
"... Factor analysis, principal component analysis, mixtures of gaussian clusters, vector quantization, Kalman filter models, and hidden Markov models can all be unified as variations of unsupervised learning under a single basic generative model. This is achieved by collecting together disparate observa ..."
Abstract

Cited by 260 (17 self)
 Add to MetaCart
Factor analysis, principal component analysis, mixtures of gaussian clusters, vector quantization, Kalman filter models, and hidden Markov models can all be unified as variations of unsupervised learning under a single basic generative model. This is achieved by collecting together disparate observations and derivations made by many previous authors and introducing a new way of linking discrete and continuous state models using a simple nonlinearity. Through the use of other nonlinearities, we show how independent component analysis is also a variation of the same basic generative model. We show that factor analysis and mixtures of gaussians can be implemented in autoencoder neural networks and learned using squared error plus the same regularization term. We introduce a new model for static data, known as sensible principal component analysis, as well as a novel concept of spatially adaptive observation noise. We also review some of the literature involving global and local mixtures of the basic models and provide pseudocode for inference and learning for all the basic models.
Observable Operator Models for Discrete Stochastic Time Series
, 1999
"... A widely used class of models for stochastic systems is Hidden Markov models. Systems which can be modeled by hidden Markov models are a proper subclass of linearly dependent processes, a class of stochastic systems known from mathematical investigations carried out over the last four decades. This ..."
Abstract

Cited by 66 (2 self)
 Add to MetaCart
A widely used class of models for stochastic systems is Hidden Markov models. Systems which can be modeled by hidden Markov models are a proper subclass of linearly dependent processes, a class of stochastic systems known from mathematical investigations carried out over the last four decades. This article provides a novel, simple characterization of linearly dependent processes, called observable operator models. The mathematical properties of observable operator models lead to a constructive learning algorithm for the identification of linearly dependent processes. The core of the algorithm has a time complexity of O(N + nmÂ³), where N is the size of training data, n is the number of distinguishable outcomes of observations, and m is model state space dimension.
Risksensitive Generalizations of Minimum Variance Estimation and Control
, 1997
"... We define here risksensitive filtering as minimising the expected value of the exponential of an estimation error (quadratic) cost scaled by a risksensitive parameter. Such filtering is a generalization of standard riskneutral filtering in that as the risksensitive parameter approaches zero, ris ..."
Abstract

Cited by 12 (6 self)
 Add to MetaCart
We define here risksensitive filtering as minimising the expected value of the exponential of an estimation error (quadratic) cost scaled by a risksensitive parameter. Such filtering is a generalization of standard riskneutral filtering in that as the risksensitive parameter approaches zero, riskneutral (minimum error variance) filtering is achieved. Also taking small noise limits, a differential game associated with H1 filtering results. In this paper, the risksensitive nonlinear stochastic filtering problem is studied in both continuous and discretetime for quite general finitedimensional signal models, including also discrete state hidden Markov models (HMMs). The risksensitive estimates are expressed in terms of the socalled information state of the model given by the Zakai equation which is linear. In the linear Gaussian signal model case, the risksensitive (minimum exponential variance) estimates are identical to the minimum variance Kalman filter state estimates, and ...
RISK SENSITIVE GENERALIZATION OF MINIMUM VARIANCE ESTIMATION AND CONTROL *
"... Abstract. In this paper, the risksensitive nonlinear stochastic filtering problem is addressed in both continuous and discretetime for quite general finitedimensional signal models, including also discrete state hidden Markov models (HMMs). The risk sensitive estimates are expressed in terms of t ..."
Abstract
 Add to MetaCart
Abstract. In this paper, the risksensitive nonlinear stochastic filtering problem is addressed in both continuous and discretetime for quite general finitedimensional signal models, including also discrete state hidden Markov models (HMMs). The risk sensitive estimates are expressed in terms of the socalled information state of the model given by the Zakai equation which is linear. In the linear Gaussian signal model case, the risksensitive (minimum exponential variance) estimates are identical to the minimum variance Kalman filter state estimates, and are thus given by a finite dimensional estimator. The estimates are also finite dimensional for discretestate HMMs, but otherwise, in general, are infinite dimensional. In the small noise limit, these estimates (including the minimum variance estimates) have an interpretation in terms of a worst case deterministic noise estimation problem given from a differential game. The related control task, that is the risksensitive generalization of minimumvariance control is studied for the discretetime models. This is motivated by the need for robustness in the widely used (risk neutral) minimum variance control, including adaptive control, of systems which are minimum phase, that is having stable inverses.