Results 1  10
of
7,387,597
The Nature of Statistical Learning Theory
, 1999
"... Statistical learning theory was introduced in the late 1960’s. Until the 1990’s it was a purely theoretical analysis of the problem of function estimation from a given collection of data. In the middle of the 1990’s new types of learning algorithms (called support vector machines) based on the deve ..."
Abstract

Cited by 12976 (32 self)
 Add to MetaCart
Statistical learning theory was introduced in the late 1960’s. Until the 1990’s it was a purely theoretical analysis of the problem of function estimation from a given collection of data. In the middle of the 1990’s new types of learning algorithms (called support vector machines) based
Maximum likelihood from incomplete data via the EM algorithm
 JOURNAL OF THE ROYAL STATISTICAL SOCIETY, SERIES B
, 1977
"... A broadly applicable algorithm for computing maximum likelihood estimates from incomplete data is presented at various levels of generality. Theory showing the monotone behaviour of the likelihood and convergence of the algorithm is derived. Many examples are sketched, including missing value situat ..."
Abstract

Cited by 11807 (17 self)
 Add to MetaCart
situations, applications to grouped, censored or truncated data, finite mixture models, variance component estimation, hyperparameter estimation, iteratively reweighted least squares and factor analysis.
Is imitation learning the route to humanoid robots?
 TRENDS IN COGNITIVE SCIENCES
, 1999
"... This review investigates two recent developments in artificial intelligence and neural computation: learning from imitation and the development of humanoid robots. It is
postulated that the study of imitation learning offers a promising route to gain new insights into mechanisms of perceptual motor ..."
Abstract

Cited by 306 (18 self)
 Add to MetaCart
This review investigates two recent developments in artificial intelligence and neural computation: learning from imitation and the development of humanoid robots. It is
postulated that the study of imitation learning offers a promising route to gain new insights into mechanisms of perceptual motor
SemiSupervised Learning Literature Survey
, 2006
"... We review the literature on semisupervised learning, which is an area in machine learning and more generally, artificial intelligence. There has been a whole
spectrum of interesting ideas on how to learn from both labeled and unlabeled data, i.e. semisupervised learning. This document is a chapter ..."
Abstract

Cited by 757 (8 self)
 Add to MetaCart
We review the literature on semisupervised learning, which is an area in machine learning and more generally, artificial intelligence. There has been a whole
spectrum of interesting ideas on how to learn from both labeled and unlabeled data, i.e. semisupervised learning. This document is a
What Can Economists Learn from Happiness Research?
 FORTHCOMING IN JOURNAL OF ECONOMIC LITERATURE
, 2002
"... Happiness is generally considered to be an ultimate goal in life; virtually everybody wants to be happy. The United States Declaration of Independence of 1776 takes it as a selfevident truth that the “pursuit of happiness” is an “unalienable right”, comparable to life and liberty. It follows that e ..."
Abstract

Cited by 517 (24 self)
 Add to MetaCart
Happiness is generally considered to be an ultimate goal in life; virtually everybody wants to be happy. The United States Declaration of Independence of 1776 takes it as a selfevident truth that the “pursuit of happiness” is an “unalienable right”, comparable to life and liberty. It follows that economics is – or should be – about individual happiness. In particular, the question is how do economic growth, unemployment and inflation, as well as institutional factors such as good governance, affect individual wellbeing? In addition to this intrinsic interest, there are three major reasons for economists to consider happiness. The first is economic policy. At the microlevel, it is often impossible to make a Paretooptimal proposal, because a social action entails costs for some individuals. Hence an evaluation of the net effects, in terms of individual utilities, is needed. On an aggregate level, economic policy must deal with tradeoffs, especially those between unemployment and
Learning the Kernel Matrix with SemiDefinite Programming
, 2002
"... Kernelbased learning algorithms work by embedding the data into a Euclidean space, and then searching for linear relations among the embedded data points. The embedding is performed implicitly, by specifying the inner products between each pair of points in the embedding space. This information ..."
Abstract

Cited by 780 (22 self)
 Add to MetaCart
problems in machine learning. In this paper we show how the kernel matrix can be learned from data via semidefinite programming (SDP) techniques. When applied
Reinforcement Learning I: Introduction
, 1998
"... In which we try to give a basic intuitive sense of what reinforcement learning is and how it differs and relates to other fields, e.g., supervised learning and neural networks, genetic algorithms and artificial life, control theory. Intuitively, RL is trial and error (variation and selection, search ..."
Abstract

Cited by 5500 (120 self)
 Add to MetaCart
In which we try to give a basic intuitive sense of what reinforcement learning is and how it differs and relates to other fields, e.g., supervised learning and neural networks, genetic algorithms and artificial life, control theory. Intuitively, RL is trial and error (variation and selection
Ensemble Methods in Machine Learning
 MULTIPLE CLASSIFIER SYSTEMS, LBCS1857
, 2000
"... Ensemble methods are learning algorithms that construct a set of classifiers and then classify new data points by taking a (weighted) vote of their predictions. The original ensemble method is Bayesian averaging, but more recent algorithms include errorcorrecting output coding, Bagging, and boostin ..."
Abstract

Cited by 607 (3 self)
 Add to MetaCart
Ensemble methods are learning algorithms that construct a set of classifiers and then classify new data points by taking a (weighted) vote of their predictions. The original ensemble method is Bayesian averaging, but more recent algorithms include errorcorrecting output coding, Bagging
Probabilistic Visual Learning for Object Representation
, 1996
"... We present an unsupervised technique for visual learning which is based on density estimation in highdimensional spaces using an eigenspace decomposition. Two types of density estimates are derived for modeling the training data: a multivariate Gaussian (for unimodal distributions) and a Mixtureof ..."
Abstract

Cited by 705 (15 self)
 Add to MetaCart
We present an unsupervised technique for visual learning which is based on density estimation in highdimensional spaces using an eigenspace decomposition. Two types of density estimates are derived for modeling the training data: a multivariate Gaussian (for unimodal distributions) and a Mixture
Results 1  10
of
7,387,597