• Documents
  • Authors
  • Tables
  • Log in
  • Sign up
  • MetaCart
  • DMCA
  • Donate

CiteSeerX logo

Tools

Sorted by:
Try your query at:
Semantic Scholar Scholar Academic
Google Bing DBLP
Results 1 - 10 of 2,632
Next 10 →

Dynamic Bayesian Networks: Representation, Inference and Learning

by Kevin Patrick Murphy , 2002
"... Modelling sequential data is important in many areas of science and engineering. Hidden Markov models (HMMs) and Kalman filter models (KFMs) are popular for this because they are simple and flexible. For example, HMMs have been used for speech recognition and bio-sequence analysis, and KFMs have bee ..."
Abstract - Cited by 770 (3 self) - Add to MetaCart
been used for problems ranging from tracking planes and missiles to predicting the economy. However, HMMs and KFMs are limited in their “expressive power”. Dynamic Bayesian Networks (DBNs) generalize HMMs by allowing the state space to be represented in factored form, instead of as a single discrete

Generalization in Reinforcement Learning: Successful Examples Using Sparse Coarse Coding

by Richard S. Sutton - Advances in Neural Information Processing Systems 8 , 1996
"... On large problems, reinforcement learning systems must use parameterized function approximators such as neural networks in order to generalize between similar situations and actions. In these cases there are no strong theoretical results on the accuracy of convergence, and computational results have ..."
Abstract - Cited by 433 (20 self) - Add to MetaCart
the control tasks they attempted, and for one that is significantly larger. The most important differences are that we used sparse-coarse-coded function approximators (CMACs) whereas they used mostly global function approximators, and that we learned online whereas they learned offline. Boyan and Moore

On-Line Q-Learning Using Connectionist Systems

by G. A. Rummery, M. Niranjan , 1994
"... Reinforcement learning algorithms are a powerful machine learning technique. However, much of the work on these algorithms has been developed with regard to discrete finite-state Markovian problems, which is too restrictive for many real-world environments. Therefore, it is desirable to extend these ..."
Abstract - Cited by 381 (1 self) - Add to MetaCart
Reinforcement learning algorithms are a powerful machine learning technique. However, much of the work on these algorithms has been developed with regard to discrete finite-state Markovian problems, which is too restrictive for many real-world environments. Therefore, it is desirable to extend

Online learning for matrix factorization and sparse coding

by Julien Mairal, Francis Bach, Jean Ponce, Guillermo Sapiro , 2010
"... Sparse coding—that is, modelling data vectors as sparse linear combinations of basis elements—is widely used in machine learning, neuroscience, signal processing, and statistics. This paper focuses on the large-scale matrix factorization problem that consists of learning the basis set in order to ad ..."
Abstract - Cited by 330 (31 self) - Add to MetaCart
to adapt it to specific data. Variations of this problem include dictionary learning in signal processing, non-negative matrix factorization and sparse principal component analysis. In this paper, we propose to address these tasks with a new online optimization algorithm, based on stochastic approximations

Adaptive Subgradient Methods for Online Learning and Stochastic Optimization

by John Duchi, Elad Hazan, Yoram Singer , 2010
"... Stochastic subgradient methods are widely used, well analyzed, and constitute effective tools for optimization and online learning. Stochastic gradient methods ’ popularity and appeal are largely due to their simplicity, as they largely follow predetermined procedural schemes. However, most common s ..."
Abstract - Cited by 311 (3 self) - Add to MetaCart
Stochastic subgradient methods are widely used, well analyzed, and constitute effective tools for optimization and online learning. Stochastic gradient methods ’ popularity and appeal are largely due to their simplicity, as they largely follow predetermined procedural schemes. However, most common

An analysis of temporal-difference learning with function approximation

by John N. Tsitsiklis, Benjamin Van Roy - IEEE Transactions on Automatic Control , 1997
"... We discuss the temporal-difference learning algorithm, as applied to approximating the cost-to-go function of an infinite-horizon discounted Markov chain. The algorithm weanalyze updates parameters of a linear function approximator on-line, duringasingle endless trajectory of an irreducible aperiodi ..."
Abstract - Cited by 313 (8 self) - Add to MetaCart
We discuss the temporal-difference learning algorithm, as applied to approximating the cost-to-go function of an infinite-horizon discounted Markov chain. The algorithm weanalyze updates parameters of a linear function approximator on-line, duringasingle endless trajectory of an irreducible

Visual Tracking with Online Multiple Instance Learning

by Boris Babenko, Ming-hsuan Yang, Serge Belongie , 2009
"... In this paper, we address the problem of learning an adaptive appearance model for object tracking. In particular, a class of tracking techniques called “tracking by detection” have been shown to give promising results at realtime speeds. These methods train a discriminative classifier in an online ..."
Abstract - Cited by 261 (19 self) - Add to MetaCart
In this paper, we address the problem of learning an adaptive appearance model for object tracking. In particular, a class of tracking techniques called “tracking by detection” have been shown to give promising results at realtime speeds. These methods train a discriminative classifier in an online

Learning attractor landscapes for learning motor primitives

by Auke Jan Ijspeert, Jun Nakanishi, Stefan Schaal - in Advances in Neural Information Processing Systems , 2003
"... Many control problems take place in continuous state-action spaces, e.g., as in manipulator robotics, where the control objective is often defined as finding a desired trajectory that reaches a particular goal state. While reinforcement learning offers a theoretical framework to learn such control p ..."
Abstract - Cited by 195 (28 self) - Add to MetaCart
Many control problems take place in continuous state-action spaces, e.g., as in manipulator robotics, where the control objective is often defined as finding a desired trajectory that reaches a particular goal state. While reinforcement learning offers a theoretical framework to learn such control

Examining Social Presence in Online Courses in Relation to Students’ Perceived Learning and Satisfaction

by Jennifer C. Richardson, Karen Swan - JOURNAL OF ASYNCHRONOUS LEARNING NETWORKS , 2003
"... Research has demonstrated that social presence not only affects outcomes but also student, and possibly instructor, satisfaction with a course [1]. Teacher immediacy behaviors and the presence of others are especially important issues for those involved in delivering online education. This study exp ..."
Abstract - Cited by 185 (8 self) - Add to MetaCart
explored the role of social presence in online learning environments and its relationship to students’ perceptions of learning and satisfaction with the instructor. The participants for this study were students who completed Empire State College’s (ESC) online learning courses in the spring of 2000

Fast Kernel Classifiers With Online And Active Learning

by Antoine Bordes, Seyda Ertekin, Jason Weston, Léon Bottou - JOURNAL OF MACHINE LEARNING RESEARCH , 2005
"... Very high dimensional learning systems become theoretically possible when training examples are abundant. The computing cost then becomes the limiting factor. Any efficient learning algorithm should at least take a brief look at each example. But should all examples be given equal attention? This ..."
Abstract - Cited by 153 (18 self) - Add to MetaCart
? This contribution proposes an empirical answer. We first present an online SVM algorithm based on this premise. LASVM yields competitive misclassification rates after a single pass over the training examples, outspeeding state-of-the-art SVM solvers. Then we show how active example selection can yield faster
Next 10 →
Results 1 - 10 of 2,632
Powered by: Apache Solr
  • About CiteSeerX
  • Submit and Index Documents
  • Privacy Policy
  • Help
  • Data
  • Source
  • Contact Us

Developed at and hosted by The College of Information Sciences and Technology

© 2007-2019 The Pennsylvania State University