• Documents
  • Authors
  • Tables
  • Log in
  • Sign up
  • MetaCart
  • DMCA
  • Donate

CiteSeerX logo

Tools

Sorted by:
Try your query at:
Semantic Scholar Scholar Academic
Google Bing DBLP
Results 1 - 10 of 960,130
Next 10 →

Locally weighted learning

by Christopher G. Atkeson, Andrew W. Moore , Stefan Schaal - ARTIFICIAL INTELLIGENCE REVIEW , 1997
"... This paper surveys locally weighted learning, a form of lazy learning and memorybased learning, and focuses on locally weighted linear regression. The survey discusses distance functions, smoothing parameters, weighting functions, local model structures, regularization of the estimates and bias, ass ..."
Abstract - Cited by 594 (53 self) - Add to MetaCart
This paper surveys locally weighted learning, a form of lazy learning and memorybased learning, and focuses on locally weighted linear regression. The survey discusses distance functions, smoothing parameters, weighting functions, local model structures, regularization of the estimates and bias

Semi-Supervised Learning Literature Survey

by Xiaojin Zhu , 2006
"... We review the literature on semi-supervised learning, which is an area in machine learning and more generally, artificial intelligence. There has been a whole spectrum of interesting ideas on how to learn from both labeled and unlabeled data, i.e. semi-supervised learning. This document is a chapter ..."
Abstract - Cited by 757 (8 self) - Add to MetaCart
We review the literature on semi-supervised learning, which is an area in machine learning and more generally, artificial intelligence. There has been a whole spectrum of interesting ideas on how to learn from both labeled and unlabeled data, i.e. semi-supervised learning. This document is a

Machine Learning in Automated Text Categorization

by Fabrizio Sebastiani - ACM COMPUTING SURVEYS , 2002
"... The automated categorization (or classification) of texts into predefined categories has witnessed a booming interest in the last ten years, due to the increased availability of documents in digital form and the ensuing need to organize them. In the research community the dominant approach to this p ..."
Abstract - Cited by 1658 (22 self) - Add to MetaCart
to this problem is based on machine learning techniques: a general inductive process automatically builds a classifier by learning, from a set of preclassified documents, the characteristics of the categories. The advantages of this approach over the knowledge engineering approach (consisting in the manual

Boosting a Weak Learning Algorithm By Majority

by Yoav Freund , 1995
"... We present an algorithm for improving the accuracy of algorithms for learning binary concepts. The improvement is achieved by combining a large number of hypotheses, each of which is generated by training the given learning algorithm on a different set of examples. Our algorithm is based on ideas pr ..."
Abstract - Cited by 516 (15 self) - Add to MetaCart
presented by Schapire in his paper "The strength of weak learnability", and represents an improvement over his results. The analysis of our algorithm provides general upper bounds on the resources required for learning in Valiant's polynomial PAC learning framework, which are the best general

The Elements of Statistical Learning -- Data Mining, Inference, and Prediction

by Trevor Hastie, Robert Tibshirani, Jerome Friedman
"... ..."
Abstract - Cited by 1320 (13 self) - Add to MetaCart
Abstract not found

Instance-based learning algorithms

by David W. Aha, Dennis Kibler, Marc K. Albert - Machine Learning , 1991
"... Abstract. Storing and using specific instances improves the performance of several supervised learning algorithms. These include algorithms that learn decision trees, classification rules, and distributed networks. However, no investigation has analyzed algorithms that use only specific instances to ..."
Abstract - Cited by 1359 (18 self) - Add to MetaCart
Abstract. Storing and using specific instances improves the performance of several supervised learning algorithms. These include algorithms that learn decision trees, classification rules, and distributed networks. However, no investigation has analyzed algorithms that use only specific instances

A Meta-Analytic Review of Experiments Examining the Effects of Extrinsic Rewards on Intrinsic Motivation

by Edward L. Deci, Richard Koestner, Richard M. Ryan
"... A meta-analysis of 128 studies examined the effects of extrinsic rewards on intrinsic motivation. As predicted, engagement-contingent, completion-contingent, and performance-contingent rewards signifi-cantly undermined free-choice intrinsic motivation (d =-0.40,-0.36, and-0.28, respectively), as did ..."
Abstract - Cited by 602 (16 self) - Add to MetaCart
this study's methods, analyses, and results differed from the previous ones. By 1971, hundreds of studies within the operant tradition (Skin-ner, 1953) had established that extrinsic rewards can control be-havior. When administered closely subsequent to a behavior, re-wards were reliably found

Learning to rank using gradient descent

by Chris Burges, Tal Shaked, Erin Renshaw, Matt Deeds, Nicole Hamilton, Greg Hullender - In ICML , 2005
"... We investigate using gradient descent methods for learning ranking functions; we propose a simple probabilistic cost function, and we introduce RankNet, an implementation of these ideas using a neural network to model the underlying ranking function. We present test results on toy data and on data f ..."
Abstract - Cited by 510 (17 self) - Add to MetaCart
We investigate using gradient descent methods for learning ranking functions; we propose a simple probabilistic cost function, and we introduce RankNet, an implementation of these ideas using a neural network to model the underlying ranking function. We present test results on toy data and on data

Monitoring the future: National survey results on drug use

by Lloyd D. Johnston, Ph. D, Patrick M. O’malley, Ph. D, Jerald G. Bachman, Ph. D, John E. Schulenberg, Ph. D - I: Secondary school students (NIH Publication No. 05-5726). Bethesda, MD: National Institute on Drug Abuse , 2005
"... by ..."
Abstract - Cited by 502 (21 self) - Add to MetaCart
Abstract not found

Kodaira-Spencer theory of gravity and exact results for quantum string amplitudes

by M. Bershadsky, S. Cecotti, H. Ooguri, C. Vafa - Commun. Math. Phys , 1994
"... We develop techniques to compute higher loop string amplitudes for twisted N = 2 theories with ĉ = 3 (i.e. the critical case). An important ingredient is the discovery of an anomaly at every genus in decoupling of BRST trivial states, captured to all orders by a master anomaly equation. In a particu ..."
Abstract - Cited by 545 (60 self) - Add to MetaCart
We develop techniques to compute higher loop string amplitudes for twisted N = 2 theories with ĉ = 3 (i.e. the critical case). An important ingredient is the discovery of an anomaly at every genus in decoupling of BRST trivial states, captured to all orders by a master anomaly equation. In a
Next 10 →
Results 1 - 10 of 960,130
Powered by: Apache Solr
  • About CiteSeerX
  • Submit and Index Documents
  • Privacy Policy
  • Help
  • Data
  • Source
  • Contact Us

Developed at and hosted by The College of Information Sciences and Technology

© 2007-2019 The Pennsylvania State University