• Documents
  • Authors
  • Tables
  • Log in
  • Sign up
  • MetaCart
  • DMCA
  • Donate

CiteSeerX logo

Advanced Search Include Citations

Tools

Sorted by:
Try your query at:
Semantic Scholar Scholar Academic
Google Bing DBLP
Results 1 - 10 of 32,088
Next 10 →

Combining labeled and unlabeled data with co-training

by Avrim Blum, Tom Mitchell , 1998
"... We consider the problem of using a large unlabeled sample to boost performance of a learning algorithm when only a small set of labeled examples is available. In particular, we consider a setting in which the description of each example can be partitioned into two distinct views, motivated by the ta ..."
Abstract - Cited by 1633 (28 self) - Add to MetaCart
data, but our goal is to use both views together to allow inexpensive unlabeled data to augment amuch smaller set of labeled examples. Speci cally, the presence of two distinct views of each example suggests strategies in which two learning algorithms are trained separately on each view, and then each

Addressing the Curse of Imbalanced Training Sets: One-Sided Selection

by Miroslav Kubat, Stan Matwin - In Proceedings of the Fourteenth International Conference on Machine Learning , 1997
"... Adding examples of the majority class to the training set can have a detrimental effect on the learner's behavior: noisy or otherwise unreliable examples from the majority class can overwhelm the minority class. The paper discusses criteria to evaluate the utility of classifiers induced f ..."
Abstract - Cited by 234 (1 self) - Add to MetaCart
Adding examples of the majority class to the training set can have a detrimental effect on the learner's behavior: noisy or otherwise unreliable examples from the majority class can overwhelm the minority class. The paper discusses criteria to evaluate the utility of classifiers induced

Learning generative visual models from few training examples: an incremental Bayesian approach tested on 101 object categories

by Li Fei-fei , 2004
"... Abstract — Current computational approaches to learning visual object categories require thousands of training images, are slow, cannot learn in an incremental manner and cannot incorporate prior information into the learning process. In addition, no algorithm presented in the literature has been te ..."
Abstract - Cited by 784 (16 self) - Add to MetaCart
are learnt incrementally in a Bayesian manner. Our incremental algorithm is compared experimentally to an earlier batch Bayesian algorithm, as well as to one based on maximum-likelihood. The incremental and batch versions have comparable classification performance on small training sets, but incremental

Sequential minimal optimization: A fast algorithm for training support vector machines

by John C. Platt - Advances in Kernel Methods-Support Vector Learning , 1999
"... This paper proposes a new algorithm for training support vector machines: Sequential Minimal Optimization, or SMO. Training a support vector machine requires the solution of a very large quadratic programming (QP) optimization problem. SMO breaks this large QP problem into a series of smallest possi ..."
Abstract - Cited by 461 (3 self) - Add to MetaCart
possible QP problems. These small QP problems are solved analytically, which avoids using a time-consuming numerical QP optimization as an inner loop. The amount of memory required for SMO is linear in the training set size, which allows SMO to handle very large training sets. Because matrix computation

Support Vector Machine Active Learning with Applications to Text Classification

by Simon Tong , Daphne Koller - JOURNAL OF MACHINE LEARNING RESEARCH , 2001
"... Support vector machines have met with significant success in numerous real-world learning tasks. However, like most machine learning algorithms, they are generally applied using a randomly selected training set classified in advance. In many settings, we also have the option of using pool-based acti ..."
Abstract - Cited by 735 (5 self) - Add to MetaCart
Support vector machines have met with significant success in numerous real-world learning tasks. However, like most machine learning algorithms, they are generally applied using a randomly selected training set classified in advance. In many settings, we also have the option of using pool

Training Support Vector Machines: an Application to Face Detection

by Edgar Osuna, Robert Freund, Federico Girosi , 1997
"... We investigate the application of Support Vector Machines (SVMs) in computer vision. SVM is a learning technique developed by V. Vapnik and his team (AT&T Bell Labs.) that can be seen as a new method for training polynomial, neural network, or Radial Basis Functions classifiers. The decision sur ..."
Abstract - Cited by 727 (1 self) - Add to MetaCart
global optimality, and can be used to train SVM's over very large data sets. The main idea behind the decomposition is the iterative solution of sub-problems and the evaluation of optimality conditions which are used both to generate improved iterative values, and also establish the stopping

On The Size of Training Set and

by The Benefit From, Zhi-hua Zhou, Dan Wei, Gang Li, Honghua Dai - In Proceedings of the 8th Pacific-Asia Conference on Knowledge Discovery and Data Mining (PAKDD’04 , 2004
"... In this paper, the impact of the size of the training set on the benefit from ensemble, i.e. the gains obtained by employing ensemble learning paradigms, is empirically studied. Experiments on Bagged/ Boosted J4.8 decision trees with/without pruning show that enlarging the training set tends to ..."
Abstract - Add to MetaCart
In this paper, the impact of the size of the training set on the benefit from ensemble, i.e. the gains obtained by employing ensemble learning paradigms, is empirically studied. Experiments on Bagged/ Boosted J4.8 decision trees with/without pruning show that enlarging the training set tends

Training Set Construction Methods

by Tomas Borovicka
"... In order to build a classification or regression model, learning algorithms use datasets to set up its parameters and estimate model performance. Training set construction is a part of data preparation. This important phase is often underestimated in data mining process. However, choose the appropri ..."
Abstract - Add to MetaCart
In order to build a classification or regression model, learning algorithms use datasets to set up its parameters and estimate model performance. Training set construction is a part of data preparation. This important phase is often underestimated in data mining process. However, choose

Optimal Brain Damage

by Yann Le Cun, John S. Denker, Sara A. Sola , 1990
"... We have used information-theoretic ideas to derive a class of practical and nearly optimal schemes for adapting the size of a neural network. By removing unimportant weights from a network, several improvements can be expected: better generalization, fewer training examples required, and improved sp ..."
Abstract - Cited by 510 (5 self) - Add to MetaCart
speed of learning and/or classification. The basic idea is to use second-derivative information to make a tradeoff between network complexity and training set error. Experiments confirm the usefulness of the methods on a real-world application.

Wrappers for Feature Subset Selection

by Ron Kohavi, George H. John - AIJ SPECIAL ISSUE ON RELEVANCE , 1997
"... In the feature subset selection problem, a learning algorithm is faced with the problem of selecting a relevant subset of features upon which to focus its attention, while ignoring the rest. To achieve the best possible performance with a particular learning algorithm on a particular training set, a ..."
Abstract - Cited by 1569 (3 self) - Add to MetaCart
In the feature subset selection problem, a learning algorithm is faced with the problem of selecting a relevant subset of features upon which to focus its attention, while ignoring the rest. To achieve the best possible performance with a particular learning algorithm on a particular training set
Next 10 →
Results 1 - 10 of 32,088
Powered by: Apache Solr
  • About CiteSeerX
  • Submit and Index Documents
  • Privacy Policy
  • Help
  • Data
  • Source
  • Contact Us

Developed at and hosted by The College of Information Sciences and Technology

© 2007-2019 The Pennsylvania State University