Results 1  10
of
22,795
A training algorithm for optimal margin classifiers
 PROCEEDINGS OF THE 5TH ANNUAL ACM WORKSHOP ON COMPUTATIONAL LEARNING THEORY
, 1992
"... A training algorithm that maximizes the margin between the training patterns and the decision boundary is presented. The technique is applicable to a wide variety of classifiaction functions, including Perceptrons, polynomials, and Radial Basis Functions. The effective number of parameters is adjust ..."
Abstract

Cited by 1865 (43 self)
 Add to MetaCart
A training algorithm that maximizes the margin between the training patterns and the decision boundary is presented. The technique is applicable to a wide variety of classifiaction functions, including Perceptrons, polynomials, and Radial Basis Functions. The effective number of parameters
An Improved Training Algorithm for Support Vector Machines
, 1997
"... We investigate the problem of training a Support Vector Machine (SVM) [1, 2, 7] on a very large date base (e.g. 50,000 data points) in the case in which the number of support vectors is also very large (e.g. 40,000). Training a SVM is equivalent to solving a linearly constrained quadratic programmin ..."
Abstract

Cited by 339 (1 self)
 Add to MetaCart
We investigate the problem of training a Support Vector Machine (SVM) [1, 2, 7] on a very large date base (e.g. 50,000 data points) in the case in which the number of support vectors is also very large (e.g. 40,000). Training a SVM is equivalent to solving a linearly constrained quadratic
Training Algorithms for Linear Text Classifiers
, 1996
"... Systems for text retrieval, routing, categorization and other IR tasks rely heavily on linear classifiers. We propose that two machine learning algorithms, the WidrowHoff and EG algorithms, be used in training linear text classifiers. In contrast to most IR methods, theoretical analysis provides pe ..."
Abstract

Cited by 276 (12 self)
 Add to MetaCart
Systems for text retrieval, routing, categorization and other IR tasks rely heavily on linear classifiers. We propose that two machine learning algorithms, the WidrowHoff and EG algorithms, be used in training linear text classifiers. In contrast to most IR methods, theoretical analysis provides
A Sequential Algorithm for Training Text Classifiers
, 1994
"... The ability to cheaply train text classifiers is critical to their use in information retrieval, content analysis, natural language processing, and other tasks involving data which is partly or fully textual. An algorithm for sequential sampling during machine learning of statistical classifiers was ..."
Abstract

Cited by 631 (10 self)
 Add to MetaCart
The ability to cheaply train text classifiers is critical to their use in information retrieval, content analysis, natural language processing, and other tasks involving data which is partly or fully textual. An algorithm for sequential sampling during machine learning of statistical classifiers
Object Detection with Discriminatively Trained Part Based Models
"... We describe an object detection system based on mixtures of multiscale deformable part models. Our system is able to represent highly variable object classes and achieves stateoftheart results in the PASCAL object detection challenges. While deformable part models have become quite popular, their ..."
Abstract

Cited by 1422 (49 self)
 Add to MetaCart
is a reformulation of MISVM in terms of latent variables. A latent SVM is semiconvex and the training problem becomes convex once latent information is specified for the positive examples. This leads to an iterative training algorithm that alternates between fixing latent values for positive examples
Training algorithms
"... acoustic model Acoustic and pronunciation model: Pr(o  w) = ∑ Pr(o  d)Pr(d  c)Pr(c  p)Pr(p  w). d,c,p Pr(o  d) : observation seq. ← distribution seq. Pr(d  c) : distribution seq. ← CD phone seq. Pr(c  p) : CD phone seq. ← phoneme seq. Pr(p  w) : phoneme seq. ← word seq. Language model: seq ..."
Abstract
 Add to MetaCart
acoustic model Acoustic and pronunciation model: Pr(o  w) = ∑ Pr(o  d)Pr(d  c)Pr(c  p)Pr(p  w). d,c,p Pr(o  d) : observation seq. ← distribution seq. Pr(d  c) : distribution seq. ← CD phone seq. Pr(c  p) : CD phone seq. ← phoneme seq. Pr(p  w) : phoneme seq. ← word seq. Language model: seq. Pr(w) , distribution over word
Discriminative Training Methods for Hidden Markov Models: Theory and Experiments with Perceptron Algorithms
, 2002
"... We describe new algorithms for training tagging models, as an alternative to maximumentropy models or conditional random fields (CRFs). The algorithms rely on Viterbi decoding of training examples, combined with simple additive updates. We describe theory justifying the algorithms through a modific ..."
Abstract

Cited by 660 (13 self)
 Add to MetaCart
We describe new algorithms for training tagging models, as an alternative to maximumentropy models or conditional random fields (CRFs). The algorithms rely on Viterbi decoding of training examples, combined with simple additive updates. We describe theory justifying the algorithms through a
Training Linear SVMs in Linear Time
, 2006
"... Linear Support Vector Machines (SVMs) have become one of the most prominent machine learning techniques for highdimensional sparse data commonly encountered in applications like text classification, wordsense disambiguation, and drug design. These applications involve a large number of examples n ..."
Abstract

Cited by 549 (6 self)
 Add to MetaCart
as well as a large number of features N, while each example has only s << N nonzero features. This paper presents a CuttingPlane Algorithm for training linear SVMs that provably has training time O(sn) for classification problems and O(sn log(n)) for ordinal regression problems. The algorithm
Instancebased learning algorithms
 Machine Learning
, 1991
"... Abstract. Storing and using specific instances improves the performance of several supervised learning algorithms. These include algorithms that learn decision trees, classification rules, and distributed networks. However, no investigation has analyzed algorithms that use only specific instances to ..."
Abstract

Cited by 1389 (18 self)
 Add to MetaCart
databases, its performance degrades rapidly with the level of attribute noise in training instances. Therefore, we extended it with a significance test to distinguish noisy instances. This extended algorithm's performance degrades gracefully with increasing noise levels and compares favorably with a
Inductive learning algorithms and representations for text categorization,”
 in Proceedings of the International Conference on Information and Knowledge Management,
, 1998
"... ABSTRACT Text categorization the assignment of natural language texts to one or more predefined categories based on their content is an important component in many information organization and management tasks. We compare the effectiveness of five different automatic learning algorithms for text ..."
Abstract

Cited by 652 (8 self)
 Add to MetaCart
ABSTRACT Text categorization the assignment of natural language texts to one or more predefined categories based on their content is an important component in many information organization and management tasks. We compare the effectiveness of five different automatic learning algorithms for text
Results 1  10
of
22,795