• Documents
  • Authors
  • Tables
  • Log in
  • Sign up
  • MetaCart
  • DMCA
  • Donate

CiteSeerX logo

Advanced Search Include Citations

Tools

Sorted by:
Try your query at:
Semantic Scholar Scholar Academic
Google Bing DBLP
Results 1 - 10 of 2,068
Next 10 →

Margin maximizing loss functions

by Saharon Rosset, Ji Zhu, Trevor Hastie - In NIPS , 2004
"... Margin maximizing properties play an important role in the analysis of classification models, such as boosting and support vector machines. Margin maximization is theoretically interesting because it facilitates generalization error analysis, and practically interesting because it presents a clear g ..."
Abstract - Cited by 26 (4 self) - Add to MetaCart
Margin maximizing properties play an important role in the analysis of classification models, such as boosting and support vector machines. Margin maximization is theoretically interesting because it facilitates generalization error analysis, and practically interesting because it presents a clear

Margin maximizing discriminant analysis

by András Kocsor, Kornél Kovács, Csaba Szepesvári - In Proceedings of the 15th European Conference on Machine Learning , 2004
"... Abstract. We propose a new feature extraction method called Margin Maximizing Discriminant Analysis (MMDA) which seeks to extract features suitable for classification tasks. MMDA is based on the principle that an ideal feature should convey the maximum information about the class labels and it shoul ..."
Abstract - Cited by 10 (2 self) - Add to MetaCart
Abstract. We propose a new feature extraction method called Margin Maximizing Discriminant Analysis (MMDA) which seeks to extract features suitable for classification tasks. MMDA is based on the principle that an ideal feature should convey the maximum information about the class labels

Efficient Margin Maximizing with Boosting

by Gunnar Rätsch, Manfred K. Warmuth , 2003
"... AdaBoost produces a linear combination of base hypotheses and predicts with the sign of this linear combination. It has been observed that the generalization error of the algorithm continues to improve even after all examples are classified correctly by the current signed linear combination, whic ..."
Abstract - Cited by 50 (7 self) - Add to MetaCart
AdaBoost produces a linear combination of base hypotheses and predicts with the sign of this linear combination. It has been observed that the generalization error of the algorithm continues to improve even after all examples are classified correctly by the current signed linear combination, which can be viewed as hyperplane in feature space where the base hypotheses form the features.

MARGIN: Maximal Frequent Subgraph Mining

by Lini T Thomas, Satyanarayana R Valluri, Kamalakar Karlapalem
"... The exponential number of possible subgraphs makes the problem of frequent subgraph mining a challenge. Maximal frequent mining has triggered much interest since the size of the set of maximal frequent subgraphs is much smaller to that of the set of frequent subgraphs. We propose an algorithm that m ..."
Abstract - Cited by 23 (1 self) - Add to MetaCart
The exponential number of possible subgraphs makes the problem of frequent subgraph mining a challenge. Maximal frequent mining has triggered much interest since the size of the set of maximal frequent subgraphs is much smaller to that of the set of frequent subgraphs. We propose an algorithm

ENTROPY REGULARIZATION AND SOFT MARGIN Maximization

by Karen A. Glocer , 2009
"... ..."
Abstract - Cited by 1 (0 self) - Add to MetaCart
Abstract not found

Online Relative Margin Maximization for Statistical Machine Translation

by Vladimir Eidelman, Yuval Marton, Philip Resnik
"... Recent advances in large-margin learning have shown that better generalization can be achieved by incorporating higher order information into the optimization, such as the spread of the data. However, these solutions are impractical in complex structured prediction problems such as statistical machi ..."
Abstract - Cited by 2 (0 self) - Add to MetaCart
machine translation. We present an online gradient-based algorithm for relative margin maximization, which bounds the spread of the projected data while maximizing the margin. We evaluate our optimizer on Chinese-English and Arabic-English translation tasks, each with small and large feature sets

Max-margin Markov networks

by Ben Taskar, Carlos Guestrin, Daphne Koller , 2003
"... In typical classification tasks, we seek a function which assigns a label to a single object. Kernel-based approaches, such as support vector machines (SVMs), which maximize the margin of confidence of the classifier, are the method of choice for many such tasks. Their popularity stems both from the ..."
Abstract - Cited by 604 (15 self) - Add to MetaCart
In typical classification tasks, we seek a function which assigns a label to a single object. Kernel-based approaches, such as support vector machines (SVMs), which maximize the margin of confidence of the classifier, are the method of choice for many such tasks. Their popularity stems both from

Entropy and margin maximization for structured output learning

by Patrick Pletscher, Cheng Soon Ong, Joachim M. Buhmann - In Proceedings of the 2010 European conference on Machine learning and knowledge discovery in databases: Part III, ECML PKDD’10 , 2010
"... Abstract. We consider the problem of training discriminative struc-tured output predictors, such as conditional random fields (CRFs) and structured support vector machines (SSVMs). A generalized loss func-tion is introduced, which jointly maximizes the entropy and the margin of the solution. The CRF ..."
Abstract - Cited by 9 (1 self) - Add to MetaCart
Abstract. We consider the problem of training discriminative struc-tured output predictors, such as conditional random fields (CRFs) and structured support vector machines (SSVMs). A generalized loss func-tion is introduced, which jointly maximizes the entropy and the margin of the solution

Large Margin Classification Using the Perceptron Algorithm

by Yoav Freund, Robert E. Schapire - Machine Learning , 1998
"... We introduce and analyze a new algorithm for linear classification which combines Rosenblatt 's perceptron algorithm with Helmbold and Warmuth's leave-one-out method. Like Vapnik 's maximal-margin classifier, our algorithm takes advantage of data that are linearly separable with large ..."
Abstract - Cited by 521 (2 self) - Add to MetaCart
We introduce and analyze a new algorithm for linear classification which combines Rosenblatt 's perceptron algorithm with Helmbold and Warmuth's leave-one-out method. Like Vapnik 's maximal-margin classifier, our algorithm takes advantage of data that are linearly separable

A training algorithm for optimal margin classifiers

by Bernhard E. Boser, et al. - PROCEEDINGS OF THE 5TH ANNUAL ACM WORKSHOP ON COMPUTATIONAL LEARNING THEORY , 1992
"... A training algorithm that maximizes the margin between the training patterns and the decision boundary is presented. The technique is applicable to a wide variety of classifiaction functions, including Perceptrons, polynomials, and Radial Basis Functions. The effective number of parameters is adjust ..."
Abstract - Cited by 1865 (43 self) - Add to MetaCart
A training algorithm that maximizes the margin between the training patterns and the decision boundary is presented. The technique is applicable to a wide variety of classifiaction functions, including Perceptrons, polynomials, and Radial Basis Functions. The effective number of parameters
Next 10 →
Results 1 - 10 of 2,068
Powered by: Apache Solr
  • About CiteSeerX
  • Submit and Index Documents
  • Privacy Policy
  • Help
  • Data
  • Source
  • Contact Us

Developed at and hosted by The College of Information Sciences and Technology

© 2007-2019 The Pennsylvania State University