Results 1 - 10
of
1,535
Divergence measures based on the Shannon entropy
- IEEE Transactions on Information theory
, 1991
"... Abstract-A new class of information-theoretic divergence measures based on the Shannon entropy is introduced. Unlike the well-known Kullback divergences, the new measures do not require the condition of absolute continuity to be satisfied by the probability distributions in-volved. More importantly, ..."
Abstract
-
Cited by 666 (0 self)
- Add to MetaCart
Abstract-A new class of information-theoretic divergence measures based on the Shannon entropy is introduced. Unlike the well-known Kullback divergences, the new measures do not require the condition of absolute continuity to be satisfied by the probability distributions in-volved. More importantly
PAC-Bayesian Learning and Domain Adaptation
"... In machine learning, Domain Adaptation (DA) arises when the distribution gen-erating the test (target) data differs from the one generating the learning (source) data. It is well known that DA is an hard task even under strong assumptions [1], among which the covariate-shift where the source and tar ..."
Abstract
- Add to MetaCart
based on the minimization of the proposed bound. Doing so, we seek for a ρ-weighted majority vote that takes into account a trade-off between three quantities. The first two quantities being, as usual in the PAC-Bayesian ap-
Simplified PAC-Bayesian margin bounds
- In COLT
, 2003
"... Abstract. The theoretical understanding of support vector machines is largely based on margin bounds for linear classifiers with unit-norm weight vectors and unit-norm feature vectors. Unit-norm margin bounds have been proved previously using fat-shattering arguments and Rademacher complexity. Recen ..."
Abstract
-
Cited by 64 (3 self)
- Add to MetaCart
. Recently Langford and Shawe-Taylor proved a dimensionindependent unit-norm margin bound using a relatively simple PAC-Bayesian argument. Unfortunately, the Langford-Shawe-Taylor bound is stated in a variational form making direct comparison to fat-shattering bounds difficult. This paper provides
Pac-Bayesian generic chaining
- ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS
, 2004
"... There exist many different generalization error bounds for classification. Each of these bounds contains an improvement over the others for certain situations. Our goal is to combine these different improvements into a single bound. In particular we combine the PAC-Bayes approach introduced by McAll ..."
Abstract
-
Cited by 3 (2 self)
- Add to MetaCart
There exist many different generalization error bounds for classification. Each of these bounds contains an improvement over the others for certain situations. Our goal is to combine these different improvements into a single bound. In particular we combine the PAC-Bayes approach introduced by Mc
Simplified PAC-Bayesian Margin Bounds
"... Abstract. The theoretical understanding of support vector machines is largely based on margin bounds for linear classifiers with unit-norm weight vectors and unit-norm feature vectors. Unit-norm margin bounds have been proved previously using fat-shattering arguments and Rademacher complexity. Recen ..."
Abstract
- Add to MetaCart
solution to the variational problem implicit in the Langford-Shawe-Taylor bound and shows that the PAC-Bayesian margin bounds are significantly tighter. Because a PAC-Bayesian bound is derived from a particular prior distribution over hypotheses, a PAC-Bayesian margin bound also seems to provide insight
A PAC-Bayesian Margin Bound for Linear Classifiers
, 2002
"... We present a bound on the generalisation error of linear classifiers in terms of a refined margin quantity on the training sample. The result is obtained in a PAC-Bayesian framework and is based on geometrical arguments in the space of linear classifiers. The new bound constitutes an exponential imp ..."
Abstract
-
Cited by 37 (3 self)
- Add to MetaCart
We present a bound on the generalisation error of linear classifiers in terms of a refined margin quantity on the training sample. The result is obtained in a PAC-Bayesian framework and is based on geometrical arguments in the space of linear classifiers. The new bound constitutes an exponential
PAC-Bayesian AUC classification and scoring
"... We develop a scoring and classification procedure based on the PAC-Bayesian ap-proach and the AUC (Area Under Curve) criterion. We focus initially on the class of linear score functions. We derive PAC-Bayesian non-asymptotic bounds for two types of prior for the score parameters: a Gaussian prior, a ..."
Abstract
-
Cited by 1 (0 self)
- Add to MetaCart
We develop a scoring and classification procedure based on the PAC-Bayesian ap-proach and the AUC (Area Under Curve) criterion. We focus initially on the class of linear score functions. We derive PAC-Bayesian non-asymptotic bounds for two types of prior for the score parameters: a Gaussian prior
Combining PAC-Bayesian and Generic Chaining Bounds
, 2007
"... There exist many different generalization error bounds in statistical learning theory. Each of these bounds contains an improvement over the others for certain situations or algorithms. Our goal is, first, to underline the links between these bounds, and second, to combine the different improvements ..."
Abstract
-
Cited by 9 (1 self)
- Add to MetaCart
improvements into a single bound. In particular we combine the PAC-Bayes approach introduced by McAllester (1998), which is interesting for randomized predictions, with the optimal union bound provided by the generic chaining technique developed by Fernique and Talagrand (see Talagrand, 1996), in a way
PAC-Bayesian Analysis of Co-clustering and Beyond
"... We derive PAC-Bayesian generalization bounds for supervised and unsupervised learning models based on clustering, such as co-clustering, matrix tri-factorization, graphical models, graph clustering, and pairwise clustering. 1 We begin with the analysis of co-clustering, which is a widely used approa ..."
Abstract
-
Cited by 14 (7 self)
- Add to MetaCart
We derive PAC-Bayesian generalization bounds for supervised and unsupervised learning models based on clustering, such as co-clustering, matrix tri-factorization, graphical models, graph clustering, and pairwise clustering. 1 We begin with the analysis of co-clustering, which is a widely used
A Better Variance Control For Pac-Bayesian Classification
- LABORATOIRE DE PROBABILITÉS ET MODÈLES ALÉATOIRES, UNIVERSITÉS PARIS 6 AND PARIS 7
, 2004
"... The common method to understand and improve classification rules is to prove bounds on the generalization error. Here we provide localized data-based PAC-bounds for the di#erence between the risk of any two randomized estimators. We derive from these bounds two types of algorithms: the first one use ..."
Abstract
-
Cited by 12 (6 self)
- Add to MetaCart
The common method to understand and improve classification rules is to prove bounds on the generalization error. Here we provide localized data-based PAC-bounds for the di#erence between the risk of any two randomized estimators. We derive from these bounds two types of algorithms: the first one
Results 1 - 10
of
1,535