## Supervised Learning of Quantizer Codebooks by Information Loss Minimization (2007)

Citations: | 33 - 0 self |

### BibTeX

@MISC{Lazebnik07supervisedlearning,

author = {Svetlana Lazebnik and Maxim Raginsky},

title = {Supervised Learning of Quantizer Codebooks by Information Loss Minimization},

year = {2007}

}

### OpenURL

### Abstract

This paper proposes a technique for jointly quantizing continuous features and the posterior distributions of their class labels based on minimizing empirical information loss, such that the index K of the quantizer region to which a given feature X is assigned approximates a sufficient statistic for its class label Y. We derive an alternating minimization procedure for simultaneously learning codebooks in the Euclidean feature space and in the simplex of posterior class distributions. The resulting quantizer can be used to encode unlabeled points outside the training set and to predict their posterior class distributions, and has an elegant interpretation in terms of lossless source coding. The proposed method is extensively validated on synthetic and real datasets, and is applied to two diverse problems: learning discriminative visual vocabularies for bag-of-features image classification, and image segmentation.