## Strong Entropy Concentration, Game Theory and Algorithmic Randomness (2001)

Citations: | 3 - 2 self |

### BibTeX

@MISC{Grünwald01strongentropy,

author = {Peter Grünwald},

title = {Strong Entropy Concentration, Game Theory and Algorithmic Randomness},

year = {2001}

}

### OpenURL

### Abstract

. We give a characterization of Maximum Entropy/Minimum Relative Entropy inference by providing two `strong entropy concentration ' theorems. These theorems unify and generalize Jaynes' `concentration phenomenon' and Van Campenhout and Cover's `conditional limit theorem'. The theorems characterize exactly in what sense a `prior' distribution Q conditioned on a given constraint and the distribution ~ P minimizing D(P jjQ) over all P satisfying the constraint are `close' to each other. We show how our theorems are related to `universal models ' for exponential families, thereby establishing a link with Rissanen's MDL/stochastic complexity. We then apply our theorems to establish the relationship (A) between entropy concentration and a game-theoretic characterization of Maximum Entropy Inference due to Topse and others; (B) between maximum entropy distributions and sequences that are random (in the sense of Martin-Lof/Kolmogorov) with respect to the given constraint. These two applications have strong implications for the use of Maximum Entropy distributions in sequential prediction tasks, both for the logarithmic loss and for general loss functions. We identify circumstances under which Maximum Entropy predictions are almost optimal. 1