## Model Selection by Normalized Maximum Likelihood (2005)

Citations: | 12 - 3 self |

### BibTeX

@MISC{Myung05modelselection,

author = {Jay I. Myung and Daniel J. Navarro and Mark A. Pitt},

title = {Model Selection by Normalized Maximum Likelihood},

year = {2005}

}

### Years of Citing Articles

### OpenURL

### Abstract

The Minimum Description Length (MDL) principle is an information theoretic approach to inductive inference that originated in algorithmic coding theory. In this approach, data are viewed as codes to be compressed by the model. From this perspective, models are compared on their ability to compress a data set by extracting useful information in the data apart from random noise. The goal of model selection is to identify the model, from a set of candidate models, that permits the shortest description length (code) of the data. Since Rissanen originally formalized the problem using the crude ‘two-part code ’ MDL method in the 1970s, many significant strides have been made, especially in the 1990s, with the culmination of the development of the refined ‘universal code’ MDL method, dubbed Normalized Maximum Likelihood (NML). It represents an elegant solution to the model selection problem. The present paper provides a tutorial review on these latest developments with a special focus on NML. An application example of NML in cognitive modeling is also provided.