• Documents
  • Authors
  • Tables
  • Log in
  • Sign up
  • MetaCart
  • DMCA
  • Donate

CiteSeerX logo

Advanced Search Include Citations
Advanced Search Include Citations

DMCA

A comparison of event models for Naive Bayes text classification (1998)

Cached

  • Download as a PDF

Download Links

  • [www.cs.cmu.edu]
  • [www.cs.cmu.edu]
  • [www-2.cs.cmu.edu]
  • [faculty.cs.byu.edu]
  • [www.cs.cmu.edu]
  • [www.kamalnigam.com]
  • [ssli.ee.washington.edu]
  • [dns2.icar.cnr.it]
  • [courses.washington.edu]
  • [www.cs.cmu.edu]

  • Save to List
  • Add to Collection
  • Correct Errors
  • Monitor Changes
by Andrew McCallum , Kamal Nigam
Citations:1023 - 26 self
  • Summary
  • Citations
  • Active Bibliography
  • Co-citation
  • Clustered Documents
  • Version History

BibTeX

@MISC{McCallum98acomparison,
    author = {Andrew McCallum and Kamal Nigam},
    title = { A comparison of event models for Naive Bayes text classification},
    year = {1998}
}

Share

Facebook Twitter Reddit Bibsonomy

OpenURL

 

Abstract

Recent work in text classification has used two different first-order probabilistic models for classification, both of which make the naive Bayes assumption. Some use a multi-variate Bernoulli model, that is, a Bayesian Network with no dependencies between words and binary word features (e.g. Larkey and Croft 1996; Koller and Sahami 1997). Others use a multinomial model, that is, a uni-gram language model with integer word counts (e.g. Lewis and Gale 1994; Mitchell 1997). This paper aims to clarify the confusion by describing the differences and details of these two models, and by empirically comparing their classification performance on five text corpora. We find that the multi-variate Bernoulli performs well with small vocabulary sizes, but that the multinomial performs usually performs even better at larger vocabulary sizes---providing on average a 27% reduction in error over the multi-variate Bernoulli model at any vocabulary size.

Keyphrases

event model    naive bayes text classification    multi-variate bernoulli model    vocabulary size    different first-order probabilistic model    binary word feature    multinomial performs    bayesian network    classification performance    text classification    integer word    text corpus    uni-gram language model    multinomial model    naive bayes assumption    small vocabulary size    recent work    multi-variate bernoulli performs   

Powered by: Apache Solr
  • About CiteSeerX
  • Submit and Index Documents
  • Privacy Policy
  • Help
  • Data
  • Source
  • Contact Us

Developed at and hosted by The College of Information Sciences and Technology

© 2007-2019 The Pennsylvania State University